mentor

skill
Security Audit
Warn
Health Warn
  • License — License: MIT
  • Description — Repository has a description
  • Active repo — Last push 6 days ago
  • Low visibility — Only 5 GitHub stars
Code Pass
  • Code scan — Scanned 2 files during light audit, no dangerous patterns found
Permissions Pass
  • Permissions — No dangerous permissions requested
Purpose
This tool is a Claude skill designed to generate structured, official-first learning paths for technical topics. It guides users through a phased learning journey, prioritizing official documentation and ranking resources by reliability.

Security Assessment
Overall Risk: Low. The tool is essentially a prompt definition and JSON schema rather than executable software. It does not execute shell commands, access sensitive local data, or request dangerous system permissions. The code scan checked two files and found no dangerous patterns or hardcoded secrets. While the skill instructs the AI to perform web searches to verify source freshness, this is standard, safe internet usage entirely contained within the AI's environment.

Quality Assessment
The project is actively maintained, with its last push occurring only six days ago. It is fully open-source under the permissive MIT license, making it free for personal and commercial use. The repository is well-organized and includes a JSON schema, evaluations, and clear examples, demonstrating solid, developer-friendly design. However, it currently has low community visibility with only five GitHub stars, meaning it has not been extensively peer-reviewed by a wide audience.

Verdict
Safe to use.
SUMMARY

A Claude skill that generates structured, official-first learning paths for technical topics

README.md

Mentor

License: MIT

A Claude skill that generates structured, official-first learning paths for technical topics.

What It Does

Given a topic like "Terraform", "Google Cloud Run", or "React hooks", Mentor generates:

  • A learner profile (inferred or defaulted)
  • A dependency-ordered core learning path
  • Optional exploration branches
  • Self-assessment checkpoints
  • An explicit skip-list (avoid for now)
  • Follow-on topic suggestions

How It Works

flowchart TD
    A["User asks to learn a topic"] --> B{"Classify topic scope"}
    B -->|Broad| C["4-phase path\n(Mental Model → Core Mechanics → Applied Patterns → Go Deeper)"]
    B -->|Narrow| D["Compressed path\n(Orientation → Core → Applied → Reference)"]
    C --> E{"Infer learner profile"}
    D --> E
    E --> F{"Clarification needed?"}
    F -->|"Would change first 5 resources"| G["Ask 1 short question"]
    F -->|No| H["Build path with defaults"]
    G --> H
    H --> I["Research & verify sources\n(web search for freshness)"]
    I --> J["Apply source ranking\nTier 1 → 2 → 3 → 4"]
    J --> K["Assemble output"]

Source ranking: official docs first, vendor/maintainer material second, official sample repos third, community content only when it fills a real gap.

Output Structure

block-beta
    columns 1
    block:core["Core Path"]
        columns 4
        p1["Phase 1\nMental Model"]
        p2["Phase 2\nCore Mechanics"]
        p3["Phase 3\nApplied Patterns"]
        p4["Phase 4\nGo Deeper"]
    end
    block:branches["Exploration Branches (optional, non-sequential)"]
        columns 3
        b1["Hands-on\nPractice"]
        b2["Architecture\n& Team Usage"]
        b3["Deep\nInternals"]
    end
    block:supporting["Supporting Sections"]
        columns 4
        s1["Checkpoints"]
        s2["Avoid for Now"]
        s3["Next Topics"]
        s4["Navigator's Note"]
    end

Every core resource includes: source tier, engagement mode (Read now / Skim / Hands-on / Bookmark as reference), exact URL, sequencing rationale, and effort estimate.

Repository Structure

flowchart LR
    root["mentor/"] --> skill["SKILL.md\nskill definition"]
    root --> refs["references/"]
    root --> examples["examples/"]
    root --> evals["evals/"]
    refs --> schema["schema.json\nJSON output contract"]
    examples --> ex1["example-output-rust.md"]
    examples --> ex2["example-output-cloud-run.md"]
    examples --> ex3["example-output-terraform-modules.md"]
    examples --> ex4["example-output-react-server-components.md"]
    evals --> evj["evals.json\n9 test cases with assertions"]
  • SKILL.md is the source of truth. All behavior rules, output format, source ranking, and anti-patterns live here.
  • references/schema.json defines the machine-readable output contract for JSON mode.
  • examples/ contains gold-standard outputs demonstrating broad topics, narrow topics, clarification triggers, and user-background adaptation.
  • evals/evals.json contains 9 test cases with machine-gradable assertions covering structure, source quality, dedup, mode variety, and more.

Design Principle

Build the shortest credible path to competence from trustworthy sources, while preserving room for exploration.

Target Runtime

This skill is designed for Claude.ai and Claude Code. It uses the Claude skill format (SKILL.md with YAML frontmatter) and relies on Claude's web search capability to verify resource URLs and freshness.

Installation

Option 1: Download the .skill file (recommended)

Download mentor.skill from the latest release, then:

Platform Command / Action
Claude Code claude skill add mentor.skill
Claude.ai Open a Project → Settings → Skills → Upload mentor.skill

Option 2: Clone the repository

git clone https://github.com/ayhammouda/mentor.git ~/.claude/skills/mentor

Option 3: Manual copy

Copy the mentor/ directory into your Claude skills location (e.g., ~/.claude/skills/mentor/).

For more details on installing and managing skills, see the official documentation:

Usage

Ask Claude to learn something:

  • "I want to learn Kubernetes"
  • "learning path for Terraform modules"
  • "teach me Rust, I'm a Go developer"

Mentor activates automatically and generates a structured learning path.

For JSON output, ask explicitly: "Give me a learning path for Docker in JSON format"

Contributing

Contributions welcome! See CONTRIBUTING.md for guidelines on proposing changes, adding examples, and strengthening eval coverage.

License

MIT

Reviews (0)

No results found