phdtaketaketake

skill
Guvenlik Denetimi
Uyari
Health Uyari
  • License — License: MIT
  • Description — Repository has a description
  • Active repo — Last push 0 days ago
  • Low visibility — Only 5 GitHub stars
Code Gecti
  • Code scan — Scanned 12 files during light audit, no dangerous patterns found
Permissions Gecti
  • Permissions — No dangerous permissions requested

Bu listing icin henuz AI raporu yok.

SUMMARY

Connection-first PhD advisor matcher — find the right advisor by network strength, not h-index. Quantitative 4.0 scoring across Connection, Publication, Experience, GPA.

README.md

phdtaketaketake

Connection-first PhD advisor matcher, packaged as a Claude Code skill (also works with Codex CLI / Cursor / any LLM coding agent that can read SKILL.md).
Find the right advisor by network strength, not h-index.

中文 · English

Install

git clone https://github.com/powerofjinbo/phdtaketaketake.git \
  ~/.claude/skills/phdtaketaketake

cd ~/.claude/skills/phdtaketaketake
pip install -e .

Claude Code auto-discovers the skill on next session. For other agents, see Use with non-Claude agents below.

Don't have Claude Code? Install at claude.com/code.

Use

In any Claude Code session, describe what you want in plain English (or Chinese):

"I'm applying for Physics PhDs this fall — here's my CV [paste]. Find advisors that match."

"我是 SJTU 材料系本科,研究方向 2D 材料 photodetector,GPA 88/100,求美国 PhD 申请定位。"

The agent will:

  1. Build your profile (asks for any missing key info)
  2. Web-research candidate advisors at your target schools matching your research direction
  3. Verify connection edges to your current advisor (co-author papers, academic genealogy, joint collaborations)
  4. Run scripts/match.py for deterministic 4.0-scale scoring
  5. Present ranked candidates with per-dimension breakdown and cited sources

Output per candidate

  • Match score (0–4.0) + application_strength (0–4.0, not a probability) with ±confidence band
  • 5-tier label: Reach · Target · Match · Safe · Far Reach
  • Per-dimension: Connection / Publication / Experience / GPA
  • Why matched — cited from real searches: e.g., "co-authored 4 papers with Prof. Wang in 2022–2024 (per Google Scholar) · same ATLAS collaboration since 2017"

Architecture: no static cache, real data only

There is no bundled cache of advisors. PhD-advisor data is too dynamic and too vast for static datasets to be useful. Instead:

Component Role
The agent (Claude / Codex / Cursor / …) Deep research: find candidates, verify connections, estimate signals — all from real web sources, never fabricated
scripts/match.py Pure-Python deterministic scoring — takes the agent's findings and applies the 4.0-scale formulas
data/journals/<field>.yaml, references/*.md, docs/scoring.md Authoritative project opinions on tiers / formulas / schema

Cardinal rule: real data only

Every connection edge, every candidate fact must trace back to a real source the agent actually fetched (Google Scholar / OpenAlex / INSPIRE-HEP / PubMed / Math Genealogy / faculty pages / etc.). Fabrication is strictly forbidden — students use these rankings for real decisions, and made-up data is worse than no data. Missing signals are honest; the matcher widens its confidence band accordingly.

Allowed sources + forbidden behaviors are enumerated in references/data_integrity.md.

Coverage

The scoring engine itself is field-agnostic, but the calibration is best-supported for fields the project was designed against.

Coverage
🟢 Best-supported physics / HEP, materials science (MSE) — bundled data/journals/<field>.yaml tier files; the scoring system was originally calibrated against these subcultures
🟡 Extensible chemistry, biology, CS, math, EE, chemical engineering, earth science — the agent uses references/journal_tiers.md cross-field guidance + its training knowledge; the confidence band is wider
⚠️ Field-specific caveats CS is conference-first (different venue hierarchy); biology has co-first authorship conventions; math has a slower publication pace; clinical fields use multi-center RCT-driven prestige

Quality of agent retrieval scales the result quality — and data is always fresh, since nothing is cached.

Adding a tier YAML for a new field: see CONTRIBUTING.md. PRs welcome.

How it differs

CSrankings h-index ranking phdtaketaketake
Data freshness static static ✅ real-time agent retrieval
Personalized ✅ student profile → candidate matching
Connection-first ✅ #1 ranking signal
Big-collab paper aware ✅ ATLAS/CMS-style 5+ author rule
Multi-STEM ❌ CS only partial ✅ universal (any field)

Scoring philosophy

Four dimensions, all on a 4.0 scale (matching GPA), tier-adaptively weighted by school competitiveness:

  • Connection (C) — paths between candidate ↔ your current advisor (co-author / genealogy / joint collaborations / committee)
  • Publication (P) — journal tier × author position decay; 5+ author papers handled specially for big-collaboration physics
  • Experience (E) — lab × duration × output, output-weighted (50%)
  • GPA (G) — direct on 4.0; percentage / 4.3 / 4.5 / UK honours all normalized

application_strength = match_score + tier_adjustment + pi_recruiting_signal, clipped to [0, 4.0].

Full formulas: docs/scoring.md · Skill instructions: SKILL.md · Profile + CandidateAdvisor schema: references/profile_schema.md.

Use with non-Claude agents

The skill is designed Claude-Code-native but the underlying matcher is plain Python and the workflow instructions in SKILL.md are framework-agnostic. To use with another agent:

  • Codex CLI / OpenCode: drop a symlink at the repo root: ln -s SKILL.md AGENTS.md. Codex auto-reads AGENTS.md.
  • Cursor: copy SKILL.md content into .cursorrules at your project root.
  • Other: tell the agent "follow the workflow in SKILL.md" — most modern coding agents read it and execute the deep-research + scripts/match.py flow correctly.

Example session

See docs/example_session.md for a walk-through.

License

MIT — see LICENSE.

Yorumlar (0)

Sonuc bulunamadi