phdtaketaketake
Health Uyari
- License — License: MIT
- Description — Repository has a description
- Active repo — Last push 0 days ago
- Low visibility — Only 5 GitHub stars
Code Gecti
- Code scan — Scanned 12 files during light audit, no dangerous patterns found
Permissions Gecti
- Permissions — No dangerous permissions requested
Bu listing icin henuz AI raporu yok.
Connection-first PhD advisor matcher — find the right advisor by network strength, not h-index. Quantitative 4.0 scoring across Connection, Publication, Experience, GPA.
phdtaketaketake
Connection-first PhD advisor matcher, packaged as a Claude Code skill (also works with Codex CLI / Cursor / any LLM coding agent that can read SKILL.md).
Find the right advisor by network strength, not h-index.
中文 · English
Install
git clone https://github.com/powerofjinbo/phdtaketaketake.git \
~/.claude/skills/phdtaketaketake
cd ~/.claude/skills/phdtaketaketake
pip install -e .
Claude Code auto-discovers the skill on next session. For other agents, see Use with non-Claude agents below.
Don't have Claude Code? Install at claude.com/code.
Use
In any Claude Code session, describe what you want in plain English (or Chinese):
"I'm applying for Physics PhDs this fall — here's my CV [paste]. Find advisors that match."
"我是 SJTU 材料系本科,研究方向 2D 材料 photodetector,GPA 88/100,求美国 PhD 申请定位。"
The agent will:
- Build your profile (asks for any missing key info)
- Web-research candidate advisors at your target schools matching your research direction
- Verify connection edges to your current advisor (co-author papers, academic genealogy, joint collaborations)
- Run
scripts/match.pyfor deterministic 4.0-scale scoring - Present ranked candidates with per-dimension breakdown and cited sources
Output per candidate
- Match score (0–4.0) + application_strength (0–4.0, not a probability) with ±confidence band
- 5-tier label: Reach · Target · Match · Safe · Far Reach
- Per-dimension: Connection / Publication / Experience / GPA
- Why matched — cited from real searches: e.g., "co-authored 4 papers with Prof. Wang in 2022–2024 (per Google Scholar) · same ATLAS collaboration since 2017"
Architecture: no static cache, real data only
There is no bundled cache of advisors. PhD-advisor data is too dynamic and too vast for static datasets to be useful. Instead:
| Component | Role |
|---|---|
| The agent (Claude / Codex / Cursor / …) | Deep research: find candidates, verify connections, estimate signals — all from real web sources, never fabricated |
scripts/match.py |
Pure-Python deterministic scoring — takes the agent's findings and applies the 4.0-scale formulas |
data/journals/<field>.yaml, references/*.md, docs/scoring.md |
Authoritative project opinions on tiers / formulas / schema |
Cardinal rule: real data only
Every connection edge, every candidate fact must trace back to a real source the agent actually fetched (Google Scholar / OpenAlex / INSPIRE-HEP / PubMed / Math Genealogy / faculty pages / etc.). Fabrication is strictly forbidden — students use these rankings for real decisions, and made-up data is worse than no data. Missing signals are honest; the matcher widens its confidence band accordingly.
Allowed sources + forbidden behaviors are enumerated in references/data_integrity.md.
Coverage
The scoring engine itself is field-agnostic, but the calibration is best-supported for fields the project was designed against.
| Coverage | |
|---|---|
| 🟢 Best-supported | physics / HEP, materials science (MSE) — bundled data/journals/<field>.yaml tier files; the scoring system was originally calibrated against these subcultures |
| 🟡 Extensible | chemistry, biology, CS, math, EE, chemical engineering, earth science — the agent uses references/journal_tiers.md cross-field guidance + its training knowledge; the confidence band is wider |
| ⚠️ Field-specific caveats | CS is conference-first (different venue hierarchy); biology has co-first authorship conventions; math has a slower publication pace; clinical fields use multi-center RCT-driven prestige |
Quality of agent retrieval scales the result quality — and data is always fresh, since nothing is cached.
Adding a tier YAML for a new field: see CONTRIBUTING.md. PRs welcome.
How it differs
| CSrankings | h-index ranking | phdtaketaketake | |
|---|---|---|---|
| Data freshness | static | static | ✅ real-time agent retrieval |
| Personalized | ❌ | ❌ | ✅ student profile → candidate matching |
| Connection-first | ❌ | ❌ | ✅ #1 ranking signal |
| Big-collab paper aware | ❌ | ❌ | ✅ ATLAS/CMS-style 5+ author rule |
| Multi-STEM | ❌ CS only | partial | ✅ universal (any field) |
Scoring philosophy
Four dimensions, all on a 4.0 scale (matching GPA), tier-adaptively weighted by school competitiveness:
- Connection (C) — paths between candidate ↔ your current advisor (co-author / genealogy / joint collaborations / committee)
- Publication (P) — journal tier × author position decay; 5+ author papers handled specially for big-collaboration physics
- Experience (E) — lab × duration × output, output-weighted (50%)
- GPA (G) — direct on 4.0; percentage / 4.3 / 4.5 / UK honours all normalized
application_strength = match_score + tier_adjustment + pi_recruiting_signal, clipped to [0, 4.0].
Full formulas: docs/scoring.md · Skill instructions: SKILL.md · Profile + CandidateAdvisor schema: references/profile_schema.md.
Use with non-Claude agents
The skill is designed Claude-Code-native but the underlying matcher is plain Python and the workflow instructions in SKILL.md are framework-agnostic. To use with another agent:
- Codex CLI / OpenCode: drop a symlink at the repo root:
ln -s SKILL.md AGENTS.md. Codex auto-readsAGENTS.md. - Cursor: copy
SKILL.mdcontent into.cursorrulesat your project root. - Other: tell the agent "follow the workflow in
SKILL.md" — most modern coding agents read it and execute the deep-research +scripts/match.pyflow correctly.
Example session
See docs/example_session.md for a walk-through.
License
MIT — see LICENSE.
Yorumlar (0)
Yorum birakmak icin giris yap.
Yorum birakSonuc bulunamadi