researchclaw-skill
Health Warn
- License — License: MIT
- Description — Repository has a description
- Active repo — Last push 0 days ago
- Low visibility — Only 6 GitHub stars
Code Pass
- Code scan — Scanned 12 files during light audit, no dangerous patterns found
Permissions Pass
- Permissions — No dangerous permissions requested
This tool is an automation wrapper for AutoResearchClaw, a 23-stage pipeline that autonomously generates research papers from a single topic prompt. It primarily provides setup automation, interactive configuration, and self-healing hooks rather than acting as a standalone application.
Security Assessment
The overall risk is rated as Low. A light code scan of 12 files found no dangerous patterns, hardcoded secrets, or malicious code. The tool does not request inherently dangerous permissions. However, because it is written in Shell, it inherently executes system commands. Furthermore, the underlying pipeline requires significant capabilities: it executes Python code to run simulated experiments and makes external network requests to fetch real papers from academic databases like arXiv and Semantic Scholar. There is no evidence that it accesses sensitive personal data, but you should be aware of its network and code-execution footprint.
Quality Assessment
The project is actively maintained, with its most recent push occurring today. It is properly licensed under the standard MIT license. The primary concern is its extremely low community visibility. With only 6 GitHub stars, the tool has not been broadly tested or vetted by the open-source community. Additionally, developers must independently trust the upstream dependency (AutoResearchClaw), as this wrapper simply orchestrates it.
Verdict
Use with caution: the wrapper itself is safe, but its low community adoption and heavy reliance on an external pipeline require you to manually verify the upstream dependency before integrating it into your environment.
Turn your coding agent into a one-command autonomous research paper generator. Wraps AutoResearchClaw's 23-stage pipeline with setup automation, interactive config, error diagnosis, and self-healing hooks.
Turn your coding agent into a one-command autonomous research paper generator.
This skill requires AutoResearchClaw to be installed. It is a wrapper that simplifies setup, configuration, execution, and troubleshooting — it does not replace the upstream project. Install AutoResearchClaw first, then install this skill on top.
What is AutoResearchClaw?
AutoResearchClaw is a fully autonomous 23-stage research pipeline by aiming-lab. You give it a research topic. It gives you a conference-grade LaTeX paper.
What happens in those 23 stages:
| Phase | What It Does |
|---|---|
| Research Scoping | Parses your topic, decomposes it into sub-problems |
| Literature Discovery | Searches arXiv, Semantic Scholar, OpenAlex for real papers |
| Knowledge Synthesis | Clusters findings, identifies gaps, generates testable hypotheses |
| Experiment Design | Plans experiments, generates hardware-aware Python code |
| Experiment Execution | Runs code in a sandbox with self-healing (up to 10 repair cycles) |
| Analysis & Decision | Multi-agent result analysis — autonomously decides to proceed, refine, or pivot |
| Paper Writing | Generates 5,000-6,500 word paper with multi-agent peer review |
| Finalization | 4-layer citation verification, LaTeX export, PDF compilation |
The pipeline includes 3 human-approval gates, anti-fabrication guards, and a self-learning system (MetaClaw) that gets smarter with each run.
The problem: AutoResearchClaw is powerful but painful to set up. 55+ issues filed in the first 5 days — most about setup failures, config confusion, and Stage 10 crashes.
This skill solves that.
Quick Install
npx skills add OthmanAdi/researchclaw-skill --skill researchclaw -g
中文版 / Chinese:
npx skills add OthmanAdi/researchclaw-skill --skill researchclaw-cn -g
Works with Claude Code, Cursor, Codex, Gemini CLI, OpenClaw, and any agent supporting the Agent Skills spec.
Prerequisites
You must install AutoResearchClaw before using this skill:
pip install researchclaw
Or from source (recommended for latest features):
git clone https://github.com/aiming-lab/AutoResearchClaw.git
cd AutoResearchClaw
pip install -e ".[all]"
You also need:
- Python 3.11+
- An LLM API key (OpenAI, Anthropic, DeepSeek, or any OpenAI-compatible provider)
- Docker (optional — only for sandbox experiment mode)
- LaTeX (optional — only for PDF compilation)
Commands
| Command | What It Does |
|---|---|
/researchclaw |
Show help and suggest next step based on what's missing |
/researchclaw:setup |
Detect and install all prerequisites (asks before installing anything) |
/researchclaw:config |
Interactive wizard — generates a working config.yaml in 3 question batches |
/researchclaw:run |
Pre-flight validation + pipeline execution + auto-diagnosis on failure |
/researchclaw:status |
Show progress: Stage X/23 — [stage name] — [running/failed/complete] |
/researchclaw:resume |
Resume from last successful stage (with workarounds for known upstream bugs) |
/researchclaw:diagnose |
Pattern-match errors from logs and surface concrete fixes |
/researchclaw:validate |
Run all checks without starting the pipeline |
How It Works
You: /researchclaw:config
↓ answers 3 batches of questions (topic, LLM, experiment mode)
↓ generates config.yaml
You: /researchclaw:run "Attention mechanisms for time series forecasting"
↓ pre-flight validation (config, API key, Docker, disk space)
↓ launches AutoResearchClaw pipeline
↓ monitors 23 stages automatically
↓ auto-diagnoses failures via PostToolUse hook
↓ reports results with paper location
You: artifacts/rc-20260320-141523-a7f3/deliverables/paper.pdf
Custom Hooks
This skill includes 4 Claude Code hooks that run automatically — no configuration needed:
| Hook | Event | What It Does |
|---|---|---|
| Error Scanner | PostToolUse |
Scans output of any researchclaw command for 10 known error patterns (HTTP 401, Stage 10, OOM, Docker, LaTeX, rate limits) and surfaces auto-diagnosis |
| Config Backup | PreToolUse |
Creates timestamped backup of config.yaml before any overwrite |
| Artifact Guard | PreToolUse |
Blocks accidental deletion of artifacts/ directory |
| Completion Notify | Notification |
Logs pipeline completion/failure + sends desktop notification |
What This Skill Does NOT Do
Honesty is a core principle. This skill:
- Does not replace AutoResearchClaw — It wraps the official CLI. You must install the upstream project first.
- Does not start Docker — It checks if Docker is running, but cannot start the daemon for you
- Does not provide API keys — You must supply your own LLM API keys
- Does not fix network issues — If your firewall blocks arXiv or Semantic Scholar, the skill tells you but cannot fix it
- Does not guarantee paper quality — Output depends on the LLM model, topic complexity, and experiment mode
- Does not modify upstream code — Zero changes to AutoResearchClaw's codebase
The skill's PostToolUse hook catches these automatically, but here's the reference:
| Error Pattern | Cause | Fix |
|---|---|---|
HTTP 401 / AuthenticationError |
Invalid or expired API key | Check config.yaml → llm.api_key_env or the env var it points to |
HTTP 429 / RateLimitError |
API rate limit hit | Wait 60 seconds and /researchclaw:resume, or switch model |
Stage 10 failure |
Code generation produced invalid Python | Use a stronger model (gpt-4o, claude-sonnet-4-20250514) or switch to simulated mode |
Docker errors |
Docker not running or permission denied | Run docker info to check; may need sudo usermod -aG docker $USER |
pdflatex not found |
LaTeX not installed | sudo apt-get install texlive-full (Linux) or brew install --cask mactex (macOS) |
quality_score < threshold |
Quality gate too strict | Lower quality.min_score in config (default 2.0 is very strict, try 3.0-4.0) |
MemoryError / OOM |
Insufficient RAM | Use simulated experiment mode or reduce max_concurrent_stages |
ConnectionError |
Network issue with arXiv/Semantic Scholar | Check internet; try curl https://api.semanticscholar.org/graph/v1/paper/search?query=test |
YAML parse error |
Malformed config file | Run python3 -c "import yaml; yaml.safe_load(open('config.yaml'))" |
ModuleNotFoundError |
Missing Python dependency | Run pip install researchclaw[all] |
AutoResearchClaw supports ACP (Agent Client Protocol), which lets your coding agent (Claude Code, Copilot CLI, Gemini CLI) act as the LLM backend for all 23 stages. No separate API key needed.
llm:
provider: "acp"
acp:
agent: "claude"
cwd: "."
In ACP mode, the pipeline maintains a persistent session with your agent across all stages. The agent remembers context from literature review when designing experiments, and from experiments when writing the paper.
Pipeline Stage Reference (All 23 Stages)| # | Stage | Phase | Gate? |
|---|---|---|---|
| 1 | Topic Initialization | Research Scoping | |
| 2 | Problem Decomposition | Research Scoping | |
| 3 | Search Strategy | Literature Discovery | |
| 4 | Literature Collection | Literature Discovery | |
| 5 | Literature Screening | Literature Discovery | Human Approval |
| 6 | Knowledge Extraction | Literature Discovery | |
| 7 | Synthesis | Knowledge Synthesis | |
| 8 | Hypothesis Generation | Knowledge Synthesis | |
| 9 | Experiment Design | Experiment Design | Human Approval |
| 10 | Code Generation | Experiment Design | |
| 11 | Resource Planning | Experiment Design | |
| 12 | Experiment Execution | Execution | |
| 13 | Iterative Refinement | Execution | |
| 14 | Result Analysis | Analysis | |
| 15 | Research Decision | Analysis | PROCEED / REFINE / PIVOT |
| 16 | Paper Outline | Paper Writing | |
| 17 | Paper Draft | Paper Writing | |
| 18 | Peer Review | Paper Writing | |
| 19 | Paper Revision | Paper Writing | |
| 20 | Quality Gate | Finalization | Human Approval |
| 21 | Knowledge Archive | Finalization | |
| 22 | Export & Publish | Finalization | |
| 23 | Citation Verification | Finalization |
Gate stages (5, 9, 20) pause for human approval. Skip with --auto-approve.
Stage 15 can autonomously loop: REFINE goes back to Stage 13, PIVOT goes back to Stage 8.
Project Structure
skills/ # Publishing source (skills.sh + npx skills add)
├── researchclaw/
│ ├── SKILL.md # Main skill definition (English)
│ ├── assets/
│ │ └── config-template.yaml # Config generation template
│ ├── references/
│ │ ├── pipeline-stages.md # All 23 stages documented
│ │ ├── config-reference.md # Every config field explained
│ │ ├── troubleshooting.md # 10 error patterns with fixes
│ │ └── README-CN.md # Chinese reference
│ └── scripts/
│ ├── check-prereqs.sh # JSON prerequisite report
│ ├── post-run-check.sh # PostToolUse error scanner
│ ├── pre-config-write.sh # PreToolUse config backup
│ ├── pre-delete-guard.sh # PreToolUse artifact guard
│ └── notify-completion.sh # Completion notification
└── researchclaw-cn/
├── SKILL.md # Chinese skill definition
└── scripts/ # Shared hook scripts
.claude/
├── hooks.json # 4 Claude Code hooks (local dev)
└── skills/ # Local install mirror
├── researchclaw/
└── researchclaw-cn/
tests/
└── test-skill.sh # 58-test self-validation suite
docs/
└── zh-CN/
└── README.md # Full Chinese documentation
media/
├── banner.png # Repo banner
├── logo.png # Logo (researcher lobster)
└── logo_wizard.png # Logo variant (wizard lobster)
Validation
Evaluated using SkillCheck and a manual security audit following the same methodology used for planning-with-files.
SkillCheck (Free Tier)
| Check | Result |
|---|---|
| Frontmatter structure (name, description, allowed-tools) | Pass |
Name format (^[a-z][a-z0-9-]*[a-z0-9]$) |
Pass |
| Description WHAT (action verb) + WHEN (trigger phrase) | Pass |
| Directory structure matches name field | Pass |
| Subdirectories follow spec (references/, scripts/, assets/) | Pass |
| Naming quality (descriptive compound, not generic) | Pass |
| No contradictions in instructions | Pass |
| No ambiguous terms | Pass |
| Output format specified | Pass |
Strengths detected (6/6): Example section, error handling, trigger phrases, output format, structured instructions, prerequisites documented.
Security Audit
| Check | Result | Detail |
|---|---|---|
| WebFetch/WebSearch in allowed-tools | Pass | Not present (the vector fixed in planning-with-files v2.21.0) |
| Unrestricted Bash | Pass | All Bash patterns scoped (e.g. Bash(python*), no wildcard Bash(*)) |
| Hardcoded credentials | Pass | API keys referenced only via env vars |
| Script injection (eval/exec) | Pass | Zero dynamic execution in all 5 scripts |
| Strict mode | Pass | All scripts use set -euo pipefail |
| PII in examples | Pass | None found |
| Artifact deletion guard | Pass (strength) | pre-delete-guard.sh blocks rm *artifacts* |
| Config backup on overwrite | Pass (strength) | pre-config-write.sh creates timestamped backups |
Testing
bash tests/test-skill.sh
Validates file structure (21 checks), content quality (20 checks), script syntax (5 checks), hooks configuration (4 checks), and config template (8 checks).
58/58 tests passing.
System Requirements
| Component | Required | Notes |
|---|---|---|
| AutoResearchClaw | Yes | pip install researchclaw — this skill does not work without it |
| Python | 3.11+ | Core requirement |
| pip or uv | Yes | Package installation |
| Git | Yes | Cloning upstream repo |
| LLM API key | Yes | OpenAI, Anthropic, DeepSeek, or any OpenAI-compatible provider |
| Docker | For sandbox mode | Not needed for simulated mode |
| LaTeX | For PDF output | texlive-full recommended |
| RAM | 16 GB minimum | 32 GB+ recommended for full pipeline |
| Disk | 10 GB free | 50 GB+ for large runs with Docker |
Upstream
This skill wraps AutoResearchClaw v0.3.x by aiming-lab. All research pipeline functionality comes from the upstream project. This skill adds setup automation, interactive configuration, error diagnosis, and hooks — it does not fork or modify the upstream codebase.
Inspired by Karpathy's autoresearch philosophy and the wave of autonomous research tools that followed.
Contributing
- Run
bash tests/test-skill.shbefore submitting - Ensure all 58 tests pass
- Add tests for new functionality
- Keep the honesty policy — never fabricate capabilities
License
MIT — same as AutoResearchClaw upstream.
Reviews (0)
Sign in to leave a review.
Leave a reviewNo results found