opencode-swarm
Health Gecti
- License — License: MIT
- Description — Repository has a description
- Active repo — Last push 0 days ago
- Community trust — 227 GitHub stars
Code Basarisiz
- rm -rf — Recursive force deletion command in .github/workflows/ci.yml
Permissions Gecti
- Permissions — No dangerous permissions requested
This is an MCP plugin for OpenCode that orchestrates a team of 11 specialized AI agents to plan, code, review, and test software projects autonomously.
Security Assessment
Overall risk: Medium. The tool uses a `bunx` or `npm install` command to install, register as a plugin, and modify local configurations automatically. While the automated scanner found no dangerous permissions requested, a recursive force deletion command (`rm -rf`) was flagged in the continuous integration (`.github/workflows/ci.yml`) file. This indicates the presence of highly destructive shell commands in the repository's workflow, though this is a common CI cleanup practice. The tool executes shell commands, writes to the local filesystem (saving session state to a `.swarm/` directory), and heavily relies on network requests to communicate with external AI model APIs. No hardcoded secrets were detected.
Quality Assessment
The project is in excellent health. It is actively maintained, with its most recent push occurring today. It uses the highly permissive MIT license, has a clear description, and is accompanied by thorough documentation. It shows solid community trust for a niche tool, currently boasting over 200 GitHub stars. The developers claim extensive test coverage, which aligns with the tool's core focus on quality assurance and gated pipelines.
Verdict
Use with caution — the plugin is highly active, well-documented, and open-source, but users should be aware that it executes commands and makes automated local file changes.
Architect-centric agentic swarm plugin for OpenCode. Hub-and-spoke orchestration with SME consultation, code generation, and QA review.
OpenCode Swarm
Your AI writes the code. Swarm makes sure it actually works.
OpenCode Swarm is a plugin for OpenCode that turns a single AI coding session into an architect-led team of 11 specialized agents. One agent writes the code. A different agent reviews it. Another writes and runs tests. Another checks security. Nothing ships until every required gate passes.
bunx opencode-swarm install
This single command installs the package, registers it as an OpenCode plugin, disables conflicting default agents, and creates a ready-to-edit config at
~/.config/opencode/opencode-swarm.json. Requires Bun (bun --versionto check). If you must use npm:npm install -g opencode-swarm && opencode-swarm install.
Why Swarm?
Most AI coding tools let one model write code and ask that same model whether the code is good. That misses too much. Swarm separates planning, implementation, review, testing, and documentation into specialized internal roles — and enforces gated execution so agents never mutate the codebase in parallel.
Key Features
- 🏗️ 11 specialized agents — architect, coder, reviewer, test engineer, critic, critic_sounding_board, critic_drift_verifier, explorer, SME, docs, designer
- 🔒 Gated pipeline — code never ships without reviewer + test engineer approval
- 🔄 Phase completion gates — completion-verify and drift verifier gates enforced before phase completion
- 🔁 Resumable sessions — all state saved to
.swarm/; pick up any project any day - 🌐 20 languages — TypeScript, Python, Go, Rust, Java, Kotlin, C/C++, C#, Ruby, Swift, Dart, PHP, JavaScript, CSS, Bash, PowerShell, INI, Regex
- 🛡️ Built-in security — SAST, secrets scanning, dependency audit per task
- 🆓 Free tier — works with OpenCode Zen's free model roster
- ⚙️ Fully configurable — override any agent's model, disable agents, tune guardrails
You select a Swarm architect once in the OpenCode GUI. The architect coordinates all other agents automatically — you never manually switch between internal roles. If you use the default OpenCode
Build/Planmodes, the plugin is bypassed entirely.
What Actually Happens
You say:
Build me a JWT auth system.
Swarm then:
- Clarifies only what it cannot infer.
- Scans the codebase to understand what already exists.
- Consults domain experts when needed and caches the guidance.
- Writes a phased implementation plan.
- Sends that plan through a critic gate before coding starts.
- Executes one task at a time through the QA pipeline:
- coder writes code
- automated checks run
- reviewer checks correctness
- test engineer writes and runs tests
- architect runs regression sweep
- failures loop back with structured feedback
- After each phase, docs and retrospectives are updated.
All project state lives in .swarm/ — plans, evidence, context, knowledge, and telemetry. Resumable by design. If .swarm/ already exists, the architect goes straight into RESUME → EXECUTE instead of repeating discovery.
Execution Modes
Swarm has two independent mode systems:
Session modes — toggle per session with a slash command:
| Mode | Safety | Speed | When to Use |
|---|---|---|---|
| Balanced (default) | High | Medium | Everyday development |
| Turbo | Medium | Fast | Rapid iteration; skips Stage B gates for non-Tier-3 files |
| Full-Auto | Depends on critic | Fast | Unattended multi-interaction runs |
Project mode — persistent via execution_mode config key:
| Value | Effect |
|---|---|
strict |
Maximum safety — adds slop-detector and incremental-verify hooks |
balanced (default) |
Standard hooks |
fast |
Skips compaction service — for short sessions under context pressure |
Switch session modes with /swarm turbo [on|off] or /swarm full-auto [on|off]. Set project mode in config. The two systems compose independently — see docs/modes.md.
Quick Start
→ For a complete first-run walkthrough, see Getting Started.
The 15-minute guide covers:
- Installation (
bunx opencode-swarm install) - Selecting the architect in OpenCode
- Running your first task
- Troubleshooting common issues
Commands
All 41 subcommands at a glance:
/swarm status # Current phase and task
/swarm plan [N] # Full plan or filtered by phase
/swarm agents # Registered agents and models
/swarm diagnose # Health check
/swarm evidence [task] # Test and review results
/swarm reset --confirm # Clear swarm state
See docs/commands.md for the full reference (41 commands).
The Agents
Swarm has 11 specialized agents. You don't manually switch between them — the architect coordinates automatically.
| Agent | Role |
|---|---|
| architect | Orchestrates workflow, writes plans, enforces gates |
| coder | Implements one task at a time |
| reviewer | Checks correctness and security |
| test_engineer | Writes and runs tests |
| critic | Reviews plans and challenges findings |
| explorer | Scans codebase and gathers context |
| sme | Provides domain expertise guidance |
| docs | Updates documentation to match implementation |
| designer | Generates UI scaffolds and design tokens |
| critic_sounding_board | Pre-escalation pushback to the architect |
| critic_drift_verifier | Verifies implementation matches plan |
Run /swarm status and /swarm agents to see what's active.
How It Compares
| Feature | Swarm | oh-my-opencode | get-shit-done |
|---|---|---|---|
| Multiple specialized agents | ✅ 11 agents | ❌ | ❌ |
| Plan reviewed before coding | ✅ | ❌ | ❌ |
| Every task reviewed + tested | ✅ | ❌ | ❌ |
| Different model for review vs. code | ✅ | ❌ | ❌ |
| Resumable sessions | ✅ | ❌ | ❌ |
| Built-in security scanning | ✅ | ❌ | ❌ |
| Learns from mistakes | ✅ | ❌ | ❌ |
LLM Provider Guide
Swarm works with any provider supported by OpenCode.
Free Tier (OpenCode Zen)
No API key required. Excellent starting point:
{
"agents": {
"coder": { "model": "opencode/minimax-m2.5-free" },
"reviewer": { "model": "opencode/big-pickle" },
"explorer": { "model": "opencode/trinity-large-preview-free" }
}
}
Paid Providers
For production, mix providers by role:
| Agent | Recommended | Why |
|---|---|---|
| architect | OpenCode UI selection | Needs strongest reasoning |
| coder | minimax-coding-plan/MiniMax-M2.5 | Fast, accurate code generation |
| reviewer | zai-coding-plan/glm-5 | Different training from coder |
| test_engineer | minimax-coding-plan/MiniMax-M2.5 | Same strengths as coder |
| explorer | google/gemini-2.5-flash | Fast read-heavy analysis |
| sme | kimi-for-coding/k2p5 | Strong domain expertise |
Provider Formats
| Provider | Format | Example |
|---|---|---|
| OpenCode Zen | opencode/<model> |
opencode/trinity-large-preview-free |
| Anthropic | anthropic/<model> |
anthropic/claude-sonnet-4-20250514 |
google/<model> |
google/gemini-2.5-flash |
|
| Z.ai | zai-coding-plan/<model> |
zai-coding-plan/glm-5 |
| MiniMax | minimax-coding-plan/<model> |
minimax-coding-plan/MiniMax-M2.5 |
| Kimi | kimi-for-coding/<model> |
kimi-for-coding/k2p5 |
Model Fallback
Automatic fallback to a secondary model on transient errors:
{
"agents": {
"coder": {
"model": "anthropic/claude-sonnet-4-20250514",
"fallback_models": ["opencode/gpt-5-nano"]
}
}
}
See docs/configuration.md for full configuration reference.
Advanced Topics (Technical Detail)
Process Remediation Model (PRM)
Swarm monitors agent trajectories and injects course-correction guidance before loops form. Detects five failure patterns:
- Repetition Loop — Same agent performs the same action repeatedly
- Ping-Pong — Agents hand off back and forth without progress
- Expansion Drift — Plan scope grows beyond original task
- Stuck-on-Test — Coder and tests fail in a loop
- Context Thrashing — Agent requests increasingly large file sets
When detected, escalation levels trigger:
- Level 1: Advisory guidance injected
- Level 2: Architect alert sent
- Level 3: Hard stop directive
Configure via:
{
"prm": {
"enabled": true,
"pattern_thresholds": {
"repetition_loop": 2,
"ping_pong": 4,
"expansion_drift": 3,
"stuck_on_test": 3,
"context_thrash": 5
}
}
}
Note: Some configuration fields (
max_trajectory_lines,escalation_enabled) are defined in schema but not yet enforced at runtime.
Persistent Memory
.swarm/plan-ledger.jsonl — authoritative source of truth (v6.44 durability model)
.swarm/context.md — technical decisions and cached SME guidance
.swarm/evidence/ — review/test results per task
.swarm/telemetry.jsonl — session observability events (fire-and-forget, never blocks execution)
.swarm/curator-summary.json — phase-level intelligence and drift reports
Guardrails & Circuit Breakers
Every agent runs inside a circuit breaker that prevents runaway behavior:
| Signal | Default Limit |
|---|---|
| Tool calls | 200 |
| Duration | 30 min |
| Same tool repeated | 10x |
| Consecutive errors | 5 |
Limits reset per task. Per-agent overrides available in config.
File Authority (Per-Agent Write Permissions)
Each agent can only write to specific paths:
- architect — Everything (except
.swarm/plan.md,.swarm/plan.json) - coder —
src/,tests/,docs/,scripts/ - reviewer —
.swarm/evidence/ - test_engineer —
tests/,.swarm/evidence/ - explorer, sme — Read-only
Override via authority.rules in config.
Quality Gates
Built-in tools verify every task before it ships:
- syntax_check — Tree-sitter validation (12 languages)
- placeholder_scan — Catches TODOs, stubs, incomplete code
- sast_scan — 63+ security rules, 9 languages (offline)
- sbom_generate — Dependency tracking (CycloneDX)
- quality_budget — Complexity, duplication, test ratio limits
All tools run locally. No Docker, no network calls.
Context Budget Guard
Monitors how much context Swarm injects to prevent overflow:
- Warning threshold (70%) — Advisory when context reaches ~2800 tokens
- Critical threshold (90%) — Alert at ~3600 tokens with
/swarm handoffrecommendation - Non-nagging — One-time alerts per session
Disable entirely with context_budget.enabled: false.
File Locking for Concurrent Safety
Hard lock on plan.json (serialized writes), advisory lock on events.jsonl (append-only log). Stale locks auto-expire via proper-lockfile.
Agent Categories
Agents are classified into four categories for the monitor server /metadata endpoint:
| Category | Agents |
|---|---|
orchestrator |
architect |
pipeline |
explorer, coder, test_engineer |
qa |
reviewer, critic, critic_sounding_board, critic_drift_verifier |
support |
sme, docs, designer |
Use getAgentCategory(agentName) from src/config/agent-categories.ts to resolve an agent's category at runtime.
Full Execution Pipeline (Technical Detail)
The Pipeline
Every task goes through this sequence. No exceptions, no overrides.
MODE: EXECUTE (per task)
│
├── 5a. @coder implements (ONE task only)
├── 5b. diff + imports (contract + dependency analysis + semantic diff context)
│ └── @system-enhancer injects AST-based semantic diff summary with blast radius
│ into @reviewer context (up to 10 files, conditional on declared scope)
├── 5c. syntax_check (parse validation)
├── 5d. placeholder_scan (catches TODOs, stubs, incomplete code)
├── 5e. lint fix → lint check
├── 5f. build_check (does it compile?)
├── 5g. pre_check_batch (4 parallel: lint, secretscan, SAST, quality budget)
├── 5h. @reviewer (correctness pass)
├── 5i. @reviewer (security pass, if security-sensitive files changed)
├── 5j. @test_engineer (verification tests + coverage ≥70%)
├── 5k. @test_engineer (adversarial tests)
├── 5l. architect regression sweep (scope:"graph" to find cross-task test regressions)
├── 5l-ter. test drift detection (conditional — fires when changes involve command behaviour,
│ parsing/routing logic, user-visible output, public contracts, assertion-heavy areas,
│ or helper lifecycle changes; validates tests still align with current behaviour)
├── 5m. ⛔ Pre-commit checklist (all 4 items required, no override)
└── 5n. Task marked complete, evidence written
If any step fails, the coder gets structured feedback and retries. After 5 failures on the same task, it escalates to you.
Architect Workflow Modes
The architect moves through these modes automatically:
| Mode | What It Means |
|---|---|
RESUME |
Existing .swarm/ state was found, so Swarm continues where it left off |
CLARIFY |
Swarm asks for missing information it cannot infer |
DISCOVER |
Explorer scans the codebase; co-change dark matter analysis runs automatically to detect hidden file couplings (v6.41) |
CONSULT |
SME agents provide domain guidance |
PLAN |
Architect writes or updates the phased plan (includes CODEBASE REALITY CHECK on brownfield projects) |
CRITIC-GATE |
Critic reviews the plan before execution |
EXECUTE |
Tasks are implemented one at a time through the QA pipeline |
PHASE-WRAP |
A phase closes out, including: explorer rescan, docs update, context.md update, write_retro, evidence check, sbom_generate, @critic_drift_verifier delegation (drift check — blocking gate), write_drift_evidence call with verdict, mandatory gate evidence verification (completion-verify.json + drift-verifier.json both required), then phase_complete |
CODEBASE REALITY CHECK (v6.29.2): Before any planning, the Architect dispatches Explorer to verify the current state of every referenced item. Produces a CODEBASE REALITY REPORT with statuses: NOT STARTED, PARTIALLY DONE, ALREADY COMPLETE, or ASSUMPTION INCORRECT. This prevents planning against stale assumptions. Skipped for greenfield projects with no existing codebase references.
Phase Completion Gates (v6.33.4): Before a phase can be marked complete, two mandatory gates are enforced: (1) completion-verify — deterministic check that plan task identifiers exist in source files, and (2) critic_drift_verifier evidence — verification that the drift verifier approved the implementation. Both gates are automatically bypassed when turbo mode is active.
Important
A second or later run does not necessarily look like a first run.
If .swarm/plan.md already exists, the architect may enter RESUME and then go directly into EXECUTE. That is expected and does not mean Swarm stopped using agents.
Use /swarm status if you are unsure what Swarm is doing.
Release automation uses release-please and requires conventional commit prefixes such as fix: or feat: on changes merged to main.
plan.md: Your Project Roadmap
# Project: Auth System
Current Phase: 2
## Phase 1: Foundation [COMPLETE]
- [x] Task 1.1: Create user model [SMALL]
- [x] Task 1.2: Add password hashing [SMALL]
- [x] Task 1.3: Database migrations [MEDIUM]
## Phase 2: Core Auth [IN PROGRESS]
- [x] Task 2.1: Login endpoint [MEDIUM]
- [ ] Task 2.2: JWT generation [MEDIUM] (depends: 2.1) ← CURRENT
- Acceptance: Returns valid JWT with user claims, 15-minute expiry
- Attempt 1: REJECTED — missing expiration claim
- [ ] Task 2.3: Token validation middleware [MEDIUM]
context.md: What's Been Decided
## Technical Decisions
- bcrypt cost factor: 12
- JWT TTL: 15 minutes; refresh TTL: 7 days
## SME Guidance (cached, never re-asked)
### security (Phase 1)
- Never log tokens or passwords
- Rate-limit login: 5 attempts / 15 min per IP
### api (Phase 1)
- Return 401 for invalid credentials (not 404)
Evidence Bundles
Every completed task writes structured evidence to .swarm/evidence/:
| Type | What It Captures |
|---|---|
| review | Verdict, risk level, specific issues |
| test | Pass/fail counts, coverage %, failure messages |
| diff | Files changed, additions/deletions |
| retrospective | Phase metrics, lessons learned, error taxonomy classification (injected into next phase) |
| secretscan | Secret scan results: findings count, files scanned, skipped files (v6.33) |
| completion-verify | Deterministic gate: verifies plan task identifiers exist in source files (written automatically by completion-verify tool; required before phase_complete) |
| drift-verifier | Phase-close drift gate: critic_drift_verifier verdict (APPROVED/NEEDS_REVISION) and summary (written by architect via write_drift_evidence; required before phase_complete) |
telemetry.jsonl: Session Observability
Swarm emits structured JSONL events to .swarm/telemetry.jsonl for observability tooling (dashboards, alerting, audit logs). Events are fire-and-forget — failures never affect execution.
{"timestamp":"2026-03-25T14:30:00.000Z","event":"session_started","sessionId":"abc123","agentName":"architect"}
{"timestamp":"2026-03-25T14:30:05.000Z","event":"delegation_begin","sessionId":"abc123","agentName":"coder","taskId":"1.1"}
{"timestamp":"2026-03-25T14:31:00.000Z","event":"delegation_end","sessionId":"abc123","agentName":"coder","taskId":"1.1","result":"success"}
{"timestamp":"2026-03-25T14:31:10.000Z","event":"gate_passed","sessionId":"abc123","gate":"reviewer","taskId":"1.1"}
{"timestamp":"2026-03-25T14:32:00.000Z","event":"phase_changed","sessionId":"abc123","oldPhase":1,"newPhase":2}
| Event | When Emitted |
|---|---|
session_started |
New agent session created |
session_ended |
Session ends (reason: normal, timeout, error) |
agent_activated |
Agent identity confirmed via chat.message |
delegation_begin |
Task dispatched to a sub-agent |
delegation_end |
Sub-agent returns (success, rejected, error) |
task_state_changed |
Task workflow state transitions |
gate_passed |
Evidence written to .swarm/evidence/{taskId}.json |
gate_failed |
Gate check blocked task completion |
phase_changed |
Phase completed and new phase started |
budget_updated |
Context budget crossed warning/critical threshold |
hard_limit_hit |
Tool call/duration/repetition limit reached |
revision_limit_hit |
Coder revision limit exceeded |
loop_detected |
Repetitive tool call pattern detected |
scope_violation |
Architect wrote outside declared scope |
qa_skip_violation |
QA gate skipped without valid reason |
model_fallback |
Transient error triggered model fallback |
heartbeat |
30-second throttled keep-alive signal |
File rotates automatically at 10MB to .swarm/telemetry.jsonl.1.
The save_plan tool requires an explicit target workspace path. It does not fall back to process.cwd().
Explicit Workspace Requirement
- The
working_directoryparameter must be provided - Providing no value or relying on implicit directory resolution will result in deterministic failure
Failure Conditions
| Condition | Behavior |
|---|---|
Missing (undefined / null) |
Fails with: "Target workspace is required" |
| Empty or whitespace-only | Fails with: "Target workspace cannot be empty or whitespace" |
Path traversal (..) |
Fails with: "Target workspace cannot contain path traversal" |
Usage Contract
When using save_plan, always pass a valid working_directory:
save_plan({
title: "My Project",
swarm_id: "mega",
phases: [{ id: 1, name: "Setup", tasks: [{ id: "1.1", description: "Initialize project" }] }],
working_directory: "/path/to/project" // Required - no fallback
})
Guardrails and Circuit Breakers
Every agent runs inside a circuit breaker that kills runaway behavior before it burns your credits.
| Signal | Default Limit | What Happens |
|---|---|---|
| Tool calls | 200 | Agent is stopped |
| Duration | 30 min | Agent is stopped |
| Same tool repeated | 10x | Agent is warned, then stopped |
| Consecutive errors | 5 | Agent is stopped |
Limits reset per task. A coder working on Task 2.3 is not penalized for tool calls made during Task 2.2.
Architect Self-Coding Block
If the architect writes files directly instead of delegating to the coder, a hard block fires:
| Write count | Behavior |
|---|---|
| 1–2 | Warning injected into next architect message |
| ≥ 3 | Error thrown with SELF_CODING_BLOCK — identifies file paths written and count |
The counter resets only when a coder delegation is dispatched. This is a hard enforcement — not advisory.
Per-agent overrides:
{
"guardrails": {
"profiles": {
"coder": { "max_tool_calls": 500, "max_duration_minutes": 60 },
"explorer": { "max_tool_calls": 50 }
}
}
}
File Authority (Per-Agent Write Permissions)
Swarm enforces per-agent file write authority — each agent can only write to specific paths. By default, these rules are hardcoded, but you can override them via config.
Default Rules
| Agent | Can Write | Blocked | Zones |
|---|---|---|---|
architect |
Everything (except plan files) | .swarm/plan.md, .swarm/plan.json |
generated |
coder |
src/, tests/, docs/, scripts/ |
.swarm/ (entire directory) |
generated, config |
reviewer |
.swarm/evidence/, .swarm/outputs/ |
src/, .swarm/plan.md, .swarm/plan.json |
generated |
test_engineer |
tests/, .swarm/evidence/ |
src/, .swarm/plan.md, .swarm/plan.json |
generated |
explorer |
Read-only | Everything | — |
sme |
Read-only | Everything | — |
docs |
docs/, .swarm/outputs/ |
— | generated |
designer |
docs/, .swarm/outputs/ |
— | generated |
critic |
.swarm/evidence/ |
— | generated |
Prefixed Agents
Prefixed agents (e.g., paid_coder, mega_reviewer, local_architect) inherit defaults from their canonical base agent via stripKnownSwarmPrefix. The lookup order is:
- Exact match for the prefixed name (if explicitly defined in user config)
- Fall back to the canonical agent's defaults (e.g.,
paid_coder→coder)
{
"authority": {
"rules": {
"coder": { "allowedPrefix": ["src/", "lib/"] },
"paid_coder": { "allowedPrefix": ["vendor/", "plugins/"] }
}
}
}
In this example, paid_coder gets its own explicit rule, while other prefixed coders (e.g., mega_coder) fall back to coder.
Runtime Enforcement
Architect direct writes are enforced at runtime via toolBefore hook. This tracks writes to source code paths outside .swarm/ and protects .swarm/plan.md and .swarm/plan.json from direct modification.
Configuration
Override default rules in .opencode/opencode-swarm.json:
{
"authority": {
"enabled": true,
"rules": {
"coder": {
"allowedPrefix": ["src/", "lib/", "scripts/"],
"blockedPrefix": [".swarm/"],
"blockedZones": ["generated"]
},
"explorer": {
"readOnly": false,
"allowedPrefix": ["notes/", "scratch/"]
}
}
}
}
Rule Fields
| Field | Type | Description |
|---|---|---|
readOnly |
boolean | If true, agent cannot write anywhere |
blockedExact |
string[] | Exact file paths that are blocked |
allowedExact |
string[] | Exact file paths that are allowed (overrides prefix/glob restrictions) |
blockedPrefix |
string[] | Path prefixes that are blocked (e.g., .swarm/) |
allowedPrefix |
string[] | Only these path prefixes are allowed. Omit to remove restriction; set [] to deny all |
blockedGlobs |
string[] | Glob patterns that are blocked (uses picomatch: **, *, ?) |
allowedGlobs |
string[] | Glob patterns that are allowed (uses picomatch: **, *, ?) |
blockedZones |
string[] | File zones to block: production, test, config, generated, docs, build |
Merge Behavior
- User rules override hardcoded defaults for the specified agent
- Scalar fields (
readOnly) — user value replaces default - Array fields (
blockedPrefix,allowedPrefix, etc.) — user array replaces entirely (not merged) - If a field is omitted in the user rule for a known agent (one with hardcoded defaults), the default value for that field is preserved
- If a field is omitted in the user rule for a custom agent (not in the defaults list), that field is
undefined— there are no defaults to inherit allowedPrefix: []explicitly denies all writes; omittingallowedPrefixentirely means no allowlist restriction is applied (all paths are evaluated against blocklist rules only)- Setting
enabled: falseignores all custom rules and uses hardcoded defaults
Custom Agents
Custom agents (not in the defaults list) start with no rules. Their write authority depends entirely on what you configure:
- Not in config at all — agent is denied with
Unknown agent(no rule exists; this is not the same as "blocked from all writes") - In config without
allowedPrefix— no allowlist restriction applies; only anyblockedPrefix,blockedZones, orreadOnlyrules you explicitly set will enforce limits - In config with
allowedPrefix: []— all writes are denied
To safely restrict a custom agent, always set allowedPrefix explicitly:
{
"authority": {
"rules": {
"my_custom_agent": {
"allowedPrefix": ["plugins/", "extensions/"],
"blockedZones": ["generated"]
}
}
}
}
Advanced Examples
Glob Pattern Support
Use glob patterns for complex path matching:
{
"authority": {
"rules": {
"coder": {
"allowedGlobs": ["src/**/*.ts", "tests/**/*.test.ts"],
"blockedGlobs": ["src/**/*.generated.ts", "**/*.d.ts"],
"allowedExact": ["src/index.ts", "package.json"]
},
"docs_agent": {
"allowedGlobs": ["docs/**/*.md", "*.md"],
"blockedExact": [".swarm/plan.md"]
}
}
}
}
Glob Pattern Features:
**— Match any number of directories:src/**/*.tsmatches all TypeScript files in src/ and subdirectories*— Match any characters except path separators:*.mdmatches all Markdown files in current directory?— Match single character:test?.jsmatchestest1.js,testa.js- Uses picomatch for cross-platform compatibility
Path Normalization and Symlinks:
Paths are resolved via realpathSync before matching, which resolves symlinks and prevents path-traversal escapes. However, if a symlink's target does not exist, realpathSync throws and the fallback returns the symlink's own path (unresolved). A dangling symlink inside an allowedPrefix directory will therefore pass prefix-based checks even if its intended target is outside the project. Use blockedExact or blockedGlobs to deny known dangling-symlink paths explicitly.
Evaluation Order:
readOnlycheck (if true, deny all writes)blockedExact(exact path matches, highest priority)blockedGlobs(glob pattern matches)allowedExact(exact path matches, overrides prefix/glob restrictions)allowedGlobs(glob pattern matches)blockedPrefix(prefix matches)allowedPrefix(prefix matches)blockedZones(zone classification)
The Context Budget Guard monitors how much context Swarm is injecting into the conversation. It helps prevent context overflow before it becomes a problem.
Default Behavior
- Enabled automatically — No setup required. Swarm starts tracking context usage right away.
- What it measures — Only the context that Swarm injects (plan, context, evidence, retrospectives). It does not count your chat history or the model's responses.
- Warning threshold (0.7 ratio) — When swarm-injected context reaches ~2800 tokens (70% of 4000), the architect receives a one-time advisory warning. This is informational — execution continues normally.
- Critical threshold (0.9 ratio) — When context reaches ~3600 tokens (90% of 4000), the architect receives a critical alert with a recommendation to run
/swarm handoff. This is also one-time only. - Non-nagging — Alerts fire once per session, not repeatedly. You won't be pestered every turn.
- Who sees warnings — Only the architect receives these warnings. Other agents are unaware of the budget.
To disable entirely, set context_budget.enabled: false in your swarm config.
Configuration Reference
| Key | Type | Default | Description |
|---|---|---|---|
context_budget.enabled |
boolean | true |
Enable or disable the context budget guard entirely |
context_budget.max_injection_tokens |
number | 4000 |
Token budget for swarm-injected context per turn. This is NOT the model's context window — it's the swarm plugin's own contribution |
context_budget.warn_threshold |
number | 0.7 |
Ratio (0.0-1.0) of max_injection_tokens that triggers a warning advisory |
context_budget.critical_threshold |
number | 0.9 |
Ratio (0.0-1.0) of max_injection_tokens that triggers a critical alert with handoff recommendation |
context_budget.enforce |
boolean | true |
When true, enforces budget limits and may trigger handoffs |
context_budget.prune_target |
number | 0.7 |
Ratio (0.0-1.0) of context to preserve when pruning occurs |
context_budget.preserve_last_n_turns |
number | 4 |
Number of recent turns to preserve when pruning |
context_budget.recent_window |
number | 10 |
Number of turns to consider as "recent" for scoring |
context_budget.tracked_agents |
string[] | ['architect'] |
Agents to track for context budget warnings |
context_budget.enforce_on_agent_switch |
boolean | true |
Enforce budget limits when switching agents |
context_budget.model_limits |
record | { default: 128000 } |
Per-model token limits (model name -> max tokens) |
context_budget.tool_output_mask_threshold |
number | 2000 |
Threshold for masking tool outputs (chars) |
context_budget.scoring.enabled |
boolean | false |
Enable context scoring/ranking |
context_budget.scoring.max_candidates |
number | 100 |
Maximum items to score (10-500) |
context_budget.scoring.weights |
object | { recency: 0.3, ... } |
Scoring weights for priority |
context_budget.scoring.decision_decay |
object | { mode: 'exponential', half_life_hours: 24 } |
Decision relevance decay |
context_budget.scoring.token_ratios |
object | { prose: 0.25, code: 0.4, ... } |
Token cost multipliers |
Example Configurations
Minimal (disable):
{
"context_budget": {
"enabled": false
}
}
Default (reference):
{
"context_budget": {
"enabled": true,
"max_injection_tokens": 4000,
"warn_threshold": 0.7,
"critical_threshold": 0.9,
"enforce": true,
"prune_target": 0.7,
"preserve_last_n_turns": 4,
"recent_window": 10,
"tracked_agents": ["architect"],
"enforce_on_agent_switch": true,
"model_limits": { "default": 128000 },
"tool_output_mask_threshold": 2000,
"scoring": {
"enabled": false,
"max_candidates": 100,
"weights": { "recency": 0.3, "relevance": 0.4, "importance": 0.3 },
"decision_decay": { "mode": "exponential", "half_life_hours": 24 },
"token_ratios": { "prose": 0.25, "code": 0.4, "json": 0.6, "logs": 0.1 }
}
}
}
Aggressive (for long-running sessions):
{
"context_budget": {
"enabled": true,
"max_injection_tokens": 2000,
"warn_threshold": 0.5,
"critical_threshold": 0.75,
"enforce": true,
"prune_target": 0.6,
"preserve_last_n_turns": 2,
"recent_window": 5,
"tracked_agents": ["architect"],
"enforce_on_agent_switch": true,
"model_limits": { "default": 128000 },
"tool_output_mask_threshold": 1500,
"scoring": {
"enabled": true,
"max_candidates": 50,
"weights": { "recency": 0.5, "relevance": 0.3, "importance": 0.2 },
"decision_decay": { "mode": "linear", "half_life_hours": 12 },
"token_ratios": { "prose": 0.2, "code": 0.35, "json": 0.5, "logs": 0.05 }
}
}
}
What This Does NOT Do
- Does NOT prune chat history — Your conversation with the model is untouched
- Does NOT modify tool outputs — What tools return is unchanged
- Does NOT block execution — The guard is advisory only; it warns but never stops the pipeline
- Does NOT interact with compaction.auto — Separate feature with separate configuration
- Only measures swarm's injected context — Not the full context window, just what Swarm adds
Built-in Tools
| Tool | What It Does |
|---|---|
| syntax_check | Tree-sitter validation across 12 languages |
| placeholder_scan | Catches TODOs, FIXMEs, stubs, placeholder text |
| sast_scan | Offline security analysis, 63+ rules, 9 languages |
| sbom_generate | CycloneDX dependency tracking, 8 ecosystems |
| build_check | Runs your project's native build/typecheck |
| incremental_verify | Post-coder typecheck for TS/JS, Go, Rust, C# (v6.29.2) |
| quality_budget | Enforces complexity, duplication, and test ratio limits |
| pre_check_batch | Runs lint, secretscan, SAST, and quality budget in parallel (~15s vs ~60s sequential) |
| phase_complete | Enforces phase completion, verifies required agents, requires a valid retrospective evidence bundle, logs events, and resets state; appends to events.jsonl with file locking |
| mutation_test | Applies LLM-generated mutation patches to source files and runs tests to measure kill rate; verdict is pass/warn/fail based on configurable thresholds; used by the mutation_test gate (opt-in, off by default) |
| generate_mutants | Architect-only: generates LLM-based mutation patches (5–10 per function across 6 types: off-by-one, null substitution, operator swap, guard removal, branch swap, side-effect deletion) for direct consumption by the mutation_test tool; returns SKIP verdict on LLM failure rather than throwing |
| write_mutation_evidence | Architect-only: writes mutation gate results atomically to .swarm/evidence/{phase}/mutation-gate.json; accepts verdict (PASS/WARN/FAIL/SKIP), kill rate metrics, and optional survived mutant details; normalizes uppercase-to-lowercase before persisting |
All tools run locally. No Docker, no network calls, no external APIs.
Optional enhancement: Semgrep (if on PATH).
Gate Configuration
{
"gates": {
"syntax_check": { "enabled": true },
"placeholder_scan": { "enabled": true },
"sast_scan": { "enabled": true },
"quality_budget": {
"enabled": true,
"max_complexity_delta": 5,
"min_test_to_code_ratio": 0.3
}
}
}
File Locking for Concurrent Write Safety
Swarm uses file locking to protect shared state files from concurrent write corruption. The locking strategy differs by file: plan.json uses hard locking (write blocked on contention), while events.jsonl uses advisory locking (write proceeds with a warning on contention).
Locking Implementation
- Library:
proper-lockfilewithretries: 0(fail-fast — no polling retries) - Scope: Each tool acquires an exclusive lock on the target file before writing
- Agents: Lock is tagged with the current agent name and task context for diagnostics
Protected Files
| File | Tool | Lock Key |
|---|---|---|
.swarm/plan.json |
update_task_status |
plan.json |
.swarm/events.jsonl |
phase_complete |
events.jsonl |
Lock Semantics
The two protected tools use different strategies:
update_task_status — Hard lock on plan.json
When two calls contend for plan.json:
- Exactly one call wins — only the first to acquire the lock proceeds
- Winner writes — the lock holder writes to the file, then releases the lock
- Losers receive
success: false— withrecovery_guidance: "retry"and an error message identifying the lock holder
{
"success": false,
"message": "Task status write blocked: plan.json is locked by architect (task: update-task-status-1.1-1234567890)",
"errors": ["Concurrent plan write detected — retry after the current write completes"],
"recovery_guidance": "Wait a moment and retry update_task_status. The lock will expire automatically if the holding agent fails."
}
What the caller should do: Retry update_task_status after a short delay.
phase_complete — Advisory lock on events.jsonl
When two calls contend for events.jsonl:
- Lock is attempted — if acquired, write is serialized
- If lock unavailable — a warning is added to the result and the write proceeds anyway
- Both callers return
success: true— duplicate concurrent appends are possible butevents.jsonlis an append-only log and duplicate phase entries do not corrupt state
This asymmetry is intentional: plan.json stores mutable structured JSON where concurrent overwrites produce malformed files; events.jsonl is an append-only log where a duplicate entry is a recoverable nuisance.
Lock Recovery
If a lock-holding agent crashes or hangs, the lock file will eventually expire (handled by proper-lockfile stale-lock cleanup). On the next retry, the call will succeed. Swarm does not auto-retry on lock contention — the architect receives the error and decides when to retry.
Config file location: ~/.config/opencode/opencode-swarm.json (global) or .opencode/opencode-swarm.json (project). Project config merges over global.
{
"agents": {
"architect": { "model": "anthropic/claude-opus-4-6" },
"coder": { "model": "minimax-coding-plan/MiniMax-M2.5", "fallback_models": ["minimax-coding-plan/MiniMax-M2.1"] },
"explorer": { "model": "minimax-coding-plan/MiniMax-M2.1" },
"sme": { "model": "kimi-for-coding/k2p5" },
"critic": { "model": "zai-coding-plan/glm-5" },
"reviewer": { "model": "zai-coding-plan/glm-5", "fallback_models": ["opencode/big-pickle"] },
"test_engineer": { "model": "minimax-coding-plan/MiniMax-M2.5" },
"docs": { "model": "zai-coding-plan/glm-4.7-flash" },
"designer": { "model": "kimi-for-coding/k2p5" }
},
"guardrails": {
"max_tool_calls": 200,
"max_duration_minutes": 30,
"profiles": {
"coder": { "max_tool_calls": 500 }
}
},
"authority": {
"enabled": true,
"rules": {
"coder": {
"allowedPrefix": ["src/", "lib/"],
"blockedPrefix": [".swarm/"],
"blockedZones": ["generated"]
}
}
},
"review_passes": {
"always_security_review": false,
"security_globs": ["**/*auth*", "**/*crypto*", "**/*session*"]
},
"automation": {
"mode": "manual",
"capabilities": {
"plan_sync": true,
"phase_preflight": false,
"config_doctor_on_startup": false,
"config_doctor_autofix": false,
"evidence_auto_summaries": true,
"decision_drift_detection": true
}
},
"knowledge": {
"enabled": true,
"swarm_max_entries": 100,
"hive_max_entries": 1000,
"auto_promote_days": 30,
"max_inject_count": 5,
"dedup_threshold": 0.6,
"scope_filter": ["global"],
"hive_enabled": true,
"rejected_max_entries": 200,
"validation_enabled": true,
"evergreen_confidence": 0.8,
"evergreen_utility": 0.5,
"low_utility_threshold": 0.2,
"min_retrievals_for_utility": 3,
"schema_version": "v6.17"
}
}
Automation
Mode Detection (v6.13)
Swarm now explicitly distinguishes five architect modes:
DISCOVER— After the explorer finishes scanning the codebase.PLAN— When the architect writes or updates the plan.EXECUTE— During task implementation (the normal pipeline).PHASE-WRAP— After all tasks in a phase are completed, before docs are updated.UNKNOWN— Fallback when the current state does not match any known mode.
Each mode determines which injection blocks are added to the LLM prompt (e.g., plan cursor is injected in PLAN, tool output truncation in EXECUTE, etc.).
Default mode: manual. No background automation — all actions require explicit slash commands.
Modes:
manual— No background automation. All actions via slash commands (default).hybrid— Background automation for safe operations, manual for sensitive ones.auto— Full background automation.
Capability defaults:
plan_sync:true— Background plan synchronization usingfs.watchwith debounced writes (300ms) and 2-second polling fallbackphase_preflight:false— Phase preflight checks before agent execution (opt-in)config_doctor_on_startup:false— Validate configuration on startupconfig_doctor_autofix:false— Auto-fix for config doctor (opt-in, security-sensitive)evidence_auto_summaries:true— Automatic summaries for evidence bundlesdecision_drift_detection:true— Detect drift between planned and actual decisions
Plan Cursor (v6.13)
The plan_cursor config compresses the plan that is injected into the LLM context.
{
"plan_cursor": {
"enabled": true,
"max_tokens": 1500,
"lookahead_tasks": 2
}
}
- enabled – When
true(default) Swarm injects a compact plan cursor instead of the fullplan.md. - max_tokens – Upper bound on the number of tokens emitted for the cursor (default 1500). The cursor contains the current phase summary, the full current task, and up to
lookahead_tasksupcoming tasks. Earlier phases are reduced to one‑line summaries. - lookahead_tasks – Number of future tasks to include in full detail (default 2). Set to
0to show only the current task.
Disabling ("enabled": false) falls back to the pre‑v6.13 behavior of injecting the entire plan text.
Tool Output Truncation (v6.13)
Control the size of tool outputs that are sent back to the LLM.
{
"tool_output": {
"truncation_enabled": true,
"max_lines": 150,
"per_tool": {
"diff": 200,
"symbols": 100
}
}
}
- truncation_enabled – Global switch (default true).
- max_lines – Default line limit for any tool output.
- per_tool – Overrides
max_linesfor specific tools. Thediffandsymbolstools are truncated by default because their outputs can be very large.
When truncation is active, a footer is appended:
---
[output truncated to {maxLines} lines – use `tool_output.per_tool.<tool>` to adjust]
Summarization Settings
Control how tool outputs are summarized for LLM context.
{
"summaries": {
"threshold_bytes": 102400,
"exempt_tools": ["retrieve_summary", "task", "read"]
}
}
- threshold_bytes – Output size threshold in bytes before summarization is triggered (default 102400 = 100KB).
- exempt_tools – Tools whose outputs are never summarized. Defaults to
["retrieve_summary", "task", "read"]to prevent re-summarization loops.
Note: The
retrieve_summarytool supports paginated retrieval viaoffsetandlimitparameters to fetch large summarized outputs in chunks.
Disabling Agents
{
"sme": { "disabled": true },
"designer": { "disabled": true },
"test_engineer": { "disabled": true }
}
All Slash Commands
| Command | Description |
|---|---|
/swarm status |
Current phase, task progress, agent count |
/swarm plan [N] |
Full plan or filtered by phase |
/swarm agents |
Registered agents with models and permissions |
/swarm history |
Completed phases with status |
/swarm config |
Current resolved configuration |
/swarm diagnose |
Health check for .swarm/ files and config |
/swarm export |
Export plan and context as portable JSON |
/swarm evidence [task] |
Evidence bundles for a task or all tasks |
/swarm archive [--dry-run] |
Archive old evidence with retention policy |
/swarm benchmark |
Performance benchmarks |
/swarm retrieve [id] |
Retrieve auto-summarized tool outputs (supports offset/limit pagination) |
/swarm reset --confirm |
Clear swarm state files |
/swarm reset-session |
Clear session state files in .swarm/session/ (preserves plan and context) |
/swarm preflight |
Run phase preflight checks |
/swarm config doctor [--fix] |
Config validation with optional auto-fix |
/swarm doctor tools |
Tool registration coherence and binary readiness check |
/swarm sync-plan |
Force plan.md regeneration from plan.json |
/swarm specify [description] |
Generate or import a feature specification |
/swarm clarify [topic] |
Clarify and refine an existing feature specification |
/swarm analyze |
Analyze spec.md vs plan.md for requirement coverage gaps |
/swarm close [--prune-branches] |
Idempotent session close-out: retrospectives, lesson curation, evidence archive, context.md reset, config-backup cleanup, optional branch pruning |
/swarm write-retro |
Write a phase retrospective manually |
/swarm handoff |
Generate a handoff summary for context-budget-critical sessions |
/swarm simulate |
Simulate plan execution without writing code |
/swarm promote |
Promote swarm-scoped knowledge to hive (global) knowledge |
/swarm evidence summary |
Generate a summary across all evidence bundles with completion ratio and blockers |
/swarm knowledge |
List knowledge entries |
/swarm knowledge migrate |
Migrate knowledge entries to the current format |
/swarm knowledge quarantine [id] |
Move a knowledge entry to quarantine |
/swarm knowledge restore [id] |
Restore a quarantined knowledge entry |
/swarm turbo |
Enable turbo mode for the current session (bypasses QA gates) |
/swarm full-auto |
Toggle Full-Auto Mode for the current session [on |
/swarm checkpoint |
Save a git checkpoint for the current state |
Supported Languages
Full Tier-1 support: TypeScript, JavaScript, Python, Go, Rust
Tier-2 support: Java, Kotlin, C#, C/C++, Swift
Tier-3 support: Dart, Ruby, PHP/Laravel
All binaries optional. Missing tools produce soft warnings, never hard-fail.
Testing
6,000+ tests. Unit, integration, adversarial, and smoke. Run with:
bun test
Design Principles
- Plan before code. Critic approves the plan before a single line is written.
- One task at a time. Coder gets one task and full context. Nothing else.
- Review everything immediately. Correctness, security, tests, adversarial tests. Every task.
- Different models catch different bugs. Blind spots of the coder are the reviewer's strength.
- Save everything to disk. Resume any project any day from
.swarm/state. - Document failures. Rejections and retries recorded. After 5 failures, escalate to you.
Documentation
- Getting Started — 15-minute first-run guide
- Documentation Index — navigate all docs
- Installation Guide — comprehensive reference
- Architecture Deep Dive — control model, pipeline, tools
- Design Rationale — why every major decision
- Commands Reference — all 41
/swarmsubcommands - Modes Guide — session modes (Turbo, Full-Auto) and project modes (strict/balanced/fast)
- Configuration — all config keys and examples
- Planning Guide — task format, phase structure, sizing
License
MIT
Stop hoping your agents figure it out. Start shipping code that actually works.
Yorumlar (0)
Yorum birakmak icin giris yap.
Yorum birakSonuc bulunamadi