Harlo
Health Uyari
- License — License: Apache-2.0
- Description — Repository has a description
- Active repo — Last push 0 days ago
- Low visibility — Only 6 GitHub stars
Code Gecti
- Code scan — Scanned 12 files during light audit, no dangerous patterns found
Permissions Gecti
- Permissions — No dangerous permissions requested
Bu listing icin henuz AI raporu yok.
Harlo — Your AI coach. Built on USD composition semantics for persistent cognitive state management. Patent Pending.
Patent Pending | Apache 2.0 | Patent Details
Your AI coach. Watches your patterns, predicts your crashes, backs off during
flow, and tells you when to stop before you burn out. Built on USD composition
semantics for persistent, local-first cognitive state management.
Your memory, your device. Harlo stores all state locally as composable USD
layers — no cloud dependency, no data mining, no rented access to your own mind.
Status
PRODUCTION LIVE — Harlo v3.4.0-path-c
1,172 tests passing · Real OpenUSD canonical persistence · USD-Lite runtime tier
8/8 phase gates passed · 19 D-block decisions clean (D1-D19)
Substrate-unified with sister project Moneta · P1 CIP defensible
458 organic observations collected · 5 sprints shipped · Path C closed (Step 3)
| Sprint | Tests | What Shipped |
|---|---|---|
| S1 State Machine | 84 | Pydantic schemas, MockCogExec DAG (networkx), 7 pure computation functions, 26-invariant validator, 10K synthetic trajectories via Profile-Driven Markov Biasing, XGBoost predictor (100% per-field accuracy), Bridge integration |
| S2 OpenExec | -- | USD 26.03 built from source with PXR_BUILD_EXEC=ON. C++ Exec libraries compile. Circuit-breaker triggered: zero Python bindings in v26.03 source. MockCogExec continues to serve. |
| S3 Hydra Delegates | 85 | HdCognitiveDelegate ABC, DelegateRegistry (capability matching), HdClaude + HdClaudeCode, compute_routing (requirements not names), OOB consent tokens (HMAC-signed, TTL), sublayer-per-delegate concurrency, CognitiveEngine singleton, 20-exchange e2e |
| S4 Real USD | 59 | CognitiveStage wrapping pxr.Usd.Stage, stage_factory toggle, .usda files on disk with time-sampled CognitiveObservation, delegate sublayer .usda files, backend parity verified (mock = real USD) |
| S5 Production | 22 | Graceful degradation (independent failure isolation), health check endpoint, kill switches (ENGINE_ENABLED, USE_REAL_USD, OBSERVATION_LOGGING, PREDICTION_ENABLED), first session verified, production docs |
| Path C Step 3 v3.4.0 | +39 | Real OpenUSD as canonical persistence (codeless schema, 21 prim types under harlo plugin separate from Moneta); USD-Lite engine preserved as fast in-memory runtime tier (Fabric pattern); sync layer per D4 policy table; migration script for USD-Lite v1 → real USD; substrate-unified with sister project Moneta. P1 CIP framing now defensible. |
Architecture · Path C (Fabric Pattern)
v3.4.0-path-c introduced codeless OpenUSD schemas as canonical
persistence while preserving the existing USD-Lite engine as a fast
in-memory runtime tier. Path C — the Fabric pattern — separates
the two tiers so each can win at what it's good at: real OpenUSD owns
durability and patent claims; USD-Lite owns hot-path latency.
Fabric pattern
flowchart TB
subgraph PERSISTENCE["PERSISTENCE LAYER · canonical truth"]
SCHEMA["HarloSchema.usda<br/>21 prim types · codeless"]:::substrate
PLUG["plugInfo.json<br/>harlo namespace"]:::substrate
DISK[".usda files on disk<br/>via pxr.Usd.Stage"]:::substrate
end
subgraph SYNCLAYER["SYNC LAYER · write-side dispatch"]
WT["write_through<br/>SessionPrim · GateStatusPrim<br/>MerkleRootPrim · MotorPrim"]:::substrate
CP["checkpoint<br/>TracePrim · CompositionLayerPrim<br/>SkillPrim · intake/multipliers"]:::substrate
end
subgraph RUNTIME["RUNTIME LAYER · hot-path reads"]
ENGINE["USD-Lite engine<br/>regex parser · sub-ms reads"]:::runtime
DC["21 dataclass prim types<br/>Python in-memory"]:::runtime
end
MIG["migrate_path_c.py<br/>USD-Lite v1 → real USD<br/>idempotent · CLI"]:::substrate
PERSISTENCE -->|"sync at boundaries"| SYNCLAYER
SYNCLAYER --> RUNTIME
MIG -.->|"upgrade path"| PERSISTENCE
classDef substrate fill:#1a2332,stroke:#4a90a4,color:#e8eef2
classDef runtime fill:#d4af37,stroke:#8b7115,color:#1a2332
The persistence layer is the canonical truth. The runtime layer is the
fast tier that tests and live sessions exercise. The sync layer routes
mutations between them based on a per-prim policy table. Reads always
hit the runtime tier; persistence is touched only at sync boundaries
(Constitution Law 4).
The [substrate] extra activates the persistence layer:
pip install -e .[substrate] # Pulls usd-core 26.5; activates persistence/
Core Harlo runs without [substrate] — pxr stays optional per
Constitution Law 3.
Schema · IsA hierarchy
The codeless schema in schema/HarloSchema.usda declares 21 prim types
in a 3-tier IsA hierarchy parallel to containment (D2):
flowchart TB
Typed["Typed · USD root"]:::substrate
HP["HarloPrim · abstract"]:::substrate
HC["HarloContainer · abstract"]:::substrate
Typed --> HP
HP --> HC
BS["BrainStage"]:::substrate
AP["AssociationPrim"]:::substrate
CP["CompositionPrim"]:::substrate
EP["ElenchusPrim"]:::substrate
ICP["InquiryContainerPrim"]:::substrate
MCP["MotorContainerPrim"]:::substrate
SCP["SkillsContainerPrim"]:::substrate
CPP["CognitiveProfilePrim"]:::substrate
HC --> BS
HC --> AP
HC --> CP
HC --> EP
HC --> ICP
HC --> MCP
HC --> SCP
HC --> CPP
TP["TracePrim"]:::runtime
CLP["CompositionLayerPrim"]:::runtime
GSP["GateStatusPrim"]:::runtime
MRP["MerkleRootPrim"]:::runtime
SP["SessionPrim"]:::runtime
IP["InquiryPrim"]:::runtime
MP["MotorPrim"]:::runtime
SkP["SkillPrim"]:::runtime
MuP["MultipliersPrim"]:::runtime
IHP["IntakeHistoryPrim"]:::runtime
HP --> TP
HP --> CLP
HP --> GSP
HP --> MRP
HP --> SP
HP --> IP
HP --> MP
HP --> SkP
HP --> MuP
HP --> IHP
APIB["APISchemaBase · USD"]:::substrate
PROV["Provenance · applied API"]:::substrate
APIB --> PROV
PROV -.->|"attaches to"| CLP
classDef substrate fill:#1a2332,stroke:#4a90a4,color:#e8eef2
classDef runtime fill:#d4af37,stroke:#8b7115,color:#1a2332
- Two abstract bases:
HarloPrim(root of every Harlo type) andHarloContainer(parent of structural composites). - Eight concrete container types:
BrainStageplus seven subsystem
containers (Association, Composition, Elenchus, Inquiry, Motor,
Skills, CognitiveProfile). - Ten concrete leaf types holding the actual cognitive-state
attributes. - One singleApply API schema (
Provenance, per D10) that attaches
origin metadata to host prims without cluttering the IsA tree.
Five enum types use lower-case allowedTokens per Constitution Cmd 11:SourceType, VerificationState, RetrievalPath, MotorGateStatus,ArcType. Cross-plugin: zero collisions with sister project Moneta'sMonetaMemory typeName (D3 verified).
Sync layer · per-prim policy
The sync layer at python/harlo/sync/ routes writes per the D4 policy
table:
flowchart LR
START["BrainStage<br/>write"]:::substrate
DECISION{"Prim type<br/>policy?"}:::substrate
WT["write_through<br/>MotorPrim · GateStatusPrim<br/>MerkleRootPrim · SessionPrim"]:::substrate
CP["checkpoint<br/>TracePrim · CompositionLayerPrim<br/>SkillPrim · intake/multipliers<br/>InquiryPrim"]:::substrate
INMEM["InjectionPrim<br/>D5 · session-scoped"]:::runtime
OUT_WT["immediate<br/>sync to disk"]:::substrate
OUT_CP["deferred sync<br/>at checkpoint"]:::substrate
OUT_INMEM["no persistence"]:::runtime
START --> DECISION
DECISION -->|"write-through"| WT
DECISION -->|"checkpoint"| CP
DECISION -.->|"in-memory-only"| INMEM
WT --> OUT_WT
CP --> OUT_CP
INMEM -.-> OUT_INMEM
classDef substrate fill:#1a2332,stroke:#4a90a4,color:#e8eef2
classDef runtime fill:#d4af37,stroke:#8b7115,color:#1a2332
- write_through — synchronous persistence on every mutation. Used
for consistency-critical prims (SessionPrim,GateStatusPrim,MerkleRootPrim) and the safety-criticalMotorPrim(D4 ruling). - checkpoint — deferred persistence; callers mark prim paths dirty
during the session and flush explicitly. Used for high-write-rate
prims to keep per-mutation persistence cost bounded. - In-memory only —
InjectionPrimis session-scoped per D5
(evicted from disk; runtime dataclass retained for/injectcommand
flows).
Containers inherit policy from their dominant child type. The migration
script (python/harlo/migrate_path_c.py) converts existing USD-Lite
text-format captures to real-USD format; read-tolerant on input,
idempotent on already-migrated files.
Tech Stack
- USD 26.03 — Cognitive state stored in real
.usdafiles. Time-sampled. Human-readable. Git-trackable. Sublayer composition via LIVRPS. - OpenExec — C++ libs built, Python bindings deferred (Pixar hasn't shipped them yet). Architecture is OpenExec-native; implementation catches up later.
- Hydra Delegates — The
Hdprefix is a naming convention, not an import. Pure Python. Any LLM implements the interface, registers, done. - XGBoost — MultiOutputRegressor predicting momentum, burnout, energy, burst from 111-feature sliding window. Trained on 10K synthetic trajectories (278K exchanges).
- Python 3.12 (USD) / 3.14 (project) — Dual venv. Real USD on 3.12, graceful mock fallback on 3.14.
- Rust — Hippocampus crate via PyO3. 1-bit SDR encoding, XOR popcount kNN, lazy decay. Sub-2ms recall.
- MCP — 8 tools over stdio. Works with Claude Desktop, Claude Code, any MCP client.
Architecture
System Layers
%%{init: {'theme': 'dark', 'themeVariables': {'primaryColor': '#1a1a2e', 'primaryTextColor': '#e0e0e0', 'primaryBorderColor': '#7c3aed', 'lineColor': '#7c3aed', 'secondaryColor': '#16213e', 'tertiaryColor': '#0f3460'}}}%%
graph TB
USER["You · Claude Desktop / Claude Code"]:::user
subgraph MCP["MCP Server · 8 Tools · stdio"]
direction LR
COACH["twin_coach"]:::tool
STORE["twin_store"]:::tool
RECALL["twin_recall"]:::tool
QPE["query_past_experience"]:::tool
PATTERNS["twin_patterns"]:::tool
SESSION["twin_session_status"]:::tool
RESOLVE["resolve_verifications"]:::tool
RECAL["trigger_recalibration"]:::tool
end
subgraph ENGINE["CognitiveEngine · Production Singleton"]
direction TB
DAG["MockCogExec · networkx DAG\nburst → energy → momentum\n→ burnout → allostasis\n+ injection_gain · context_budget · routing"]:::engine
DELEGATES["Hydra Delegates\nHdClaude · HdClaudeCode\ncapability-matched routing"]:::engine
PREDICT["XGBoost Predictor\n3-step window · 111 features\n→ momentum · burnout · energy · burst"]:::engine
end
subgraph STAGE["USD Stage · .usda on Disk"]
direction LR
ROOT["harlo.usda\nTime-sampled state\nCanonical prim hierarchy"]:::usd
CLAUDE_SUB["delegates/claude.usda\nInteractive opinions"]:::usd
CODE_SUB["delegates/claude_code.usda\nBatch opinions"]:::usd
end
subgraph MEMORY["Core Twin · Biologically-Architected Memory"]
direction TB
HOT["Hot Tier · FTS5\n< 0.2ms store"]:::memory
WARM["Warm Tier · SDR Hamming\nRust PyO3 · < 2ms recall"]:::memory
ELENCHUS["Elenchus · GVR\ntrace-excluded verify"]:::memory
HEBBIAN["Hebbian · dual-mask\nSDR evolution"]:::memory
COMPOSITION["Composition · Merkle\nLIVRPS resolution"]:::memory
end
BUFFER["Observation Buffer\nanchor 20% · organic 80%\n458 observations"]:::buffer
USER --> MCP
MCP --> ENGINE
ENGINE --> STAGE
ENGINE --> BUFFER
STAGE --> ENGINE
MCP --> MEMORY
MEMORY --> MCP
ENGINE -->|"enriched context"| USER
classDef user fill:#7c3aed,stroke:#a78bfa,color:#fff,font-weight:bold
classDef tool fill:#0f3460,stroke:#3b82f6,color:#93c5fd
classDef engine fill:#1e3a5f,stroke:#60a5fa,color:#bfdbfe,font-weight:bold
classDef usd fill:#1a4a3a,stroke:#22c55e,color:#bbf7d0,font-weight:bold,stroke-width:3px
classDef memory fill:#2e1a4a,stroke:#a78bfa,color:#ddd6fe
classDef buffer fill:#4a3a1a,stroke:#f59e0b,color:#fde68a
Exchange Loop
Every MCP tool call flows through this 7-step pipeline:
%%{init: {'theme': 'dark', 'themeVariables': {'primaryColor': '#1a1a2e', 'primaryTextColor': '#e0e0e0', 'primaryBorderColor': '#7c3aed', 'lineColor': '#7c3aed'}}}%%
graph LR
CALL["MCP Tool Call"]:::input
subgraph PIPELINE["CognitiveEngine · Per-Exchange Pipeline"]
direction LR
S1["1 · Author\nBuild observation\nfrom tool context"]:::step
S2["2 · Evaluate\nDAG: burst → energy\n→ momentum → burnout\n→ allostasis"]:::step
S3["3 · Route\ncompute_routing →\ncapability requirements"]:::step
S4["4 · Delegate\nSync → Execute\n→ CommitResources\nto sublayer"]:::step
S5["5 · Observe\nEmit to buffer\nanchor/organic split"]:::step
S6["6 · Predict\nXGBoost forecast\nauthor to /prediction"]:::step
S7["7 · Save\n.usda to disk\ngraceful on failure"]:::step
S1 --> S2 --> S3 --> S4 --> S5 --> S6 --> S7
end
RESPONSE["Enriched Response\ncognitive_context\ndelegate_id · expert\nprediction"]:::output
CALL --> PIPELINE --> RESPONSE
classDef input fill:#7c3aed,stroke:#a78bfa,color:#fff,font-weight:bold
classDef step fill:#1e3a5f,stroke:#60a5fa,color:#bfdbfe
classDef output fill:#22c55e,stroke:#4ade80,color:#fff,font-weight:bold
Cognitive State Machines
Five state machines evaluated via topologically-sorted DAG on every exchange:
%%{init: {'theme': 'dark', 'themeVariables': {'primaryColor': '#1a1a2e', 'primaryTextColor': '#e0e0e0', 'primaryBorderColor': '#7c3aed', 'lineColor': '#7c3aed'}}}%%
stateDiagram-v2
direction LR
state Momentum {
direction LR
[*] --> COLD_START
CRASHED --> COLD_START: always
COLD_START --> BUILDING: tasks >= threshold
BUILDING --> ROLLING: coherence + velocity
ROLLING --> PEAK: exchanges + burst
PEAK --> CRASHED: burnout >= ORANGE
}
state Burnout {
direction LR
[*] --> GREEN
GREEN --> YELLOW: frustration or duration
YELLOW --> ORANGE: sustained frustration
ORANGE --> RED: extreme frustration
note right of RED: ANY -> RED via exogenous override
}
state Energy {
direction LR
[*] --> MEDIUM
HIGH --> MEDIUM: natural decay
MEDIUM --> LOW: session length
LOW --> DEPLETED: continued work
note right of DEPLETED: Burst suspends decay\nDebt applies on exit
}
state Burst {
direction LR
[*] --> NONE_B
NONE_B --> DETECTED: velocity + coherence
DETECTED --> PROTECTED: sustained
PROTECTED --> WINDING: exchange threshold
WINDING --> EXIT_PREP: exit threshold
EXIT_PREP --> NONE_B: next exchange
}
Hydra Delegate Pattern
The DAG outputs what's needed. The registry selects who fulfills it. The DAG never names a specific LLM.
%%{init: {'theme': 'dark', 'themeVariables': {'primaryColor': '#1a1a2e', 'primaryTextColor': '#e0e0e0', 'primaryBorderColor': '#7c3aed', 'lineColor': '#7c3aed'}}}%%
graph TB
ROUTING["compute_routing\nOutputs: requirements\nNOT delegate names"]:::route
subgraph REQUIREMENTS["Capability Requirements"]
direction LR
REQ_TASKS["supported_tasks\nreasoning · coaching\ncode_generation"]:::req
REQ_LATENCY["latency_max\nrealtime · interactive\nbatch"]:::req
REQ_CODING["requires_coding\ntrue / false"]:::req
REQ_CTX["context_budget\nlight · medium · heavy"]:::req
end
subgraph SAFETY["Safety Overrides"]
direction LR
RED["RED burnout\n-> force restorer\nconsent ignored"]:::red
ORANGE["ORANGE + no consent\n-> force restorer"]:::orange
CONSENT["OOB Consent\nHMAC-signed\nTTL · revocable"]:::consent
end
subgraph REGISTRY["DelegateRegistry · Capability Match"]
direction TB
MATCH["Filter → Sort → Select\nprefer lower latency\nthen higher context"]:::registry
subgraph DELEGATES["Registered Delegates"]
direction LR
CLAUDE["HdClaude\nreasoning · coaching\nanalysis · exploration\ninteractive · 200K"]:::claude
CODE["HdClaudeCode\nimplementation · debugging\ncode_generation · testing\nbatch · 200K"]:::code
FUTURE["Your Delegate\nimplement interface\nregister · done"]:::future
end
end
subgraph SUBLAYERS["Per-Delegate .usda Sublayers"]
direction LR
SUB_C["claude.usda\nSTRONGEST"]:::sub
SUB_CC["claude_code.usda"]:::sub
end
ROUTING --> REQUIREMENTS
ROUTING --> SAFETY
REQUIREMENTS --> REGISTRY
SAFETY --> REGISTRY
MATCH --> DELEGATES
DELEGATES -->|"Sync/Execute/Commit"| SUBLAYERS
classDef route fill:#0f3460,stroke:#3b82f6,color:#93c5fd,font-weight:bold
classDef req fill:#1e3a5f,stroke:#60a5fa,color:#bfdbfe
classDef red fill:#7f1d1d,stroke:#ef4444,color:#fff,font-weight:bold
classDef orange fill:#5c1a1a,stroke:#ef4444,color:#fca5a5
classDef consent fill:#4a3a1a,stroke:#f59e0b,color:#fde68a
classDef registry fill:#2e1a4a,stroke:#a78bfa,color:#ddd6fe
classDef claude fill:#0f3460,stroke:#3b82f6,color:#93c5fd,font-weight:bold
classDef code fill:#1a3a4a,stroke:#06b6d4,color:#a5f3fc,font-weight:bold
classDef future fill:#1a1a2e,stroke:#6b7280,color:#9ca3af,stroke-dasharray: 5 5
classDef sub fill:#1a4a3a,stroke:#22c55e,color:#bbf7d0,stroke-width:2px
Prediction Pipeline
From synthetic autoresearch to live organic observations:
%%{init: {'theme': 'dark', 'themeVariables': {'primaryColor': '#1a1a2e', 'primaryTextColor': '#e0e0e0', 'primaryBorderColor': '#7c3aed', 'lineColor': '#7c3aed'}}}%%
graph TB
subgraph SYNTHETIC["Autoresearch · Sprint 1"]
direction TB
GEN["Trajectory Generator\n7 profiles · Markov Biasing\nnormal 40% · deep_work 15%\nstruggling 15% · recovery 10%\ninjection 10% · crisis 5% · mobile 5%"]:::gen
TRAJ["10,000 sessions\n278,577 exchanges\n0 invariant violations"]:::gen
GEN --> TRAJ
end
subgraph BUFFER["Observation Buffer · SQLite"]
direction LR
ANCHOR["Anchor Partition\n20% · locked synthetic\nbaseline coverage"]:::anchor
ORGANIC["Organic Partition\n80% · surprise-weighted\nlive session data"]:::organic
end
subgraph TRAINING["XGBoost Training"]
direction TB
WINDOW["3-step sliding window\n111 features per sample"]:::train
ENCODE["Ordinal: momentum, burnout, energy\nOne-Hot: action_type, injection_profile\nDrop: exchange_index, session_id"]:::train
MODEL["MultiOutputRegressor\nXGBRegressor(reg:squarederror)\nRound + clamp to valid range"]:::train
WINDOW --> ENCODE --> MODEL
end
subgraph LIVE["Live Prediction · Per Exchange"]
direction TB
OBS_WIN["Last 3 observations\nfrom current session"]:::live
PRED["Predict: momentum\nburnout · energy · burst"]:::live
AUTHOR["Author to\n/prediction/forecast\non USD stage"]:::live
OBS_WIN --> PRED --> AUTHOR
end
TRAJ --> ANCHOR
TRAJ --> TRAINING
ORGANIC -->|"retrain"| TRAINING
MODEL --> LIVE
classDef gen fill:#2e1a4a,stroke:#a78bfa,color:#ddd6fe
classDef anchor fill:#1a4a3a,stroke:#22c55e,color:#bbf7d0,stroke-width:3px
classDef organic fill:#4a3a1a,stroke:#f59e0b,color:#fde68a
classDef train fill:#0f3460,stroke:#3b82f6,color:#93c5fd
classDef live fill:#1a4a3a,stroke:#22c55e,color:#bbf7d0,font-weight:bold
Graceful Degradation
Every component fails independently. The MCP server never crashes.
| Component Failure | Fallback | Logged |
|---|---|---|
| USD import fails | MockUsdStage (dict) | WARNING |
| Model file missing | Prediction disabled | WARNING |
| DB locked | Memory queue (max 100) | WARNING |
| DAG evaluation fails | Default computed values | ERROR |
| Delegate cycle fails | Empty context returned | ERROR |
| Stage save fails | Queued for next exchange | WARNING |
| Engine disabled | Pre-Sprint 3 MCP behavior | -- |
Project Structure
src/ Cognitive State Machine + Production Engine
├── cognitive_engine.py Production singleton: DAG → route → delegate → observe → predict
├── cognitive_stage.py Real pxr.Usd.Stage wrapper (.usda on disk)
├── mock_usd_stage.py Dict-based fallback stage
├── stage_factory.py Backend toggle: USE_REAL_USD
├── mock_cogexec.py networkx DAG evaluator (topological sort)
├── schemas.py Pydantic IntEnum ordinals + CognitiveObservation
├── delegate_base.py HdCognitiveDelegate ABC (Hydra pattern)
├── delegate_registry.py Capability-matching selection
├── delegate_claude.py Interactive reasoning delegate
├── delegate_claude_code.py Implementation/code delegate
├── consent.py OOB consent tokens (HMAC, TTL, revocable)
├── engine_config.py Kill switches + paths
├── usd_bootstrap.py USD 26.03 sys.path setup
├── computations/ Pure functions (no internal counters)
│ ├── compute_momentum.py CRASHED→COLD_START→BUILDING→ROLLING→PEAK
│ ├── compute_burnout.py GREEN→YELLOW→ORANGE→RED + exogenous override
│ ├── compute_energy.py Adrenaline masking, RED degradation, exercise recovery
│ ├── compute_injection_gain.py Anchor = 1.0 ALWAYS (structural immunity)
│ ├── compute_context_budget.py Hysteresis: promote >4.2x, demote <3.8x
│ ├── compute_burst.py 5-phase hyperfocus lifecycle
│ ├── compute_allostasis.py 6-weight composite + trend detection
│ └── compute_routing.py Capability requirements (NOT delegate names)
├── trajectory_generator.py 10K sessions via Profile-Driven Markov Biasing
├── validator.py 26 invariants (INV-01 to INV-26)
├── train_predictor.py XGBoost MultiOutputRegressor
├── predict.py 3-step window inference
├── bridge.py Exchange loop coordinator (simulation)
└── observation_buffer.py SQLite priority queue (anchor 20% / organic 80%)
python/harlo/ Core Twin: MCP server + biologically-architected memory
├── mcp_server.py 8 MCP tools over stdio
├── migrate_path_c.py Path C migration script (USD-Lite v1 → real USD)
├── brainstem/ Lossless translation (14 adapter files)
├── elenchus/ Verification engine (GVR, trace-excluded)
├── elenchus_v8/ Deferred verification (Actor-side)
├── composition/ Merkle stages, LIVRPS resolution
├── hebbian/ Dual-mask SDR evolution, reconstruction
├── hot_store/ L1 Hot Tier (FTS5, zero-encoding)
├── modulation/ Allostatic load, gain, burst detection
├── motor/ Basal Ganglia gate (inhibit-default)
├── inquiry/ DMN (apophenia guard, sincerity gate)
├── coach/ System prompt projection
├── encoder/ ONNX BGE + LSH → 2048-bit SDR
├── trust/ Continuous [0,1] trust ledger
├── intake/ Neuropsych-informed cognitive profile
├── skills/ Incremental competence tracking
├── session/ Session lifecycle management
├── sync/ Path C sync layer (write-side dispatch)
│ ├── policy.py Per-prim policy table (D4)
│ ├── write_through.py Synchronous persist strategy
│ └── checkpoint.py Deferred-flush strategy
└── usd_lite/ 21 prim dataclasses, .usda serialization
└── persistence/ Path C real-USD writer/reader
├── writer.py BrainStage → real-USD .usda via pxr
└── reader.py real-USD .usda → BrainStage
schema/ Path C codeless schema artifacts
├── HarloSchema.usda 21 prim types, IsA hierarchy, allowedTokens
├── plugInfo.json harlo namespace plugin registration
└── generatedSchema.usda Compiled form (hand-authored)
crates/hippocampus/ Rust hot path (SDR, XOR search, lazy decay, apoptosis)
data/stages/ Real .usda files (your cognitive state)
├── brain.usda Path C root stage (real-USD via pxr)
├── harlo.usda Sprint 4 root stage (vendored USD path)
└── delegates/ Per-delegate sublayers
harness/path_c/ Path C surgery harness (Mile 1 → Mile 3)
├── 01_KICKOFF.md, 02_CONSTITUTION.md, 03_HANDOFF.md, 04_DEEP_THINK_BRIEF.md
├── 05_DECISIONS.md (D1-D5), 06_DECISIONS_PHASE_1.md (D6-D14),
│ 07_DECISIONS_PHASE_4.md (D15-D19)
├── blocker_decisions.md Codec-blocker resolution log
├── memory_hypothesis.md, substrate_pin.md, baseline_resolution.md
├── tracking_issues.md TI-001 (closed-on-arrival)
└── baseline_tests.txt, baseline_latency.json, phase_3_latency.json,
phase_6_latency.json
Quick Start
git clone <repo-url> && cd harlo
python3.12 -m venv .venv312 && source .venv312/bin/activate
# Core install (no real-USD persistence)
pip install -e .
# Path C real-USD persistence (optional, requires Python 3.12)
pip install -e .[substrate] # Pulls usd-core 26.5
# Test-suite dev dependencies (sentence_transformers, anthropic, pytest)
pip install -e .[dev]
# Health check
python scripts/health_check.py
# First session (10-exchange simulation)
python scripts/first_session.py
# Migrate an existing USD-Lite stage to real-USD format (Path C)
python -m harlo.migrate_path_c data/stages/your_stage.usda --output data/stages/brain.usda
On Windows, if pip install -e .[substrate] fails on a .pyd file lock
during the maturin rebuild (D13 documented quirk), close any process
holding python/harlo/hippocampus.cp312-win_amd64.pyd open, or install
the substrate dep directly:
pip install "usd-core>=24.05" # Same end state; bypasses editable rebuild
Environment variables:
ENGINE_ENABLED=1 # Master kill switch
USE_REAL_USD=1 # Real pxr.Usd.Stage (requires Python 3.12)
OBSERVATION_LOGGING=1 # Emit observations per exchange
PREDICTION_ENABLED=1 # XGBoost predictions
The 33 Rules
The architecture is constrained by 33 inviolable rules covering biological fidelity (0W idle, 1-bit SDRs, lazy decay), verification integrity (trace exclusion, max 3 GVR cycles, verified-only consolidation), inquiry safeguards (apophenia guard, sincerity gate, rupture & repair), motor safety (inhibition default, one action at a time, RED kills everything), and Hebbian constraints (Merkle isolation, dual masks not XOR, homeostatic plasticity). These aren't guidelines — they're structural constraints enforced by 1,172 tests (Path C v3.4.0; +39 since Mile 1 baseline). See CLAUDE.md for the full specification and harness/path_c/ for the Path C surgery harness (D1–D19 decisions log, phase gate audits).
Philosophy
Your memory, your device. Harlo stores all state locally as composable USD layers. Cloud models provide reasoning; your machine provides memory and safety. Nothing leaves your device without explicit action.
License
Licensed under the Apache License 2.0. Patent Pending.
Aspects of this architecture are the subject of pending US patent applications.
The Apache 2.0 license includes a patent grant for users of this software.
See PATENTS.md for details.
Yorumlar (0)
Yorum birakmak icin giris yap.
Yorum birakSonuc bulunamadi