epistemic-memory
Health Gecti
- License — License: MIT
- Description — Repository has a description
- Active repo — Last push 0 days ago
- Community trust — 14 GitHub stars
Code Gecti
- Code scan — Scanned 2 files during light audit, no dangerous patterns found
Permissions Gecti
- Permissions — No dangerous permissions requested
This project provides a structured protocol and set of templates for AI memory. It allows AI tools to track user beliefs and traits using a system that scores confidence, tracks contradictions, and monitors how beliefs decay over time.
Security Assessment
Overall risk: Low. The tool does not request any dangerous permissions, execute background shell commands, or make external network requests. It is a protocol rather than a software library, meaning it consists of templates and standards rather than executable code. A light code scan of its 2 files found no dangerous patterns, no hardcoded secrets, and no mechanisms for accessing sensitive data without explicit user consent.
Quality Assessment
The project is actively maintained, with its most recent push happening today. It is properly licensed under the standard MIT license and includes a clear, detailed description of its goals. While community trust is currently modest at 14 GitHub stars, this is expected for a niche conceptual tool. The codebase is minimal, transparent, and easy to audit manually.
Verdict
Safe to use.
What should a machine remember about a person? A protocol for AI memory that models who you are — with confidence, decay, and contradiction tracking.
Epistemic Memory
A protocol for AI memory that models who you are — not just what you use.
An AI that remembers your name and your language preferences is a tool. One that understands how you think — your cognitive style, what motivates you, where you contradict yourself — is a thinking partner. It pushes back when you need it, frames things the way you'll actually hear them, and notices when something's off before you do. The gap between those two experiences is enormous, and it comes down to what the memory system is willing to model.
Why This Exists
Every AI with persistent memory is already building a psychographic profile of you — personality traits, cognitive style, motivations, working patterns. It's just doing it badly. No confidence tracking, no decay, no contradiction tolerance, and no way for you to see or challenge any of it. "You're a deep thinker" carries the same weight as "you use dark mode." A preference from six months ago is treated as current gospel. Contradictions get smoothed over into a flattering caricature that drifts from who you actually are.
The deeper your AI understands how you think, the more useful the collaboration becomes. But that deeper understanding needs to be honest — held as hypotheses with varying confidence, not settled facts. It needs to weaken when evidence goes stale. And it needs to be yours to see, challenge, and correct.
Epistemic Memory is a protocol for doing this right: treating every stored belief as a weighted hypothesis with explicit confidence, decay, and contradiction tracking. The difference between this and adtech-style profiling is consent — every belief is visible, and the subject holds the keys.
What This Is
- A protocol, not a library. No code, no dependencies, no build step.
- Templates for storing user beliefs with per-belief metadata (confidence, permanence, dates).
- A skill (
/mirror) for Claude Code that surfaces what the AI thinks it knows — as a prose portrait, an audit dashboard, or a structured interview. - Drop-in adoption. Copy the files into your Claude Code project and start using it.
The Core Idea
Every belief about the user carries metadata:
## Builder Philosophy
`conf:0.75 | first:2026-03-10 | confirmed:2026-03-25 | challenged:— | perm:durable`
Builds for herself first, monetizes if it generalizes.
| Field | What it means |
|---|---|
conf |
Confidence (0.0–1.0). Must be earned through evidence, not assigned by how insightful it sounds. |
first |
When this belief was first recorded. |
confirmed |
Last session where behavior was consistent with this belief. |
challenged |
Last session where something contradicted it. — if never. |
perm |
How stable this trait is: stable (decade), durable (year), situational (month), unknown. |
Confidence and permanence are independent. You can be certain about something that will change (high conf, situational), or uncertain about something deep (low conf, stable).
10 Components
- Per-belief metadata — no belief exists as bare text
- Confidence that must be earned — observable behavior outranks interpretive narrative
- Permanence classes — separates "how sure" from "how stable"
- Dormancy decay — confidence attenuates when the system goes unused
- Tensions log — contradictions stay on record instead of being resolved away
- Self-report vs. behavior rule — when stated and observed diverge, hold both
- Silent consistency — behavioral patterns count as evidence without explicit narration (enforced by the Observe skill)
- Session counter — mechanical trigger for periodic reviews (every 10 sessions)
- Maintenance cost bound — only update beliefs relevant to the current session
- Evolving epistemology — the framework itself improves over time
Full rationale, adversarial review findings, and limitations: docs/design.md
Quick Start
- Copy
memory/into your Claude Code project memory directory - Copy
.claude/skills/into your project's.claude/skills/(includes Mirror and Observe) - Copy
.claude/hooks/session-counter.shinto your project's.claude/hooks/ - Register the hook in your
.claude/settings.json:{ "hooks": { "SessionStart": [ { "type": "command", "command": "bash .claude/hooks/session-counter.sh" } ] } } - Add the snippet from
CLAUDE.md.exampleto your project'sCLAUDE.md - Start a conversation — beliefs accumulate naturally as you work together
- Run
/mirrorto see what the AI thinks it knows about you
The Mirror Skill
Four modes for inspecting and refining the model:
| Command | What it does |
|---|---|
/mirror |
Prose portrait — "here's who I think you are" |
/mirror audit |
Epistemological dashboard — confidence distributions, decay, drift |
/mirror gut-check |
Quick validation of highest and lowest confidence beliefs |
/mirror interview |
Structured conversation that probes blind spots and tests stale beliefs |
The Observe Skill
Silent consistency (component 7) needs enforcement — the system can't rely on the LLM remembering to notice behavioral patterns. The Observe skill runs automatically at conversation end alongside your harvest process. It watches for:
- Judgment calls where reasonable people would disagree
- Pushback on suggestions, revealing priorities or values
- Behavioral contradictions with the existing profile
- Process signals — what the user lingers on, skips, or returns to
Each observation is logged as dated evidence with a strength rating (single, confirming, complicating, contradicting). Observations accumulate as raw evidence; they get synthesized into the portrait during /mirror or periodic review. The separation is intentional — observations are cheap and frequent, portrait updates are expensive and should require pattern confirmation across multiple sessions.
Full specification: .claude/skills/observe/SKILL.md
Session Counter Hook
The session counter (component 8) gives periodic reviews a mechanical trigger. A bash hook increments a counter at every session start and alerts you when a review is due.
Setup requires registering the hook in .claude/settings.json — see Quick Start above. The counter file lives at memory/user/session-counter.json and tracks session count, last session date, and the next review threshold.
Why Not Just Use MEMORY.md?
Claude Code's built-in memory stores facts: "prefers dark mode," "uses Python," "name is Alex." Epistemic Memory stores hypotheses: beliefs held at different weights, with mechanisms to weaken stale ones, flag contradictions, and periodically ask "am I still right about this?" The difference is between a filing cabinet and a living model.
FAQ
Is this only for Claude Code?
The protocol is general — any AI system with persistent memory can adopt the belief format and update rules. The Mirror skill is Claude Code specific, but the memory structure works anywhere.
How many beliefs before this is useful?
5–10 is enough. The value shows up as soon as you have beliefs at different confidence levels and the system starts distinguishing what it knows well from what it's guessing.
What if I don't want the Mirror skill?
The protocol stands alone. Add metadata to your existing memory files and follow the update rules. The skill is optional tooling on top.
How is this different from RAG or vector memory?
RAG retrieves relevant context. This is about the quality of what's stored — whether the system knows what it knows, what it's guessing, and what's gone stale. They're complementary, not competing.
Built From Use
This is a framework and starter kit, extracted from a personal system used in daily Claude Code sessions. The repo ships clean — session counter at zero, example profile fictional — so you build your own from scratch. It's not a theoretical proposal; it was stress-tested through adversarial review before release. See docs/design.md for the full story, including what broke in v1 and how it was fixed.
License
Yorumlar (0)
Yorum birakmak icin giris yap.
Yorum birakSonuc bulunamadi