m_flow
Health Pass
- License — License: Apache-2.0
- Description — Repository has a description
- Active repo — Last push 0 days ago
- Community trust — 224 GitHub stars
Code Pass
- Code scan — Scanned 12 files during light audit, no dangerous patterns found
Permissions Pass
- Permissions — No dangerous permissions requested
This tool is a cognitive memory engine for AI models that focuses on reasoning and association. It organizes unstructured text into a four-level layered hierarchy to retrieve connected concepts rather than just matching surface-level similarities.
Security Assessment
Risk: Low. The automated code scan evaluated 12 files and found no dangerous patterns. It does not request dangerous permissions, execute shell commands, or contain hardcoded secrets. Because it acts as a memory engine, it will naturally process the data and queries fed into it, but the underlying infrastructure does not expose your system to unauthorized access or malicious execution.
Quality Assessment
The project is in excellent shape and exhibits strong signs of active maintenance. It received its last code push today, meaning the repository is highly active. It also benefits from strong community trust, currently boasting 224 GitHub stars. Furthermore, it is well-documented and fully compliant, utilizing the permissive Apache-2.0 license which makes it straightforward for legal integration into almost any project. The testing standards also appear rigorous, with a passing badge indicating 963 successful tests.
Verdict
Safe to use.
A memory engine that focuses on reasoning and association.
M-flow
A memory engine that focuses on reasoning and association.
M-flow 是一款重构检索架构、侧重推理与联想的记忆引擎。
m-flow.ai ·
flowelement.ai ·
Quick Start ·
Architecture ·
Examples ·
OpenClaw Skill ·
Contact
What is M-flow?
M-flow is not a vector database or a RAG framework. It is a cognitive memory engine — a system that doesn't just store text, but understands it, distilling incoming knowledge into structured layers and weaving them into the right place of a persistent architecture.
Where traditional retrieval matches queries by surface similarity, M-flow reasons through relationships — tracing connections, weighing context, and reaching answers the way cognition does.
Most AI memory systems fall into two categories:
- Context Window — The model re-reads everything from scratch. Exhaustive, linear, forgetful.
- Similarity Search — External memory exists, but retrieval is surface-level — matching by vector shape, blind to structure. Approximate, lossy.
M-flow introduces a third paradigm: a Cognitive Engine where memory becomes a mind. A query at any granularity finds a precise anchor, then expands outward — surfacing related context and associative content across different levels of detail.
How It Works
M-flow organizes knowledge into a four-level Cone Graph — a layered hierarchy from abstract summaries to atomic facts:
| Level | What it captures | Query example |
|---|---|---|
| Episode | A bounded semantic focus — an incident, decision process, or workflow | "What happened with the tech stack decision?" |
| Facet | One dimension of that Episode — a topical cross-section | "What were the performance targets?" |
| FacetPoint | An atomic assertion or fact derived from a Facet | "Was the P99 target under 500ms?" |
| Entity | A named thing — person, tool, metric — linked across all Episodes | "Tell me about GPT-4o" → surfaces all related contexts |
Retrieval is graph-routed: the system casts a wide net across all levels, projects hits into the knowledge graph, propagates cost along every possible path, and scores each Episode by its tightest chain of evidence. One strong path is enough — the way a single association triggers an entire memory.
For the full technical deep-dive, see Retrieval Architecture | 检索架构详解(中文)
Benchmarks
All benchmarks use gpt-5-mini (answer) + gpt-4o-mini (judge) + top-k=10.
LoCoMo-10
| System | LLM-Judge |
|---|---|
| M-flow | 81.8% |
| Cognee Cloud | 79.4% |
| Zep Cloud | 73.4% |
| Supermemory Cloud | 64.4% |
LongMemEval
| System | LLM-Judge |
|---|---|
| M-flow | 89% |
| Supermemory Cloud | 74% |
| Mem0 | 71% |
| Zep Cloud | 61% |
| Cognee | 57% |
Per-category breakdowns, reproduction scripts, raw data, and methodology for all systems: mflow-benchmarks
Features
| Episodic + Procedural memory | Hierarchical recall for facts and step-by-step knowledge |
| 5 retrieval modes | Episodic, Procedural, Triplet Completion, Lexical, Cypher |
| 50+ file formats | PDFs, DOCX, HTML, Markdown, images, audio, and more |
| Multi-DB support | LanceDB, Neo4j, PostgreSQL/pgvector, ChromaDB, KùzuDB, Pinecone |
| LLM-agnostic | OpenAI, Anthropic, Mistral, Groq, Ollama, LLaMA-Index, LangChain |
| Precise summarization | Preserves all factual details (dates, numbers, names) at the cost of lower compression — RAG context will be longer but more accurate |
| MCP server | Expose memory as Model Context Protocol tools for any IDE |
| CLI & Web UI | Interactive console, knowledge graph visualization, config wizard |
Retrieval modes: Episodic is the primary retrieval mode — it uses graph-routed Bundle Search for best accuracy and is used in all benchmarks. Triplet Completion is a simpler vector-based mode suited for customization and secondary development. See Retrieval Architecture for details.
Quick Start
One-Command Setup (Docker)
git clone https://github.com/FlowElement-ai/m_flow.git && cd m_flow
./quickstart.sh
The script checks your environment, configures API keys interactively, and starts the full stack (backend + frontend). On Windows, use .\quickstart.ps1.
Install via pip
pip install mflow-ai # or: uv pip install mflow-ai
export LLM_API_KEY="sk-..."
Install from Source
git clone https://github.com/FlowElement-ai/m_flow.git && cd m_flow
pip install -e . # editable install for development
Run
import asyncio
import m_flow
async def main():
await m_flow.add("M-flow builds persistent memory for AI agents.")
await m_flow.memorize()
results = await m_flow.search("How does M-flow work?")
for r in results:
print(r)
asyncio.run(main())
CLI
mflow add "M-flow builds persistent memory for AI agents."
mflow memorize
mflow search "How does M-flow work?"
mflow -ui # Launch the local web console
Architecture Overview
┌───────────────┐ ┌───────────────┐ ┌───────────────┐
│ Data Input │────▶│ Extract │────▶│ Memorize │
│ (50+ formats)│ │ (chunking, │ │ (KG build, │
│ │ │ parsing) │ │ embeddings) │
└───────────────┘ └───────────────┘ └───────┬───────┘
│
┌───────────────┐ ┌───────▼───────┐
│ Search │◀────│ Load │
│ (graph-routed│ │ (graph + │
│ bundle search│ │ vector DB) │
└───────────────┘ └───────────────┘
Project Layout
m_flow/ Core Python library & API
├── api/ FastAPI routers (add, memorize, search, …)
├── cli/ Command-line interface (`mflow`)
├── adapters/ DB adapters (graph, vector, cache)
├── core/ Domain models (Episode, Facet, FacetPoint, …)
├── memory/ Memory processing (episodic, procedural)
├── retrieval/ Search & retrieval algorithms
├── pipeline/ Composable pipeline tasks & orchestration
├── auth/ Authentication & multi-tenancy
├── shared/ Logging, settings, cross-cutting utilities
└── tests/ Unit & integration tests
m_flow-frontend/ Next.js web console
m_flow-mcp/ Model Context Protocol server
mflow_workers/ Distributed execution helpers (Modal, workers)
examples/ Runnable example scripts
docs/ Architecture & design documents
Development
git clone https://github.com/FlowElement-ai/m_flow.git && cd m_flow
uv sync --dev --all-extras --reinstall
# Test
PYTHONPATH=. uv run pytest m_flow/tests/unit/ -v
# Lint
uv run ruff check . && uv run ruff format .
See CONTRIBUTING.md for the full contributor guide.
Deployment
Docker
docker compose up # Backend only
docker compose --profile ui up # Backend + frontend
docker compose --profile neo4j up # Backend + Neo4j
MCP Server
cd m_flow-mcp
uv sync --dev --all-extras
uv run python src/server.py --transport sse
Testing
PYTHONPATH=. pytest m_flow/tests/unit/ -v # ~963 test cases
PYTHONPATH=. pytest m_flow/tests/integration/ -v # Needs .env with API keys
Contributing
We welcome contributions! Please see CONTRIBUTING.md for guidelines, and our Code of Conduct for community standards.
License
M-flow is licensed under the Apache License 2.0.
Copyright 2026 Junting Hua
Licensed under the Apache License, Version 2.0.
You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0
Reviews (0)
Sign in to leave a review.
Leave a reviewNo results found