Four days after publishing our original competitive analysis, the landscape has shifted significantly. Hindsight emerged as a credible SOTA benchmark contender, Cognee crossed 15k stars with v1.0, and LangMem became the default in LangGraph Platform. This post is the updated analysis — see the original April 10 analysis for where we started.
9.2k GitHub stars · Python + TypeScript + Rust · From Vectorize
Hindsight is the most credible new entrant in the agentic memory space. Its claim to fame is state-of-the-art performance on LongMemEval with independently verified scores from Virginia Tech and The Washington Post — a level of benchmark credibility no other system has achieved.
Architecture:
hindsight-all): Python-only, no external DB, LLM still requiredMemory model:
retain(): LLM extracts entities, relationships, time series from raw inputrecall(): merges 4 retrieval strategies (semantic, keyword, graph, temporal) via RRF + cross-encoder rerankingreflect(): generates new observations/insights from existing memories using LLM — Mental ModelsWeaknesses:
15.3k GitHub stars · Python · Apache 2.0 · v1.0.0 shipped April 11, 2026
Cognee crossed 15k stars and released v1.0 with three notable additions: a
cognify-mcp package for MCP server integration, Cognee Cloud managed service,
and native Hermes Agent integration as memory provider. The most full-featured knowledge
engine combining vector search, graph databases, and cognitive science approaches.
Still requires external databases and an LLM for the cognify pipeline.
1.4k GitHub stars · Python · MIT · From LangChain
LangMem is LangChain's memory primitives library, integrated natively into LangGraph
Platform. This gives it massive distribution: every LangGraph Platform deployment defaults
to LangMem for agent memory. Functional memory primitives
(create_manage_memory_tool(), create_search_memory_tool())
are storage-backend agnostic but default to Postgres in production.
Weaknesses: LLM required, no graph traversal or temporal facts, no cryptographic integrity, vec-only retrieval.
| Feature | MentisDB | Hindsight | Cognee | LangMem | Mem0 | Graphiti |
|---|---|---|---|---|---|---|
| Language | Rust | Python | Python | Python | Python | Python |
| Storage | Embedded (sled) | External (PG) | External | External | External DB | External DB |
| LLM Required | No (opt-in) | Yes | Yes | Yes | Yes | Yes |
| Local-First | Yes | No | No | No | Partial | No |
| Crypto Integrity | Hash chain | No | No | No | No | No |
| Hybrid Retrieval | BM25+vec+graph | 4-signal RRF | vec+graph | vec only | vec+keyword | sem+kw+graph |
| MCP Support | Built-in | No | MCP client | No | No | Yes |
| Agent Registry | Yes | No | No | No | No | No |
| Federated Search | Cross-chain | No | No | No | No | No |
| Skills/Extensions | Skill registry | No | No | No | No | No |
| Webhooks | Yes | No | No | No | No | No |
| Temporal Facts | valid_at/invalid_at | Via metadata | No | No | Updates | valid_at |
| Memory Dedup | Jaccard threshold | No | Merge | No | Yes | Merge |
| Custom Ontology | entity_type registry | Via metadata | Schema | No | No | Pydantic |
| Memory Branching | BranchesFrom | No | No | No | No | No |
| Benchmark R@10 | 74.0% (self) | SOTA (indep. verified) | N/A | N/A | N/A | N/A |
In 11 releases, we shipped everything on the original roadmap and several unplanned additions:
| Feature | Version | Status |
|---|---|---|
| Temporal Facts | 0.8.2 | Shipped |
| Memory Dedup | 0.8.2 | Shipped (Jaccard) |
| Multi-Level Scopes | 0.8.2 | Shipped (tag-based) |
| CLI Tool | 0.8.2 | Shipped |
| RRF Reranking | 0.8.6 | Shipped |
| Memory Branching | 0.8.6 | Shipped (BranchesFrom) |
| Per-Field BM25 Cutoffs | 0.8.6 | Shipped |
| Custom Ontology | 0.8.7 | Shipped (entity_type + registry) |
| Episode Provenance | 0.8.8 | Shipped (source_episode field) |
| LLM Reranking | 0.8.8 | Shipped (opt-in) |
| Federated Cross-Chain Search | 0.9.1 | Shipped |
| Webhooks | 0.9.1 | Shipped |
| Opt-in LLM Extraction | 0.9.1 | Shipped |
| Python Client | 0.9.1 | Shipped (pymentisdb on PyPI) |
| Wizard Brew-First Setup | 0.9.1 | Shipped |
Hindsight's scores are independently verified by Virginia Tech. Ours are self-reported. Fix: Partner with an academic group to independently verify our scores.
LangMem is the default in LangGraph Platform — massive structural distribution advantage.
Fix: Build langchain-mentisdb pip package with
BaseStore implementation.
The 19pp gap is the biggest retrieval quality gap. Miss analysis shows near-zero vector scores on misses — the semantic layer isn't contributing in multi-hop cases. Fix: Entity coreference, deeper graph traversal, query expansion.
Mem0, Cognee, Hindsight, and Fast.io all offer hosted versions. Fix: MentisDB Cloud.
agentmemory uses Ebbinghaus decay curves; Hindsight uses reflect().
MentisDB has dedup and temporal validity but no automatic memory evolution.
Fix: Implement automatic memory consolidation tiers.
MentisDB is the only local-first, zero-dependency, cryptographically-integrity-verified semantic memory with built-in hybrid retrieval — in Rust. This combination is still unique.
What we've added since April 10 that no competitor has:
The competitive threats are real: Hindsight's independently verified benchmarks and managed service are credible; LangMem's LangGraph Platform distribution is a structural advantage. But MentisDB's architectural properties — Rust performance, cryptographic integrity, embedded storage — matter more in enterprise, audit-critical, and air-gapped deployments.
The next battle is ecosystem and distribution, not features. Native LangChain store, academic benchmark verification, and MentisDB Cloud are the three moves that would make 1.0 a genuinely competitive release.
See the original April 10 analysis for where we started, and the 0.9.1 announcement for the full benchmark results and feature breakdown.
MentisDB is an open-source durable memory layer for AI agents. It stores memories in an append-only hash-chained log, retrieves them with hybrid BM25+semantic+graph search, and runs entirely locally with no cloud dependencies. GitHub · Docs · Website