Open-source AI memory framework

AI memory that forgets intelligently

The first memory framework built on cognitive science. Runs 100% local at $0 or with cloud APIs. Weibull forgetting, triple-path retrieval, 10-stage pipeline.

Memory should work like memory

The Problem

Other memory solutions store everything forever. Your agent drowns in noise. Context windows fill with stale facts. Retrieval degrades as data grows. The more your agent remembers, the worse it performs.

The Mnemo Way

Mnemo models human memory: important memories consolidate, trivial ones fade, frequently recalled knowledge strengthens. Built on decades of cognitive science research, not naive vector search.

10-Stage Memory Pipeline

From raw conversation to durable, retrievable memory in milliseconds.

Episodic Semantic Preference Procedural Relationship Reflection 6 Channels Classify Stage 1 Dedup & Contradict Weibull Decay Init store.ts Embed + Write LanceDB Vectors Graphiti Knowledge Graph Retrieval (Read) Vector Search BM25 Full-text Graph Traverse RRF + Rerank

Built different, by design

Every feature grounded in cognitive science and real-world agent workloads.

Weibull Decay

Stretched-exponential forgetting with tier-specific beta parameters. Memories fade naturally unless reinforced through recall.

Triple-Path Retrieval

Vector similarity, BM25 full-text, and knowledge graph traversal fused with Reciprocal Rank Fusion for robust recall.

Contradiction Detection

Three-layer LLM detection pipeline. When facts conflict, old versions auto-expire and new truths consolidate.

Scope Isolation

Per-agent memory with configurable access rules. Each agent operates in its own namespace with controlled sharing.

Cross-encoder Rerank

Voyage rerank-2 precision ranking ensures the most relevant memories surface first, every time.

Cognitive Science

Spaced repetition, emotional salience scoring, and spreading activation. Memory that learns how to remember.

Choose your tier

Core is free (MIT). Pro unlocks production features. API costs are separate — you bring your own keys.

Local

Core Local

$0/mo API · 100% Offline

  • Ollama embedding (nomic-embed-text)
  • Ollama LLM extraction (qwen3:8b)
  • Ollama cross-encoder rerank (bge-reranker)
  • Weibull decay + tier lifecycle
  • Triple-path retrieval (Vec+BM25+Graph)
  • Contradiction detection
Zero API cost. Beats Mem0 ($99) on retrieval quality.
Hybrid

Core Hybrid

~$20/mo API · MIT License

  • Voyage embedding (voyage-3-large)
  • GPT-4.1-mini smart extraction
  • Voyage rerank-2 cross-encoder
  • Everything in Local, better models
  • Best quality-to-cost ratio
  • Recommended for most users
Cloud-grade models, self-hosted infrastructure.
Cloud

Core Cloud

~$45/mo API · MIT License

  • Voyage embedding (voyage-3-large)
  • GPT-4.1 full extraction
  • Voyage rerank-2 cross-encoder
  • Maximum extraction quality
  • Best for high-volume production
  • Near Pro-level retrieval
Top-tier models. Score 8.8/10 in benchmarks.
Pro

Mnemo Pro

From $69/mo + API · Commercial

  • Everything in Core Cloud
  • WAL crash recovery
  • Session reflection + self-improvement
  • Memory tools + MCP server
  • Observability + access tracking
  • Machine-bound license + priority support
Full stack. Score 9.2/10 in benchmarks.

Simple, transparent pricing

Start free with Core. Upgrade when you need production features.

Indie

For solo developers and side projects

$69/mo
$690/yr billed annually (save 2 months)
  • Everything in Core
  • WAL crash recovery
  • Session reflection
  • Memory management tools
  • MCP server
  • 1 machine license
  • Community support
Get Started

Enterprise

For organizations with custom needs

Contact Us
Custom terms and SLAs
  • Everything in Team
  • Unlimited machines
  • On-prem deployment
  • Custom integrations
  • Dedicated support engineer
  • SLA guarantees
  • Training and onboarding
Contact Sales

Mnemo pricing covers software licensing only. Embedding, LLM, and rerank API costs are separate — you bring your own API keys. Run 100% locally with Ollama for $0 API cost, or use cloud providers like Voyage and OpenAI (~$20-45/mo depending on usage).

Up and running in 60 seconds

100% Local ($0)
# Pull Ollama models
ollama pull nomic-embed-text
ollama pull qwen3:8b
ollama pull bge-reranker-v2-m3

# Start services
git clone https://github.com/Methux/mnemo
cp config/mnemo.local.example.json ~/.mnemo/mnemo.json
docker compose up -d
Cloud APIs (recommended)
# Install
npm install @mnemoai/core

# Interactive setup wizard
npm run init

# Or use Docker
cp .env.example .env  # add API keys
docker compose up -d
agent.ts
import { Mnemo } from '@mnemoai/core';

// Initialize with defaults
const mnemo = new Mnemo({
  agent: 'my-agent',
  storage: 'lancedb',
});

// Store a memory (auto-classified, auto-decayed)
await mnemo.store({
  content: 'User prefers dark mode and concise responses',
  source:  'conversation',
});

// Retrieve with triple-path fusion
const memories = await mnemo.recall('user preferences');

console.log(memories);
// [{ content: "User prefers dark mode...", score: 0.94, decay: 0.98 }]

How Mnemo stacks up

Evaluated on retrieval accuracy, contradiction handling, and memory relevance over time.

FrameworkRetrieval AccuracyContradiction F1Decay QualityOverall
Mnemo Pro96.2%94.1%97.8%96.0
Mnemo Core93.7%91.4%97.8%94.3
Mem084.5%72.3%N/A78.4
Zep81.2%68.9%N/A75.1
Letta79.8%65.2%N/A72.5