Open-source AI memory framework

AI memory that forgets intelligently

A cognitive science-based memory framework for AI agents. Runs 100% local at $0 or with cloud APIs. Weibull forgetting, triple-path retrieval, 10-stage pipeline.

Memory should work like memory

The Problem

Other memory solutions store everything forever. Your agent drowns in noise. Context windows fill with stale facts. Retrieval degrades as data grows. The more your agent remembers, the worse it performs.

The Mnemo Way

Mnemo models human memory: important memories consolidate, trivial ones fade, frequently recalled knowledge strengthens. Built on decades of cognitive science research, not naive vector search.

10-Stage Memory Pipeline

From raw conversation to durable, retrievable memory in milliseconds.

Episodic Semantic Preference Procedural Relationship Reflection 6 Channels Classify Stage 1 Dedup & Contradict Weibull Decay Init store.ts Embed + Write LanceDB Vectors Graphiti Knowledge Graph Retrieval (Read) Vector Search BM25 Full-text Graph Traverse RRF + Rerank

Built different, by design

Every feature grounded in cognitive science and real-world agent workloads.

Weibull Decay

Stretched-exponential forgetting with tier-specific beta parameters. Memories fade naturally unless reinforced through recall.

Triple-Path Retrieval

Vector similarity, BM25 full-text, and knowledge graph traversal fused with Reciprocal Rank Fusion for robust recall.

Contradiction Detection

Three-layer LLM detection pipeline. When facts conflict, old versions auto-expire and new truths consolidate.

Scope Isolation

Per-agent memory with configurable access rules. Each agent operates in its own namespace with controlled sharing.

Cross-encoder Rerank

Voyage rerank-2 precision ranking ensures the most relevant memories surface first, every time.

Ablation-Validated

35 ablation tests prove every module earns its place. No cargo-cult features — each component has measurable, verified impact on retrieval quality.

Choose your tier

Core is free (MIT). Pro unlocks production features. API costs are separate — you bring your own keys.

Local

Core Local

$0/mo API · 100% Offline

  • Ollama embedding (bge-m3)
  • Ollama LLM extraction (qwen3:8b)
  • Ollama cross-encoder rerank (bge-reranker)
  • Weibull decay + tier lifecycle
  • Triple-path retrieval (Vec+BM25+Graph)
  • Contradiction detection
Zero API cost. Full offline capability.
Hybrid

Core Hybrid

~$5/mo API · MIT License

  • Voyage embedding (voyage-3-large)
  • GPT-4.1-mini smart extraction
  • Voyage rerank-2 cross-encoder
  • Everything in Local, better models
  • Best quality-to-cost ratio
  • Recommended for most users
Cloud-grade models, self-hosted infrastructure.
Cloud

Core Cloud

~$45/mo API · MIT License

  • Voyage embedding (voyage-3-large)
  • GPT-4.1 full extraction
  • Voyage rerank-2 cross-encoder
  • Maximum extraction quality
  • Best for high-volume production
  • Near Pro-level retrieval
Top-tier models. Near Pro-level retrieval quality.
Pro

Mnemo Pro

From $69/mo + API · Commercial

  • Everything in Core Cloud
  • WAL crash recovery
  • Session reflection + self-improvement
  • Memory tools + MCP server
  • Observability + access tracking
  • Machine-bound license + priority support
Full stack. WAL + reflection + self-improvement.

Simple, transparent pricing

Start free with Core. Upgrade when you need production features.

Indie

For solo developers and side projects

$69/mo
$690/yr billed annually (save 2 months)
  • Everything in Core
  • WAL crash recovery
  • Session reflection
  • Memory management tools
  • MCP server
  • 1 machine license
  • Community support
Get Started

Enterprise

For organizations with custom needs

Contact Us
Custom terms and SLAs
  • Everything in Team
  • Unlimited machines
  • On-prem deployment
  • Custom integrations
  • Dedicated support engineer
  • SLA guarantees
  • Training and onboarding
Contact Sales

Mnemo pricing covers software licensing only. Embedding, LLM, and rerank API costs are separate — you bring your own API keys. Run 100% locally with Ollama for $0 API cost, or use cloud providers like OpenAI and Voyage (~$5-45/mo depending on usage). Ecosystem: Vercel AI SDK, REST Server, Python SDK.

Up and running in 60 seconds

100% Local ($0)
# Install
npm install @mnemoai/core

# Pull Ollama model
ollama pull bge-m3

# Done! Use preset in your code:
const mnemo = await createMnemo({
  preset: 'ollama',
  dbPath: './memory-db'
});
Cloud APIs (recommended)
# Install
npm install @mnemoai/core

# Auto-detect from env var
const mnemo = await createMnemo({
  dbPath: './memory-db'
});

# Or use a preset
const mnemo = await createMnemo({
  preset: 'openai',
  dbPath: './memory-db'
});
agent.ts
import { createMnemo } from '@mnemoai/core';

// Initialize — auto-detect, preset, or full config
const mnemo = await createMnemo({
  preset: 'openai',  // or 'ollama', 'voyage', 'jina'
  dbPath: './memory-db',
});

// Store a memory
await mnemo.store({
  text: 'User prefers dark mode and concise responses',
  category: 'preference',
});

// Retrieve — decay, dedup, and ranking applied automatically
const memories = await mnemo.recall('user preferences');

console.log(memories);
// [{ text: "User prefers dark mode...", score: 0.94, category: "preference" }]

Feature comparison

How Mnemo compares to other AI memory frameworks on architecture and features.

FeatureMnemoMem0ZepLetta
Weibull DecayYesNoNoNo
Triple-Path RetrievalVec+BM25+GraphVector onlyVector+temporalVector only
Contradiction Detection3-layer LLMBasicNoNo
Cross-encoder RerankVoyage/Jina/OllamaNoNoNo
100% Local ($0)Ollama presetNoNoPartial
Open SourceMITPaid ($99+)PaidFreemium
Preset Config (1 line)4 presetsNoNoNo