Sponsored by Deepsite.site

Remembra

Created By
remembra-ai14 days ago
Persistent memory layer for AI agents with entity resolution, PII detection, AES-256-GCM encryption at rest, and hybrid search. Self-hosted. 100% on LoCoMo benchmark. Works with Claude Code, Cursor, VS Code, Windsurf, JetBrains, and more.
Overview

Remembra Logo

Remembra

The memory layer for AI that actually works.
Persistent memory with entity resolution, temporal decay, and graph-aware recall.
Self-host in minutes. No vendor lock-in.

PyPI npm GitHub Stars License: MIT Documentation

DocumentationWebsiteQuick StartWhy Remembra?TwitterDiscord


🚀 What's New in v0.9.0

  • ⏳ Temporal Knowledge Graph — Bi-temporal relationships with valid_from, valid_to, and point-in-time queries. Ask "Where did Alice work in January 2022?"
  • 🛠️ 11 MCP Tools — 6 new tools: update_memory, search_entities, list_memories, share_memory, timeline, relationships_at
  • 📊 Entity Graph Visualization — Interactive force-directed graph with flowing particle effects on relationship edges
  • 🔄 Contradiction Detection — New relationships automatically supersede old ones with full history preserved
  • 🔐 AES-256-GCM Field Encryption — Encrypt memory content at rest with OWASP-compliant key derivation
  • 🛡️ Enterprise Security Suite — PII detection, anomaly monitoring, audit logging

The Problem

Every AI app needs memory. Your chatbot forgets users between sessions. Your agent can't recall decisions from yesterday. Your assistant asks the same questions over and over.

Existing solutions have tradeoffs:

  • Mem0: Graph features require $249/mo plan; limited self-hosting documentation
  • Zep: Academic approach, complex deployment
  • Letta: Research-grade, not production-ready
  • LangChain Memory: Too basic, no persistence

The Solution

from remembra import Memory

memory = Memory(user_id="user_123")

# Store — entities and facts extracted automatically
memory.store("Had a meeting with Sarah from Acme Corp. She prefers email over Slack.")

# Recall — semantic search finds relevant memories
result = memory.recall("How should I contact Sarah?")
print(result.context)
# → "Sarah from Acme Corp prefers email over Slack."

# It knows "Sarah" and "Acme Corp" are entities. It builds relationships.
# It persists across sessions, reboots, context windows. Forever.

⚡ Quick Start (2 Minutes)

One Command Install

curl -sSL https://raw.githubusercontent.com/remembra-ai/remembra/main/quickstart.sh | bash

That's it. Remembra + Qdrant + Ollama start locally. No API keys needed.

Or with Docker Compose directly:

git clone https://github.com/remembra-ai/remembra && cd remembra
docker compose -f docker-compose.quickstart.yml up -d

Try it:

# Store a memory
curl -X POST http://localhost:8787/api/v1/memories \
  -H "Content-Type: application/json" \
  -d '{"content": "Alice is CEO of Acme Corp", "user_id": "demo"}'

# Recall it
curl -X POST http://localhost:8787/api/v1/memories/recall \
  -H "Content-Type: application/json" \
  -d '{"query": "Who runs Acme?", "user_id": "demo"}'

Connect to Claude (MCP)

Claude Desktop — add to ~/Library/Application Support/Claude/claude_desktop_config.json:

{
  "mcpServers": {
    "remembra": {
      "command": "remembra-mcp",
      "env": {
        "REMEMBRA_URL": "http://localhost:8787",
        "REMEMBRA_USER_ID": "default"
      }
    }
  }
}

Claude Code:

claude mcp add remembra -e REMEMBRA_URL=http://localhost:8787 -- remembra-mcp

Cursor — add to .cursor/mcp.json:

{
  "mcpServers": {
    "remembra": {
      "command": "remembra-mcp",
      "env": {
        "REMEMBRA_URL": "http://localhost:8787"
      }
    }
  }
}

Now ask Claude: "Remember that Alice is CEO of Acme Corp" — then later: "Who runs Acme?"

Python SDK

pip install remembra
from remembra import Memory

memory = Memory(user_id="user_123")
memory.store("Had a meeting with Sarah from Acme Corp. She prefers email over Slack.")
result = memory.recall("How should I contact Sarah?")
print(result.context)  # "Sarah from Acme Corp prefers email over Slack."

TypeScript SDK

npm install remembra
import { Remembra } from 'remembra';

const memory = new Remembra({ url: 'http://localhost:8787' });
await memory.store('User prefers dark mode');
const result = await memory.recall('preferences');

🔥 Why Remembra?

Feature Comparison

FeatureRemembraMem0Zep/GraphitiLettaEngram
One-Command Installcurl | bash✅ pip✅ pip⚠️ Complex✅ brew
Bi-Temporal Relationships✅ Point-in-time⚠️ Basic
Entity Resolution✅ Free💰 $249/mo
Conflict Detection✅ Auto-supersede
PII Detection✅ Built-in
Hybrid Search✅ BM25+Vector
6 Embedding Providers✅ Hot-swap❌ (1-2)❌ (1)
Plugin System
Sleep-Time Compute
Self-Host + Billing✅ Stripe
Memory Spaces✅ Multi-tenant
MCP Server✅ 11 Tools
PricingFree / $49 / $199$19 → $249$25+FreeFree
LicenseMITApache 2.0Apache 2.0Apache 2.0MIT

Core Features

🧠 Smart Extraction — LLM-powered fact extraction from raw text

👥 Entity Resolution — "Adam", "Mr. Smith", "my husband" → same person

⏱️ Temporal Memory — TTL, decay curves, historical queries

🔍 Hybrid Search — Semantic + keyword for accurate recall

🔒 Security — PII detection, anomaly monitoring, audit logs

📊 Dashboard — Visual memory browser, entity graphs, analytics


📊 Benchmark Results

Tested on the LoCoMo benchmark (Snap Research, ACL 2024) — the standard academic benchmark for AI memory systems.

CategoryAccuracyQuestions
Single-hop (direct recall)100%37
Multi-hop (cross-session reasoning)100%32
Temporal (time-based queries)100%13
Open-domain (world knowledge + memory)100%70
Overall (memory categories)100%152

Scored with LLM judge (GPT-4o-mini). Adversarial detection not yet implemented. Run your own: python benchmarks/locomo_runner.py --data /tmp/locomo/data/locomo10.json


📖 Documentation

ResourceDescription
Quick StartGet running in minutes
Python SDKFull Python reference
TypeScript SDKJavaScript/TypeScript guide
MCP ServerTool reference + setup guides for 11 tools
REST APIAPI reference
Self-HostingDocker deployment guide

🛠️ MCP Server

Give any AI coding tool persistent memory with one command. Works with Claude Code, Cursor, VS Code + Copilot, Windsurf, JetBrains, Zed, OpenAI Codex, and any MCP-compatible client.

pip install remembra[mcp]
claude mcp add remembra -e REMEMBRA_URL=http://localhost:8787 -- remembra-mcp

Available Tools (11 total):

ToolDescription
store_memorySave facts, decisions, context
recall_memoriesSemantic search across memories
update_memoryUpdate content without delete+recreate
forget_memoriesGDPR-compliant deletion
list_memoriesBrowse stored memories
search_entitiesSearch the entity graph
share_memoryCross-agent memory sharing via Spaces
timelineTemporal browsing by entity and date
relationships_atPoint-in-time relationship queries
ingest_conversationAuto-extract from chat history
health_checkVerify connection

🏗️ Architecture

┌─────────────────────────────────────────────────────────────┐
│                    Your Application                          │
├──────────┬──────────────┬───────────────────────────────────┤
│ Python   │ TypeScript   │ MCP Server (Claude/Cursor)        │
│ SDK      │ SDK          │ remembra-mcp                      │
├──────────┴──────────────┴───────────────────────────────────┤
│                   Remembra REST API                          │
├──────────────┬──────────────┬───────────────┬───────────────┤
│  Extraction  │   Entities   │   Retrieval   │   Security    │
│  (LLM)       │  (Graph)     │ (Hybrid)      │  (PII/Audit)  │
├──────────────┴──────────────┴───────────────┴───────────────┤
│                    Storage Layer                             │
│         Qdrant (vectors) + SQLite (metadata/graph)          │
└─────────────────────────────────────────────────────────────┘

🤝 Contributing

We welcome contributions! See CONTRIBUTING.md for guidelines.

# Clone
git clone https://github.com/remembra-ai/remembra
cd remembra

# Install dev dependencies
pip install -e ".[dev]"

# Run tests
pytest

# Start dev server
remembra-server --reload

📄 License

MIT License — Use it however you want.


⭐ Star History

If Remembra helps you, please star the repo! It helps others discover the project.

Star History Chart


Built with ❤️ by DolphyTech
remembra.devdocstwitterdiscord

Recommend Servers
TraeBuild with Free GPT-4.1 & Claude 3.7. Fully MCP-Ready.
Tavily Mcp
Howtocook Mcp基于Anduin2017 / HowToCook (程序员在家做饭指南)的mcp server,帮你推荐菜谱、规划膳食,解决“今天吃什么“的世纪难题; Based on Anduin2017/HowToCook (Programmer's Guide to Cooking at Home), MCP Server helps you recommend recipes, plan meals, and solve the century old problem of "what to eat today"
ChatWiseThe second fastest AI chatbot™
Visual Studio Code - Open Source ("Code - OSS")Visual Studio Code
CursorThe AI Code Editor
Baidu Map百度地图核心API现已全面兼容MCP协议,是国内首家兼容MCP协议的地图服务商。
Playwright McpPlaywright MCP server
MCP AdvisorMCP Advisor & Installation - Use the right MCP server for your needs
Y GuiA web-based graphical interface for AI chat interactions with support for multiple AI models and MCP (Model Context Protocol) servers.
MiniMax MCPOfficial MiniMax Model Context Protocol (MCP) server that enables interaction with powerful Text to Speech, image generation and video generation APIs.
Jina AI MCP ToolsA Model Context Protocol (MCP) server that integrates with Jina AI Search Foundation APIs.
RedisA Model Context Protocol server that provides access to Redis databases. This server enables LLMs to interact with Redis key-value stores through a set of standardized tools.
WindsurfThe new purpose-built IDE to harness magic
Zhipu Web SearchZhipu Web Search MCP Server is a search engine specifically designed for large models. It integrates four search engines, allowing users to flexibly compare and switch between them. Building upon the web crawling and ranking capabilities of traditional search engines, it enhances intent recognition capabilities, returning results more suitable for large model processing (such as webpage titles, URLs, summaries, site names, site icons, etc.). This helps AI applications achieve "dynamic knowledge acquisition" and "precise scenario adaptation" capabilities.
AiimagemultistyleA Model Context Protocol (MCP) server for image generation and manipulation using fal.ai's Stable Diffusion model.
Amap Maps高德地图官方 MCP Server
EdgeOne Pages MCPAn MCP service designed for deploying HTML content to EdgeOne Pages and obtaining an accessible public URL.
BlenderBlenderMCP connects Blender to Claude AI through the Model Context Protocol (MCP), allowing Claude to directly interact with and control Blender. This integration enables prompt assisted 3D modeling, scene creation, and manipulation.
DeepChatYour AI Partner on Desktop
Serper MCP ServerA Serper MCP Server