Rememb
Persistent memory for AI agents with sectioned entries (project/user/context/etc), semantic search, CLI, and per-project scope. Local JSON, zero config, no server required.
README
<!-- mcp-name: io.github.LuizEduPP/rememb -->

AI agents forget everything between sessions. rememb gives them persistent memory — local, portable, and works with any agent.

The problem
Every dev using AI professionally hits this wall:
Session 1: "We're using PostgreSQL, auth at src/auth/, prefer async patterns."
Session 2: Agent starts from zero. You explain everything again.
Session 3: Same thing.
Existing solutions (Mem0, Zep, Letta) require servers, API keys, and cloud accounts.
You just want the agent to remember your project.
Install
pip install rememb[mcp] # Recommended — includes MCP server
pip install rememb # CLI only
pip install rememb[mcp,semantic,pdf] # All features
Quick Start
With MCP (recommended)
Zero friction. No CLI commands. Native IDE integration.
1. Add to your IDE's MCP config:
{
"mcpServers": {
"rememb": {
"command": "rememb",
"args": ["mcp"]
}
}
}
2. Restart your IDE.
The agent now automatically reads memory at session start, writes when learning something new, and searches when needed.
Without MCP
rememb rules # Print generic rules for AI agents
Copy the output to your editor's rules file (.windsurfrules, .cursorrules, CLAUDE.md, etc.)
How it works
.rememb/
entries.json ← structured memory (project, actions, systems, user, context)
meta.json ← project metadata
A JSON file in your project. Your agent reads it at the start of every session.
User: "We're using PostgreSQL, auth at src/auth/, async patterns"
Agent: [rememb_write] → Saved
[New session]
Agent: [rememb_read] → Context loaded
Agent: "I see you're using PostgreSQL with auth at src/auth/..."
Search uses local semantic embeddings (no API, no cloud). Falls back to keyword search if embeddings aren't available.
Memory sections
| Section | What to store |
|---|---|
project |
Tech stack, architecture, goals |
actions |
What was done, decisions made |
systems |
Services, modules, integrations |
requests |
User preferences, recurring asks |
user |
Name, style, expertise, preferences |
context |
Anything else relevant |
CLI
rememb init # Initialize memory store
rememb write "text" # Add entry (--section, --tags)
rememb read # List all entries (--section, --agent)
rememb search "query" # Semantic/keyword search (--top)
rememb edit <id> # Update entry (--content, --section, --tags)
rememb delete <id> # Remove entry
rememb clear --yes # Delete all entries
rememb import <folder> # Import .md/.txt/.pdf files
rememb rules # Show generic rules for AI agents
Design
- Local first — plain JSON file in your project
- Portable — copy
.rememb/anywhere, it works - Agnostic — any agent, any IDE (MCP or CLI)
- No lock-in — no servers, no API keys, no accounts
Contributing
git clone https://github.com/LuizEduPP/Rememb
cd rememb
pip install -e ".[dev]"
PRs welcome. Issues welcome. Stars welcome. 🌟
License
MIT
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.