cachly-mcp-server

cachly-mcp-server

Persistent AI memory for Claude Code, Cursor, GitHub Copilot & Windsurf โ€” sessions, lessons learned, semantic search, and team brain. 38 MCP tools. Free tier, EU servers.

Category
Visit Server

README

๐Ÿง  cachly AI Brain โ€” MCP Server

Persistent memory for Claude Code, Cursor, GitHub Copilot & Windsurf.
Your AI remembers every lesson, every fix, every architecture decision โ€” forever.

<p align="center"> <a href="https://github.com/cachly-dev/mcp-server/stargazers"> <img src="https://img.shields.io/github/stars/cachly-dev/mcp-server?style=social" alt="GitHub Stars" /> </a> ย  <a href="https://www.npmjs.com/package/@cachly-dev/mcp-server"> <img src="https://img.shields.io/npm/v/@cachly-dev/mcp-server?color=red&logo=npm" alt="npm version" /> </a> ย  <a href="https://www.npmjs.com/package/@cachly-dev/mcp-server"> <img src="https://img.shields.io/npm/dw/@cachly-dev/mcp-server?color=blue&label=weekly%20installs" alt="npm downloads" /> </a> ย  <a href="https://cachly.dev"> <img src="https://img.shields.io/badge/Free%20tier-โ‚ฌ0%2Fmo-brightgreen" alt="Free tier" /> </a> ย  <a href="https://cachly.dev/legal"> <img src="https://img.shields.io/badge/GDPR-EU%20only-green" alt="GDPR: EU only" /> </a> ย  <img src="https://img.shields.io/badge/License-Apache--2.0-yellow" alt="License: Apache-2.0" /> </p>


The Problem

Every morning, you open your AI coding assistant. It doesn't remember yesterday.
You explain your architecture. You explain the deployment process. You explain the bug you fixed last week.

The average developer wastes 45 minutes/day re-establishing context. That's โ‚ฌ15,000+ in lost productivity per engineer per year.

The Fix โ€” One Command

npx @cachly-dev/mcp-server@latest setup

The interactive wizard:

  1. Signs you in (free, no credit card required)
  2. Picks or creates your AI Brain instance
  3. Auto-detects Cursor, Windsurf, VS Code, Claude Code, Continue.dev
  4. Writes the correct MCP config for every detected editor
  5. Creates CLAUDE.md with memory rules pre-filled

Result: Your AI remembers everything. Always.


Quick Start (Manual)

Step 1 โ€” Get your free credentials at cachly.dev (25 MB free, forever, no credit card).

Step 2 โ€” Add to your editor's MCP config:

<details> <summary><b>Claude Code</b> (<code>~/.claude/mcp.json</code> or <code>.mcp.json</code>)</summary>

{
  "mcpServers": {
    "cachly": {
      "type": "stdio",
      "command": "npx",
      "args": ["-y", "@cachly-dev/mcp-server"],
      "env": {
        "CACHLY_JWT": "your-jwt-token",
        "CACHLY_INSTANCE_ID": "your-instance-id"
      }
    }
  }
}

</details>

<details> <summary><b>Cursor / Windsurf / VS Code</b> (<code>.cursor/mcp.json</code> / <code>.mcp.json</code>)</summary>

{
  "servers": {
    "cachly": {
      "type": "stdio",
      "command": "npx",
      "args": ["-y", "@cachly-dev/mcp-server"],
      "env": {
        "CACHLY_JWT": "your-jwt-token",
        "CACHLY_INSTANCE_ID": "your-instance-id"
      }
    }
  }
}

</details>

Step 3 โ€” Add to CLAUDE.md / .github/copilot-instructions.md:

## AI Brain Rules (cachly)
- Call session_start BEFORE reading any files or making changes
- Call learn_from_attempts AFTER every fix, deploy, or discovery
- Call session_end when closing the window

With vs. Without cachly

Situation Without cachly With cachly
Session start "What's your architecture?" "Ready. 23 lessons, last session: deployed API."
Known bug hits again Re-researches from scratch "You fixed this on March 12, here's the exact command"
After holiday / team handoff Context dead Fully briefed in < 10 seconds
Repeated LLM patterns Pays for each re-discovery Cached by meaning, ~1,200 tokens saved per hit
New team member Weeks to onboard session_start gives full context instantly

38 MCP Tools

๐Ÿง  Session & Memory (most used)

Tool What it does
session_start Full briefing: last session summary, open failures, recent lessons, brain health
session_end Save what you built, auto-extract lessons from summary
learn_from_attempts Store structured lessons after any fix, deploy, or discovery
recall_best_solution Best known solution for a topic โ€” with success/failure history
remember_context Cache architecture findings, decisions, file summaries
smart_recall BM25+ full-text search across all brain data
session_handoff Hand off remaining tasks to next window, with context

โš™๏ธ Instance Management

Tool What it does
list_instances List all your cache instances
create_instance Spin up a new instance (free or paid)
get_connection_string Get the redis:// URL for your app
delete_instance Remove an instance
get_real_time_stats Memory, hit rate, ops/sec

๐Ÿ—„๏ธ Cache Operations

Tool What it does
cache_get / cache_set / cache_delete Standard cache operations
cache_mget / cache_mset Bulk pipeline (single round-trip)
cache_lock_acquire / cache_lock_release Distributed Redlock-lite
cache_stream_set / cache_stream_get LLM token stream caching

๐Ÿ” Semantic Cache

Tool What it does
semantic_search Find cached entries by meaning (pgvector HNSW, EU-hosted)
semantic_warmup Pre-warm cache with prompt/response pairs
detect_namespace Auto-classify prompt into code/qa/summary/translation/creative

๐Ÿ‘ฅ Team Brain

Tool What it does
team_learn / team_recall Share lessons across the team
global_learn / global_recall Cross-project universal lessons
list_orgs / create_org Manage team organizations
invite_member Invite a developer to your org by email

Pricing

Tier RAM Price Best for
Free 25 MB โ‚ฌ0/mo forever Dev & side projects
Dev 200 MB โ‚ฌ19/mo Individual developers
Pro 900 MB โ‚ฌ49/mo Teams
Speed 900 MB + Dragonfly + Semantic Cache โ‚ฌ79/mo AI-heavy workloads
Business 7 GB โ‚ฌ199/mo Scale-ups

โœ… All plans: German servers ยท GDPR-compliant ยท 99.9% SLA ยท No credit card for Free tier


Environment Variables

Variable Default Description
CACHLY_JWT โ€” Required. Your API token from cachly.dev
CACHLY_INSTANCE_ID โ€” Default instance UUID (optional if you pass per-call)
CACHLY_API_URL https://api.cachly.dev Override for self-hosted
CACHLY_NO_TELEMETRY unset Set to 1 to disable anonymous usage pings
CACHLY_NO_UPDATE_CHECK unset Set to 1 to disable the version-check on startup

Links

Recommended Servers

playwright-mcp

playwright-mcp

A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.

Official
Featured
TypeScript
Magic Component Platform (MCP)

Magic Component Platform (MCP)

An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.

Official
Featured
Local
TypeScript
Audiense Insights MCP Server

Audiense Insights MCP Server

Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.

Official
Featured
Local
TypeScript
VeyraX MCP

VeyraX MCP

Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.

Official
Featured
Local
graphlit-mcp-server

graphlit-mcp-server

The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.

Official
Featured
TypeScript
Kagi MCP Server

Kagi MCP Server

An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.

Official
Featured
Python
E2B

E2B

Using MCP to run code via e2b.

Official
Featured
Neon Database

Neon Database

MCP server for interacting with Neon Management API and databases

Official
Featured
Exa Search

Exa Search

A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.

Official
Featured
Qdrant Server

Qdrant Server

This repository is an example of how to create a MCP server for Qdrant, a vector search engine.

Official
Featured