hive
Context infrastructure for AI-assisted development — on-demand Obsidian vault access via MCP
README
hive-vault
<!-- mcp-name: io.github.mlorentedev/hive-vault -->
Your AI coding assistant forgets everything between sessions. Hive fixes that.
Every session, your assistant loads 800+ lines of static context. Most of it is irrelevant. You pay the full token cost every time. And next session? It starts from zero again.
Hive is an MCP server that connects your AI assistant to an Obsidian vault. Instead of loading everything upfront, it queries only what's needed — architecture decisions, lessons learned, project context — all on demand via MCP.
The numbers:
| Metric | Without Hive | With Hive |
|---|---|---|
| Context loaded per session | ~800 lines (static) | ~50 lines (on demand) |
| Token cost for context | 100% every session | 6% average per query |
| Knowledge retained between sessions | 0% | 100% (in vault) |
| Time to find past decisions | Manual search | vault_search in seconds |
Measured on a real vault with 19 projects, 200+ files. See benchmarks.
Install (30 seconds)
One command. No cloning, no venv, no config files. Use user scope (-s user) so Hive works across all your projects — that's where cross-project knowledge shines.
Claude Code:
claude mcp add -s user hive -- uvx --upgrade hive-vault
Gemini CLI:
gemini mcp add -s user hive-vault uvx -- --upgrade hive-vault
OpenAI Codex CLI — add to ~/.codex/config.toml:
[mcp_servers.hive-vault]
command = "uvx"
args = ["--upgrade", "hive-vault"]
GitHub Copilot (VS Code) — add to .vscode/mcp.json:
{
"servers": {
"hive-vault": {
"command": "uvx",
"args": ["--upgrade", "hive-vault"]
}
}
}
Other MCP clients (Cursor, Windsurf, etc.): point your client at uvx --upgrade hive-vault via stdio transport.
Then ask your assistant:
"Use vault_list_projects to see my vault"
That's it. You're running.
What You Get
14 Vault Tools — your knowledge, on demand
| Tool | What it does |
|---|---|
vault_query |
Load project context, tasks, roadmap, lessons — or any file by path |
vault_search |
Full-text search with metadata filters and regex support |
vault_smart_search |
Ranked results with relevance scoring (status + recency + match density) |
session_briefing |
One call = tasks + lessons + git log + health. Start every session here |
vault_list_projects |
See all projects in your vault |
vault_list_files |
Browse project structure with glob pattern filtering |
vault_health |
File counts, staleness metrics, coverage gaps per project |
vault_recent |
What changed in the last N days (via git + frontmatter) |
vault_update |
Write to vault with YAML validation + auto git commit |
vault_create |
Create files with auto-generated frontmatter + auto git commit |
vault_patch |
Surgical find-and-replace with ambiguity rejection + auto git commit |
capture_lesson |
Capture a lesson inline — deduplicates, appends to 90-lessons.md |
vault_summarize |
Small files returned directly, large files delegated for compression |
vault_usage |
Tool call analytics — which tools, which projects, how many tokens |
3 Worker Tools — delegate to cheaper models
| Tool | What it does |
|---|---|
delegate_task |
Route tasks to Ollama (free, local) or OpenRouter (free/paid cloud) |
list_models |
See all available models across providers |
worker_status |
Budget remaining, connectivity, usage stats |
Routing: Ollama first (free) → OpenRouter free tier → OpenRouter paid ($1/mo cap) → reject.
Your primary model handles architecture. Cheaper models handle boilerplate.
Before / After
Before Hive — static CLAUDE.md:
# My Project
## Architecture
[200 lines of decisions you made 3 months ago]
## Standards
[150 lines of coding patterns]
## Lessons
[100 lines of past bugs]
## Tasks
[50 lines of backlog]
# ...loaded every single session, whether relevant or not
With Hive — dynamic, on demand:
# Only when the assistant needs architecture context:
vault_query(project="my-project", section="context")
# Only when searching for a past decision:
vault_search(query="database migration strategy")
# Start of session — just the essentials:
session_briefing(project="my-project")
Configure Your Vault
Default vault path: ~/Projects/knowledge. To change it:
# Claude Code
claude mcp add -s user hive -e VAULT_PATH=/path/to/vault -- uvx --upgrade hive-vault
# Gemini CLI
gemini mcp add -s user -e VAULT_PATH=/path/to/vault hive-vault uvx -- --upgrade hive-vault
Enable Worker Delegation (optional)
claude mcp add -s user hive \
-e VAULT_PATH=/path/to/vault \
-e HIVE_OLLAMA_ENDPOINT=http://your-ollama:11434 \
-e OPENROUTER_API_KEY=sk-or-... \
-- uvx --upgrade hive-vault
All Configuration
| Variable | Default | Description |
|---|---|---|
VAULT_PATH |
~/Projects/knowledge |
Path to your Obsidian vault |
HIVE_OLLAMA_ENDPOINT |
http://localhost:11434 |
Ollama API endpoint |
HIVE_OLLAMA_MODEL |
qwen2.5-coder:7b |
Default Ollama model |
HIVE_OPENROUTER_API_KEY |
— | OpenRouter API key (also reads OPENROUTER_API_KEY) |
HIVE_OPENROUTER_MODEL |
qwen/qwen3-coder:free |
Default free tier model |
HIVE_OPENROUTER_PAID_MODEL |
qwen/qwen3-coder |
Paid tier model |
HIVE_OPENROUTER_BUDGET |
1.0 |
Monthly budget cap (USD) |
HIVE_VAULT_SCOPES |
{"projects": "10_projects", "meta": "00_meta"} |
JSON mapping of scope names to vault subdirectories |
See full configuration reference for all 15 environment variables.
Recommended Workflow
The highest-value setup combines three tools:
- Obsidian — local-first knowledge base with 1M+ community, Markdown native, no lock-in
- Obsidian Git — auto-commits your vault changes on a schedule (version history for free)
- Hive — bridges your vault to any AI coding assistant via MCP
Your assistant writes lessons and decisions to the vault → Obsidian Git auto-commits → next session, everything is there. No manual sync. No context lost.
Hive works with any directory of Markdown files — Obsidian is recommended, not required.
Vault Structure
For best results, follow this layout:
~/Projects/knowledge/
├── 00_meta/patterns/ # cross-project patterns
├── 10_projects/
│ ├── my-project/
│ │ ├── 00-context.md # vault_query section="context"
│ │ ├── 10-roadmap.md # vault_query section="roadmap"
│ │ ├── 11-tasks.md # vault_query section="tasks"
│ │ ├── 90-lessons.md # vault_query section="lessons"
│ │ └── 30-architecture/ # any path works with vault_query path="..."
│ └── another-project/
└── ...
Make Your Assistant Use Hive Consistently
MCP tools don't activate on their own. Add this to your project's CLAUDE.md (or equivalent):
## Vault & Knowledge (Hive MCP)
When hive-vault MCP is available:
- `session_briefing(project="myproject")` — start every session here
- `vault_query(project="myproject", section="context")` — project overview
- `vault_search(query="...")` — find past decisions
- `capture_lesson(...)` — capture insights inline, don't wait until session end
Without these instructions, your assistant uses Hive inconsistently. With them, it uses Hive every session, predictably.
Resources & Prompts
5 MCP Resources for auto-discoverable data:
| URI | Description |
|---|---|
hive://projects |
All vault projects with file counts |
hive://health |
Vault health metrics |
hive://projects/{project}/context |
Project context |
hive://projects/{project}/tasks |
Project backlog |
hive://projects/{project}/lessons |
Lessons learned |
4 MCP Prompts for guided workflows:
| Prompt | Description |
|---|---|
retrospective |
End-of-session review → extract lessons to vault |
delegate |
Structured protocol for worker delegation |
vault_sync |
Post-sprint vault sync — reconcile docs with shipped code |
benchmark |
Estimate token savings from Hive in the current session |
Architecture
MCP Host (Claude Code, Gemini CLI, Codex CLI, Cursor, ...)
└── hive-vault (MCP server, stdio)
├── Vault Tools (14) ── Obsidian vault (Markdown + YAML frontmatter)
│ query, search, smart_search, list_files, patch,
│ update, create, capture_lesson, summarize,
│ session_briefing, recent, usage, health, list_projects
│
└── Worker Tools (3) ── Task delegation + routing:
delegate_task 1. Ollama (local, free)
list_models 2. OpenRouter free tier
worker_status 3. OpenRouter paid ($1/mo cap)
4. Reject → host handles it
Development
See CONTRIBUTING.md for setup, code standards, and PR workflow.
git clone https://github.com/mlorentedev/hive.git
cd hive
make install # create venv + install deps
make check # lint + typecheck + test (265 tests, 92% coverage)
License
MIT
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.