LLM Wiki Kit
Enables creation of persistent, compounding knowledge bases using Karpathy's LLM Wiki pattern with LLM-maintained markdown wikis. Supports automated ingestion, cross-referencing, synthesis, and linting of sources as an alternative to traditional RAG systems.
README
š llm-wiki-kit
Stop re-explaining your research to your AI agent every session.
llm-wiki-kit gives your AI agent a persistent, structured memory that compounds over time. Drop PDFs, URLs, YouTube videos ā your agent builds a wiki, connects the dots, and remembers everything across sessions.
Based on Karpathy's LLM Wiki pattern. Works with Claude, Codex, Cursor, Windsurf, and any MCP-compatible agent.
The Problem
Every time you start a new chat:
You: "Remember that paper on speculative decoding I shared last week?"
Agent: "I don't have access to previous conversations..."
You: *sighs, re-uploads PDF, re-explains context*
You're constantly re-teaching your agent things it should already know.
The Solution
With llm-wiki-kit, your agent maintains its own knowledge base:
You: "What did we learn about speculative decoding?"
Agent: *searches wiki* "Based on the 3 papers you've shared, the Eagle
architecture shows the best efficiency tradeoffs because..."
The wiki persists. Cross-references build up. Your agent gets smarter with every source you add.
ā” Quickstart (2 minutes)
1. Install
pip install "llm-wiki-kit[all] @ git+https://github.com/iamsashank09/llm-wiki-kit.git"
2. Initialize a wiki
mkdir my-research && cd my-research
llm-wiki-kit init --agent claude
3. Connect your agent
Add to Claude Desktop config (claude_desktop_config.json):
{
"mcpServers": {
"llm-wiki-kit": {
"command": "llm-wiki-kit",
"args": ["serve", "--root", "/path/to/my-research"]
}
}
}
<details> <summary><b>Other agents (Codex, Cursor, Windsurf)</b></summary>
OpenAI Codex
codex mcp add llm-wiki-kit -- llm-wiki-kit serve --root /path/to/my-research
Cursor
Add to .cursor/mcp.json:
{
"mcpServers": {
"llm-wiki-kit": {
"command": "llm-wiki-kit",
"args": ["serve", "--root", "/path/to/my-research"]
}
}
}
Windsurf
Add to ~/.codeium/windsurf/mcp_config.json:
{
"mcpServers": {
"llm-wiki-kit": {
"command": "llm-wiki-kit",
"args": ["serve", "--root", "/path/to/my-research"]
}
}
}
</details>
4. Use it
You: "Ingest this paper: raw/attention-is-all-you-need.pdf"
Agent: *creates wiki pages, cross-references concepts, updates index*
You: "Now ingest https://youtube.com/watch?v=kCc8FmEb1nY"
Agent: *extracts transcript, links to existing transformer concepts*
You: "How does the attention mechanism in the paper relate to Karpathy's explanation?"
Agent: *searches wiki, synthesizes answer from both sources*
Your agent now has persistent memory that survives across sessions.
š„ What Makes This Different
| Feature | Why It Matters |
|---|---|
| Multi-format ingest | PDFs, URLs, YouTube, markdown ā just drop it in |
| Auto cross-referencing | Agent builds [[wiki links]] between related concepts |
| Persistent across sessions | Start fresh chats without losing context |
| Full-text search | Agent finds relevant pages instantly (SQLite FTS5) |
| Health checks | wiki_lint catches broken links, orphan pages, contradictions |
| Zero lock-in | It's just markdown files in a folder ā view in Obsidian, VS Code, anywhere |
| Works with any MCP agent | Claude, Codex, Cursor, Windsurf, and more |
š„ Supported Sources
Your agent can ingest anything:
| Drop this... | Get this... |
|---|---|
raw/paper.pdf |
Extracted text, page markers, metadata |
https://arxiv.org/abs/... |
Clean article content, auto-saved to raw/ |
https://youtube.com/watch?v=... |
Full transcript with timestamps |
raw/notes.md |
Direct markdown ingestion |
Install what you need:
pip install "llm-wiki-kit[pdf]" # PDF support
pip install "llm-wiki-kit[web]" # URL extraction
pip install "llm-wiki-kit[youtube]" # YouTube transcripts
pip install "llm-wiki-kit[all]" # Everything
š§ How It Works
āāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāā
ā YOU ā
ā "Ingest this paper. How does it relate to X?" ā
āāāāāāāāāāāāāāāāāāāāāāāāā¬āāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāā
ā
āāāāāāāāāāāāāāāāāāāāāāāāā¼āāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāā
ā WIKI (agent-maintained) ā
ā ā
ā āāāāāāāāāāāāāāāā āāāāāāāāāāāāāāāā āāāāāāāāāāāāāāāā ā
ā ā concepts/ ā ā sources/ ā ā synthesis/ ā ā
ā ā attention.md āāā⤠paper-1.md āāāāŗ cache.md ā ā
ā ā [[linked]] ā ā [[linked]] ā ā [[linked]] ā ā
ā āāāāāāāāāāāāāāāā āāāāāāāāāāāāāāāā āāāāāāāāāāāāāāāā ā
ā ā
ā + index.md (table of contents) ā
ā + log.md (what happened when) ā
āāāāāāāāāāāāāāāāāāāāāāāāā¬āāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāā
ā
āāāāāāāāāāāāāāāāāāāāāāāāā¼āāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāā
ā RAW SOURCES (immutable) ā
ā paper.pdf, article.html, transcript.md ā
āāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāāā
The agent reads raw sources, writes wiki pages, and maintains the connections. You never touch the wiki directly ā the agent does all the work.
š Available Tools
Your agent gets these MCP tools:
| Tool | What it does |
|---|---|
wiki_ingest |
Process any source (file, URL, YouTube) |
wiki_write_page |
Create or update a wiki page |
wiki_read_page |
Read a specific page |
wiki_search |
Full-text search across all pages |
wiki_lint |
Find broken links, orphans, empty pages |
wiki_status |
Overview: page count, sources, recent activity |
wiki_log |
Append to the operation log |
š” Use Cases
Research: Feed papers into your wiki over weeks. Ask synthesis questions that span all your reading.
Technical onboarding: Ingest a codebase's docs. Your agent answers architecture questions from accumulated context.
Competitive intel: Add market reports, earnings calls, news. Agent maintains a living landscape that updates as you add more.
Learning: Watch YouTube tutorials, read blog posts. Agent builds a personalized wiki of everything you've studied.
Book notes: Ingest chapters as you read. Agent tracks characters, themes, plot threads, and connections.
š Pro Tips
- Use Obsidian to visualize your wiki's graph ā it's just a folder of markdown files
- Git init your wiki directory ā get version history for free
- Let the agent link aggressively ā the value compounds in the connections
- Run lint periodically ā catches contradictions and gaps in your knowledge base
- Start small ā even 5-10 sources produce a surprisingly useful wiki
š¦ Development
git clone https://github.com/iamsashank09/llm-wiki-kit
cd llm-wiki-kit
uv venv && source .venv/bin/activate
uv pip install -e ".[all]"
š Credits
Based on the LLM Wiki idea by Andrej Karpathy.
š License
MIT ā do whatever you want with it.
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.