Dolphin MCP
Provides AI-powered semantic code search across multiple repositories, allowing natural language queries to find code chunks and retrieve specific file contents through Dolphin AI embeddings.
README
dolphin-mcp
MCP server for Dolphin semantic code search. Conforms to MCP spec.
Quick Start
No installation needed - use bunx:
bunx dolphin-mcp
Configuration
Continue.dev
Add to config.yaml:
mcpServers:
- name: Dolphin-KB
command: bunx
args:
- dolphin-mcp
env:
DOLPHIN_API_URL: "http://127.0.0.1:7777"
# Optional: Performance optimization for parallel snippet fetching
MAX_CONCURRENT_SNIPPET_FETCH: "8"
SNIPPET_FETCH_TIMEOUT_MS: "2000"
SNIPPET_FETCH_RETRY_ATTEMPTS: "1"
Claude Desktop
Add to claude_desktop_config.json:
{
"mcpServers": {
"dolphin-kb": {
"command": "bunx",
"args": ["dolphin-mcp"],
"env": {
"DOLPHIN_API_URL": "http://127.0.0.1:7777",
"MAX_CONCURRENT_SNIPPET_FETCH": "8",
"SNIPPET_FETCH_TIMEOUT_MS": "2000",
"SNIPPET_FETCH_RETRY_ATTEMPTS": "1"
}
}
}
}
Environment Variables
| Variable | Default | Description |
|---|---|---|
DOLPHIN_API_URL |
http://127.0.0.1:7777 |
Dolphin API endpoint |
LOG_LEVEL |
info |
Logging level (debug, info, warn, error) |
Parallel Snippet Fetching Configuration
These variables control the performance optimization for parallel snippet fetching in search_knowledge:
| Variable | Default | Description | Recommended Range |
|---|---|---|---|
MAX_CONCURRENT_SNIPPET_FETCH |
8 |
Maximum parallel snippet requests | 4-12 |
SNIPPET_FETCH_TIMEOUT_MS |
2000 |
Timeout per snippet request (ms) | 1500-3000 |
SNIPPET_FETCH_RETRY_ATTEMPTS |
1 |
Retry attempts for failed requests | 0-3 |
Configuration Presets
Conservative (recommended for limited resources):
MAX_CONCURRENT_SNIPPET_FETCH=4
SNIPPET_FETCH_TIMEOUT_MS=1500
SNIPPET_FETCH_RETRY_ATTEMPTS=1
Recommended (balanced performance):
MAX_CONCURRENT_SNIPPET_FETCH=8
SNIPPET_FETCH_TIMEOUT_MS=2000
SNIPPET_FETCH_RETRY_ATTEMPTS=1
Performance (maximum throughput):
MAX_CONCURRENT_SNIPPET_FETCH=10
SNIPPET_FETCH_TIMEOUT_MS=3000
SNIPPET_FETCH_RETRY_ATTEMPTS=2
Available Tools
search_knowledge
Semantically query code and docs across indexed repositories and return ranked snippets with citations.
{
"query": "string (required)",
"repos": ["string"],
"path_prefix": ["string"],
"top_k": "number (1-100)",
"max_snippets": "number",
"embed_model": "small | large",
"score_cutoff": "number"
}
fetch_chunk
Fetch a chunk by chunk_id and return fenced code with citation.
{
"chunk_id": "string (required)"
}
fetch_lines
Fetch a file slice [start, end] inclusive from disk and return fenced code with citation.
{
"repo": "string (required)",
"path": "string (required)",
"start": "number (required, 1-indexed)",
"end": "number (required, inclusive)"
}
get_vector_store_info
Report namespaces, dims, limits, and approximate counts.
{}
open_in_editor
Compute a vscode://file URI for a repo path and optional position.
{
"repo": "string (required)",
"path": "string (required)",
"line": "number (1-indexed)",
"column": "number (1-indexed)"
}
Installation (Optional)
If you prefer installing globally:
bun install -g dolphin-mcp
Then use dolphin-mcp instead of bunx dolphin-mcp.
Requirements
- Bun >= 1.0.0 - Install
- Dolphin API running on configured endpoint
License
MIT - see LICENSE file for details.
Links
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.