Nexus-MCP
A fully local, open-source MCP server that replaces Azure AI Foundry, enabling model management, inference, evaluation, and RAG using Ollama without any cloud dependencies.
README
#Nexus-MCP ( Foundry-Reverse )
Newer Foundry MCP reverse-engineered to work on Ollama
A fully local, open-source alternative to the Azure AI Foundry MCP Server, powered entirely by Ollama.
No Azure subscription, no API keys, no cloud required.
Features
| Category | Tools |
|---|---|
| Health | health_check |
| Model Management | list_models, get_model_info, pull_model, delete_model, list_running_models, compare_models |
| Inference | generate, chat |
| Evaluation | evaluate_response, evaluate_agent |
| Knowledge / RAG | create_index, list_indexes, add_document, query_knowledge, delete_index |
Plus MCP Resources (ollama://models, ollama://running, ollama://indexes) and reusable Prompts (summarize, rag_answer, code_review).
Requirements
Quick Start
# 1. Clone
git clone https://github.com/deadSwank001/Foundry-Reverse.git
cd Foundry-Reverse
# 2. Copy and edit configuration
cp .env.example .env
# 3. Run the MCP server (stdio transport)
uv run foundry-reverse
VS Code / Copilot Integration
Copy mcp.json to your VS Code workspace .vscode/mcp.json (or user-level MCP config), then restart VS Code.
{
"mcpServers": {
"foundry-reverse": {
"command": "uv",
"args": ["run", "foundry-reverse"],
"env": {
"OLLAMA_BASE_URL": "http://localhost:11434"
}
}
}
}
Configuration
| Variable | Default | Description |
|---|---|---|
OLLAMA_BASE_URL |
http://localhost:11434 |
Ollama API endpoint |
OLLAMA_TIMEOUT |
120 |
Request timeout in seconds |
EMBED_MODEL |
nomic-embed-text |
Ollama model used for embeddings |
JUDGE_MODEL |
(first available) | Ollama model used as evaluator |
KNOWLEDGE_STORE_PATH |
.foundry_knowledge.json |
Path for the local RAG store |
Development
# Install deps (including dev)
uv sync --all-groups
# Run tests
uv run pytest
Architecture
src/foundry_reverse/
├── __init__.py # Package version
├── server.py # FastMCP server – all tools, resources, prompts
├── ollama_client.py # Async Ollama REST API client
├── evaluation.py # LLM-as-judge evaluation helpers
└── knowledge.py # In-memory vector store (cosine similarity + Ollama embeddings)
Comparison with Azure AI Foundry MCP
| Feature | Azure Foundry MCP | Foundry-Reverse |
|---|---|---|
| Model catalog | Azure AI model registry | Local Ollama models |
| Inference | Azure OpenAI / serverless | Ollama (/api/generate, /api/chat) |
| Embeddings | Azure OpenAI embeddings | Ollama (/api/embeddings) |
| Vector search | Azure AI Search | In-memory cosine similarity |
| Evaluation | Azure AI Evaluation SDK | Local LLM-as-judge |
| Auth | Azure Service Principal / keys | None (local only) |
| Cost | Pay-per-token | Free |
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.