Model Context Protocol Server

Model Context Protocol Server

A middleware server that intelligently routes AI model queries to appropriate data sources, providing contextual information to enhance AI responses.

Category
Visit Server

README

Model Context Protocol (MCP) Server

A FastAPI-based server implementing the Model Context Protocol for providing contextual information to AI models. This server acts as a middleware between AI models and various data sources, intelligently routing queries to the most appropriate data provider.

Features

  • Intelligent query routing based on query analysis
  • Support for multiple data sources (Database, GraphQL, REST)
  • Integration with Ollama models (Mistral, Qwen, Llama2)
  • Environment-aware configuration (Development/Production)
  • Comprehensive logging and error handling
  • Health check endpoints
  • Mock data support for development

Prerequisites

  • Python 3.8+
  • Ollama installed and running locally
  • Required Ollama models:
    • mistral
    • qwen
    • llama2

Installation

  1. Clone the repository:
git clone <repository-url>
cd mcp-server
  1. Create and activate a virtual environment:
python -m venv .venv
source .venv/bin/activate  # On Windows: .venv\Scripts\activate
  1. Install dependencies:
pip install -r requirements.txt
  1. Create a .env file:
cp .env.example .env
  1. Update the .env file with your configuration:
ENVIRONMENT=development
OLLAMA_MODEL=mistral
OLLAMA_BASE_URL=http://localhost:11434

Running the Server

  1. Start Ollama (if not already running):
ollama serve
  1. Start the MCP server:
python main.py

The server will be available at http://localhost:8000

API Endpoints

Get Context

curl -X POST http://localhost:8000/context \
  -H "Content-Type: application/json" \
  -d '{
    "query": "Tell me about iPhone 15",
    "model": "mistral"
  }'

List Available Models

curl http://localhost:8000/models

Health Check

curl http://localhost:8000/health

Project Structure

mcp-server/
├── context_providers/     # Data source providers
│   ├── database.py       # Database provider
│   ├── graphql.py        # GraphQL provider
│   ├── rest.py          # REST API provider
│   └── provider_factory.py
├── model_providers/      # AI model providers
│   ├── base.py          # Base model provider
│   ├── ollama.py        # Ollama integration
│   └── provider_factory.py
├── main.py              # FastAPI application
├── query_analyzer.py    # Query analysis logic
├── logger_config.py     # Logging configuration
├── requirements.txt     # Project dependencies
└── README.md           # Project documentation

Development

Adding New Providers

  1. Create a new provider class in the appropriate directory
  2. Implement the required interface methods
  3. Register the provider in the factory

Adding New Models

  1. Add the model to the AVAILABLE_MODELS dictionary in model_providers/ollama.py
  2. Update the model validation logic if needed

Contributing

  1. Fork the repository
  2. Create a feature branch
  3. Commit your changes
  4. Push to the branch
  5. Create a Pull Request

License

This project is licensed under the MIT License - see the LICENSE file for details.

Recommended Servers

playwright-mcp

playwright-mcp

A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.

Official
Featured
TypeScript
Magic Component Platform (MCP)

Magic Component Platform (MCP)

An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.

Official
Featured
Local
TypeScript
Audiense Insights MCP Server

Audiense Insights MCP Server

Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.

Official
Featured
Local
TypeScript
VeyraX MCP

VeyraX MCP

Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.

Official
Featured
Local
graphlit-mcp-server

graphlit-mcp-server

The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.

Official
Featured
TypeScript
Kagi MCP Server

Kagi MCP Server

An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.

Official
Featured
Python
E2B

E2B

Using MCP to run code via e2b.

Official
Featured
Neon Database

Neon Database

MCP server for interacting with Neon Management API and databases

Official
Featured
Exa Search

Exa Search

A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.

Official
Featured
Qdrant Server

Qdrant Server

This repository is an example of how to create a MCP server for Qdrant, a vector search engine.

Official
Featured