MCP API Gateway
A unified local API gateway providing caching, rate limiting, and full Model Context Protocol compatibility for AI agent integration. It enables users to aggregate multiple API endpoints into a single gateway with built-in observability and customizable eviction strategies.
README
MCP API Gateway
A unified local API gateway with caching, rate limiting, and full MCP (Model Context Protocol) compatibility.
Features
- π Unified API Aggregation - Manage multiple API endpoints through a single gateway
- πΎ Multi-Strategy Caching - LRU, LFU, FIFO, and TTL cache eviction policies
- β‘ Rate Limiting - Token bucket and sliding window algorithms
- π MCP Protocol - Full Model Context Protocol support for AI agent integration
- π Observability - Built-in statistics and metrics
- π Retry Logic - Automatic retry with exponential backoff
Installation
# Clone the repository
git clone https://github.com/bandageok/mcp-api-gateway.git
cd mcp-api-gateway
# Install dependencies
pip install -r requirements.txt
# Or install directly
pip install aiohttp pyyaml
Quick Start
1. Create a Configuration File
python gateway.py --create-config
This creates a config.yaml with sample endpoints:
host: localhost
port: 8080
cache:
enabled: true
max_size: 1000
ttl: 300
strategy: lru
rate_limit:
enabled: true
requests_per_minute: 60
apis:
- name: github-api
url: https://api.github.com
method: GET
2. Run the Gateway
# With config file
python gateway.py -c config.yaml
# Or with command line arguments
python gateway.py --host 0.0.0.0 --port 8080
3. Use the Gateway
# Call an API endpoint
curl http://localhost:8080/api/github-api/users/bandageok
# Check health
curl http://localhost:8080/health
# Get statistics
curl http://localhost:8080/stats
# Clear cache
curl -X DELETE http://localhost:8080/cache/clear
# Get configuration
curl http://localhost:8080/config
MCP Protocol Integration
The gateway provides full MCP protocol support for AI agents:
MCP Tools
{
"jsonrpc": "2.0",
"id": 1,
"method": "tools/list",
"params": {}
}
Response:
{
"jsonrpc": "2.0",
"id": 1,
"result": {
"tools": [
{
"name": "github-api",
"description": "Call GET https://api.github.com",
"inputSchema": {
"type": "object",
"properties": {
"params": {"type": "object"},
"data": {"type": "object"}
}
}
}
]
}
}
Call a Tool
{
"jsonrpc": "2.0",
"id": 2,
"method": "tools/call",
"params": {
"name": "github-api",
"arguments": {
"params": {"path": "/users/bandageok"}
}
}
}
MCP Resources
{
"jsonrpc": "2.0",
"id": 3,
"method": "resources/list",
"params": {}
}
Configuration Options
| Option | Type | Default | Description |
|---|---|---|---|
host |
string | localhost | Host to bind to |
port |
int | 8080 | Port to bind to |
debug |
bool | false | Enable debug mode |
log_level |
string | INFO | Logging level |
cache.enabled |
bool | true | Enable caching |
cache.max_size |
int | 1000 | Maximum cache entries |
cache.ttl |
int | 300 | Cache TTL in seconds |
cache.strategy |
string | lru | Cache strategy (lru/lfu/fifo/ttl) |
rate_limit.enabled |
bool | true | Enable rate limiting |
rate_limit.requests_per_minute |
int | 60 | Rate limit threshold |
API Endpoints
| Endpoint | Method | Description |
|---|---|---|
/ |
GET | Health check |
/health |
GET | Detailed health status |
/stats |
GET | Gateway statistics |
/config |
GET | Current configuration |
/cache/clear |
DELETE | Clear the cache |
/api/{name} |
* | Proxy to configured API |
/mcp |
POST | MCP protocol endpoint |
Architecture
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β MCP API Gateway β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β βββββββββββββββ βββββββββββββββ βββββββββββββββββ β
β β Cache β βRate Limiter β β MCP Handler β β
β β (LRU/LFU) β β (Token) β β β β
β βββββββββββββββ βββββββββββββββ βββββββββββββββββ β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β API Client Pool β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β ββββββββββββ ββββββββββββ ββββββββββββ ββββββββββββ β
β β GitHub β β Weather β β Stocks β β Custom β β
β β API β β API β β API β β API β β
β ββββββββββββ ββββββββββββ ββββββββββββ ββββββββββββ β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
Use Cases
1. AI Agent Integration
Connect AI agents to external APIs through MCP:
import requests
# Initialize MCP
response = requests.post("http://localhost:8080/mcp", json={
"jsonrpc": "2.0",
"id": 1,
"method": "initialize",
"params": {}
})
# List available tools
response = requests.post("http://localhost:8080/mcp", json={
"jsonrpc": "2.0",
"id": 2,
"method": "tools/list",
"params": {}
})
2. API Rate Limiting
Protect external APIs from being overwhelmed:
rate_limit:
enabled: true
requests_per_minute: 60 # Max 60 requests per minute
3. Response Caching
Cache expensive API responses:
cache:
enabled: true
max_size: 1000
ttl: 300 # Cache for 5 minutes
strategy: lru # Evict least recently used
Examples
Python Client
import aiohttp
import asyncio
async def call_gateway():
async with aiohttp.ClientSession() as session:
# Call an API
async with session.get("http://localhost:8080/api/github-api/users/bandageok") as resp:
data = await resp.json()
print(data)
# Check stats
async with session.get("http://localhost:8080/stats") as resp:
stats = await resp.json()
print(f"Cache hit rate: {stats['cache_hit_rate']}")
asyncio.run(call_gateway())
Add Custom API Endpoint
apis:
- name: my-api
url: https://api.example.com
method: GET
headers:
Authorization: Bearer YOUR_TOKEN
timeout: 30
retry_count: 3
Performance
- Throughput: ~1000 requests/second (with caching)
- Latency: <10ms overhead (cache hit), <100ms overhead (cache miss)
- Memory: ~50MB base + cache size
License
MIT License - See LICENSE for details.
Author
- BandageOK - GitHub
Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
β Star us on GitHub if you find this useful!
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.