MCP Host CLI
Local http server that proxies requests o LLMs and uses mcp-servers if needed
VyacheslavVanin
README
MCP Host CLI
A FastAPI-based CLI application that hosts and manages MCP (Model Context Protocol) servers, providing an HTTP API for interacting with tools and resources.
Features
- Manages multiple MCP server connections
- Provides HTTP API endpoints for:
- User requests
- Tool approval workflow
- Session state management
- Supports both direct LLM API and Ollama local models
Installation
- Clone the repository:
git clone https://github.com/VyacheslavVanin/mcp-host-cli.git
cd mcp-host-cli
- Run:
uv run main.py
Configuration
Server Configuration
- Create/edit
servers_config.json
to configure your MCP servers:
{
"mcpServers": {
"server-name": {
"command": "node",
"args": ["path/to/server.js"],
"env": {
"API_KEY": "your-api-key"
}
}
}
}
Application Configuration
Configuration can be set via environment variables or command line arguments (CLI args take precedence).
Environment Variables
LLM_API_KEY
: API key for LLM service (if not using Ollama)LLM_PROVIDER
: "ollama" (default) or "openai"LLM_MODEL
: Model name (default: "qwen2.5-coder:latest")PORT
: Server port (default: 8000)OPENAI_BASE_URL
: Base URL for OpenAI-compatible API (default: "https://openrouter.ai/api/v1")USE_OLLAMA
: Set to "true" to use local Ollama models
Command Line Arguments
python main.py --model MODEL_NAME --port PORT_NUMBER --provider PROVIDER --openai-base-url URL
Where:
- PROVIDER is either "ollama" (default) or "openai"
- URL is the base URL for OpenAI-compatible API (default: "https://openrouter.ai/api/v1")
Configuration Precedence
- Command line arguments (highest priority)
- Environment variables
- Default values (lowest priority)
Examples
# Using environment variables
export LLM_MODEL="llama3:latest"
export PORT=8080
python main.py
# Using CLI arguments
python main.py --model "llama3:latest" --port 8080
# Using defaults
python main.py
API Endpoints
POST /user_request
Handle user input and return LLM response or tool approval request.
Request:
{
"input": "your question or command"
}
Response:
{
"message": "response text",
"request_id": "uuid-if-approval-needed",
"requires_approval": true/false,
"tool": "tool-name-if-applicable"
}
POST /approve
Approve or deny a tool execution request.
Request:
{
"request_id": "uuid-from-user_request",
"approve": true/false
}
Response:
{
"message": "execution result or denial message",
"request_id": "same-request-id",
"tool": "tool-name"
}
GET /session_state
Get current chat session state including messages and pending requests.
Response:
{
"messages": [
{"role": "system/user/assistant", "content": "message text"}
],
"_pending_request_id": "uuid-or-null",
"_pending_tool_call": {
"tool": "tool-name",
"arguments": {}
}
}
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
MCP Package Docs Server
Facilitates LLMs to efficiently access and fetch structured documentation for packages in Go, Python, and NPM, enhancing software development with multi-language support and performance optimization.
Claude Code MCP
An implementation of Claude Code as a Model Context Protocol server that enables using Claude's software engineering capabilities (code generation, editing, reviewing, and file operations) through the standardized MCP interface.
@kazuph/mcp-taskmanager
Model Context Protocol server for Task Management. This allows Claude Desktop (or any MCP client) to manage and execute tasks in a queue-based system.
Linear MCP Server
Enables interaction with Linear's API for managing issues, teams, and projects programmatically through the Model Context Protocol.
mermaid-mcp-server
A Model Context Protocol (MCP) server that converts Mermaid diagrams to PNG images.
Jira-Context-MCP
MCP server to provide Jira Tickets information to AI coding agents like Cursor

Linear MCP Server
A Model Context Protocol server that integrates with Linear's issue tracking system, allowing LLMs to create, update, search, and comment on Linear issues through natural language interactions.

Sequential Thinking MCP Server
This server facilitates structured problem-solving by breaking down complex issues into sequential steps, supporting revisions, and enabling multiple solution paths through full MCP integration.