MCP Host CLI

MCP Host CLI

Local http server that proxies requests o LLMs and uses mcp-servers if needed

VyacheslavVanin

Developer Tools
Visit Server

README

MCP Host CLI

A FastAPI-based CLI application that hosts and manages MCP (Model Context Protocol) servers, providing an HTTP API for interacting with tools and resources.

Features

  • Manages multiple MCP server connections
  • Provides HTTP API endpoints for:
    • User requests
    • Tool approval workflow
    • Session state management
  • Supports both direct LLM API and Ollama local models

Installation

  1. Clone the repository:
git clone https://github.com/VyacheslavVanin/mcp-host-cli.git
cd mcp-host-cli
  1. Run:
uv run main.py

Configuration

Server Configuration

  1. Create/edit servers_config.json to configure your MCP servers:
{
  "mcpServers": {
    "server-name": {
      "command": "node",
      "args": ["path/to/server.js"],
      "env": {
        "API_KEY": "your-api-key"
      }
    }
  }
}

Application Configuration

Configuration can be set via environment variables or command line arguments (CLI args take precedence).

Environment Variables

  • LLM_API_KEY: API key for LLM service (if not using Ollama)
  • LLM_PROVIDER: "ollama" (default) or "openai"
  • LLM_MODEL: Model name (default: "qwen2.5-coder:latest")
  • PORT: Server port (default: 8000)
  • OPENAI_BASE_URL: Base URL for OpenAI-compatible API (default: "https://openrouter.ai/api/v1")
  • USE_OLLAMA: Set to "true" to use local Ollama models

Command Line Arguments

python main.py --model MODEL_NAME --port PORT_NUMBER --provider PROVIDER --openai-base-url URL

Where:

Configuration Precedence

  1. Command line arguments (highest priority)
  2. Environment variables
  3. Default values (lowest priority)

Examples

# Using environment variables
export LLM_MODEL="llama3:latest"
export PORT=8080
python main.py

# Using CLI arguments
python main.py --model "llama3:latest" --port 8080

# Using defaults
python main.py

API Endpoints

POST /user_request

Handle user input and return LLM response or tool approval request.

Request:

{
  "input": "your question or command"
}

Response:

{
  "message": "response text",
  "request_id": "uuid-if-approval-needed",
  "requires_approval": true/false,
  "tool": "tool-name-if-applicable"
}

POST /approve

Approve or deny a tool execution request.

Request:

{
  "request_id": "uuid-from-user_request",
  "approve": true/false
}

Response:

{
  "message": "execution result or denial message",
  "request_id": "same-request-id",
  "tool": "tool-name"
}

GET /session_state

Get current chat session state including messages and pending requests.

Response:

{
  "messages": [
    {"role": "system/user/assistant", "content": "message text"}
  ],
  "_pending_request_id": "uuid-or-null",
  "_pending_tool_call": {
    "tool": "tool-name",
    "arguments": {}
  }
}

Recommended Servers

playwright-mcp

playwright-mcp

A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.

Official
Featured
TypeScript
Magic Component Platform (MCP)

Magic Component Platform (MCP)

An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.

Official
Featured
Local
TypeScript
MCP Package Docs Server

MCP Package Docs Server

Facilitates LLMs to efficiently access and fetch structured documentation for packages in Go, Python, and NPM, enhancing software development with multi-language support and performance optimization.

Featured
Local
TypeScript
Claude Code MCP

Claude Code MCP

An implementation of Claude Code as a Model Context Protocol server that enables using Claude's software engineering capabilities (code generation, editing, reviewing, and file operations) through the standardized MCP interface.

Featured
Local
JavaScript
@kazuph/mcp-taskmanager

@kazuph/mcp-taskmanager

Model Context Protocol server for Task Management. This allows Claude Desktop (or any MCP client) to manage and execute tasks in a queue-based system.

Featured
Local
JavaScript
Linear MCP Server

Linear MCP Server

Enables interaction with Linear's API for managing issues, teams, and projects programmatically through the Model Context Protocol.

Featured
JavaScript
mermaid-mcp-server

mermaid-mcp-server

A Model Context Protocol (MCP) server that converts Mermaid diagrams to PNG images.

Featured
JavaScript
Jira-Context-MCP

Jira-Context-MCP

MCP server to provide Jira Tickets information to AI coding agents like Cursor

Featured
TypeScript
Linear MCP Server

Linear MCP Server

A Model Context Protocol server that integrates with Linear's issue tracking system, allowing LLMs to create, update, search, and comment on Linear issues through natural language interactions.

Featured
JavaScript
Sequential Thinking MCP Server

Sequential Thinking MCP Server

This server facilitates structured problem-solving by breaking down complex issues into sequential steps, supporting revisions, and enabling multiple solution paths through full MCP integration.

Featured
Python