mcp-inception

mcp-inception

A TypeScript-based server that allows calling other MCP clients from your own MCP client, facilitating task delegation and context window offloading for enhanced multi-agent interactions.

tanevanwifferen

Remote Shell Execution
AI Integration Systems
Visit Server

Tools

execute_map_reduce_mcp_client

Process multiple items in parallel then sequentially reduce the results to a single output.

execute_mcp_client

Offload certain tasks to AI. Used for research purposes, do not use for code editing or anything code related. Only used to fetch data.

execute_parallel_mcp_client

Execute multiple AI tasks in parallel, with responses in JSON key-value pairs.

README

Disclaimer

Ok this is a difficult one. Will take some setting up unfortunately. However, if you manage to make this more straightforward, please send me PR's.

mcp-inception MCP Server

Call another mcp client from your mcp client. Delegate tasks, offload context windows. An agent for your agent!

This is a TypeScript-based MCP server that implements a simple LLM query system.

  • MCP Server and Client in one
  • Made with use of mcp-client-cli
  • Offload context windows
  • Delegate tasks
  • Parallel and map-reduce execution of tasks

<a href="https://glama.ai/mcp/servers/hedrd1hxv5"><img width="380" height="200" src="https://glama.ai/mcp/servers/hedrd1hxv5/badge" alt="Inception Server MCP server" /></a>

Features

Tools

  • execute_mcp_client - Ask a question to a separate LLM, ignore all the intermediate steps it takes when querying it's tools, and return the output.
    • Takes question as required parameters
    • Returns answer, ignoring all the intermediate context
  • execute_parallel_mcp_client - Takes a list of inputs and a main prompt, and executes the prompt in parallel for each string in the input. E.G. get the time of 6 major cities right now - London, Paris, Tokyo, Rio, New York, Sidney.
    • takes main prompt "What is the time in this city?"
    • takes list of inputs, London Paris etc
    • runs the prompt in parallel for each input
    • note: wait for this before using this feature
  • execute_map_reduce_mcp_client - Process multiple items in parallel and then sequentially reduce the results to a single output.
    • Takes mapPrompt with {item} placeholder for individual item processing
    • Takes reducePrompt with {accumulator} and {result} placeholders for combining results
    • Takes list of items to process
    • Optional initialValue for the accumulator
    • Processes items in parallel, then sequentially reduces results
    • Example use case: Analyze multiple documents, then synthesize key insights from all documents into a summary

Development

Dependencies:

  • Install mcp-client-cli
    • Also install the config file, and the mcp servers it needs in ~/.llm/config.json
  • create a bash file somewhere that activates the venv and executes the llm executable
#!/bin/bash
source ./venv/bin/activate
llm --no-confirmations

install package

Install dependencies:

npm install

Build the server:

npm run build

For development with auto-rebuild:

npm run watch

Installation

To use with Claude Desktop, add the server config:

On MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json On Windows: %APPDATA%/Claude/claude_desktop_config.json

{
  "mcpServers": {
    "mcp-inception": {
      "command": "node",
      "args": ["~/Documents/Cline/MCP/mcp-inception/build/index.js"], // build/index.js from this repo
      "disabled": false,
      "autoApprove": [],
      "env": {
        "MCP_INCEPTION_EXECUTABLE": "./run_llm.sh", // bash file from Development->Dependencies
        "MCP_INCEPTION_WORKING_DIR": "/mcp-client-cli working dir"
      }
    }
  }
}

Debugging

Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the MCP Inspector, which is available as a package script:

npm run inspector

The Inspector will provide a URL to access debugging tools in your browser.

Recommended Servers

playwright-mcp

playwright-mcp

A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.

Official
Featured
TypeScript
E2B

E2B

Using MCP to run code via e2b.

Official
Featured
Neon Database

Neon Database

MCP server for interacting with Neon Management API and databases

Official
Featured
Qdrant Server

Qdrant Server

This repository is an example of how to create a MCP server for Qdrant, a vector search engine.

Official
Featured
AIO-MCP Server

AIO-MCP Server

🚀 All-in-one MCP server with AI search, RAG, and multi-service integrations (GitLab/Jira/Confluence/YouTube) for AI-enhanced development workflows. Folk from

Featured
Local
React MCP

React MCP

react-mcp integrates with Claude Desktop, enabling the creation and modification of React apps based on user prompts

Featured
Local
Atlassian Integration

Atlassian Integration

Model Context Protocol (MCP) server for Atlassian Cloud products (Confluence and Jira). This integration is designed specifically for Atlassian Cloud instances and does not support Atlassian Server or Data Center deployments.

Featured
Any OpenAI Compatible API Integrations

Any OpenAI Compatible API Integrations

Integrate Claude with Any OpenAI SDK Compatible Chat Completion API - OpenAI, Perplexity, Groq, xAI, PyroPrompts and more.

Featured
Exa MCP

Exa MCP

A Model Context Protocol server that enables AI assistants like Claude to perform real-time web searches using the Exa AI Search API in a safe and controlled manner.

Featured
MySQL Server

MySQL Server

Allows AI assistants to list tables, read data, and execute SQL queries through a controlled interface, making database exploration and analysis safer and more structured.

Featured