Telegram Bridge MCP

Telegram Bridge MCP

Bridges AI assistants to a Telegram bot to enable two-way messaging, interactive confirmations, and live status updates. It supports automatic voice transcription via local Whisper models and provides secure, single-user communication for MCP-compatible hosts.

Category
Visit Server

README

Telegram Bridge MCP

Unblock your agent workflow through Telegram

Telegram Bridge MCP

A Model Context Protocol server that bridges AI assistants to a Telegram bot — enabling two-way messaging, interactive confirmations, live status updates, and automatic voice transcription.

Works with any MCP-compatible AI host: VS Code Copilot, Claude Desktop, and others.

[!NOTE] Pre-release: This project is functional but has not yet been widely tested in production. Expect rough edges and possible breaking changes.


What it does

Once configured, your AI assistant can:

  • Send messages to your Telegram chat — plain text, formatted Markdown, photos
  • Ask questions and wait for your reply — as free text or button choices
  • Post live status updates — an in-place checklist that updates as tasks progress
  • React to messages — emoji reactions instead of noise text
  • Transcribe voice messages — speak your reply; it arrives as text
  • Send and receive files — send documents/photos from disk or URL; receive any file type and download on demand
  • Receive all of this in real time — long-polling, no webhooks, no public URL needed

Prerequisites

  • Node.js 18+nodejs.org
  • pnpm — install once via: npm install -g pnpm

If you prefer npm, you can substitute all pnpm commands with their npm equivalents (npm install, npm run build, etc.). The project works with either.


Quick Start

1. Clone and install

git clone https://github.com/electricessence/Telegram-Bridge-MCP.git
cd Telegram-Bridge-MCP
pnpm install
pnpm build

2. Create a Telegram bot

Open Telegram, message @BotFather, and run /newbot. Copy the token it gives you.

3. Pair the bot to your account

pnpm pair

This interactive wizard:

  1. Verifies your bot token
  2. Generates a one-time pairing code
  3. Waits for you to send that code to your bot in Telegram
  4. Captures your user ID and chat ID
  5. Writes everything to .env

4. Configure your MCP host

VS Code — add to .vscode/mcp.json:

{
  "servers": {
    "telegram": {
      "type": "stdio",
      "command": "node",
      "args": ["dist/index.js"],
      "cwd": "/absolute/path/to/telegram-bridge-mcp",
      "env": {
        "BOT_TOKEN": "YOUR_TOKEN",
        "ALLOWED_USER_ID": "YOUR_USER_ID",
        "ALLOWED_CHAT_ID": "YOUR_CHAT_ID"
      }
    }
  }
}

Claude Desktop — add to claude_desktop_config.json:

{
  "mcpServers": {
    "telegram": {
      "command": "node",
      "args": ["/absolute/path/to/telegram-bridge-mcp/dist/index.js"],
      "env": {
        "BOT_TOKEN": "YOUR_TOKEN",
        "ALLOWED_USER_ID": "YOUR_USER_ID",
        "ALLOWED_CHAT_ID": "YOUR_CHAT_ID"
      }
    }
  }
}

5. Start a session

Paste the contents of LOOP-PROMPT.md into your AI assistant's chat. It will connect, announce itself over Telegram, and wait for your instructions.


Tools

High-level (use these 99% of the time)

Tool What it does
get_agent_guide Loads the behavioral guide — call this at session start
set_topic Sets a default title prepended to all outbound messages as [Title] — e.g. [Refactor Agent]. Useful when multiple VS Code instances share one Telegram chat so you can tell which agent sent what. Pass empty string to clear.
notify Silent or audible notification with title, body, and severity
ask Sends a question; blocks until you reply with text
choose Sends a question with buttons; blocks until you tap one
send_confirmation Yes/No prompt wired to wait_for_callback_query
update_status Live in-place checklist — updates as steps complete

Messaging

send_message · edit_message_text · forward_message · delete_message · pin_message · send_chat_action · show_typing · cancel_typing

Files

send_document · send_photo · send_video · send_audio · send_voice · download_file

Interaction primitives

wait_for_message · wait_for_callback_query · answer_callback_query

Info & utilities

get_me · get_chat · set_commands · set_reaction · get_updates · restart_server

set_commands — registers (or clears) the bot's slash-command menu in the active chat. Pass [{command, description}, ...] to show commands in Telegram's autocomplete; pass [] to remove the menu.


Resources

Three guides are available as MCP resources — any MCP client can read them directly:

Resource URI Contents
telegram-bridge-mcp://agent-guide Behavioral guide for AI assistants
telegram-bridge-mcp://setup-guide Full bot setup walkthrough
telegram-bridge-mcp://formatting-guide Markdown/MarkdownV2/HTML reference

Security

The server enforces a strict two-layer security model:

  • ALLOWED_USER_ID — Inbound updates from any other user are silently discarded before the assistant ever sees them. Prevents message injection.
  • ALLOWED_CHAT_ID — Outbound tool calls to any other chat are rejected immediately. Prevents misdirected messages.

The server is designed for single-user, single-chat use — chat_id is never a tool parameter; it is resolved from config transparently.

See SETUP.md for the full security model and threat analysis.


Voice Transcription

All message-receiving tools (wait_for_message, ask, choose, get_updates) automatically transcribe voice messages using a local Whisper model via @huggingface/transformers (ONNX Runtime).

  • No external API calls
  • No ffmpeg required
  • Model weights are downloaded once on first use and cached locally

Configure via environment variables:

WHISPER_MODEL=onnx-community/whisper-base   # default
WHISPER_CACHE_DIR=/path/to/cache            # optional

Development

pnpm build          # Compile TypeScript
pnpm dev            # Watch mode
pnpm test           # Run tests
pnpm coverage       # Test coverage report
pnpm pair           # Re-run pairing wizard

License

MIT — see LICENSE.

Recommended Servers

playwright-mcp

playwright-mcp

A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.

Official
Featured
TypeScript
Audiense Insights MCP Server

Audiense Insights MCP Server

Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.

Official
Featured
Local
TypeScript
Magic Component Platform (MCP)

Magic Component Platform (MCP)

An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.

Official
Featured
Local
TypeScript
VeyraX MCP

VeyraX MCP

Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.

Official
Featured
Local
Kagi MCP Server

Kagi MCP Server

An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.

Official
Featured
Python
graphlit-mcp-server

graphlit-mcp-server

The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.

Official
Featured
TypeScript
E2B

E2B

Using MCP to run code via e2b.

Official
Featured
Neon Database

Neon Database

MCP server for interacting with Neon Management API and databases

Official
Featured
Exa Search

Exa Search

A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.

Official
Featured
Qdrant Server

Qdrant Server

This repository is an example of how to create a MCP server for Qdrant, a vector search engine.

Official
Featured