Gladia MCP
Official Model Context Protocol server that enables interaction with powerful Speech-to-Text and Audio Intelligence APIs, allowing clients like Claude Desktop to transcribe audio, analyze speech, translate content, and more.
README
Gladia MCP
<div class="title-block" style="text-align: center;" align="center">
</div>
<p align="center"> Official Gladia <a href="https://github.com/modelcontextprotocol">Model Context Protocol (MCP)</a> server that enables interaction with powerful Speech-to-Text and Audio Intelligence APIs. This server allows MCP clients like <a href="https://www.anthropic.com/claude">Claude Desktop</a>, <a href="https://www.cursor.so">Cursor</a>, <a href="https://codeium.com/windsurf">Windsurf</a>, <a href="https://github.com/openai/openai-agents-python">OpenAI Agents</a> and others to transcribe audio, analyze speech, translate content, and more. </p>
Features
- Audio transcription with speaker diarization
- Real-time speech-to-text
- Audio intelligence capabilities:
- Translation
- Summarization
- Named Entity Recognition
- Sentiment Analysis
- Content Moderation
- Chapterization
- Audio to LLM integration
- Async API with FastAPI
- Easy-to-use CLI interface
- Configurable logging
- CORS support
- Health check endpoint
Quickstart with Claude Desktop
- Get your API key from Gladia. There is a free tier available.
- Install
uv(Python package manager), install withcurl -LsSf https://astral.sh/uv/install.sh | shor see theuvrepo for additional install methods. - Go to Claude > Settings > Developer > Edit Config > claude_desktop_config.json to include the following:
{
"mcpServers": {
"Gladia": {
"command": "uvx",
"args": ["gladia-mcp"],
"env": {
"GLADIA_API_KEY": "<insert-your-api-key-here>"
}
}
}
}
If you're using Windows, you will have to enable "Developer Mode" in Claude Desktop to use the MCP server. Click "Help" in the hamburger menu at the top left and select "Enable Developer Mode".
Other MCP clients
For other clients like Cursor and Windsurf, run:
pip install gladia-mcppython -m gladia_mcp --api-key={{PUT_YOUR_API_KEY_HERE}} --printto get the configuration. Paste it into appropriate configuration directory specified by your MCP client.
Example usage
Try asking Claude:
- "Transcribe this audio file and identify different speakers"
- "Convert this recording to text and translate it to Spanish"
- "Analyze the sentiment and emotions in this speech"
- "Extract key topics and create chapters from this long audio file"
- "Transcribe this conversation and summarize the main points"
Optional features
You can add the GLADIA_MCP_BASE_PATH environment variable to the claude_desktop_config.json to specify the base path MCP server should look for and output files specified with relative paths.
Contributing
If you want to contribute or run from source:
- Clone the repository:
git clone https://github.com/gladia/gladia-mcp
cd gladia-mcp
- Create a virtual environment and install dependencies using uv:
uv venv
source .venv/bin/activate
uv pip install -e ".[dev]"
- Copy
.env.exampleto.envand add your Gladia API key:
cp .env.example .env
# Edit .env and add your API key
- Run the tests to make sure everything is working:
./scripts/test.sh
# Or with options
./scripts/test.sh --verbose --fail-fast
-
Install the server in Claude Desktop:
mcp install gladia_mcp/server.py -
Debug and test locally with MCP Inspector:
mcp dev gladia_mcp/server.py
API Endpoints
Health Check
GET /health
Transcribe Audio
POST /transcribe
Parameters:
file: Audio file (multipart/form-data)diarization: Enable speaker diarization (boolean, optional)language: Language code (string, optional)
Example using curl:
curl -X POST "http://localhost:8000/transcribe" \
-H "accept: application/json" \
-H "Content-Type: multipart/form-data" \
-F "file=@audio.wav" \
-F "diarization=true"
Troubleshooting
Logs when running with Claude Desktop can be found at:
- Windows:
%APPDATA%\Claude\logs\mcp-server-gladia.log - macOS:
~/Library/Logs/Claude/mcp-server-gladia.log
MCP Gladia: spawn uvx ENOENT
If you encounter the error "MCP Gladia: spawn uvx ENOENT", confirm its absolute path by running this command in your terminal:
which uvx
Once you obtain the absolute path (e.g., /usr/local/bin/uvx), update your configuration to use that path (e.g., "command": "/usr/local/bin/uvx"). This ensures that the correct executable is referenced.
Development
Running Tests
pytest
Code Style
The project follows PEP 8 style guide. Use flake8 for linting:
flake8 gladia_mcp
License
MIT License
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.