fal
MCP server for interacting with fal.ai models and services. Uses the latest streaming MCP support.
README
fal MCP Server
A Model Context Protocol (MCP) server for interacting with fal.ai models and services. This project was inspired by am0y's MCP server, but updated to use the latest streaming MCP support.
Features
- List all available fal.ai models
- Search for specific models by keywords
- Get model schemas
- Generate content using any fal.ai model
- Support for both direct and queued model execution
- Queue management (status checking, getting results, cancelling requests)
- File upload to fal.ai CDN
- Full streaming support via HTTP transport
Requirements
- Python 3.12+
- fastmcp
- httpx
- aiofiles
- A fal.ai API key
Installation
- Clone this repository:
git clone https://github.com/derekalia/fal.git
cd fal
- Install the required packages:
# Using uv (recommended)
uv sync
# Or using pip
pip install fastmcp httpx aiofiles
Usage
Running the Server Locally
-
Get your fal.ai API key from fal.ai
-
Start the MCP server with HTTP transport:
./run_http.sh YOUR_FAL_API_KEY
The server will start and display connection information in your terminal.
- Connect to it from your LLM IDE (Claude Code or Cursor) by adding to your configuration:
{
"Fal": {
"url": "http://127.0.0.1:6274/mcp/"
}
}
Development Mode (with MCP Inspector)
For testing and debugging, you can run the server in development mode:
fastmcp dev main.py
This will:
- Start the server on a random port
- Launch the MCP Inspector web interface in your browser
- Allow you to test all tools interactively with a web UI
The Inspector URL will be displayed in the terminal (typically http://localhost:PORT).
Environment Variables
The run_http.sh script automatically handles all environment variables for you. If you need to customize:
PORT: Server port for HTTP transport (default: 6274)
Setting API Key Permanently
If you prefer to set your API key permanently instead of passing it each time:
- Create a
.envfile in the project root:
echo 'FAL_KEY="YOUR_FAL_API_KEY_HERE"' > .env
- Then run the server without the API key argument:
./run_http.sh
For manual setup:
FAL_KEY: Your fal.ai API key (required)MCP_TRANSPORT: Transport mode -stdio(default) orhttp
Available Tools
models(page=None, total=None)- List available models with optional paginationsearch(keywords)- Search for models by keywordsschema(model_id)- Get OpenAPI schema for a specific modelgenerate(model, parameters, queue=False)- Generate content using a modelresult(url)- Get result from a queued requeststatus(url)- Check status of a queued requestcancel(url)- Cancel a queued requestupload(path)- Upload a file to fal.ai CDN
License
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.