FineData MCP Server
Enables AI agents to scrape any website by providing tools for JavaScript rendering, antibot bypass, and automatic captcha solving. It supports synchronous, asynchronous, and batch scraping operations with built-in proxy rotation.
README
FineData MCP Server
MCP (Model Context Protocol) server for FineData web scraping API.
Enables AI agents like Claude, Cursor, and GPT to scrape any website with:
- Antibot Bypass - Cloudflare, DataDome, PerimeterX, and more
- JavaScript Rendering - Full browser rendering with Playwright
- Captcha Solving - reCAPTCHA, hCaptcha, Cloudflare Turnstile, Yandex
- Proxy Rotation - 87K+ datacenter, residential, and mobile proxies
- Smart Retry - Automatic retries with block detection
Installation
Using uvx (Recommended)
# Install uv if you haven't already
curl -LsSf https://astral.sh/uv/install.sh | sh
# Run directly with uvx
FINEDATA_API_KEY=fd_xxx uvx finedata-mcp
Using pip
pip install finedata-mcp
# Run
FINEDATA_API_KEY=fd_xxx finedata-mcp
Using npx
npx -y @finedata/mcp-server
Configuration
Cursor IDE
Step 1: Open Cursor Settings → MCP
Or create/edit ~/.cursor/mcp.json:
macOS/Linux:
mkdir -p ~/.cursor && nano ~/.cursor/mcp.json
Windows:
%USERPROFILE%\.cursor\mcp.json
Step 2: Add FineData MCP server:
{
"mcpServers": {
"finedata": {
"command": "uvx",
"args": ["finedata-mcp"],
"env": {
"FINEDATA_API_KEY": "fd_your_api_key_here"
}
}
}
}
Step 3: Restart Cursor
Step 4: Test by asking the agent:
"Scrape https://example.com and show me the title"
Alternative: Using npx (if uv not installed)
{
"mcpServers": {
"finedata": {
"command": "npx",
"args": ["-y", "@finedata/mcp-server"],
"env": {
"FINEDATA_API_KEY": "fd_your_api_key_here"
}
}
}
}
Note: npx requires Python 3.10+ and uv/pipx installed. uvx is recommended.
Claude Desktop
Step 1: Open config file:
macOS: ~/Library/Application Support/Claude/claude_desktop_config.json
Windows: %APPDATA%\Claude\claude_desktop_config.json
Step 2: Add MCP server:
{
"mcpServers": {
"finedata": {
"command": "uvx",
"args": ["finedata-mcp"],
"env": {
"FINEDATA_API_KEY": "fd_your_api_key_here"
}
}
}
}
Step 3: Restart Claude Desktop
Environment Variables
| Variable | Required | Description |
|---|---|---|
FINEDATA_API_KEY |
Yes | Your FineData API key |
FINEDATA_API_URL |
No | API URL (default: https://api.finedata.ai) |
FINEDATA_TIMEOUT |
No | Default timeout in seconds (default: 60) |
Available Tools
scrape_url
Scrape content from any web page with antibot bypass.
scrape_url(
url: "https://example.com",
use_js_render: false, # Enable Playwright for SPAs
use_residential: false, # Use residential proxy
use_undetected: false, # Use Undetected Chrome
solve_captcha: false, # Auto-solve captchas
timeout: 60 # Timeout in seconds
)
Token costs:
- Base request: 1 token
- Antibot bypass: +2 tokens
- JS rendering: +5 tokens
- Residential proxy: +3 tokens
- Captcha solving: +10 tokens
scrape_async
Submit an async scraping job for long-running requests.
scrape_async(
url: "https://heavy-site.com",
use_js_render: true,
timeout: 120,
callback_url: "https://your-webhook.com/callback"
)
Returns a job_id for status polling.
get_job_status
Get the status of an async scraping job.
get_job_status(job_id: "550e8400-e29b-41d4-a716-446655440000")
Statuses: pending, processing, completed, failed, cancelled
batch_scrape
Scrape multiple URLs in a single batch (up to 100 URLs).
batch_scrape(
urls: ["https://example.com/1", "https://example.com/2"],
use_js_render: false,
callback_url: "https://your-webhook.com/batch-done"
)
get_usage
Get current API token usage.
get_usage()
Examples
Basic Scraping
Ask Claude or your AI agent:
"Scrape https://example.com and show me the content"
JavaScript Rendered Page
"Scrape https://spa-website.com with JavaScript rendering enabled"
Protected Site with Captcha
"Scrape https://protected-site.com using residential proxy and captcha solving"
Batch Scraping
"Scrape these URLs: https://example.com/1, https://example.com/2, https://example.com/3"
Pricing
FineData uses token-based pricing. Each feature adds tokens:
| Feature | Tokens |
|---|---|
| Base request | 1 |
| Antibot (TLS fingerprinting) | +2 |
| JS Rendering (Playwright) | +5 |
| Undetected Chrome | +5 |
| Residential Proxy | +3 |
| Mobile Proxy | +4 |
| reCAPTCHA / hCaptcha | +10 |
| Cloudflare Turnstile | +12 |
| Yandex SmartCaptcha | +15 |
Get your API key and free trial tokens at finedata.ai.
Troubleshooting
"No module named finedata_mcp"
Install uv and use uvx:
curl -LsSf https://astral.sh/uv/install.sh | sh
"externally-managed-environment" on macOS
This happens with Homebrew Python. Use uvx instead of pip:
{
"command": "uvx",
"args": ["finedata-mcp"]
}
MCP server not appearing in Cursor
- Check
~/.cursor/mcp.jsonsyntax (valid JSON) - Ensure
FINEDATA_API_KEYis set - Restart Cursor completely
- Check Cursor Output → MCP for errors
Support
- Documentation: https://docs.finedata.ai
- Email: support@finedata.ai
- Issues: https://github.com/quality-network/finedata-mcp/issues
License
MIT
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.