Oxylabs MCP
A scraper tool that leverages the Oxylabs Web Scraper API to fetch and process web content with flexible options for parsing and rendering pages, enabling efficient content extraction from complex websites.
oxylabs
Tools
oxylabs_scraper
Scrape url using Oxylabs Web Api
oxylabs_web_unblocker
Scrape url using Oxylabs Web Unblocker
README
MCP Server for Oxylabs Scraper
A Model Context Protocol (MCP) server that enables AI assistants like Claude to seamlessly access web data through Oxylabs' powerful web scraping technology.
📖 Overview
The Oxylabs MCP server provides a bridge between AI models and the web. It enables them to scrape any URL, render JavaScript-heavy pages, extract and format content for AI use, bypass anti-scraping measures, and access geo-restricted web data from 195+ countries.
This implementation leverages the Model Context Protocol (MCP) to create a secure, standardized way for AI assistants to interact with web content.
✨ Key Features
<details> <summary><strong> Scrape content from any site</strong></summary> <br>
- Extract data from any URL, including complex single-page applications
- Fully render dynamic websites using headless browser support
- Choose full JavaScript rendering, HTML-only, or none
- Emulate Mobile and Desktop viewports for realistic rendering
</details>
<details> <summary><strong> Automatically get AI-ready data</strong></summary> <br>
- Automatically clean and convert HTML to Markdown for improved readability
- Use automated parsers for popular targets like Google, Amazon, and etc.
</details>
<details> <summary><strong> Bypass blocks & geo-restrictions</strong></summary> <br>
- Bypass sophisticated bot protection systems with high success rate
- Reliably scrape even the most complex websites
- Get automatically rotating IPs from a proxy pool covering 195+ countries
</details>
<details> <summary><strong> Flexible setup & cross-platform support</strong></summary> <br>
- Set rendering and parsing options if needed
- Feed data directly into AI models or analytics tools
- Works on macOS, Windows, and Linux
</details>
<details> <summary><strong> Built-in error handling and request management</strong></summary> <br>
- Comprehensive error handling and reporting
- Smart rate limiting and request management
</details>
💡 Example Queries
When you've set up the MCP server with Claude, you can make requests like:
- Could you scrape
https://www.google.com/search?q=ai
page? - Scrape
https://www.amazon.de/-/en/Smartphone-Contract-Function-Manufacturer-Exclusive/dp/B0CNKD651V
with parse enabled - Scrape
https://www.amazon.de/-/en/gp/bestsellers/beauty/ref=zg_bs_nav_beauty_0
with parse and render enabled - Use web unblocker with render to scrape
https://www.bestbuy.com/site/top-deals/all-electronics-on-sale/pcmcat1674241939957.c
✅ Prerequisites
Before you begin, make sure you have:
- Oxylabs Account: Obtain your username and password from Oxylabs (1-week free trial available)
Basic Usage
Via Smithery CLI:
- Node.js (v16+)
npx
command-line tool
Via uv:
uv
package manager – install it using this guide
Local/Dev Setup
- Python 3.12+
uv
package manager – install it using this guide
🧩 API Parameters
The Oxylabs MCP server supports these parameters:
Parameter | Description | Values |
---|---|---|
url |
The URL to scrape | Any valid URL |
parse |
Enable structured data extraction | True or False |
render |
Use headless browser rendering | html or None |
⚙️ Basic Setup Instructions
Install via Smithery
Automatically install Oxylabs MCP server for Claude Desktop via Smithery:
npx -y @smithery/cli install @oxylabs/oxylabs-mcp --client claude
Install using uv in Claude Desktop
With uv
installed, this method will automatically set up the Oxylabs MCP server in Claude Desktop. Navigate to Claude → Settings → Developer → Edit Config and edit your claude_desktop_config.json
file as follows:
{
"mcpServers": {
"oxylabs_scraper": {
"command": "uvx",
"args": ["oxylabs-mcp"],
"env": {
"OXYLABS_USERNAME": "YOUR_USERNAME_HERE",
"OXYLABS_PASSWORD": "YOUR_PASSWORD_HERE"
}
}
}
}
[!TIP] If you run into errors, try using the full path to
uvx
in thecommand
field. For example,/Users/my-user/.local/bin/uvx
.
💻 Local/Dev Setup Instructions
Clone repository
git clone <git:url>
Install dependencies
Install MCP server dependencies:
cd mcp-server-oxylabs
# Create virtual environment and activate it
uv venv
source .venv/bin/activate # MacOS/Linux
# OR
.venv/Scripts/activate # Windows
# Install dependencies
uv sync
Setup with Claude Desktop
Navigate to Claude → Settings → Developer → Edit Config and edit your claude_desktop_config.json
file as follows:
{
"mcpServers": {
"oxylabs_scraper": {
"command": "uv",
"args": [
"--directory",
"/<Absolute-path-to-folder>/oxylabs-mcp",
"run",
"oxylabs-mcp"
],
"env": {
"OXYLABS_USERNAME": "YOUR_USERNAME_HERE",
"OXYLABS_PASSWORD": "YOUR_PASSWORD_HERE"
}
}
}
}
🐞 Debugging
make run
Then access MCP Inspector at http://localhost:5173
. You may need to add your username and password as environment variables in the inspector under OXYLABS_USERNAME
and OXYLABS_PASSWORD
.
🛠️ Technical Details
This server provides two main tools:
- oxylabs_scraper: Uses Oxylabs Web Scraper API for general website scraping
- oxylabs_web_unblocker: Uses Oxylabs Web Unblocker for hard-to-access websites
Web Scraper API supports JavaScript rendering, parsed structured data, and cleaned HTML in Markdown format. Web Unblocker offers JavaScript rendering and cleaned HTML, but doesn’t return parsed data.
[!WARNING] Usage with the MCP Inspector is affected by an ongoing issue with the Python SDK for MCP, see: https://github.com/modelcontextprotocol/python-sdk/pull/85. For Claude, a forked version of the SDK is used as a temporary fix.
License
This project is licensed under the MIT License.
About Oxylabs
Established in 2015, Oxylabs is a market-leading web intelligence collection platform, driven by the highest business, ethics, and compliance standards, enabling companies worldwide to unlock data-driven insights.
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Mult Fetch MCP Server
A versatile MCP-compliant web content fetching tool that supports multiple modes (browser/node), formats (HTML/JSON/Markdown/Text), and intelligent proxy detection, with bilingual interface (English/Chinese).
Persistent Knowledge Graph
An implementation of persistent memory for Claude using a local knowledge graph, allowing the AI to remember information about users across conversations with customizable storage location.
Hyperbrowser MCP Server
Welcome to Hyperbrowser, the Internet for AI. Hyperbrowser is the next-generation platform empowering AI agents and enabling effortless, scalable browser automation. Built specifically for AI developers, it eliminates the headaches of local infrastructure and performance bottlenecks, allowing you to
React MCP
react-mcp integrates with Claude Desktop, enabling the creation and modification of React apps based on user prompts
Exa MCP
A Model Context Protocol server that enables AI assistants like Claude to perform real-time web searches using the Exa AI Search API in a safe and controlled manner.
Perplexity Chat MCP Server
MCP Server for the Perplexity API.
Web Research Server
A Model Context Protocol server that enables Claude to perform web research by integrating Google search, extracting webpage content, and capturing screenshots.
Browser Use (used by Deploya.dev)
AI-driven browser automation server that implements the Model Context Protocol to enable natural language control of web browsers for tasks like navigation, form filling, and visual interaction.
PubMedSearch
A Model Content Protocol server that provides tools to search and retrieve academic papers from PubMed database.