
LinkedIn Profile Scraper MCP Server
MCP server that fetches LinkedIn profile information using the Fresh LinkedIn Profile Data API, allowing users to retrieve profile data in JSON format by providing a LinkedIn profile URL.
Tools
get_profile
Get LinkedIn profile data for a given profile URL. Args: linkedin_url: The LinkedIn profile URL.
README
LinkedIn Profile Scraper MCP Server
This MCP server uses the Fresh LinkedIn Profile Data API to fetch LinkedIn profile information. It is implemented as a model context protocol (MCP) server and exposes a single tool, get_profile
, which accepts a LinkedIn profile URL and returns the profile data in JSON format.
Features
- Fetch Profile Data: Retrieves LinkedIn profile information including skills and other settings (with most additional details disabled).
- Asynchronous HTTP Requests: Uses
httpx
for non-blocking API calls. - Environment-based Configuration: Reads the
RAPIDAPI_KEY
from your environment variables usingdotenv
.
Prerequisites
- Python 3.7+ – Ensure you are using Python version 3.7 or higher.
- MCP Framework: Make sure the MCP framework is installed.
- Required Libraries: Install
httpx
,python-dotenv
, and other dependencies. - RAPIDAPI_KEY: Obtain an API key from RapidAPI and add it to a
.env
file in your project directory (or set it in your environment).
Installation
-
Clone the Repository:
git clone https://github.com/AIAnytime/Awesome-MCP-Server cd linkedin_profile_scraper
-
Install Dependencies:
uv add mcp[cli] httpx requests
-
Set Up Environment Variables:
Create a
.env
file in the project directory with the following content:RAPIDAPI_KEY=your_rapidapi_key_here
Running the Server
To run the MCP server, execute:
uv run linkedin.py
The server will start and listen for incoming requests via standard I/O.
MCP Client Configuration
To connect your MCP client to this server, add the following configuration to your config.json
. Adjust the paths as necessary for your environment:
{
"mcpServers": {
"linkedin_profile_scraper": {
"command": "C:/Users/aiany/.local/bin/uv",
"args": [
"--directory",
"C:/Users/aiany/OneDrive/Desktop/YT Video/linkedin-mcp/project",
"run",
"linkedin.py"
]
}
}
}
Code Overview
- Environment Setup: The server uses
dotenv
to load theRAPIDAPI_KEY
required to authenticate with the Fresh LinkedIn Profile Data API. - API Call: The asynchronous function
get_linkedin_data
makes a GET request to the API with specified query parameters. - MCP Tool: The
get_profile
tool wraps the API call and returns formatted JSON data, or an error message if the call fails. - Server Execution: The MCP server is run with the
stdio
transport.
Troubleshooting
- Missing RAPIDAPI_KEY: If the key is not set, the server will raise a
ValueError
. Make sure the key is added to your.env
file or set in your environment. - API Errors: If the API request fails, the tool will return a message indicating that the profile data could not be fetched.
License
This project is licensed under the MIT License. See the LICENSE file for more details.
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.

VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.

E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.