LinkedIn MCP Server
Enables searching and scraping of LinkedIn for structured data on people, companies, and job listings. It allows AI clients to retrieve detailed profiles, experience, and activity sections using browser automation.
README
LinkedIn MCP Server
A Model Context Protocol (MCP) server for LinkedIn. Search people, companies, and jobs, scrape profiles, and retrieve structured JSON data from any MCP-compatible AI client.
Built with FastMCP, Patchright, and a clean hexagonal architecture.
Features
| Category | Tools |
|---|---|
| People | get_person_profile · search_people |
| Companies | get_company_profile · get_company_posts |
| Jobs | get_job_details · search_jobs |
| Browser | close_browser |
Person Profile Sections
The get_person_profile tool supports granular section scraping. Request only the sections you need:
- Main profile (always included) — name, headline, location, followers, connections, about, profile image
- Experience — title, company, dates, duration, description, company logo
- Education — school, degree, dates, description, school logo
- Contact info — email, phone, websites, birthday, LinkedIn URL
- Interests — people, companies, and groups followed
- Honors and awards — title, issuer, description
- Languages — language name and proficiency level
- Posts — recent activity with reactions and timestamps
- Recommendations — received and given, with author details
Company Profile Sections
- About (always included) — overview, website, industry, size, headquarters, specialties, logo
- Posts — recent feed posts with engagement metrics
- Jobs — current open positions
Job Search Filters
The search_jobs tool supports the following filters:
| Filter | Values |
|---|---|
date_posted |
past_hour, past_24_hours, past_week, past_month |
job_type |
full_time, part_time, contract, temporary, internship, other |
experience_level |
entry, associate, mid_senior, director, executive |
work_type |
on_site, remote, hybrid |
easy_apply |
true / false |
sort_by |
date, relevance |
Prerequisites
- Python 3.12 or later
- uv package manager
- A LinkedIn account for authentication
Quick Start
1. Clone and install
git clone https://github.com/eliasbiondo/linkedin-mcp-server.git
cd linkedin-mcp-server
uv sync
2. Authenticate with LinkedIn
uv run linkedin-mcp-server --login
A browser window will open. Log in to LinkedIn and the session will be persisted locally at ~/.linkedin-mcp-server/browser-data.
3. Run the server
stdio transport (default — for Claude Desktop, Cursor, and similar clients):
uv run linkedin-mcp-server
HTTP transport (for remote clients, the MCP Inspector, etc.):
uv run linkedin-mcp-server --transport streamable-http --host 0.0.0.0 --port 8000
Client Integration
Claude Desktop / Cursor
Add to your MCP configuration file:
{
"mcpServers": {
"linkedin": {
"command": "uv",
"args": [
"--directory", "/path/to/linkedin-mcp-server",
"run", "linkedin-mcp-server"
]
}
}
}
MCP Inspector
npx @modelcontextprotocol/inspector
Then connect to http://localhost:8000/mcp if using HTTP transport.
Configuration
Configuration follows a strict precedence chain: CLI args > environment variables > .env file > defaults.
CLI Arguments
| Argument | Description | Default |
|---|---|---|
--transport |
stdio or streamable-http |
stdio |
--host |
Host for HTTP transport | 127.0.0.1 |
--port |
Port for HTTP transport | 8000 |
--log-level |
DEBUG, INFO, WARNING, ERROR |
WARNING |
--login |
Open browser for LinkedIn login | — |
--logout |
Clear stored credentials | — |
--status |
Check session status | — |
Environment Variables
Create a .env file in the project root:
# Server
LINKEDIN_TRANSPORT=stdio
LINKEDIN_HOST=127.0.0.1
LINKEDIN_PORT=8000
LINKEDIN_LOG_LEVEL=WARNING
# Browser
LINKEDIN_HEADLESS=true
LINKEDIN_SLOW_MO=0
LINKEDIN_TIMEOUT=5000
LINKEDIN_VIEWPORT_WIDTH=1280
LINKEDIN_VIEWPORT_HEIGHT=720
LINKEDIN_CHROME_PATH=
LINKEDIN_USER_AGENT=
LINKEDIN_USER_DATA_DIR=~/.linkedin-mcp-server/browser-data
Architecture
The project follows a hexagonal (ports and adapters) architecture with strict layer separation:
src/linkedin_mcp_server/
├── domain/ # Core business logic — zero external dependencies
│ ├── models/ # Data models (Person, Company, Job, Search)
│ ├── parsers/ # HTML to structured data parsers
│ ├── exceptions.py # Domain exceptions
│ └── value_objects.py # Immutable configuration and content objects
├── ports/ # Abstract interfaces
│ ├── auth.py # Authentication port
│ ├── browser.py # Browser automation port
│ └── config.py # Configuration port
├── application/ # Use cases — orchestration layer
│ ├── scrape_person.py
│ ├── scrape_company.py
│ ├── scrape_job.py
│ ├── search_people.py
│ ├── search_jobs.py
│ └── manage_session.py
├── adapters/ # Concrete implementations
│ ├── driven/ # Infrastructure adapters (browser, auth, config)
│ └── driving/ # Interface adapters (CLI, MCP tools, serialization)
└── container.py # Dependency injection composition root
Design Decisions
- Ports and adapters — Domain logic is fully decoupled from infrastructure. The browser engine, MCP framework, and configuration source can all be swapped independently.
- Dependency injection — A single
Containerclass acts as the composition root and is the only place that imports concrete adapter classes. - Structured JSON output — LinkedIn HTML is parsed into typed Python dataclasses, then serialized to JSON for reliable LLM consumption.
- Session persistence — Browser state is saved to disk, so authentication is required only once.
Development
Setup
uv sync --group dev
uv run pre-commit install
Running tests
uv run pytest
With coverage:
uv run pytest --cov=linkedin_mcp_server
Linting and formatting
This project uses Ruff for both linting and formatting. Pre-commit hooks will run these automatically on each commit.
# Lint
uv run ruff check .
# Lint and auto-fix
uv run ruff check . --fix
# Format
uv run ruff format .
License
This project is licensed under the MIT License. See the LICENSE file for details.
Contributing
Contributions are welcome. Please read the contributing guide for details on the development workflow and submission process.
Disclaimer
This tool is intended for personal and educational use. Scraping LinkedIn may violate their Terms of Service. Use responsibly and at your own risk. The authors are not responsible for any misuse or consequences arising from the use of this software.
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.