JinaAI Search

JinaAI Search

Enables efficient web search integration with Jina.ai's Search API, offering clean, LLM-optimized content retrieval with support for various content types and configurable caching.

spences10

Programming Docs Access
Database Interaction
Visit Server

Tools

search

Search the web and get clean, LLM-friendly content using Jina.ai Reader. Returns top 5 results with URLs and clean content.

README

mcp-jinaai-search


⚠️ Notice

This repository is no longer maintained.

The functionality of this tool is now available in mcp-omnisearch, which combines multiple MCP tools in one unified package.

Please use mcp-omnisearch instead.


A Model Context Protocol (MCP) server for integrating Jina.ai's Search API with LLMs. This server provides efficient and comprehensive web search capabilities, optimised for retrieving clean, LLM-friendly content from the web.

<a href="https://glama.ai/mcp/servers/u6603w196t"> <img width="380" height="200" src="https://glama.ai/mcp/servers/u6603w196t/badge" /> </a>

Features

  • 🔍 Advanced web search through Jina.ai Search API
  • 🚀 Fast and efficient content retrieval
  • 📄 Clean text extraction with preserved structure
  • 🧠 Content optimised for LLMs
  • 🌐 Support for various content types including documentation
  • 🏗️ Built on the Model Context Protocol
  • 🔄 Configurable caching for performance
  • 🖼️ Optional image and link gathering
  • 🌍 Localisation support through browser locale
  • 🎯 Token budget control for response size

Configuration

This server requires configuration through your MCP client. Here are examples for different environments:

Cline Configuration

Add this to your Cline MCP settings:

{
	"mcpServers": {
		"jinaai-search": {
			"command": "node",
			"args": ["-y", "mcp-jinaai-search"],
			"env": {
				"JINAAI_API_KEY": "your-jinaai-api-key"
			}
		}
	}
}

Claude Desktop with WSL Configuration

For WSL environments, add this to your Claude Desktop configuration:

{
	"mcpServers": {
		"jinaai-search": {
			"command": "wsl.exe",
			"args": [
				"bash",
				"-c",
				"JINAAI_API_KEY=your-jinaai-api-key npx mcp-jinaai-search"
			]
		}
	}
}

Environment Variables

The server requires the following environment variable:

  • JINAAI_API_KEY: Your Jina.ai API key (required)

API

The server implements a single MCP tool with configurable parameters:

search

Search the web and get clean, LLM-friendly content using Jina.ai Reader. Returns top 5 results with URLs and clean content.

Parameters:

  • query (string, required): Search query
  • format (string, optional): Response format ("json" or "text"). Defaults to "text"
  • no_cache (boolean, optional): Bypass cache for fresh results. Defaults to false
  • token_budget (number, optional): Maximum number of tokens for this request
  • browser_locale (string, optional): Browser locale for rendering content
  • stream (boolean, optional): Enable stream mode for large pages. Defaults to false
  • gather_links (boolean, optional): Gather all links at the end of response. Defaults to false
  • gather_images (boolean, optional): Gather all images at the end of response. Defaults to false
  • image_caption (boolean, optional): Caption images in the content. Defaults to false
  • enable_iframe (boolean, optional): Extract content from iframes. Defaults to false
  • enable_shadow_dom (boolean, optional): Extract content from shadow DOM. Defaults to false
  • resolve_redirects (boolean, optional): Follow redirect chains to final URL. Defaults to true

Development

Setup

  1. Clone the repository
  2. Install dependencies:
pnpm install
  1. Build the project:
pnpm run build
  1. Run in development mode:
pnpm run dev

Publishing

  1. Create a changeset:
pnpm changeset
  1. Version the package:
pnpm version
  1. Build and publish:
pnpm release

Contributing

Contributions are welcome! Please feel free to submit a Pull Request.

License

MIT License - see the LICENSE file for details.

Acknowledgments

Recommended Servers

E2B

E2B

Using MCP to run code via e2b.

Official
Featured
Exa Search

Exa Search

A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.

Official
Featured
Neon Database

Neon Database

MCP server for interacting with Neon Management API and databases

Official
Featured
AIO-MCP Server

AIO-MCP Server

🚀 All-in-one MCP server with AI search, RAG, and multi-service integrations (GitLab/Jira/Confluence/YouTube) for AI-enhanced development workflows. Folk from

Featured
Local
Persistent Knowledge Graph

Persistent Knowledge Graph

An implementation of persistent memory for Claude using a local knowledge graph, allowing the AI to remember information about users across conversations with customizable storage location.

Featured
Local
Hyperbrowser MCP Server

Hyperbrowser MCP Server

Welcome to Hyperbrowser, the Internet for AI. Hyperbrowser is the next-generation platform empowering AI agents and enabling effortless, scalable browser automation. Built specifically for AI developers, it eliminates the headaches of local infrastructure and performance bottlenecks, allowing you to

Featured
Local
Perplexity Deep Research

Perplexity Deep Research

A server that allows AI assistants to perform web searches using Perplexity's sonar-deep-research model with citation support.

Featured
Docx Document Processing Service

Docx Document Processing Service

A powerful Word document processing service based on FastMCP, enabling AI assistants to create, edit, and manage docx files with full formatting support. Preserves original styles when editing content.

Featured
PostgreSQL Database Management Server

PostgreSQL Database Management Server

A Model Context Protocol server that enables powerful PostgreSQL database management capabilities including analysis, schema management, data migration, and monitoring through natural language interactions.

Featured
OpenRouter MCP Server

OpenRouter MCP Server

Provides integration with OpenRouter.ai, allowing access to various AI models through a unified interface.

Featured