inspect-logs-mcp
Enables LLMs to explore and analyze UK Government BEIS inspect_ai evaluation logs directly from tools like Claude Code and Cursor. It provides capabilities to list logs, view evaluation summaries, and inspect conversation histories for specific samples.
README
inspect-logs-mcp
MCP server for exploring inspect_ai evaluation logs from Claude Code.
What it does
This MCP server gives Claude Code, Cursor, and other LLMs direct access to your inspect_ai evaluation logs, allowing you to:
- List logs: See all your evaluation runs with metadata
- View summaries: Get detailed evaluation results, scores, and token usage
- Inspect samples: View full conversation history for any sample
- Search & filter: Find logs by task, model, date, or status
- Compare runs: Side-by-side comparison of two evaluation runs
- Aggregate stats: Get statistics across multiple runs
Installation
Currently only source installation is supported.
git clone https://github.com/PranshuSrivastava/inspect-logs-mcp.git
cd inspect-logs-mcp
pip install -e .
Configuration
To use with Claude code in your current directory, run the following command:
claude mcp add --transport stdio inspect-logs inspect-logs-mcp
To use with Claude code in your global configuration, run the following command:
claude mcp add --scope user --transport stdio inspect-logs inspect-logs-mcp
To use with Cursor, Antigravity or other IDEs, add the following to your mcp config file:
{
"mcpServers": {
"inspect-logs": {
"command": "inspect-logs-mcp",
"env": {
"INSPECT_LOG_DIR": "./logs"
}
}
}
}
Environment Variables
| Variable | Description | Default |
|---|---|---|
INSPECT_LOG_DIR |
Directory containing .eval log files |
./logs |
INSPECT_LOGS_MCP_DEFAULT_LIMIT |
Default number of logs limit | 50 |
INSPECT_LOGS_MCP_MAX_LIMIT |
Maximum number of logs limit | 500 |
How it Works
The server reads .eval files (which are ZIP archives containing JSON) using the inspect_ai.log API. All file operations happen in memory - no files are extracted to disk, so your logs directory stays clean.
Requirements
- Python 3.10+
- inspect-ai >= 0.3.70
- mcp >= 1.0.0
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.