Chronulus AI Forecasting & Prediction Agents
Enables integration of Chronulus AI Forecasting & Prediction Agents with Claude, allowing users to access AI-powered forecasting and prediction capabilities directly through Claude's interface.
ChronulusAI
README
<div align="center"> <img width="150px" src="https://www.chronulus.com/brand-assets/chronulus-logo-blue-on-alpha-square.png" alt="Chronulus AI"> <h1 align="center">MCP Server for Chronulus</h1> <h3 align="center">Chat with Chronulus AI Forecasting & Prediction Agents in Claude</h3> </div>
Quickstart: Claude for Desktop
Install
Claude for Desktop is currently available on macOS and Windows.
Install Claude for Desktop here
Configuration
Follow the general instructions here to configure the Claude desktop client.
You can find your Claude config at one of the following locations:
- macOS:
~/Library/Application Support/Claude/claude_desktop_config.json
- Windows:
%APPDATA%\Claude\claude_desktop_config.json
Then choose one of the following methods that best suits your needs and add it to your claude_desktop_config.json
<details> <summary>Using pip</summary>
(Option 1) Install release from PyPI
pip install chronulus-mcp
(Option 2) Install from Github
git clone https://github.com/ChronulusAI/chronulus-mcp.git
cd chronulus-mcp
pip install .
{
"mcpServers": {
"chronulus-agents": {
"command": "python",
"args": ["-m", "chronulus_mcp"],
"env": {
"CHRONULUS_API_KEY": "<YOUR_CHRONULUS_API_KEY>"
}
}
}
}
Note, if you get an error like "MCP chronulus-agents: spawn python ENOENT",
then you most likely need to provide the absolute path to python
.
For example /Library/Frameworks/Python.framework/Versions/3.11/bin/python3
instead of just python
</details>
<details> <summary>Using docker</summary>
Here we will build a docker image called 'chronulus-mcp' that we can reuse in our Claude config.
git clone https://github.com/ChronulusAI/chronulus-mcp.git
cd chronulus-mcp
docker build . -t 'chronulus-mcp'
In your Claude config, be sure that the final argument matches the name you give to the docker image in the build command.
{
"mcpServers": {
"chronulus-agents": {
"command": "docker",
"args": ["run", "-i", "--rm", "-e", "CHRONULUS_API_KEY", "chronulus-mcp"],
"env": {
"CHRONULUS_API_KEY": "<YOUR_CHRONULUS_API_KEY>"
}
}
}
}
</details>
<details> <summary>Using uvx</summary>
uvx
will pull the latest version of chronulus-mcp
from the PyPI registry, install it, and then run it.
{
"mcpServers": {
"chronulus-agents": {
"command": "uvx",
"args": ["chronulus-mcp"],
"env": {
"CHRONULUS_API_KEY": "<YOUR_CHRONULUS_API_KEY>"
}
}
}
}
Note, if you get an error like "MCP chronulus-agents: spawn uvx ENOENT", then you most likely need to either:
- install uv or
- Provide the absolute path to
uvx
. For example/Users/username/.local/bin/uvx
instead of justuvx
</details>
Additional Servers (Filesystem, Fetch, etc)
In our demo, we use third-party servers like fetch and filesystem.
For details on installing and configure third-party server, please reference the documentation provided by the server maintainer.
Below is an example of how to configure filesystem and fetch alongside Chronulus in your claude_desktop_config.json
:
{
"mcpServers": {
"chronulus-agents": {
"command": "uvx",
"args": ["chronulus-mcp"],
"env": {
"CHRONULUS_API_KEY": "<YOUR_CHRONULUS_API_KEY>"
}
},
"filesystem": {
"command": "npx",
"args": [
"-y",
"@modelcontextprotocol/server-filesystem",
"/path/to/AIWorkspace"
]
},
"fetch": {
"command": "uvx",
"args": ["mcp-server-fetch"]
}
}
}
Claude Preferences
To streamline your experience using Claude across multiple sets of tools, it is best to add your preferences to under Claude Settings.
You can upgrade your Claude preferences in a couple ways:
- From Claude Desktop:
Settings -> General -> Claude Settings -> Profile (tab)
- From claude.ai/settings:
Profile (tab)
Preferences are shared across both Claude for Desktop and Claude.ai (the web interface). So your instruction need to work across both experiences.
Below are the preferences we used to achieve the results shown in our demos:
## Tools-Dependent Protocols
The following instructions apply only when tools/MCP Servers are accessible.
### Filesystem - Tool Instructions
- Do not use 'read_file' or 'read_multiple_files' on binary files (e.g., images, pdfs, docx) .
- When working with binary files (e.g., images, pdfs, docx) use 'get_info' instead of 'read_*' tools to inspect a file.
### Chronulus Agents - Tool Instructions
- When using Chronulus, prefer to use input field types like TextFromFile, PdfFromFile, and ImageFromFile over scanning the files directly.
- When plotting forecasts from Chronulus, always include the Chronulus-provided forecast explanation below the plot and label it as Chronulus Explanation.
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.
AIO-MCP Server
🚀 All-in-one MCP server with AI search, RAG, and multi-service integrations (GitLab/Jira/Confluence/YouTube) for AI-enhanced development workflows. Folk from
Persistent Knowledge Graph
An implementation of persistent memory for Claude using a local knowledge graph, allowing the AI to remember information about users across conversations with customizable storage location.
Hyperbrowser MCP Server
Welcome to Hyperbrowser, the Internet for AI. Hyperbrowser is the next-generation platform empowering AI agents and enabling effortless, scalable browser automation. Built specifically for AI developers, it eliminates the headaches of local infrastructure and performance bottlenecks, allowing you to
React MCP
react-mcp integrates with Claude Desktop, enabling the creation and modification of React apps based on user prompts
Atlassian Integration
Model Context Protocol (MCP) server for Atlassian Cloud products (Confluence and Jira). This integration is designed specifically for Atlassian Cloud instances and does not support Atlassian Server or Data Center deployments.

Any OpenAI Compatible API Integrations
Integrate Claude with Any OpenAI SDK Compatible Chat Completion API - OpenAI, Perplexity, Groq, xAI, PyroPrompts and more.