MCP Server for Spinnaker
A Model Context Protocol server implementation that allows AI models to interact with and manage Spinnaker deployments, pipelines, and applications through a standardized interface.
dion-hagan
README
MCP Server for Spinnaker
This package provides a Model Context Protocol (MCP) server implementation for Spinnaker integrations. It allows AI models to interact with Spinnaker deployments, pipelines, and applications through the standardized MCP interface.
AI Integration
This MCP server is a powerful example of how Anthropic's new AI model, Claude, can directly integrate with and enhance software deployment processes using the Model Context Protocol. By following MCP standards, Claude can access rich contextual information about Spinnaker applications, pipelines, and deployments, and actively manage them using well-defined tools.
Let's dive into some of the exciting possibilities this integration enables for AI-driven CI/CD:
-
Intelligent Deployment Decisions: With access to comprehensive context about the state of applications and pipelines, AI models like Claude can analyze this information to make intelligent decisions about when and how to deploy. For example, Claude could look at factors like test coverage, code churn, and historical success rates to determine the optimal time and target environment for a deployment.
-
Proactive Issue Detection and Autonomous Remediation: AI models can continuously monitor the CI/CD process, spotting potential issues before they cause problems. Imagine Claude detecting that a new version of a dependency has a known vulnerability and automatically creating a pull request to update it, or noticing that a deployment is taking longer than usual and proactively spinning up additional resources to prevent a timeout.
-
Continuous Process Optimization: With each deployment, AI models can learn and adapt, continuously optimizing the CI/CD process. Claude could analyze build and deployment logs to identify bottlenecks, then experiment with different configurations to improve speed and reliability. Over time, the entire deployment process becomes more efficient and robust.
-
Automated Root Cause Analysis and Recovery: When issues do occur, AI can rapidly diagnose the problem and even attempt to fix it autonomously. Claude could correlate errors across different parts of the system, identify the most likely root cause, and then take corrective actions like rolling back to a previous version or applying a known patch.
And these are just a few examples! As the Model Context Protocol evolves and more integrations are built, we can expect AI to take on increasingly sophisticated roles in the DevOps world. Across the entire CI/CD pipeline, AI could provide intelligent insights and recommendations, acting as a virtual assistant for product engineers.
By empowering AI to work alongside humans in the CI/CD process, MCP integrations like this Spinnaker server showcase how AI can become a proactive, intelligent partner in Developer Productivity infrastructure. It's a significant step towards more efficient, reliable, and autonomous software delivery.
Installation
npm install @airjesus17/mcp-server-spinnaker
or
yarn add @airjesus17/mcp-server-spinnaker
Usage
import { SpinnakerMCPServer } from '@airjesus17/mcp-server-spinnaker';
// Initialize the server
const server = new SpinnakerMCPServer(
'https://your-gate-url',
['app1', 'app2'], // List of applications to monitor
['prod', 'staging'] // List of environments to monitor
);
// Start the server
const port = 3000;
server.listen(port, () => {
console.log(`Spinnaker MCP Server is running on port ${port}`);
});
Available Tools
The server provides the following tools for AI models to interact with Spinnaker:
get-applications
Retrieves a list of monitored Spinnaker applications and their current state.
// Example response
{
"success": true,
"data": [
{
"name": "myapp",
"description": "My application",
"pipelines": [
{
"id": "pipeline-1",
"name": "Deploy to Production",
"status": "SUCCEEDED"
}
]
}
]
}
get-pipelines
Retrieves all pipelines for a specific application.
// Parameters
{
"application": "myapp"
}
// Example response
{
"success": true,
"data": [
{
"id": "pipeline-1",
"name": "Deploy to Production",
"status": "SUCCEEDED",
"stages": [...]
}
]
}
trigger-pipeline
Triggers a pipeline execution for a specific application.
// Parameters
{
"application": "myapp",
"pipelineId": "pipeline-1",
"parameters": {
"version": "1.2.3",
"environment": "production"
}
}
// Example response
{
"success": true,
"data": {
"ref": "01HFGH2J...",
"status": "RUNNING"
}
}
Context Updates
The server automatically maintains context about your Spinnaker deployments. The context includes:
- List of applications and their current state
- Pipeline status for each application
- Current deployments across monitored environments
- Recent pipeline executions
Context is refreshed every 30 seconds by default.
Environment Variables
The server can be configured using the following environment variables:
GATE_URL
: URL of your Spinnaker Gate serviceMCP_PORT
: Port to run the MCP server on (default: 3000)REFRESH_INTERVAL
: Context refresh interval in seconds (default: 30)
Types
The package exports TypeScript types for working with the server:
import type {
SpinnakerApplication,
SpinnakerPipeline,
SpinnakerDeployment,
SpinnakerExecution
} from '@airjesus17/mcp-server-spinnaker';
Development
To contribute to the development:
- Clone the repository
- Install dependencies:
yarn install
- Build the project:
yarn build
- Run tests:
yarn test
License
MIT License - see LICENSE for details.
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Playwright MCP Server
Provides a server utilizing Model Context Protocol to enable human-like browser automation with Playwright, allowing control over browser actions such as navigation, element interaction, and scrolling.
MCP Package Docs Server
Facilitates LLMs to efficiently access and fetch structured documentation for packages in Go, Python, and NPM, enhancing software development with multi-language support and performance optimization.
Claude Code MCP
An implementation of Claude Code as a Model Context Protocol server that enables using Claude's software engineering capabilities (code generation, editing, reviewing, and file operations) through the standardized MCP interface.
@kazuph/mcp-taskmanager
Model Context Protocol server for Task Management. This allows Claude Desktop (or any MCP client) to manage and execute tasks in a queue-based system.
Apple MCP Server
Enables interaction with Apple apps like Messages, Notes, and Contacts through the MCP protocol to send messages, search, and open app content using natural language.
contentful-mcp
Update, create, delete content, content-models and assets in your Contentful Space