AWS Common MCP Servers with CDK Deployment
Deployable MCP Servers for common AWS services (Location, S3, Aurora PG Data API) using AWS CDK.
vAirpower
README
AWS Common MCP Servers with CDK Deployment
This project provides deployable Model Context Protocol (MCP) servers for common AWS services, along with AWS Cloud Development Kit (CDK) code (Python) to provision the necessary infrastructure for running these servers, typically on ECS Fargate.
The goal is to offer reusable components that AI developers (e.g., using Amazon Bedrock, LangChain) can easily deploy into their AWS accounts to interact with AWS services through the standardized MCP interface.
Included MCP Servers
- AWS Location Service: (TypeScript) Exposes functionalities like searching places, getting place details, calculating routes, etc. Based on the aws-location-server.
- Amazon S3: (Python) Provides tools for basic S3 operations like listing buckets/objects, getting, putting, and deleting objects.
- Amazon Aurora PostgreSQL (via RDS Data API): (Python) Allows executing SQL statements against a specified Aurora PostgreSQL cluster using the secure RDS Data API.
Prerequisites
- AWS Account
- AWS CLI configured locally with appropriate permissions (for CDK deployment). Credentials should be set up via
~/.aws/credentials
or environment variables. - Node.js and npm (for the Location Service server build process and CDK)
- Python 3.9+ (for CDK and Python-based MCP servers)
- AWS CDK CLI (
npm install -g aws-cdk
) - Docker (running locally for CDK to build container images)
- Git
Specific Prerequisites for Aurora PostgreSQL Data API Server:
- An existing Aurora PostgreSQL-Compatible cluster running in your target AWS region.
- The cluster must have the RDS Data API enabled.
- An AWS Secrets Manager secret containing the database credentials (username, password) for the cluster. The ARN of this secret is required for deployment.
Project Structure
aws-mcp-infra/
├── mcp_servers/ # Source code for the MCP servers
│ ├── location_service/ # TypeScript server for AWS Location Service
│ ├── s3/ # Python server for S3
│ └── aurora_pg_data_api/ # Python server for Aurora PG Data API
├── cdk_deployment/ # AWS CDK code (Python)
│ ├── app.py # CDK App entry point
│ ├── cdk.json # CDK configuration
│ ├── requirements.txt # CDK Python dependencies
│ └── aws_mcp_infra/ # CDK Stack definitions
│ ├── __init__.py
│ ├── location_service_stack.py
│ ├── s3_stack.py
│ └── aurora_pg_data_api_stack.py
├── .gitignore
└── README.md # This file
Deployment
- Clone the Repository:
git clone https://github.com/vAirpower/aws-common-mcp-servers.git cd aws-common-mcp-servers
- Configure Aurora Prerequisites (if deploying Aurora server):
- Ensure your Aurora PostgreSQL cluster exists and has the Data API enabled.
- Create a secret in AWS Secrets Manager (same region as deployment) containing the DB credentials (e.g., keys
username
,password
). - Update the
db_cluster_arn
anddb_secret_arn
variables incdk_deployment/aws_mcp_infra/aurora_pg_data_api_stack.py
with your specific cluster ARN and the ARN of the secret you created.
- Install CDK Dependencies:
cd cdk_deployment python3 -m venv .venv source .venv/bin/activate pip install -r requirements.txt
- Bootstrap CDK (if first time using CDK in this account/region):
cdk bootstrap aws://ACCOUNT-NUMBER/REGION # Replace with your AWS account number and region
- Deploy Stacks: Deploy all stacks (creates VPC, Cluster, and all three services):
Or deploy specific stacks:cdk deploy --all --require-approval never
cdk deploy LocationServiceMcpStack S3McpStack AuroraPgDataApiMcpStack --require-approval never
Deployment will build the Docker images for each server, push them to ECR repositories created by CDK, and provision the ECS Fargate services.
Usage
Once deployed, the MCP servers run as tasks within ECS Fargate services. They are not publicly exposed by default. Interaction typically occurs from within your AWS environment:
- From Lambda Functions / Bedrock Agents: Use the AWS SDK (e.g.,
boto3
for Python) to invoke the ECS tasks or potentially interact via AWS Systems Manager Session Manager (if configured). The exact mechanism depends on how the client application is designed to communicate with MCP servers running in ECS. - Local Testing (Requires Adaptation): To test these servers locally before deployment, you would typically run them directly (
node build/index.js
orpython aurora_pg_data_api_server.py
) after installing their respective dependencies (npm install
orpip install -r requirements.txt
) and ensuring your local environment has AWS credentials configured. For the Aurora server, you'd also need to set theDB_CLUSTER_ARN
andDB_SECRET_ARN
environment variables locally.
Cleanup
To remove all deployed resources, run:
cd cdk_deployment
source .venv/bin/activate
cdk destroy --all
Recommended Servers
contentful-mcp
Update, create, delete content, content-models and assets in your Contentful Space

Supabase MCP Server
A Model Context Protocol (MCP) server that provides programmatic access to the Supabase Management API. This server allows AI models and other clients to manage Supabase projects and organizations through a standardized interface.

Azure MCP Server
Enables natural language interaction with Azure services through Claude Desktop, supporting resource management, subscription handling, and tenant selection with secure authentication.

SettleMint
Leverage SettleMint's Model Context Protocol server to seamlessly interact with enterprise blockchain infrastructure. Build, deploy, and manage smart contracts through AI-powered assistants, streamlining your blockchain development workflow for maximum efficiency.

Brev
Run, build, train, and deploy ML models on the cloud.

Story SDK MCP Server
This server provides MCP (Model Context Protocol) tools for interacting with Story's Python SDK. Features Get license terms Mint and register IP Asset with PIL Terms Mint license tokens Send $IP to a wallet Upload image to ipfs via Pinata [External] Upload ip and nft metadata via Pinata [External]

Tembo MCP Server
An MCP server that enables Claude to interact with Tembo Cloud platform API, allowing users to manage Tembo Cloud resources through natural language.

Workers MCP
A package that connects Claude Desktop and other MCP clients to Cloudflare Workers, enabling custom functionality to be accessed via natural language through the Model Context Protocol.

Appwrite MCP Server
A Model Context Protocol server that allows AI assistants to interact with Appwrite's API, providing tools to manage databases, users, functions, teams, and other resources within Appwrite projects.
MCP2Lambda
Enables AI models to interact with AWS Lambda functions via the MCP protocol, allowing access to private resources, real-time data, and custom computation in a secure environment.