FastAPI Hello World Application

FastAPI Hello World Application

A test repository created using the GitHub MCP server

xxradar

Developer Tools
Visit Server

README

FastAPI Hello World Application

A simple Hello World API built with FastAPI and MCP SSE support.

Features

  • Root endpoint that returns a Hello World message
  • Dynamic greeting endpoint that takes a name parameter
  • OpenAI integration with GPT-4o for advanced AI-powered chat completions
  • Automatic API documentation with Swagger UI

Prerequisites

  • Python 3.7+ (for local setup)
  • pip (Python package installer)
  • OpenAI API key (for the /openai endpoint)
  • Docker (optional, for containerized setup)

Setup Instructions

You can run this application either locally or using Docker.

Local Setup

1. Clone the repository

git clone https://github.com/xxradar/mcp-test-repo.git
cd mcp-test-repo

2. Create a virtual environment (optional but recommended)

# On macOS/Linux
python -m venv venv
source venv/bin/activate

# On Windows
python -m venv venv
venv\Scripts\activate

3. Install dependencies

pip install -r requirements.txt

4. Run the application

uvicorn main:app --reload

The application will start and be available at http://127.0.0.1:8000

Alternatively, you can run the application directly with Python:

python main.py

Docker Setup

1. Clone the repository

git clone https://github.com/xxradar/mcp-test-repo.git
cd mcp-test-repo

2. Build the Docker image

docker build -t fastapi-hello-world .

3. Run the Docker container

docker run -p 8000:8000 fastapi-hello-world

The application will be available at http://localhost:8000

API Endpoints

  • GET /: Returns a simple Hello World message
  • GET /hello/{name}: Returns a personalized greeting with the provided name
  • GET /openai: Returns a response from OpenAI's GPT-4o model (accepts an optional prompt query parameter)
  • GET /docs: Swagger UI documentation
  • GET /redoc: ReDoc documentation

OpenAI Integration

The /openai endpoint uses OpenAI's GPT-4o model and requires an OpenAI API key to be set as an environment variable:

Local Setup

# Set the OpenAI API key as an environment variable
export OPENAI_API_KEY=your_api_key_here

# Run the application
uvicorn main:app --reload

Docker Setup

# Run the Docker container with the OpenAI API key
docker run -p 8000:8000 -e OPENAI_API_KEY=your_api_key_here fastapi-hello-world

Example Usage

Using curl

# Get Hello World message
curl http://127.0.0.1:8000/

# Get personalized greeting
curl http://127.0.0.1:8000/hello/John

# Get OpenAI chat completion with default prompt
curl http://127.0.0.1:8000/openai

# Get OpenAI chat completion with custom prompt
curl "http://127.0.0.1:8000/openai?prompt=Tell%20me%20a%20joke%20about%20programming"

Using MCP

Connect to MCP Inspector

npx @modelcontextprotocol/inspector

Using a web browser

  • Open http://127.0.0.1:8000/ in your browser for the Hello World message
  • Open http://127.0.0.1:8000/hello/John in your browser for a personalized greeting
  • Open http://127.0.0.1:8000/openai in your browser to get a response from OpenAI with the default prompt
  • Open http://127.0.0.1:8000/openai?prompt=What%20is%20FastAPI? in your browser to get a response about FastAPI
  • Open http://127.0.0.1:8000/docs for the Swagger UI documentation

Development

To make changes to the application, edit the main.py file. The server will automatically reload if you run it with the --reload flag.

Recommended Servers

playwright-mcp

playwright-mcp

A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.

Official
Featured
TypeScript
Magic Component Platform (MCP)

Magic Component Platform (MCP)

An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.

Official
Featured
Local
TypeScript
MCP Package Docs Server

MCP Package Docs Server

Facilitates LLMs to efficiently access and fetch structured documentation for packages in Go, Python, and NPM, enhancing software development with multi-language support and performance optimization.

Featured
Local
TypeScript
Claude Code MCP

Claude Code MCP

An implementation of Claude Code as a Model Context Protocol server that enables using Claude's software engineering capabilities (code generation, editing, reviewing, and file operations) through the standardized MCP interface.

Featured
Local
JavaScript
@kazuph/mcp-taskmanager

@kazuph/mcp-taskmanager

Model Context Protocol server for Task Management. This allows Claude Desktop (or any MCP client) to manage and execute tasks in a queue-based system.

Featured
Local
JavaScript
Linear MCP Server

Linear MCP Server

Enables interaction with Linear's API for managing issues, teams, and projects programmatically through the Model Context Protocol.

Featured
JavaScript
mermaid-mcp-server

mermaid-mcp-server

A Model Context Protocol (MCP) server that converts Mermaid diagrams to PNG images.

Featured
JavaScript
Jira-Context-MCP

Jira-Context-MCP

MCP server to provide Jira Tickets information to AI coding agents like Cursor

Featured
TypeScript
Linear MCP Server

Linear MCP Server

A Model Context Protocol server that integrates with Linear's issue tracking system, allowing LLMs to create, update, search, and comment on Linear issues through natural language interactions.

Featured
JavaScript
Sequential Thinking MCP Server

Sequential Thinking MCP Server

This server facilitates structured problem-solving by breaking down complex issues into sequential steps, supporting revisions, and enabling multiple solution paths through full MCP integration.

Featured
Python