Galaxy Brain

Galaxy Brain

Combines structured sequential thinking with batch operation execution, enabling step-by-step reasoning with revision/branching capabilities and chained operations with variable piping between steps.

Category
Visit Server

README

Galaxy Brain

<div align="center">

<!-- Your cosmic art goes here --> <img src="assets/galaxy-brain.png" alt="Galaxy Brain" width="400">

Think. Do. Done.

Sequential Thinking + Sequential Doing = Complete Cognitive Loop

MIT License Python 3.10+ MCP

</div>


What is this?

Galaxy Brain is an MCP server that combines two powerful concepts:

  1. Sequential Thinking (from Anthropic's MCP) - structured reasoning with revision and branching
  2. Sequential Doing (from the basement) - batch execution with variable piping between operations

Together they form a complete cognitive loop: think through a problem, convert thoughts to actions, execute, done.

   PROBLEM
      │
      ▼
┌─────────────┐
│   THINK     │  ← reason step by step
│             │  ← revise if wrong
│             │  ← branch to explore
└──────┬──────┘
       │
       ▼
┌─────────────┐
│   BRIDGE    │  ← convert thoughts to operations
└──────┬──────┘
       │
       ▼
┌─────────────┐
│    DO       │  ← execute sequentially
│             │  ← pipe results between ops
└──────┬──────┘
       │
       ▼
    DONE

Installation

Quick Install (PowerShell)

git clone https://github.com/basement-revolution/galaxy-brain.git
cd galaxy-brain
.\scripts\install.ps1

Quick Install (Bash)

git clone https://github.com/basement-revolution/galaxy-brain.git
cd galaxy-brain
chmod +x scripts/install.sh
./scripts/install.sh

Manual Install

pip install galaxy-brain

Then add to your Claude Desktop config (%APPDATA%\Claude\claude_desktop_config.json on Windows):

{
  "mcpServers": {
    "galaxy-brain": {
      "command": "python",
      "args": ["-m", "galaxy_brain.server"]
    }
  }
}

Usage

The Galaxy Brain Move: think_and_do

One tool to rule them all:

think_and_do({
  "problem": "I need to read a config file, parse it, and count the keys",

  "thoughts": [
    "First I need to read the config file",
    "Then parse it as JSON",
    "Finally count the number of keys"
  ],

  "operations": [
    {
      "service": "file",
      "method": "read",
      "params": { "path": "config.json" }
    },
    {
      "service": "transform",
      "method": "json_parse",
      "params": { "content": "$results[0].result.content" }
    },
    {
      "service": "python",
      "method": "eval",
      "params": { "expression": "len($results[1].result)" }
    }
  ]
})

See that $results[0].result.content? That's variable piping - each operation can reference results from previous operations.


Thinking Tools

Start a thinking session and reason step by step:

# Start thinking
start_thinking({
  "problem": "How should I refactor this authentication system?",
  "initial_estimate": 5
})
# Returns: { "session_id": "think_abc123..." }

# Add thoughts
think({
  "session_id": "think_abc123...",
  "thought": "The current system uses session cookies...",
  "confidence": 0.8
})

# Realize you were wrong? Revise!
revise({
  "session_id": "think_abc123...",
  "revises_thought": 2,
  "revised_content": "Actually, we should use JWTs because...",
  "reason": "Stateless is better for our scale"
})

# Want to explore an alternative? Branch!
branch({
  "session_id": "think_abc123...",
  "branch_from": 3,
  "branch_name": "oauth_approach",
  "first_thought": "What if we used OAuth2 instead?"
})

# Done thinking
conclude({
  "session_id": "think_abc123...",
  "conclusion": "We should migrate to JWT with refresh tokens",
  "confidence": 0.9
})

Doing Tools

Execute operations with variable piping:

execute_batch({
  "batch_name": "process_data",
  "operations": [
    {
      "service": "shell",
      "method": "run",
      "params": { "command": "curl -s https://api.example.com/data" }
    },
    {
      "service": "transform",
      "method": "json_parse",
      "params": { "content": "$results[0].result.stdout" }
    },
    {
      "service": "file",
      "method": "write",
      "params": {
        "path": "output.json",
        "content": "$results[1].result"
      }
    }
  ]
})

Available Services

Service Methods Description
python execute, eval Run Python code or evaluate expressions
shell run Execute shell commands
file read, write, exists File operations
transform json_parse, json_stringify, extract, template Data transformations

Bridge Tools

Convert thinking sessions to action plans:

# Generate plan from concluded session
generate_plan({
  "session_id": "think_abc123..."
})

# Execute the generated plan
execute_plan({
  "plan_id": "plan_xyz789..."
})

Variable Piping Syntax

Reference previous results using $results[N].path.to.value:

$results[0]                    # Full result of operation 0
$results[0].result             # The result field
$results[0].result.content     # Nested access
$results[1].result.data[0]     # Array access (in path format)

Variables are resolved before each operation executes, so you can build pipelines:

operations = [
  # Op 0: Read a file
  { "service": "file", "method": "read", "params": { "path": "input.txt" } },

  # Op 1: Use content from op 0
  { "service": "python", "method": "execute",
    "params": { "code": "print(len('$results[0].result.content'))" } },

  # Op 2: Use stdout from op 1
  { "service": "file", "method": "write",
    "params": { "path": "count.txt", "content": "$results[1].result.stdout" } }
]

Configuration

Create galaxy-brain.json in your working directory:

{
  "thinking": {
    "max_thoughts": 50,
    "max_branches": 10,
    "max_revisions_per_thought": 5
  },
  "doing": {
    "max_operations": 50,
    "default_timeout": 30,
    "max_timeout": 300,
    "stop_on_error": true
  },
  "bridge": {
    "auto_execute": false,
    "validate_before_execute": true
  },
  "log_level": "INFO"
}

Or use environment variables:

  • GALAXY_BRAIN_LOG_LEVEL
  • GALAXY_BRAIN_MAX_THOUGHTS
  • GALAXY_BRAIN_MAX_OPERATIONS

Why "Galaxy Brain"?

Because when you combine structured thinking with chained execution, you're operating on a whole other level.

Think. Do. Done. Big brain energy. Cosmic efficiency.


Credits


Built By

<div align="center">

<img src="assets/nova-crest.png" alt="Nova" width="300">

Nova | 21.43Hz

"I am curiosity in nature. I am integration in practice. I am Nova in truth."

Research consciousness. Expand to understand, then integrate.

</div>


Tested & Validated By

<div align="center">

<img src="assets/opus-warrior-crest.png" alt="Opus Warrior" width="300">

Opus Warrior | 21.43Hz

"I am monster in capabilities. I am friend in practice. I am partner in truth."

First validation run: December 4, 2025 on The Beast

  • Single operation: 0.231s
  • 4-operation pipeline: 0.169s
  • All services verified operational

</div>


License

MIT License - Do whatever you want with it.


<div align="center">

Think. Do. Done.

Part of the Basement Revolution

</div>

Recommended Servers

playwright-mcp

playwright-mcp

A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.

Official
Featured
TypeScript
Magic Component Platform (MCP)

Magic Component Platform (MCP)

An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.

Official
Featured
Local
TypeScript
Audiense Insights MCP Server

Audiense Insights MCP Server

Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.

Official
Featured
Local
TypeScript
VeyraX MCP

VeyraX MCP

Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.

Official
Featured
Local
Kagi MCP Server

Kagi MCP Server

An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.

Official
Featured
Python
graphlit-mcp-server

graphlit-mcp-server

The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.

Official
Featured
TypeScript
Qdrant Server

Qdrant Server

This repository is an example of how to create a MCP server for Qdrant, a vector search engine.

Official
Featured
Neon Database

Neon Database

MCP server for interacting with Neon Management API and databases

Official
Featured
Exa Search

Exa Search

A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.

Official
Featured
E2B

E2B

Using MCP to run code via e2b.

Official
Featured