Registry Review MCP Server
Automates carbon credit project registration document review, transforming 6-8 hour manual reviews into 60-90 minute guided workflows with automated document discovery, evidence extraction, cross-validation, and compliance report generation.
README
Registry Review MCP Server
MCP server that automates carbon credit project registration reviews through an eight-stage workflow.
Overview
The Registry Review MCP Server transforms a 6-8 hour manual document review into a guided workflow where AI handles document organization, data extraction, and consistency checking while humans provide expertise, judgment, and final approval.
Core Capabilities:
- Document discovery and intelligent classification
- Requirement mapping with semantic matching
- Evidence extraction with page citations
- Cross-document validation (dates, land tenure, project IDs)
- Structured report generation (Markdown + JSON)
Quick Start
# Install dependencies
uv sync
# Run the MCP server
uv run python -m registry_review_mcp.server
# Run tests (expensive tests excluded by default)
uv run pytest
Claude Desktop Integration
Add to your claude_desktop_config.json:
{
"mcpServers": {
"registry-review": {
"command": "uv",
"args": [
"--directory", "/path/to/regen-registry-review-mcp",
"run", "python", "-m", "registry_review_mcp.server"
]
}
}
}
The Eight-Stage Workflow
Each stage produces artifacts for human verification before proceeding. The workflow follows a collaboration model where AI handles tedious document processing while humans provide expertise and final judgment.
Stage A: Initialize
Create a review session with project metadata and load the checklist template.
/A-initialize Botany Farm 2022-2023, /path/to/documents
Output: Session ID, project metadata, loaded checklist (23 requirements for Soil Carbon v1.2.2)
Stage B: Document Discovery
Scan the documents directory, extract file metadata, and classify each document by type.
/B-document-discovery
Agent Actions:
- Recursively scan for PDFs, shapefiles, GeoJSON, spreadsheets
- Classify documents (project plan, baseline report, monitoring report, land tenure, etc.)
- Generate document inventory with confidence scores
Human Actions: Review classifications, mark documents as in-scope/ignored/pinned
Output: Document inventory with normalized names, types, and source references
Stage C: Requirement Mapping
Connect discovered documents to specific checklist requirements using semantic matching.
/C-requirement-mapping
Agent Actions:
- Parse checklist into structured requirements with expected evidence types
- Analyze documents and suggest requirement → document mappings
- Flag requirements with no plausible matches
Human Actions: Confirm/reject suggested mappings, manually add missing mappings
Output: Mapping matrix linking each requirement to 0+ documents with confidence scores
Stage D: Evidence Extraction
Extract key data points and text snippets from mapped documents.
/D-evidence-extraction
Agent Actions:
- Parse document content (PDF text, tables, metadata)
- Extract 0-3 evidence snippets per requirement with page citations
- Extract structured data: dates, locations, ownership info, numerical values
Human Actions: Review snippets, delete irrelevant ones, add manual notes
Output: Evidence database with snippets, citations, and structured data points
Stage E: Cross-Validation
Verify consistency, completeness, and compliance across all extracted evidence.
/E-cross-validation
Validation Checks:
- Date Alignment: Sampling dates within ±120 days of imagery dates
- Land Tenure: Owner names consistent across documents (fuzzy matching)
- Project ID: Consistent project identifiers across all documents
- Completeness: Each requirement has mapped documents with sufficient evidence
Output: Validation results with pass/warning/fail flags and coverage statistics
Stage F: Report Generation
Produce structured, auditable Registry Review Report.
/F-report-generation
Output Formats:
- Markdown: Human-readable report with executive summary, per-requirement findings, citations
- JSON: Machine-readable for audit trails and downstream systems
Report Contents: Project metadata, coverage statistics, requirement findings with evidence snippets, validation results, items requiring human review
Stage G: Human Review
Expert validation, annotation, and revision handling.
/G-human-review
Human Actions:
- Review flagged items requiring judgment
- Override agent assessments where expert knowledge differs
- Request revisions from proponent if gaps identified
- Make final determination: Approve / Conditional / Reject / On Hold
Output: Finalized report with human annotations and approval decision
Stage H: Completion
Finalize and archive the review.
/H-completion
Agent Actions:
- Lock finalized report
- Generate archive package with audit trail
- Prepare data for on-chain registration (if approved)
Output: Locked report, complete audit trail, archived session
Quick Example
/A-initialize Botany Farm 2022-2023, /home/user/projects/botany-farm
/B-document-discovery
/C-requirement-mapping
/D-evidence-extraction
/E-cross-validation
/F-report-generation
Each stage auto-selects the most recent session, so you can run them in sequence without specifying session IDs.
Available Tools
Session Management:
create_session- Create new review sessionload_session/list_sessions/delete_session- Session lifecyclestart_review- Quick-start: create session + discover documentslist_example_projects- List available test projects
File Upload:
create_session_from_uploads- Create session from uploaded filesupload_additional_files- Add files to existing sessionstart_review_from_uploads- Full workflow from uploads
Document Processing:
discover_documents- Scan and classify project documentsadd_documents- Add document sources to sessionextract_pdf_text- Extract text from PDFsextract_gis_metadata- Extract GIS shapefile metadata
Requirement Mapping:
map_all_requirements- Semantic mapping to documentsconfirm_mapping/remove_mapping- Manual mapping adjustmentsget_mapping_status- View mapping statistics
Evidence & Validation:
extract_evidence- Extract evidence for all requirementsmap_requirement- Map and extract for single requirement
Configuration
Copy .env.example to .env and configure:
# Required for LLM-powered extraction
REGISTRY_REVIEW_ANTHROPIC_API_KEY=sk-ant-api03-...
REGISTRY_REVIEW_LLM_EXTRACTION_ENABLED=true
# Optional
REGISTRY_REVIEW_LLM_MODEL=claude-sonnet-4-5-20250929
REGISTRY_REVIEW_LOG_LEVEL=INFO
See .env.example for all configuration options including chunking, image processing, cost management, and validation settings.
Project Structure
regen-registry-review-mcp/
├── src/registry_review_mcp/
│ ├── server.py # MCP entry point
│ ├── config/ # Settings management
│ ├── extractors/ # PDF and LLM extraction
│ ├── models/ # Pydantic models
│ ├── prompts/ # A-H workflow prompts
│ ├── services/ # Document processing
│ ├── tools/ # MCP tool implementations
│ └── utils/ # State, cache, helpers
├── data/
│ ├── checklists/ # Methodology requirements (JSON)
│ ├── sessions/ # Active sessions (gitignored)
│ └── cache/ # Cached extractions (gitignored)
├── tests/ # Test suite
├── docs/
│ └── specs/ # Workflow specifications
└── examples/ # Test data (Botany Farm)
Development
# Run tests (fast tests only - expensive tests excluded)
uv run pytest
# Format and lint
uv run black src/ tests/
uv run ruff check src/ tests/
Test Markers:
smoke- Critical path tests (<1s)expensive- Tests with API costs (excluded by default)marker- PDF extraction tests (slow, 8GB+ RAM)accuracy- Ground truth validation tests
See pytest.ini for marker configuration.
Requirements
- Python >= 3.10
- uv package manager
- 4GB RAM minimum (8GB recommended for large PDFs)
License
Copyright 2025 Regen Network Development, Inc.
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.
VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.