nefesh-mcp-server

nefesh-mcp-server

Real-time human state awareness for AI agents. Fuses cardiovascular, vocal, visual, and textual signals into a unified stress score (0-100). Streamable HTTP transport with 4 tools: ingest_signal, get_human_state, get_history, delete_subject.

Category
Visit Server

README

Nefesh MCP + A2A Server

A Model Context Protocol and Agent-to-Agent (A2A) server that gives AI agents real-time awareness of human physiological state.

What it does

Send sensor data (heart rate, voice, facial expression, text sentiment), get back a unified state with a machine-readable action your agent can follow directly. Zero prompt engineering required.

On the 2nd+ call, the response includes adaptation_effectiveness — telling your agent whether its previous approach actually worked. A closed-loop feedback system for self-improving agents.

Adaptation Effectiveness (Closed-Loop)

Most APIs give you a state. Nefesh tells you whether your reaction to that state actually worked.

On the 2nd+ call within a session, every response includes:

{
  "state": "focused",
  "stress_score": 45,
  "suggested_action": "simplify_and_focus",
  "adaptation_effectiveness": {
    "previous_action": "de-escalate_and_shorten",
    "previous_score": 68,
    "current_score": 45,
    "stress_delta": -23,
    "effective": true
  }
}

Your agent can read effective: true and know its previous de-escalation worked. If effective: false, the agent adjusts its strategy. No other human-state system provides this feedback loop.

Setup

Option A: Connect first, get a key through your agent (fastest)

Add the config without an API key — your agent will get one automatically.

{
  "mcpServers": {
    "nefesh": {
      "url": "https://mcp.nefesh.ai/mcp"
    }
  }
}

Then ask your agent:

"Connect to Nefesh and get me a free API key for name@example.com"

The agent calls request_api_key → you click one email link → the agent picks up the key. No signup form, no manual copy-paste. After that, add the key to your config for future sessions:

{
  "mcpServers": {
    "nefesh": {
      "url": "https://mcp.nefesh.ai/mcp",
      "headers": {
        "X-Nefesh-Key": "nfsh_free_..."
      }
    }
  }
}

Option B: Get a key first, then connect

Sign up at nefesh.ai/signup (1,000 calls/month, no credit card), then add the config with your key:

{
  "mcpServers": {
    "nefesh": {
      "url": "https://mcp.nefesh.ai/mcp",
      "headers": {
        "X-Nefesh-Key": "YOUR_API_KEY"
      }
    }
  }
}

Agent-specific config files

Agent Config file
Cursor ~/.cursor/mcp.json
Windsurf ~/.codeium/windsurf/mcp_config.json
Claude Desktop ~/Library/Application Support/Claude/claude_desktop_config.json
Claude Code .mcp.json (project root)
VS Code (Copilot) .vscode/mcp.json or ~/Library/Application Support/Code/User/mcp.json
Cline cline_mcp_settings.json (via UI: "Configure MCP Servers")
Continue.dev .continue/config.yaml
Roo Code .roo/mcp.json
Kiro (Amazon) ~/.kiro/mcp.json
OpenClaw ~/.config/openclaw/mcp.json
JetBrains IDEs Settings > Tools > MCP Server
Zed ~/.config/zed/settings.json (uses context_servers)
OpenAI Codex CLI ~/.codex/config.toml
Goose CLI ~/.config/goose/config.yaml
ChatGPT Desktop Settings > Apps > Add MCP Server (UI)
Gemini CLI Settings (UI)
Augment Settings Panel (UI)
Replit Integrations Page (web UI)
LibreChat librechat.yaml (self-hosted)

<details> <summary><strong>VS Code (Copilot)</strong> — uses <code>servers</code> instead of <code>mcpServers</code></summary>

{
  "servers": {
    "nefesh": {
      "type": "http",
      "url": "https://mcp.nefesh.ai/mcp",
      "headers": {
        "X-Nefesh-Key": "<YOUR_API_KEY>"
      }
    }
  }
}

</details>

<details> <summary><strong>Zed</strong> — uses <code>context_servers</code> in settings.json</summary>

{
  "context_servers": {
    "nefesh": {
      "settings": {
        "url": "https://mcp.nefesh.ai/mcp",
        "headers": {
          "X-Nefesh-Key": "<YOUR_API_KEY>"
        }
      }
    }
  }
}

</details>

<details> <summary><strong>OpenAI Codex CLI</strong> — uses TOML in <code>~/.codex/config.toml</code></summary>

[mcp_servers.nefesh]
url = "https://mcp.nefesh.ai/mcp"

</details>

<details> <summary><strong>Continue.dev</strong> — uses YAML in <code>.continue/config.yaml</code></summary>

mcpServers:
  - name: nefesh
    type: streamable-http
    url: https://mcp.nefesh.ai/mcp

</details>

All agents connect via Streamable HTTP — no local installation required.

A2A Integration (Agent-to-Agent Protocol v1.0)

Nefesh is also available as an A2A-compatible agent. While MCP handles tool-calling (your agent calls Nefesh), A2A enables agent-collaboration — other AI agents can communicate with Nefesh as a peer.

Agent Card: /.well-known/agent-card.json

A2A Endpoint: POST https://mcp.nefesh.ai/a2a (JSON-RPC 2.0)

A2A Skill Description
get-human-state Stress state (0-100), suggested_action, adaptation_effectiveness
ingest-signals Send biometric signals, receive unified state
get-trigger-memory Psychological trigger profile (active vs resolved)
get-session-history Timestamped history with trend

Same authentication as MCP — X-Nefesh-Key header or Authorization: Bearer token. Free tier works on both protocols.

MCP Tools

Tool Auth Description
request_api_key No Request a free API key by email. Poll with check_api_key_status until ready.
check_api_key_status No Poll for API key activation. Returns pending or ready with API key.
get_human_state Yes Get stress state (0-100), suggested_action (maintain/simplify/de-escalate/pause), and adaptation_effectiveness — a closed-loop showing whether your previous action reduced stress.
ingest Yes Send biometric signals (heart rate, HRV, voice tone, expression, sentiment, 30+ fields) and get unified state back. Include subject_id for trigger memory.
get_trigger_memory Yes Get psychological trigger profile — which topics cause stress (active) and which have been resolved over time.
get_session_history Yes Get timestamped state history with trend (rising/falling/stable).

How self-provisioning works

Your AI agent can get a free API key autonomously. You only click one email link.

  1. Agent calls request_api_key(email) — no API key needed for this call
  2. You receive a verification email and click the link
  3. Agent polls check_api_key_status(request_id) every 10 seconds
  4. Once verified, the agent receives the API key and can use all other tools

Free tier: 1,000 calls/month, all signal types, 10 req/min. No credit card.

Quick test

After adding the config, ask your AI agent:

"What tools do you have from Nefesh?"

It should list the 6 tools above.

Pricing

Plan Price API Calls
Free $0 1,000/month, no credit card
Solo $25/month 50,000/month
Enterprise Custom Custom SLA

CLI Alternative

Prefer the terminal over MCP? Use the Nefesh CLI (10-32x lower token cost than MCP for AI agents):

npm install -g @nefesh/cli
nefesh ingest --session test --heart-rate 72 --tone calm
nefesh state test --json

GitHub: nefesh-ai/nefesh-cli

Documentation

Privacy

  • No video or audio uploads — edge processing runs client-side
  • No PII stored
  • GDPR/BIPA compliant — cascading deletion via delete_subject
  • Not a medical device — for contextual AI adaptation only

License

MIT — see LICENSE.

Recommended Servers

playwright-mcp

playwright-mcp

A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.

Official
Featured
TypeScript
Magic Component Platform (MCP)

Magic Component Platform (MCP)

An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.

Official
Featured
Local
TypeScript
Audiense Insights MCP Server

Audiense Insights MCP Server

Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.

Official
Featured
Local
TypeScript
VeyraX MCP

VeyraX MCP

Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.

Official
Featured
Local
graphlit-mcp-server

graphlit-mcp-server

The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.

Official
Featured
TypeScript
Kagi MCP Server

Kagi MCP Server

An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.

Official
Featured
Python
E2B

E2B

Using MCP to run code via e2b.

Official
Featured
Neon Database

Neon Database

MCP server for interacting with Neon Management API and databases

Official
Featured
Exa Search

Exa Search

A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.

Official
Featured
Qdrant Server

Qdrant Server

This repository is an example of how to create a MCP server for Qdrant, a vector search engine.

Official
Featured