
Omi Memories MCP Server
A Model Context Protocol server that enables access to Omi memories from a specific user account through a tool interface.
Tools
fetch-memories
README
Omi Memories MCP Server
This is a Model Context Protocol (MCP) server that provides access to Omi memories for a specific user through a tool interface.
Features
- Tool to fetch all memories for the specified user ID from the OMI App
Setup
- Install dependencies:
npm install
-
Configure your user ID:
- Open
src/server.ts
- Update the
SPECIFIC_USER_ID
constant with your user ID from the Account section of the Omira App
- Open
-
Build the TypeScript code:
npm run build
- Start the server:
npm start
Available Tools
fetch-memories
Fetches all memories for the configured user ID.
import { Client } from "@modelcontextprotocol/sdk/client/index.js";
import { StdioClientTransport } from "@modelcontextprotocol/sdk/client/stdio.js";
const transport = new StdioClientTransport({
command: "node",
args: ["dist/server.js"]
});
const client = new Client(
{
name: "example-client",
version: "1.0.0"
},
{
capabilities: {
tools: {}
}
}
);
await client.connect(transport);
// Fetch memories using the tool
const result = await client.callTool({
name: "fetch-memories",
arguments: {}
});
console.log(result.content[0].text);
Configuration
The server expects:
- The Express API to be running at
http://localhost:3000
- The user ID should configured : Update the
SPECIFIC_USER_ID
constant insrc/server.ts
to your user ID which you could get from the Account section of the Omira App.
Claude Desktop Integration
To integrate with Claude Desktop, update your Claude Desktop configuration (claude_desktop_config.json
) to include:
{
"mcpServers": {
"omi-mcp": {
"command": "node",
"args": [
"/path/to/your/mcp-server/dist/server.js"
],
"env": {
"NODE_ENV": "development"
}
}
}
}
Cursor IDE Integration
To integrate with Cursor IDE:
- Open Cursor IDE settings
- Navigate to "AI & Copilot" settings
- Under "Model Context Protocol", add a new MCP server with these settings:
{
"name": "Omi Memories",
"command": "node",
"args": [
"/path/to/your/mcp-server/dist/server.js"
],
"cwd": "/path/to/your/mcp-server",
"env": {
"NODE_ENV": "development"
}
}
Replace /path/to/your/mcp-server
with the actual path to your MCP server installation directory.
Recommended Servers
playwright-mcp
A Model Context Protocol server that enables LLMs to interact with web pages through structured accessibility snapshots without requiring vision models or screenshots.
Magic Component Platform (MCP)
An AI-powered tool that generates modern UI components from natural language descriptions, integrating with popular IDEs to streamline UI development workflow.
Audiense Insights MCP Server
Enables interaction with Audiense Insights accounts via the Model Context Protocol, facilitating the extraction and analysis of marketing insights and audience data including demographics, behavior, and influencer engagement.

VeyraX MCP
Single MCP tool to connect all your favorite tools: Gmail, Calendar and 40 more.
graphlit-mcp-server
The Model Context Protocol (MCP) Server enables integration between MCP clients and the Graphlit service. Ingest anything from Slack to Gmail to podcast feeds, in addition to web crawling, into a Graphlit project - and then retrieve relevant contents from the MCP client.
Kagi MCP Server
An MCP server that integrates Kagi search capabilities with Claude AI, enabling Claude to perform real-time web searches when answering questions that require up-to-date information.

E2B
Using MCP to run code via e2b.
Neon Database
MCP server for interacting with Neon Management API and databases
Exa Search
A Model Context Protocol (MCP) server lets AI assistants like Claude use the Exa AI Search API for web searches. This setup allows AI models to get real-time web information in a safe and controlled way.
Qdrant Server
This repository is an example of how to create a MCP server for Qdrant, a vector search engine.