- Explore MCP Servers
- graphiti-mcp-server
Graphiti Mcp Server
What is Graphiti Mcp Server
Graphiti MCP Server is a powerful knowledge graph server designed for AI agents, built using Neo4j and integrated with the Model Context Protocol (MCP).
Use cases
Use cases include enhancing AI agents with contextual knowledge, performing advanced semantic searches, and managing dynamic knowledge graphs for various applications.
How to use
To use Graphiti MCP Server, clone the repository, set up environment variables including your OpenAI API key, and start the services using Docker Compose.
Key features
Key features include dynamic knowledge graph management with Neo4j, seamless integration with OpenAI models, MCP support, Docker-ready deployment, custom entity extraction capabilities, and advanced semantic search functionality.
Where to use
Graphiti MCP Server can be used in various fields such as AI development, data analysis, and knowledge management, particularly where knowledge graphs are beneficial.
Clients Supporting MCP
The following are the main client software that supports the Model Context Protocol. Click the link to visit the official website for more information.
Overview
What is Graphiti Mcp Server
Graphiti MCP Server is a powerful knowledge graph server designed for AI agents, built using Neo4j and integrated with the Model Context Protocol (MCP).
Use cases
Use cases include enhancing AI agents with contextual knowledge, performing advanced semantic searches, and managing dynamic knowledge graphs for various applications.
How to use
To use Graphiti MCP Server, clone the repository, set up environment variables including your OpenAI API key, and start the services using Docker Compose.
Key features
Key features include dynamic knowledge graph management with Neo4j, seamless integration with OpenAI models, MCP support, Docker-ready deployment, custom entity extraction capabilities, and advanced semantic search functionality.
Where to use
Graphiti MCP Server can be used in various fields such as AI development, data analysis, and knowledge management, particularly where knowledge graphs are beneficial.
Clients Supporting MCP
The following are the main client software that supports the Model Context Protocol. Click the link to visit the official website for more information.
Content
Graphiti MCP Server 🧠
🌟 A powerful knowledge graph server for AI agents, built with Neo4j and integrated with Model Context Protocol (MCP).
🚀 Features
- 🔄 Dynamic knowledge graph management with Neo4j
- 🤖 Seamless integration with OpenAI models
- 🔌 MCP (Model Context Protocol) support
- 🐳 Docker-ready deployment
- 🎯 Custom entity extraction capabilities
- 🔍 Advanced semantic search functionality
🛠️ Installation
Prerequisites
- Docker and Docker Compose
- Python 3.10 or higher
- OpenAI API key
- Minimum 4GB RAM (recommended 8GB)
- 2GB free disk space
Quick Start 🚀
- Clone the repository:
git clone https://github.com/gifflet/graphiti-mcp-server.git
cd graphiti-mcp-server
- Set up environment variables:
cp .env.sample .env
- Edit
.envwith your configuration:
# Required for LLM operations
OPENAI_API_KEY=your_openai_api_key_here
MODEL_NAME=gpt-4o
# Optional: Custom OpenAI endpoint
# OPENAI_BASE_URL=https://api.openai.com/v1
# Neo4j Configuration (defaults work with Docker)
NEO4J_URI=bolt://neo4j:7687
NEO4J_USER=neo4j
NEO4J_PASSWORD=demodemo
EOF
- Start the services:
docker compose up -d
- Verify installation:
# Check if services are running
docker compose ps
# Check logs
docker compose logs graphiti-mcp
Alternative: Environment Variables
You can run with environment variables directly:
OPENAI_API_KEY=your_key MODEL_NAME=gpt-4o docker compose up
🔧 Configuration
Service Ports 🌐
| Service | Port | Purpose |
|---|---|---|
| Neo4j Browser | 7474 | Web interface for graph visualization |
| Neo4j Bolt | 7687 | Database connection |
| Graphiti MCP | 8000 | MCP server endpoint |
Environment Variables 🔧
| Variable | Required | Default | Description |
|---|---|---|---|
OPENAI_API_KEY |
✅ | - | Your OpenAI API key |
MODEL_NAME |
❌ | gpt-4o |
OpenAI model to use |
OPENAI_BASE_URL |
❌ | - | Custom OpenAI endpoint |
NEO4J_URI |
❌ | bolt://neo4j:7687 |
Neo4j connection URI |
NEO4J_USER |
❌ | neo4j |
Neo4j username |
NEO4J_PASSWORD |
❌ | demodemo |
Neo4j password |
Neo4j Settings 🗄️
Default configuration for Neo4j:
- Username:
neo4j - Password:
demodemo - URI:
bolt://neo4j:7687(within Docker network) - Memory settings optimized for development
Docker Environment Variables 🐳
You can run with environment variables directly:
OPENAI_API_KEY=your_key MODEL_NAME=gpt-4o docker compose up
🔌 Integration
Cursor IDE Integration 🖥️
- Configure Cursor MCP settings:
{
"mcpServers": {
"Graphiti": {
"command": "uv",
"args": [
"run",
"graphiti_mcp_server.py"
],
"env": {
"OPENAI_API_KEY": "your_key_here"
}
}
}
}
- For Docker-based setup:
{
"mcpServers": {
"Graphiti": {
"url": "http://localhost:8000/sse"
}
}
}
- Add Graphiti rules to Cursor’s User Rules (see
graphiti_cursor_rules.mdc) - Start an agent session in Cursor
Other MCP Clients
The server supports standard MCP transports:
- SSE (Server-Sent Events):
http://localhost:8000/sse - WebSocket:
ws://localhost:8000/ws - Stdio: Direct process communication
💻 Development
Local Development Setup
- Install dependencies:
# Using uv (recommended)
curl -LsSf https://astral.sh/uv/install.sh | sh
uv sync
# Or using pip
pip install -r requirements.txt
- Start Neo4j locally:
docker run -d \ --name neo4j-dev \ -p 7474:7474 -p 7687:7687 \ -e NEO4J_AUTH=neo4j/demodemo \ neo4j:5.26.0
- Run the server:
# Set environment variables
export OPENAI_API_KEY=your_key
export NEO4J_URI=bolt://localhost:7687
# Run with stdio transport
uv run graphiti_mcp_server.py
# Or with SSE transport
uv run graphiti_mcp_server.py --transport sse --use-custom-entities
Testing
# Run basic connectivity test
curl http://localhost:8000/health
# Test MCP endpoint
curl http://localhost:8000/sse
🔍 Troubleshooting
Common Issues
🐳 Docker Issues
# Clean up and restart
docker compose down -v
docker compose up --build
# Check disk space
docker system df
Logs and Debugging
# View all logs
docker compose logs -f
# View specific service logs
docker compose logs -f graphiti-mcp
docker compose logs -f neo4j
# Enable debug logging
docker compose up -e LOG_LEVEL=DEBUG
Performance Issues
- Memory: Increase Neo4j heap size in
docker-compose.yml - Storage: Monitor Neo4j data volume usage
- Network: Check for firewall blocking ports 7474, 7687, 8000
🏗️ Architecture
┌─────────────────┐ ┌──────────────────┐ ┌─────────────────┐ │ MCP Client │ │ Graphiti MCP │ │ Neo4j │ │ (Cursor) │◄──►│ Server │◄──►│ Database │ │ │ │ (Port 8000) │ │ (Port 7687) │ └─────────────────┘ └──────────────────┘ └─────────────────┘ │ ▼ ┌──────────────────┐ │ OpenAI API │ │ (LLM Client) │ └──────────────────┘
Components
- Neo4j Database: Graph storage and querying
- Graphiti MCP Server: API layer and LLM operations
- OpenAI Integration: Entity extraction and semantic processing
- MCP Protocol: Standardized AI agent communication
🤝 Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
📝 License
This project is licensed under the MIT License - see the LICENSE file for details.
🙏 Acknowledgments
- Neo4j team for the amazing graph database
- OpenAI for their powerful LLM models
- MCP community for the protocol specification
- Graphiti Core for the knowledge graph framework
Need help? Open an issue or check our troubleshooting guide above.
Dev Tools Supporting MCP
The following are the main code editors that support the Model Context Protocol. Click the link to visit the official website for more information.










