by memgraph
Provides a lightweight server implementation of the Model Context Protocol to enable interaction between Memgraph graph database and large language models.
Enables seamless communication between Memgraph and LLMs by exposing Memgraph operations (e.g., Cypher queries, schema retrieval) through the Model Context Protocol.
uv venv
and .venv\Scripts\activate
.uv add "mcp[cli]" httpx
.uv run server.py
.docker run -p 7687:7687 memgraph/memgraph-mage --schema-info-enabled=True
.--schema-info-enabled=True
.uv
and a single server.py
entry point.uv
is recommended for its speed and simplicity, but any environment that can install mcp[cli]
and httpx
will work.mcpServers
in claude_desktop_config.json
with the command uv
and arguments --directory <repo-path> run server.py
.7687
) exposed so the server can connect.--schema-info-enabled=True
; then get_schema()
will return full schema details.[!IMPORTANT]
This repository has been merged into the Memgraph AI Toolkit monorepo to avoid duplicating tools.
It will be deleted in one month—please follow the MCP integration there for all future development, and feel free to open issues or PRs in that repo.
Memgraph MCP Server is a lightweight server implementation of the Model Context Protocol (MCP) designed to connect Memgraph with LLMs.
uv
and create venv
with uv venv
. Activate virtual environment with .venv\Scripts\activate
.uv add "mcp[cli]" httpx
uv run server.py
.MacOS/Linux
code ~/Library/Application\ Support/Claude/claude_desktop_config.json
Windows
code $env:AppData\Claude\claude_desktop_config.json
Example config:
{
"mcpServers": {
"mpc-memgraph": {
"command": "/Users/katelatte/.local/bin/uv",
"args": [
"--directory",
"/Users/katelatte/projects/mcp-memgraph",
"run",
"server.py"
]
}
}
}
[!NOTE]
You may need to put the full path to the uv executable in the command field. You can get this by runningwhich uv
on MacOS/Linux orwhere uv
on Windows. Make sure you pass in the absolute path to your server.
docker run -p 7687:7687 memgraph/memgraph-mage --schema-info-enabled=True
The --schema-info-enabled
configuration setting is set to True
to allow LLM to run SHOW SCHEMA INFO
query.Run a Cypher query against Memgraph.
Get Memgraph schema information (prerequisite: --schema-info-enabled=True
).
The Memgraph MCP Server is just at its beginnings. We're actively working on expanding its capabilities and making it even easier to integrate Memgraph into modern AI workflows. In the near future, we'll be releasing a TypeScript version of the server to better support JavaScript-based environments. Additionally, we plan to migrate this project into our central AI Toolkit repository, where it will live alongside other tools and integrations for LangChain, LlamaIndex, and MCP. Our goal is to provide a unified, open-source toolkit that makes it seamless to build graph-powered applications and intelligent agents with Memgraph at the core.
Please log in to share your review and rating for this MCP.
Discover more MCP servers with similar functionality and use cases
by danny-avila
Provides a customizable ChatGPT‑like web UI that integrates dozens of AI models, agents, code execution, image generation, web search, speech capabilities, and secure multi‑user authentication, all open‑source and ready for self‑hosting.
by ahujasid
BlenderMCP integrates Blender with Claude AI via the Model Context Protocol (MCP), enabling AI-driven 3D scene creation, modeling, and manipulation. This project allows users to control Blender directly through natural language prompts, streamlining the 3D design workflow.
by pydantic
Enables building production‑grade generative AI applications using Pydantic validation, offering a FastAPI‑like developer experience.
by GLips
Figma-Context-MCP is a Model Context Protocol (MCP) server that provides Figma layout information to AI coding agents. It bridges design and development by enabling AI tools to directly access and interpret Figma design data for more accurate and efficient code generation.
by mcp-use
Easily create and interact with MCP servers using custom agents, supporting any LLM with tool calling and offering multi‑server, sandboxed, and streaming capabilities.
by sonnylazuardi
This project implements a Model Context Protocol (MCP) integration between Cursor AI and Figma, allowing Cursor to communicate with Figma for reading designs and modifying them programmatically.
by lharries
WhatsApp MCP Server is a Model Context Protocol (MCP) server for WhatsApp that allows users to search, read, and send WhatsApp messages (including media) through AI models like Claude. It connects directly to your personal WhatsApp account via the WhatsApp web multi-device API and stores messages locally in a SQLite database.
by idosal
GitMCP is a free, open-source remote Model Context Protocol (MCP) server that transforms any GitHub project into a documentation hub, enabling AI tools to access up-to-date documentation and code directly from the source to eliminate "code hallucinations."
by Klavis-AI
Klavis AI provides open-source Multi-platform Control Protocol (MCP) integrations and a hosted API for AI applications. It simplifies connecting AI to various third-party services by managing secure MCP servers and authentication.