by nkapila6
mcp-local-rag is a "primitive" RAG-like web search model context protocol (MCP) server that runs locally, enabling large language models (LLMs) to perform live web searches and retrieve up-to-date information without relying on external APIs.
mcp-local-rag is a local server implementing the Model Context Protocol (MCP) to provide Retrieval-Augmented Generation (RAG) capabilities for large language models. It allows LLMs to perform web searches and incorporate real-time information into their responses, effectively bypassing the need for external APIs for web access.
mcp-local-rag can be integrated into your MCP server configuration. There are two primary methods for installation:
uvx
: This method requires uv
to be installed. You add a specific JSON configuration to your MCP server settings, which instructs it to run mcp-local-rag
using uvx
with Python 3.10.
{
"mcpServers": {
"mcp-local-rag":{
"command": "uvx",
"args": [
"--python=3.10",
"--from",
"git+https://github.com/nkapila6/mcp-local-rag",
"mcp-local-rag"
]
}
}
}
mcp-local-rag
as a Docker container.
{
"mcpServers": {
"mcp-local-rag": {
"command": "docker",
"args": [
"run",
"--rm",
"-i",
"--init",
"-e",
"DOCKER_CONTAINER=true",
"ghcr.io/nkapila6/mcp-local-rag:latest"
]
}
}
}
Once configured, any MCP client that supports tool calling can leverage mcp-local-rag
.
"primitive" RAG-like web search model context protocol (MCP) server that runs locally. ✨ no APIs ✨
Locate your MCP config path here or check your MCP client settings.
uvx
This is the easiest and quickest method. You need to install uv for this to work. Add this to your MCP server configuration:
{
"mcpServers": {
"mcp-local-rag":{
"command": "uvx",
"args": [
"--python=3.10",
"--from",
"git+https://github.com/nkapila6/mcp-local-rag",
"mcp-local-rag"
]
}
}
}
Ensure you have Docker installed. Add this to your MCP server configuration:
{
"mcpServers": {
"mcp-local-rag": {
"command": "docker",
"args": [
"run",
"--rm",
"-i",
"--init",
"-e",
"DOCKER_CONTAINER=true",
"ghcr.io/nkapila6/mcp-local-rag:latest"
]
}
}
}
MseeP does security audits on every MCP server, you can see the security audit of this MCP server by clicking here.
The MCP server should work with any MCP client that supports tool calling. Has been tested on the below clients.
When an LLM (like Claude) is asked a question requiring recent web information, it will trigger mcp-local-rag
.
When asked to fetch/lookup/search the web, the model prompts you to use MCP server for the chat.
In the example, have asked it about Google's latest Gemma models released yesterday. This is new info that Claude is not aware about.
mcp-local-rag
performs a live web search, extracts context, and sends it back to the model—giving it fresh knowledge:
Have ideas or want to improve this project? Issues and pull requests are welcome!
This project is licensed under the MIT License.
Please log in to share your review and rating for this MCP.