by GongRzhe
Provides a simple interface to query documents through a Langflow backend using the Model Context Protocol.
Langflow Document QA Server enables document question‑answering by forwarding user queries to a Langflow flow that processes uploaded files with a language model. It implements the Model Context Protocol (MCP) so client applications can invoke the query_docs tool via stdio.
http://127.0.0.1:7860/api/v1/run/<flow-id>?stream=false).API_ENDPOINT environment variable to the copied URL.query_docs tool with a query string; the server returns Langflow’s response.query_docs tool – accepts a natural‑language query and returns answers generated by the Langflow backend.npm run watch recompiles on source changes for rapid development.npm run inspector launches a web UI to monitor stdio communication.Q: Do I need a Langflow account? A: No, Langflow can be run locally; just install it and create the required flow.
Q: Which environment variables are required?
A: Only API_ENDPOINT is needed; it defaults to a sample endpoint if omitted.
Q: How do I debug communication issues?
A: Run npm run inspector to launch the MCP Inspector, which shows request/response logs.
Q: Can I run the server on a remote machine? A: Yes, as long as the Langflow API endpoint is reachable from the server's network.
Q: What Node.js version is supported? A: The project targets the current LTS version of Node.js (≥18).
{
"mcpServers": {
"langflow-doc-qa-server": {
"command": "npx",
"args": ["-y", "@GongRzhe/Langflow-DOC-QA-SERVER"],
"env": {
"API_ENDPOINT": "http://127.0.0.1:7860/api/v1/run/480ec7b3-29d2-4caa-b03b-e74118f35fac"
}
}
}
}
A Model Context Protocol server for document Q&A powered by Langflow
This is a TypeScript-based MCP server that implements a document Q&A system. It demonstrates core MCP concepts by providing a simple interface to query documents through a Langflow backend.
http://127.0.0.1:7860/api/v1/run/<flow-id>?stream=falseAPI_ENDPOINT configurationquery_docs - Query the document Q&A system
Install dependencies:
npm install
Build the server:
npm run build
For development with auto-rebuild:
npm run watch
To use with Claude Desktop, add the server config:
On MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
{
"mcpServers": {
"langflow-doc-qa-server": {
"command": "node",
"args": [
"/path/to/doc-qa-server/build/index.js"
],
"env": {
"API_ENDPOINT": "http://127.0.0.1:7860/api/v1/run/480ec7b3-29d2-4caa-b03b-e74118f35fac"
}
}
}
}
To install Document Q&A Server for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install @GongRzhe/Langflow-DOC-QA-SERVER --client claude
The server supports the following environment variables for configuration:
API_ENDPOINT: The endpoint URL for the Langflow API service. Defaults to http://127.0.0.1:7860/api/v1/run/480ec7b3-29d2-4caa-b03b-e74118f35fac if not specified.Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the MCP Inspector, which is available as a package script:
npm run inspector
The Inspector will provide a URL to access debugging tools in your browser.
This project is licensed under the MIT License.
Please log in to share your review and rating for this MCP.
Discover more MCP servers with similar functionality and use cases
by danny-avila
Provides a customizable ChatGPT‑like web UI that integrates dozens of AI models, agents, code execution, image generation, web search, speech capabilities, and secure multi‑user authentication, all open‑source and ready for self‑hosting.
by ahujasid
BlenderMCP integrates Blender with Claude AI via the Model Context Protocol (MCP), enabling AI-driven 3D scene creation, modeling, and manipulation. This project allows users to control Blender directly through natural language prompts, streamlining the 3D design workflow.
by pydantic
Enables building production‑grade generative AI applications using Pydantic validation, offering a FastAPI‑like developer experience.
by GLips
Figma-Context-MCP is a Model Context Protocol (MCP) server that provides Figma layout information to AI coding agents. It bridges design and development by enabling AI tools to directly access and interpret Figma design data for more accurate and efficient code generation.
by mcp-use
Easily create and interact with MCP servers using custom agents, supporting any LLM with tool calling and offering multi‑server, sandboxed, and streaming capabilities.
by sonnylazuardi
This project implements a Model Context Protocol (MCP) integration between Cursor AI and Figma, allowing Cursor to communicate with Figma for reading designs and modifying them programmatically.
by lharries
WhatsApp MCP Server is a Model Context Protocol (MCP) server for WhatsApp that allows users to search, read, and send WhatsApp messages (including media) through AI models like Claude. It connects directly to your personal WhatsApp account via the WhatsApp web multi-device API and stores messages locally in a SQLite database.
by idosal
GitMCP is a free, open-source remote Model Context Protocol (MCP) server that transforms any GitHub project into a documentation hub, enabling AI tools to access up-to-date documentation and code directly from the source to eliminate "code hallucinations."
by Klavis-AI
Klavis AI provides open-source Multi-platform Control Protocol (MCP) integrations and a hosted API for AI applications. It simplifies connecting AI to various third-party services by managing secure MCP servers and authentication.