by atla-ai
The ATLA-MCP-Server is a tool for evaluating the response quality of large language models (LLMs). It provides a standardized interface that enables AI agents to assess LLM responses based on specified criteria, delivering scores and textual feedback.
The Atla MCP Server provides a standardized interface for Large Language Models (LLMs) to interact with the Atla API, which is designed for state-of-the-art LLM-as-a-judge (LLMJ) evaluation. It allows AI agents to evaluate the quality of LLM responses based on specified criteria.
To use the Atla MCP Server, you need an Atla API key. The server can be run manually using uvx
after setting the ATLA_API_KEY
environment variable. It can then be connected to various MCP clients like the OpenAI Agents SDK.
Q: What is Atla? A: Atla is a platform that provides state-of-the-art tools for evaluating and improving Large Language Models.
Q: Do I need an API key to use the server? A: Yes, an Atla API key is required to use the MCP server. You can sign up on the Atla website to get one.
Q: What are the main evaluation tools available?
A: The server provides two main tools: evaluate_llm_response
for single-criterion evaluation and evaluate_llm_response_on_multiple_criteria
for multi-criteria evaluation.
An MCP server implementation providing a standardized interface for LLMs to interact with the Atla API for state-of-the-art LLMJ evaluation.
Learn more about Atla here. Learn more about the Model Context Protocol here.
evaluate_llm_response
: Evaluate an LLM's response to a prompt using a given evaluation criteria. This function uses an Atla evaluation model under the hood to return a dictionary containing a score for the model's response and a textual critique containing feedback on the model's response.evaluate_llm_response_on_multiple_criteria
: Evaluate an LLM's response to a prompt across multiple evaluation criteria. This function uses an Atla evaluation model under the hood to return a list of dictionaries, each containing an evaluation score and critique for a given criteria.To use the MCP server, you will need an Atla API key. You can find your existing API key here or create a new one here.
We recommend using
uv
to manage the Python environment. See here for installation instructions.
Once you have uv
installed and have your Atla API key, you can manually run the MCP server using uvx
(which is provided by uv
):
ATLA_API_KEY=<your-api-key> uvx atla-mcp-server
Having issues or need help connecting to another client? Feel free to open an issue or contact us!
For more details on using the OpenAI Agents SDK with MCP servers, refer to the official documentation.
pip install openai-agents
import os
from agents import Agent
from agents.mcp import MCPServerStdio
async with MCPServerStdio(
params={
"command": "uvx",
"args": ["atla-mcp-server"],
"env": {"ATLA_API_KEY": os.environ.get("ATLA_API_KEY")}
}
) as atla_mcp_server:
...
For more details on configuring MCP servers in Claude Desktop, refer to the official MCP quickstart guide.
claude_desktop_config.json
file:{
"mcpServers": {
"atla-mcp-server": {
"command": "uvx",
"args": ["atla-mcp-server"],
"env": {
"ATLA_API_KEY": "<your-atla-api-key>"
}
}
}
}
You should now see options from atla-mcp-server
in the list of available MCP tools.
For more details on configuring MCP servers in Cursor, refer to the official documentation.
.cursor/mcp.json
file:{
"mcpServers": {
"atla-mcp-server": {
"command": "uvx",
"args": ["atla-mcp-server"],
"env": {
"ATLA_API_KEY": "<your-atla-api-key>"
}
}
}
}
You should now see atla-mcp-server
in the list of available MCP servers.
Contributions are welcome! Please see the CONTRIBUTING.md file for details.
This project is licensed under the MIT License. See the LICENSE file for details.
Reviews feature coming soon
Stay tuned for community discussions and feedback