by ferrislucas
iterm-mcp is a Model Context Protocol (MCP) server that enables Large Language Models (LLMs) to interact with and control your iTerm2 terminal session on macOS, acting as a bridge for AI models to execute commands, read output, and send control characters within the terminal environment.
iterm-mcp is a Model Context Protocol (MCP) server designed to enable Large Language Models (LLMs) to interact with and control your iTerm2 terminal session on macOS. It acts as a bridge, allowing AI models to execute commands, read output, and send control characters within the terminal environment.
iterm-mcp can be integrated with MCP clients like Claude Desktop. To use it, you need to configure your MCP client to recognize iterm-mcp as a server. This typically involves adding a JSON configuration to your client's settings, specifying the command to run npx -y iterm-mcp
. Alternatively, you can use Smithery to automatically install and configure iterm-mcp for Claude Desktop with a simple npx -y @smithery/cli install iterm-mcp --client claude
command.
npx
, making it easy to integrate with various MCP clients.write_to_terminal
(to execute commands), read_terminal_output
(to read output), and send_control_character
(to send control signals).yarn run inspector
and then yarn debug <command>
to access debugging tools in your browser.A Model Context Protocol server that provides access to your iTerm session.
Efficient Token Use: iterm-mcp gives the model the ability to inspect only the output that the model is interested in. The model typically only wants to see the last few lines of output even for long running commands.
Natural Integration: You share iTerm with the model. You can ask questions about what's on the screen, or delegate a task to the model and watch as it performs each step.
Full Terminal Control and REPL support: The model can start and interact with REPL's as well as send control characters like ctrl-c, ctrl-z, etc.
Easy on the Dependencies: iterm-mcp is built with minimal dependencies and is runnable via npx. It's designed to be easy to add to Claude Desktop and other MCP clients. It should just work.
write_to_terminal
- Writes to the active iTerm terminal, often used to run a command. Returns the number of lines of output produced by the command.read_terminal_output
- Reads the requested number of lines from the active iTerm terminal.send_control_character
- Sends a control character to the active iTerm terminal.To use with Claude Desktop, add the server config:
On macOS: ~/Library/Application Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
{
"mcpServers": {
"iterm-mcp": {
"command": "npx",
"args": [
"-y",
"iterm-mcp"
]
}
}
}
To install iTerm for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install iterm-mcp --client claude
Install dependencies:
yarn install
Build the server:
yarn run build
For development with auto-rebuild:
yarn run watch
Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the MCP Inspector, which is available as a package script:
yarn run inspector
yarn debug <command>
The Inspector will provide a URL to access debugging tools in your browser.
Please log in to share your review and rating for this MCP.
Discover more MCP servers with similar functionality and use cases
by danny-avila
Provides a customizable ChatGPT‑like web UI that integrates dozens of AI models, agents, code execution, image generation, web search, speech capabilities, and secure multi‑user authentication, all open‑source and ready for self‑hosting.
by ahujasid
BlenderMCP integrates Blender with Claude AI via the Model Context Protocol (MCP), enabling AI-driven 3D scene creation, modeling, and manipulation. This project allows users to control Blender directly through natural language prompts, streamlining the 3D design workflow.
by pydantic
Enables building production‑grade generative AI applications using Pydantic validation, offering a FastAPI‑like developer experience.
by GLips
Figma-Context-MCP is a Model Context Protocol (MCP) server that provides Figma layout information to AI coding agents. It bridges design and development by enabling AI tools to directly access and interpret Figma design data for more accurate and efficient code generation.
by mcp-use
Easily create and interact with MCP servers using custom agents, supporting any LLM with tool calling and offering multi‑server, sandboxed, and streaming capabilities.
by sonnylazuardi
This project implements a Model Context Protocol (MCP) integration between Cursor AI and Figma, allowing Cursor to communicate with Figma for reading designs and modifying them programmatically.
by lharries
WhatsApp MCP Server is a Model Context Protocol (MCP) server for WhatsApp that allows users to search, read, and send WhatsApp messages (including media) through AI models like Claude. It connects directly to your personal WhatsApp account via the WhatsApp web multi-device API and stores messages locally in a SQLite database.
by idosal
GitMCP is a free, open-source remote Model Context Protocol (MCP) server that transforms any GitHub project into a documentation hub, enabling AI tools to access up-to-date documentation and code directly from the source to eliminate "code hallucinations."
by Klavis-AI
Klavis AI provides open-source Multi-platform Control Protocol (MCP) integrations and a hosted API for AI applications. It simplifies connecting AI to various third-party services by managing secure MCP servers and authentication.