by gptme
Empowers large language models to act as personal AI assistants directly inside the terminal, providing capabilities such as code execution, file manipulation, web browsing, vision, and interactive tool usage.
Gptme enables an LLM to operate in a local command‑line environment, turning the model into an autonomous assistant that can run shell commands, execute Python snippets, edit files, browse the internet, process images, and interact with GUI applications.
pipx install gptme
(requires Python 3.10+).gptme
to open an interactive chat session./tools
, /model
).shell
, python
, patch
, browser
, vision
, or computer
to fulfil the request.-m openai/gpt-4
), workspace (-w ./myproj
), enable/disable specific tools (-t shell,patch
), or run in non‑interactive mode.llama.cpp
models.Q: Do I need an internet connection? A: Only if you use remote LLM providers (OpenAI, Anthropic, etc.) or the browser tool. Local models run entirely offline.
Q: How is privacy handled? A: When using local models, no data leaves your machine. Remote providers follow their own policies.
Q: Can I integrate Gptme into other tools? A: Yes – the web UI, REST API, and sub‑agent architecture allow embedding Gptme in custom workflows.
Q: What platforms are supported? A: Works on any OS with Python 3.10+, including Linux, macOS, and Windows (via WSL or native).
[!NOTE] These demos are very out of date (2023) and do not reflect the latest capabilities.
You can find more Demos and Examples in the documentation.
llama.cpp
stdin
or as arguments.
GPTME_TOOL_SOUNDS=true
mypy
, ruff
, and pyupgrade
.Install with pipx:
# requires Python 3.10+
pipx install gptme
Now, to get started, run:
gptme
Here are some examples:
gptme 'write an impressive and colorful particle effect using three.js to particles.html'
gptme 'render mandelbrot set to mandelbrot.png'
gptme 'suggest improvements to my vimrc'
gptme 'convert to h265 and adjust the volume' video.mp4
git diff | gptme 'complete the TODOs in this diff'
make test | gptme 'fix the failing tests'
For more, see the Getting Started guide and the Examples in the documentation.
$ gptme --help
Usage: gptme [OPTIONS] [PROMPTS]...
gptme is a chat-CLI for LLMs, empowering them with tools to run shell
commands, execute code, read and manipulate files, and more.
If PROMPTS are provided, a new conversation will be started with it. PROMPTS
can be chained with the '-' separator.
The interface provides user commands that can be used to interact with the
system.
Available commands:
/undo Undo the last action
/log Show the conversation log
/tools Show available tools
/model List or switch models
/edit Edit the conversation in your editor
/rename Rename the conversation
/fork Copy the conversation using a new name
/summarize Summarize the conversation
/replay Rerun tools in the conversation, won't store output
/impersonate Impersonate the assistant
/tokens Show the number of tokens used
/export Export conversation as HTML
/commit Ask assistant to git commit
/setup Setup gptme with completions and configuration
/help Show this help message
/exit Exit the program
Keyboard shortcuts:
Ctrl+X Ctrl+E Edit prompt in your editor
Ctrl+J Insert a new line without executing the prompt
Options:
--name TEXT Name of conversation. Defaults to generating a random
name.
-m, --model TEXT Model to use, e.g. openai/gpt-5, anthropic/claude-
sonnet-4-20250514. If only provider given then a
default is used.
-w, --workspace TEXT Path to workspace directory. Pass '@log' to create a
workspace in the log directory.
--agent-path TEXT Path to agent workspace directory.
-r, --resume Load most recent conversation.
-y, --no-confirm Skip all confirmation prompts.
-n, --non-interactive Non-interactive mode. Implies --no-confirm.
--system TEXT System prompt. Options: 'full', 'short', or something
custom.
-t, --tools TEXT Tools to allow as comma-separated list. Available:
append, browser, chats, choice, computer, gh,
ipython, morph, patch, rag, read, save, screenshot,
shell, subagent, tmux, tts, vision, youtube.
--tool-format TEXT Tool format to use. Options: markdown, xml, tool
--no-stream Don't stream responses
--show-hidden Show hidden system messages.
-v, --verbose Show verbose output.
--version Show version and configuration information
--help Show this message and exit.
Please log in to share your review and rating for this MCP.
Discover more MCP servers with similar functionality and use cases
by zed-industries
Provides real-time collaborative editing powered by Rust, enabling developers to edit code instantly across machines with a responsive, GPU-accelerated UI.
by cline
Provides autonomous coding assistance directly in the IDE, enabling file creation, editing, terminal command execution, browser interactions, and tool extension with user approval at each step.
by continuedev
Provides continuous AI assistance across IDEs, terminals, and CI pipelines, offering agents, chat, inline editing, and autocomplete to accelerate software development.
by block
Automates engineering tasks by installing, executing, editing, and testing code using any large language model, providing end‑to‑end project building, debugging, workflow orchestration, and external API interaction.
by RooCodeInc
An autonomous coding agent that lives inside VS Code, capable of generating, refactoring, debugging code, managing files, running terminal commands, controlling a browser, and adapting its behavior through custom modes and instructions.
by lastmile-ai
A lightweight, composable framework for building AI agents using Model Context Protocol and simple workflow patterns.
by wonderwhy-er
DesktopCommanderMCP is a Model Context Protocol (MCP) server that extends Claude's capabilities to include terminal control, file system search, and diff file editing. It transforms Claude into a powerful development and automation assistant by enabling AI to interact directly with your computer's file system and execute terminal commands.
by opensumi
A framework for quickly building AI native IDE products, supporting Model Context Protocol tools via an MCP server.
by evalstate
Create and interact with sophisticated AI agents and workflows using a declarative syntax, with built‑in support for MCP features, multimodal inputs, and a wide range of language model providers.