by stass
Integrates the LLDB debugger with Claude's Model Context Protocol, enabling AI‑driven control and interaction with debugging sessions.
Lldb Mcp connects the LLDB debugging engine to Claude’s Model Context Protocol, allowing natural‑language commands to start, manage, and inspect LLDB sessions. It turns AI interaction into a practical workflow for debugging C/C++ programs, core dumps, and live processes.
pip install mcp
)."mcpServers": {
"lldb-mcp": {
"command": "python3",
"args": ["/path/to/lldb-mcp/lldb_mcp.py"],
"disabled": false
}
}
--debug
flag provides detailed server logs.python3 lldb_mcp.py &
or use a process manager.lldb_help
command to retrieve LLDB’s help text.See it in acton here, automatically debugging a buffer overflow: https://x.com/full_duplex/status/1904770477698277847
LLDB-MCP is a tool that integrates the LLDB debugger with Claude's Model Context Protocol (MCP). This integration allows Claude to start, control, and interact with LLDB debugging sessions directly, enabling AI-assisted debugging workflows.
The tool provides a comprehensive set of commands for working with LLDB, including:
Clone the repository:
git clone https://github.com/stass/lldb-mcp.git
cd lldb-mcp
Install dependencies:
pip install mcp
Configure Claude to use the LLDB-MCP server:
"mcpServers": {
"lldb-mcp": {
"command": "python3",
"args": ["/path/to/lldb-mcp/lldb_mcp.py"],
"disabled": false
}
}
Once installed and configured, you can interact with LLDB through Claude using natural language.
Here are some examples of how to interact with LLDB-MCP through Claude:
lldb_start
: Start a new LLDB sessionlldb_terminate
: Terminate an LLDB sessionlldb_list_sessions
: List all active LLDB sessionslldb_load
: Load a program into LLDBlldb_attach
: Attach to a running processlldb_load_core
: Load a core dump filelldb_run
: Run the loaded programlldb_continue
: Continue program executionlldb_step
: Step to next line or instructionlldb_next
: Step over function callslldb_finish
: Execute until the current function returnslldb_kill
: Kill the running processlldb_set_breakpoint
: Set a breakpointlldb_breakpoint_list
: List all breakpointslldb_breakpoint_delete
: Delete a breakpointlldb_watchpoint
: Set a watchpoint on a variable or memory addresslldb_backtrace
: Show call stacklldb_print
: Print value of expressionlldb_examine
: Examine memorylldb_info_registers
: Display registerslldb_frame_info
: Get detailed information about a stack framelldb_disassemble
: Disassemble codelldb_process_info
: Get information about the current processlldb_thread_list
: List all threads in the current processlldb_thread_select
: Select a specific threadlldb_command
: Execute an arbitrary LLDB commandlldb_expression
: Evaluate an expression in the current framelldb_help
: Get help for LLDB commandsexample/overflow.c
contains an example C program that causes buffer overflow with certain arguments.
Compile it using cc overflow.c
and ask Claude to debug the issue with the resulting program:
I'm trying to debug my program a.out that is crashing with certain arguments, e.g. when I pass "hello".
Can you help me debug it?
lldb_command
when you need to execute LLDB commands that don't have a dedicated function--debug
flag when starting the server for detailed loggingBSD 2-clause
Please log in to share your review and rating for this MCP.
{ "mcpServers": { "lldb-mcp": { "command": "python3", "args": [ "/path/to/lldb-mcp/lldb_mcp.py" ], "env": {} } } }
Explore related MCPs that share similar capabilities and solve comparable challenges
by zed-industries
A high‑performance, multiplayer code editor designed for speed and collaboration.
by modelcontextprotocol
Model Context Protocol Servers
by modelcontextprotocol
A Model Context Protocol server for Git repository interaction and automation.
by modelcontextprotocol
A Model Context Protocol server that provides time and timezone conversion capabilities.
by cline
An autonomous coding assistant that can create and edit files, execute terminal commands, and interact with a browser directly from your IDE, operating step‑by‑step with explicit user permission.
by continuedev
Enables faster shipping of code by integrating continuous AI agents across IDEs, terminals, and CI pipelines, offering chat, edit, autocomplete, and customizable agent workflows.
by upstash
Provides up-to-date, version‑specific library documentation and code examples directly inside LLM prompts, eliminating outdated information and hallucinated APIs.
by github
Connects AI tools directly to GitHub, enabling natural‑language interactions for repository browsing, issue and pull‑request management, CI/CD monitoring, code‑security analysis, and team collaboration.
by daytonaio
Provides a secure, elastic infrastructure that creates isolated sandboxes for running AI‑generated code with sub‑90 ms startup, unlimited persistence, and OCI/Docker compatibility.