GhidrAssist Documentation

GhidrAssist is a comprehensive LLM-powered plugin for Ghidra that enhances reverse engineering workflows through intelligent automation.

GhidrAssist Main Interface

What is GhidrAssist?

GhidrAssist integrates large language models directly into your Ghidra workflow. Instead of switching between tools or copying code snippets, you can ask questions, generate explanations, and receive intelligent suggestions without leaving CodeBrowser.

The plugin supports local LLMs (Ollama, LM Studio) and cloud providers (OpenAI, Anthropic), letting you choose between privacy, cost, and performance.

Core Capabilities

GhidrAssist is organized into seven tabs, each focused on a specific aspect of LLM-assisted analysis:

Tab Purpose
Explain Generate and store function explanations with security analysis
Query Interactive chat with context macros and autonomous ReAct agent
Actions LLM-powered suggestions for renaming, retyping, and struct creation
Semantic Graph Build a knowledge graph of function relationships and security properties
RAG Manage external documentation for context-enhanced queries
Settings Configure LLM providers, MCP servers, and plugin options

Key Features

MCP Tool Integration

GhidrAssist supports the Model Context Protocol (MCP), allowing LLMs to interact with Ghidra through tool calls. When MCP is enabled, the LLM can:

See the Query Workflow for details on using MCP tools.

ReAct Autonomous Agent

For complex investigations, GhidrAssist includes a ReAct (Reasoning + Acting) agent that can autonomously:

  1. Plan an investigation based on your question
  2. Execute multiple tool calls to gather information
  3. Reflect on findings and adapt its approach
  4. Synthesize a comprehensive answer

This is useful for exploratory questions like “What does this binary do?” or “Trace the data flow from user input.”

Extended Thinking

For models that support it (Claude Sonnet 4+, OpenAI o1/o3, local reasoning models), GhidrAssist provides reasoning effort control:

Level Use Case
None Quick queries, simple questions
Low Light reasoning tasks
Medium Moderate complexity analysis
High Deep analysis, complex vulnerability research

Higher reasoning effort allows the model more “thinking time” but increases latency and cost.

Semantic Graph

Build a rich knowledge graph of your binary that captures:

The graph can be queried directly or used to enhance LLM responses.

SymGraph Cloud Sharing

Share your analysis with the community through SymGraph:

Supported LLM Providers

GhidrAssist supports a wide range of LLM providers:

Provider Type Description
Ollama Local LLM inference (recommended for privacy)
LM Studio Local models with GUI
Open WebUI Self-hosted web interface
OpenAI Platform API GPT-5+, GPT-4o, o1, o3, o4 with API key
OpenAI OAuth ChatGPT Pro/Plus subscription
Anthropic Platform API Claude models with API key
Anthropic OAuth Claude Pro/Max subscription
Anthropic CLI Claude Code CLI wrapper
LiteLLM Proxy for 100+ providers (AWS Bedrock, Azure, etc.)

Architecture Overview

GhidrAssist uses a Model-View-Controller architecture:

All LLM operations run in background tasks so the UI stays responsive, with streaming for real-time output.

Getting Started

Ready to start using GhidrAssist? See the Getting Started Guide for installation and configuration instructions.

Common Workflows

Tab Reference