Last updated: 2026-02-23

CLI

SWE-Agent

Research-driven AI agent from Princeton that automatically resolves GitHub issues using a custom Agent-Computer Interface.

About SWE-Agent

SWE-Agent is an AI agent developed by researchers at Princeton and Stanford that takes GitHub issues and attempts to automatically fix them. It uses a custom Agent-Computer Interface (ACI) that enhances LLM capabilities for browsing repositories, viewing, editing, and executing code files. SWE-Agent achieved state-of-the-art performance on the SWE-bench benchmark, demonstrating its ability to resolve real-world software engineering issues. It supports multiple LLMs and can be used for bug fixing, feature implementation, and even offensive cybersecurity research. The project has recently shifted focus to mini-swe-agent, a dramatically simplified version that achieves 65% on SWE-bench verified in just 100 lines of Python. This demonstrates that effective AI agents can be surprisingly simple when built with the right abstractions.

In-Depth Review

SWE-Agent is a research tool first and a development tool second, and understanding that distinction is key to evaluating it fairly. Its custom Agent-Computer Interface (ACI) is a genuinely clever innovation — instead of giving the LLM raw file contents, it provides structured commands for browsing, searching, and editing repositories, which dramatically improves the model's ability to navigate complex codebases. The recent top-tier conferences publication validates its approach, and the SWE-bench scores are legitimately impressive.

For daily development use, SWE-Agent has significant rough edges. Setup involves cloning the repository, configuring Docker, and managing LLM API keys, which is more involved than installing a VS Code extension. The agent works best with well-structured GitHub issues that clearly describe a bug with reproducible steps — give it a vague feature request and results are poor. The shift toward mini-swe-agent (100 lines of Python, 65%+ on SWE-bench verified) is exciting because it proves that most of SWE-Agent's original complexity was unnecessary, but it also means the main SWE-Agent codebase may receive less attention going forward.

Compared to OpenHands, SWE-Agent is more focused on issue resolution specifically while OpenHands is a broader agent platform. Compared to Devin, SWE-Agent is free and open-source but lacks the polished UX and sandboxed environment. SWE-Agent is best for researchers studying AI software engineering, teams that want to automate triage of simple bug reports, and developers interested in understanding how AI agents interact with codebases. For production daily use, OpenHands or Claude Code are more practical choices.

Key Features

  • Autonomous GitHub issue resolution
  • Custom Agent-Computer Interface (ACI)
  • Repository browsing and code navigation
  • File editing and code execution
  • Support for multiple LLM backends
  • SWE-bench evaluation framework
  • Mini-swe-agent simplified version
  • Competitive coding challenge support

Pros

  • Strong academic backing with published peer-reviewed research
  • State-of-the-art SWE-bench performance
  • Open-source with active research community

Cons

  • Research-oriented tool, not designed for daily development workflows
  • Requires technical setup and LLM API configuration
  • Limited documentation compared to commercial tools

Getting Started with SWE-Agent

1

Clone the repository: git clone https://github.com/SWE-agent/SWE-agent.git

2

Install dependencies: pip install -e . (requires Python 3.9+ and Docker)

3

Set your API key: export ANTHROPIC_API_KEY=your-key or export OPENAI_API_KEY=your-key

4

Run on a GitHub issue: python run.py --issue_url https://github.com/owner/repo/issues/123

5

Alternatively, try mini-swe-agent for a simpler 100-line version: pip install mini-swe-agent

Supported Languages

pythonjavascripttypescriptjavac++gorustruby

Pricing Details

Open Source Free
  • Full SWE-Agent and mini-swe-agent
  • all features
  • bring your own LLM API keys
With Claude Sonnet ~$1-5/issue
  • Best performance on SWE-bench
  • typical cost per issue resolution
With GPT-4o ~$1-3/issue
  • Good performance at lower cost
  • well-supported by the ACI

Best For

Researchers and advanced developers interested in automated issue resolution and AI agent capabilities for software engineering

Verdict

SWE-Agent is a pioneering research tool that proved AI can autonomously resolve real GitHub issues. It's best for researchers and advanced developers — for daily development use, tools like OpenHands or Claude Code provide a more polished experience built on similar principles.

Sources & Methodology

This page is based on public product documentation, vendor pricing pages, and hands-on product testing. Facts may change as vendors update their offerings.

READY TO START? Live Orchestration

[ HIVEOS / LAUNCH ]

Orchestrate Your AI Coding Agents

Manage multiple Claude Code sessions, monitor progress in real-time, and ship faster with HiveOS.