
SWE Atlas, Scale's benchmark suite for evaluating coding agents inside real software repositories, is now complete. With the Refactoring leaderboard now live alongside Codebase QnA and Test Writing, the suite evaluates how AI coding agents perform across a spectrum of professional software engineering tasks. Across the evaluation suite, these benchmarks measure the engineering loop across 284 tasks: understanding a system before changing it, validating that behavior holds, and maintaining structure as the codebase evolves.
Built on the foundation of SWE-Bench Pro, the standard for measuring whether agents can resolve real software issues, SWE Atlas considers the engineering work behind those resolutions: the investigation, validation, and maintenance that determine whether code changes can be trusted. Across the suite, top systems cluster in the 40s, with none crossing 50%.

The complete SWE Atlas reveals the gaps remain in the engineering loop for agents. In all three benchmarks agents struggle to investigate systems thoroughly, write tests with surgical precision, and complete refactors without leaving stale code behind. These are different failures, but they point to the same broader limitation: agents can often get part of the way there on tasks but they struggle to finish the work with the completeness real engineering requires.
Reliability is a second gap. When models try the same task three separate times, they are two to three times more likely to succeed on one attempt than succeeding on all three. Capability and consistency are improving on different timelines, and a coding agent that solves a task on one attempt out of three is not reliable enough for real life environments.

In Codebase QnA, an agent is placed inside a real repository and asked the kinds of questions engineers face in real work. The benchmark spans five categories:
Top scores sit in the low 40s, and the failure patterns are revealing. For example, GPT models tend to run experiments but miss rubric sub-questions, leaving answers incomplete. Claude models tend to reason from the source code without actually running anything, even when the prompt explicitly asks for runtime evidence. Different model families fail in different ways, but answering these questions well requires investigating the system in motion, beyond what reading the code alone can show.
Test Writing gives an agent a real repository with missing tests for an important behavior, described at a high level in the prompt. The agent has to explore the codebase, identify the right test targets, place the tests in the appropriate files, follow repository conventions, and produce a manifest of the tests it wrote. The benchmark spans three test types:
Top scores reach the mid 40s. The biggest failure mode is weak assertions: agents write tests that look comprehensive but pass equally well on broken code, so the tests would not catch regressions if a real bug were introduced.
A common pattern is volume over precision: writing more tests does not necessarily improve performance, and higher test counts correlate with higher mutation failure rates. The leading model under a common scaffold writes the fewest tests on average and lands the highest score. The benchmark rewards agents that find the right tests to write, rather than agents that simply add more tests.
The final leaderboard in this suite, Refactoring, analyzes how models reorganize code. The refactoring tasks span four types:
Success requires editing across multiple files, updating call sites, preserving existing behavior, passing existing tests, removing stale code, keeping documentation aligned, and avoiding regressions.
These are the broadest code-change tasks in SWE Atlas. Reference solutions involve roughly twice the lines of code changed and 1.7 times the file edits of SWE-Bench Pro tasks. They span production repositories across Go, TypeScript, Python, C, C++, and JavaScript. Top scores sit just under 50%.
Most models pass on test-preservation but struggle on completeness. The common failures:
Passing tests is necessary, but a complete refactor also has to leave the codebase cleaner than it found it.
A pattern shows up across all three benchmarks: stronger agents are more investigative. The gap between models that explore aggressively and ones that don't is large, and it appears in every benchmark in the suite.
In Codebase QnA, top models execute code at much higher rates than weaker ones, setting up applications, sending live requests, and performing runtime analysis to understand behavior rather than reasoning from the source alone. In Test Writing, the leading systems frontload codebase exploration, aggressively searching and reading early in the trajectory before writing any tests. In Refactoring, success correlates strongly with file-edit recall: whether the agent identified every file that needed changing instead of making incomplete or sloppy refactors.
Across SWE Atlas, investigation correlates directly with success. Agents operate through tools, and SWE Atlas measures model-plus-scaffold behavior because that is how coding agents are actually used. Models running in their native scaffolds (Claude Code, Codex CLI) perform 1.5x to 2x more exploration, search, and execution than the same models on a generic harness, and they score noticeably higher. A coding agent's capability is inseparable from the environment it operates in.
Coding ability is not a single capability. A model can be strong at producing patches and still struggle to investigate a system. It can write plausible tests while missing the behavior that matters. It can pass existing tests while leaving a refactor incomplete. These are different failure modes, and they matter in different ways for production software.
Frontier coding models have been moving fast on issue resolution. SWE-Bench Pro shows top frontier scores rising into the high 50s. SWE Atlas asks what happens around those patches: the investigation, validation, and maintenance work. The same models hover in the high 40s.
Tracking the Claude Opus 4.x series shows substantial gains on each benchmark in the suite over the past eight months:

Resolve rates roughly tripled on Codebase Q&A (from ~17% to ~49%), more than doubled on Test Writing, and climbed from under 20% to ~45% on Refactoring. Though trajectories are steep, the current ceiling is very much there.
The concern now extends beyond code generation into engineering judgment under context: knowing what to look at, deciding what evidence matters, finding all the call sites a refactor touches, writing the right tests rather than many tests, and cleaning up after the change. With SWE Atlas complete, the shape of that gap is now visible and measurable on all three leaderboards.