
Can AI actually automate jobs? To answer this important question and to help people track AI’s impact on jobs, Scale and the Center for AI Safety (CAIS) are introducing the Remote Labor Index (RLI). This is the first benchmark and public leaderboard to test AI agents on how well they can complete paid freelance work from fields like software development, design, architecture, and data analysis. We chose freelance work because it gives us the clearest possible view of what AI agents can actually do on their own.
When it comes to freelance work, each project represents a true distribution of complete, self-contained, economically valuable tasks, requiring a median of around 11.5 hours for a human professional to complete with a median value of $200. With a clear brief, price, and deliverable, the RLI evaluates if the AI can be the freelancer and complete the entire, complex, paid project from start to finish, rather than if an AI can be a helpful assistant for parts of a project.
Our research shows that the best-performing AI agents can successfully automate 2.5% of these real-world, paid freelance projects. Across 240 real-world projects spanning 23 domains, the human freelancers who originally completed this work earned a combined $143,991. In contrast, the top-performing agent, Manus, earned $1,720. This 2.5% success rate gives us a crucial starting point for tracking AI's trajectory and grounding public discussion in data.

Performance of current AI agents on the RLI. All models tested automate less than 3% of complex, end-to-end freelance projects, with the highest-scoring agent achieving a 2.5% automation rate.
To understand the high failure rate, our team conducted an analysis of the hundreds of failed submissions. We found that agent failures are not random but cluster into several key patterns, revealing why they are not ready for complex, professional work.
We found that:
45.6% of failed submissions had quality issues where the work was simply not at a professional standard a client would accept.
35.7% included incomplete or malformed deliverables, with agents submitting truncated videos, missing source files, or empty directories.
17.6% had technical and file integrity issues, producing corrupt, unusable, or empty files.
14.8% contained inconsistencies, failing to maintain visual or logical consistency across different files in the same project.
A single failed project often exhibited several of these patterns at once. For instance, in a jewelry design project, an agent was instructed to modify a provided ring image to change the diamond's cut. Instead, it ignored the provided file and submitted two entirely new, AI-generated images. This single failure demonstrates a quality issue (the images were amateurish), a failure to follow the brief (it ignored the input file), and inconsistency (the two new rings failed to match each other).

The projects that were successfully completed were not random. They were predominantly creative projects involving audio and image generation (like creating sound effects or logos) and some data or writing projects (like report generation). This reveals a clear pattern: agents excel at creating from scratch from a simple prompt, but fail at complex editing or following a precise, multi-step brief. Their generative skill is high, but their ability to act as a reliable, detail-oriented professional is low.
The RLI’s findings provide a clear, data-driven answer to the question we started with. The fear of imminent, widespread automation is not supported by the data; the 97.5% failure rate shows that AI is not yet capable of autonomously performing complex, professional work. Our analysis reveals a critical gap between AI's skill on isolated tasks and the end-to-end reliability required to fulfill a real-world client brief.
The 2.5% success, however, is just as telling. It shows that AI is already at a professional level for some generative tasks (creating images, audio, or code from scratch). The failure comes when the job requires complex editing, tool use, and following precise, multi-step specifications. This suggests the immediate impact isn't mass automation, but augmentation.
The next great challenge is about capability, reliability, and scale, building agents that can move from simple prompts to complex project execution. The Remote Labor Index provides the economically-grounded tool to guide and measure this next phase, grounding the public discussion in data, not speculation.