Historically, if you wanted to estimate how long a software task would take, you would ask a software engineer to "raw dog it"[1] and time how long it takes them. This is mostly impossible now, because almost all software engineers are highly dependent on LLMs. An engineer today who tries to do a given task without AI assistance would be significantly slower than a "pre-LLM" engineer on the same task. If you tried to use post-LLM engineer data to estimate LLM time horizons, you'd get biased results.[2]
METR might be able to hire people doing competitive programming problems or similar competitions, since they still spend a lot of time working without the help of LLMs. E.g. many high school students might be interested in such studies.
These are not professional SWEs. Professional SWE tasks look very different competitive programming problems and also it's hard to create competitive programming tasks that take 8h-30h.
The low-background steel problem
Modern steel is slightly radioactive. We did a lot of atomic testing in the 40s and 50s, and now our atmosphere has some amount of radioactive particles, which make their way into steel during production. This is mostly fine, but some scientific instruments require steel that is not radioactive. In order to get such steel, people typically scavenge for pre-1950s steel (such as the sunken Imperial German Navy). Here's a photo of the SMS Hindenburg being salvaged.
You might see an analogy here between pre-war steel and pre-LLM data. Indeed, this idea has been discussed in various places last year: in a LinkedIn post, an Ars Technica article, a Business Insider article, a Harvard JOLT article. It's a good analogy, but not the one we wish to make. Instead of talking about data, let's talk about humans.
Modern software engineers are radioactive.
Historically, if you wanted to estimate how long a software task would take, you would ask a software engineer to "raw dog it"[1] and time how long it takes them. This is mostly impossible now, because almost all software engineers are highly dependent on LLMs. An engineer today who tries to do a given task without AI assistance would be significantly slower than a "pre-LLM" engineer on the same task. If you tried to use post-LLM engineer data to estimate LLM time horizons, you'd get biased results.[2]
Let's say we wanted to extend the METR benchmark by adding more tasks in the 8-30 hour range. But when we give humans a METR task they were able to do in 8 hours in 2024, it now takes them 20 hours. Thus, we can’t just extend the METR benchmark in a naive way. Incidentally, METR has recently written about their experience with contaminated developers. In their Feb 2026 'uplift update', they say: "When surveyed, 30% to 50% of developers told us that they were choosing not to submit some tasks because they did not want to do them without AI."[3]
We're facing a new and fundamental problem with getting human data on long-horizon software/agentic tasks. Humans have gotten used to engineering with AI. We can’t compare a human’s performance on software tasks in 2026 to 2024.
Cyborg evals
How do we address this problem? We propose 'cyborg evals': studies which compare eval performance between current gen AI-assisted humans and next gen AI-assisted humans. That is, we are comparing . Cyborg evals are thus a kind of uplift study. These studies help us understand the practical improvements that new models offer, even as AI advances and human capabilities change. They are also relatively easy and cheap to do.
We can imagine some variations on this kind of study. For example, we could compare . Consider that if model N is able to do 100% of N+human, then humans are cooked.[4]. If you plot this over time, you're on your way to forecasting full automation in the eval's domain.
This is what a forecast informed by cyborg evals might look like. (Domain is software engineering.)
Cyborg evals can also tell us about the relationship between AI labor and human labor. From an employer's perspective, AI labor and human labor are both economic substitutes (AI can be used as a replacement for web devs) and complements (AI increases the productivity of and demand for senior engineers). What is the "exchange rate"?
Suppose you want to estimate the cost of completing a software task in 2026. You might want to look at the cost to complete the task with different configurations of human labor and AI labor. While researching this article, we found a few recent data points for the difficult task of making a (good) compiler. Anthropic "tasked Opus 4.6 using agent teams to build a C Compiler, and then (mostly) walked away". Paul Biggar, who has a PhD in compiler design, built a Darklang compiler in 1-2 weeks using Claude Code, a task that he says "it would have taken [] at least 2 years to do" unassisted.[5] Here's a graph of the costs of these projects; note the synergy effects.
The tradeoff between AI labor and engineer labor depends on both AI capabilities and on tasks. Cyborg evals will give us a quantitative view of this tradeoff.
What kind of tasks would be good for cyborg evals?
What do we want out of tasks for cyborg evals? One important property is something like difficulty+discernment. We want evals that are hard in the sense that they don't get easily saturated. A standard approach here is to have an eval contain tasks of varying difficulty level. This is basically what you have to do if your tasks are pass/fail. Alternatively, you can choose tasks with a wide range of ordinal outcomes. For example, Vending-Bench tests how much money models make in a simulated business scenario. Ideally, tasks can distinguish performance even into the far future.
Another desired property is something like economic usefulness. We want the results on our evals to correspond to work that is actually done; economic usefulness corresponds very well to practical usefulness and value. It gives us a good way to think about capabilities. Accordingly, tasks should be similar to what professional human (or human+AI) teams are paid to do in the real world. We think good tasks here will be those for which is easy and fun for a human+AI to solve together but hard for AI to solve by itself. These tasks form the frontier of software engineering today; this is the kind of thing people get paid to do!
Here are some specific ideas for kinds of tasks that would work well for cyborg evals. All our suggestions do well on the difficulty+discernment axis, though they vary in terms of economic usefulness.
Notably, none of the evals linked to above have published scores for cyborgs.
The Cyborg Gap
Chess serves as a nice historical illustration of a cyborg gap. In 1997, the best AI and the best humans were close to parity in chess skill. In 2005, AI was better than the best humans. From 2005 to 2007, cyborgs (or "centaurs") were better than AI. From 2010 to 2017, it's disputed whether cyborgs had an advantage. From 2017 (AlphaZero) onwards, cyborgs had no advantage over AI.
These days, humans are no longer pure, un-AI-assisted humans when they do work in the real world. We work as cyborgs. This is especially true in the domain of software engineering. How much more capable are human engineers (cyborgs) than Claude Code at the present moment? Is that capability gap shrinking or increasing? When will that gap disappear? These are questions we can answer with cyborg evals.
I swear this is a thing I heard an actual software engineer say.
One approach to addressing this problem would be to find developers who are “untouched by AI”. This is already really hard, especially if you're looking for ~professional level software engineers, and will only get harder and harder. Also, an "untouched developer” is not very representative of what the average economically productive developer looks like.
Another difficulty with extending the METR benchmark is that the duration we need for tasks is much longer now than it was even a year ago. In METR terms we need 8hr-30hr tasks. Even putting aside worries about engineer contamination, it’s simply hard and expensive to find people to take these tasks. This is discussed in METR's Feb post.
Assuming the eval isn't saturated and incorporates every domain we care about. This second assumption is big, so we might want to remove it and instead just think about this in a domain-specific way. On the other hand, you could argue that full automation in the software domain quickly becomes full automation of everything, due to recursive self-improvement.
What it means for a task to be "AI only" is fuzzy. Somebody has to prompt the AI, build its harness, (in some cases) iterate on its harness, and so on. The engineer in the Anthropic project built a custom harness for it.