Sonnet Code
← Volver a todos los artículos
AI & Machine Learning4 de mayo de 2026·8 min read

Why Your Agent Looks Great on SWE-bench and Wobbles in Production: The 37% Lab-to-Prod Gap

The number, in one paragraph

A new analysis of 12 major agentic AI benchmarks — SWE-bench, WebArena, AgentBench, GAIA, ToolLLM, and seven others — found validity issues affecting 7 out of 10 of the widely cited ones, cost-misestimation rates up to 100%, and a 37% gap between lab benchmark scores and production deployment performance for enterprise agentic systems. The same paper documented up to 50x cost variation between deployments achieving similar accuracy. Separately, the SWE-bench Verified leaderboard now shows that the same model moves 5–15 points depending on the agent scaffolding, and OpenAI has stopped reporting Verified scores citing confirmed contamination — third parties continue running the eval independently.

The headline you'll see in trade press: benchmarks don't predict production. The implication for product teams is sharper: the eval suite you trust today probably doesn't tell you what you think it does, and the gap between "good demo" and "good in production" is no longer ambiguous — it has a number on it.

What the 37% gap is actually measuring

The gap is not "lab scores are wrong." It is "lab scores do not measure the things production cares about." The Beyond Accuracy framework that produced the number explicitly enumerates what is missing from current agentic benchmarks:

  • Security against prompt injection. SWE-bench does not adversarially probe whether the agent can be tricked into modifying files outside the requested scope. Production does.
  • Compliance with organizational policies. WebArena does not test whether the agent will refuse to take an action that violates an internal access-control policy. Production does, every day.
  • Latency within SLA constraints. Most benchmarks score success regardless of how long the run took. Production users abandon at 30 seconds.
  • Graceful error handling. Benchmarks measure pass/fail. Production cares about how the agent fails — does it blow up, ask the user for help, fall back to a safer mode, or silently produce wrong output?
  • Cost per successful task. The same benchmark score can be hit at $0.20/task or $10/task depending on the scaffold. None of the public leaderboards tell you which.

When a benchmark score gets reported in the trade press, none of these dimensions are in the number. That's not a flaw in the benchmark; it's a flaw in using a single number as a procurement signal.

Why benchmarks are getting worse, not better, as a procurement input

Three forces are degrading the predictive power of public benchmarks at exactly the moment buyers most want to lean on them:

1. Contamination. The most-cited public benchmarks have been in pretraining corpuses for years. OpenAI's decision to stop reporting SWE-bench Verified is downstream of confirmed leakage. Once the model has seen the test, the score is no longer measuring generalization — it is measuring memorization. This applies to every widely-shared benchmark, not just SWE-bench.

2. Scaffold variance. A 5–15 point swing on the same model from scaffold differences means a benchmark score is a property of an entire stack, not of the model. Two vendors quoting the same benchmark number against the same model can be running materially different agents.

3. Optimization pressure. Once a benchmark becomes the leaderboard everyone optimizes for, it stops measuring the underlying capability and starts measuring capacity to fit the leaderboard. This is Goodhart's Law in software-engineering form, and it has happened to every major coding benchmark on a roughly 18-month cycle.

The combined effect: a leaderboard score in 2026 has roughly the same predictive power for production behavior that an MMLU score had in 2023. Useful as one signal among many, useless as a standalone decision input.

What "production-grade" eval actually looks like

The teams shipping agentic systems that don't wobble in production share a small number of practices, and none of them are about adding more public benchmarks:

  • A regression suite of real failures. Pulled from the team's own ticket queue, support archive, or internal bug tracker. Each example is a real workflow the agent has to handle, with a hand-graded expected outcome.
  • A multi-dimensional rubric. Accuracy is one column. Cost, latency, refusal-correctness, scope adherence, and rollback safety are the others. A run that nails accuracy at 12x the cost is not a passing run.
  • An adversarial layer. Prompt-injection probes, scope-creep probes, jailbreak probes — all included in CI. The goal is not to prove the agent is unbreakable; it is to know which prompts break it before a user finds them.
  • Production-shadow evaluation. A subset of real production traffic mirrored to a staging agent on every release. The diff between staging-agent and production-agent outputs is the most honest leaderboard there is.
  • Domain experts in the grading loop. For workloads where "correct" is contested — clinical reasoning, legal analysis, regulatory fit — the rubric and the gold answers have to be written by people who know more than the model. This is the human-in-the-loop layer that the public-benchmark leaderboards conspicuously don't cover.

This list reads as obvious in hindsight. The reason most teams don't have it is that none of these items are pre-built. You can't buy a vendor's regression suite of real failures from your own ticket queue. You have to write it.

Where we'd push back on the doom narrative

The 37% gap framing has a "benchmarks are useless" temptation that is wrong on its face. Public benchmarks are an excellent floor check — if your candidate model can't get within reach of the public leaderboard on a task family relevant to your work, it definitely won't survive your domain. SWE-bench Verified at 87.6% is meaningful, even contaminated, even noisy. It tells you the model has the underlying capability shape; it does not tell you it will hold up against your codebase.

The right framing: public benchmarks rule things out, internal evals rule things in. The teams that conflate the two are the ones that will pay the 37% gap as a production incident.

What we would do with this today

  • Audit the evals you're trusting. If your model-selection decision was driven by leaderboard scores from public benchmarks alone, write down the gap between those benchmarks and your actual workflow. Most teams find the gap is bigger than they assumed.
  • Build a small regression suite from your own ticket history. 30 examples is enough to start. Each one is a real workflow with a hand-graded gold answer. This becomes the most-used asset on the team within a quarter.
  • Add the missing dimensions to your CI. Cost per successful task, p95 latency, scope adherence, refusal correctness. The instrumentation is small and the visibility is large.
  • Hire or staff the domain expertise. For regulated workloads, the eval suite is only as good as the people writing the rubric. This is the part most engineering teams under-staff and the part that most determines whether the agent ships.

Sonnet Code's take

The 37% gap is the gap two halves of our practice exist to close, and we'll say so plainly: AI development work closes it by building the harness, instrumentation, and routing logic that makes the agent safe to ship; AI training work closes it by staffing the senior domain experts who write the regression suites, rubrics, and red-team prompts that public benchmarks don't cover. Public benchmarks tell you whether to take a meeting with a model. Production-grade evals tell you whether to ship it. The teams that win the next eighteen months will be the ones that stopped confusing the two — and the ones that staffed the eval work with people senior enough to be right about it. If your roadmap has agents on it and you're not yet sure your eval suite would catch a real production regression, that's the conversation we run most often this quarter.