Skip to content

Benchmarks Quickly Saturate as AI Advances

by Benjamin Mann on July 20, 2025

The rapid saturation of AI benchmarks masks the true pace of AI progress, requiring more ambitious testing to reveal actual capability improvements.

The Benchmark Saturation Problem

  • AI benchmarks quickly reach performance ceilings, creating a false impression of slowing progress
    • "When you release a new benchmark within like six to twelve months it immediately gets saturated"
    • This creates a recurring narrative that AI progress is plateauing, which "comes out like every six months or so and it's never been true"

Why Progress Appears to Slow

  • Task-specific saturation occurs when models reach human-level performance on specific tasks
    • "For some tasks we are saturating the amount of intelligence needed for that task"
    • Simple information extraction tasks reach 100% accuracy, creating no room to demonstrate improvement
  • Time compression distorts perception of progress
    • Model releases have accelerated from "once a year" to "every month or three months"
    • Comparing to recent releases rather than year-over-year masks the magnitude of advancement
    • "There's this weird time compression effect... like being in a near light speed journey where a day that passes for you is like five days back on earth"

Evidence That Progress Continues

  • Scaling laws continue to hold across orders of magnitude
    • "This is one of the few phenomena in the world that has held across so many orders of magnitude"
    • "If you look at fundamental laws of physics many of them don't hold across 15 orders of magnitude"
  • Post-training techniques are accelerating progress
    • Reinforcement learning has enabled continued scaling beyond traditional pretraining
    • "Progress has actually been accelerating in many ways"

The Benchmark Challenge

  • The real constraint is creating better benchmarks that reveal intelligence improvements
    • "Maybe the real constraint is like how can we come up with better benchmarks"
    • Need "better ambition of using the tools that then reveals the bumps in intelligence"
  • Effective benchmarks should:
    • Test capabilities beyond current saturation points
    • Measure progress on more complex, multi-step reasoning
    • Evaluate performance on tasks with higher ceilings

Implications for AI Development

  • Benchmark design becomes increasingly important for measuring true progress
  • Need to focus on tasks with higher intelligence ceilings to demonstrate capability jumps
  • The appearance of plateaus may lead to complacency while progress continues underneath