Skip to content
Longterm Wiki
Back

SWE-bench Pro Leaderboard - Scale AI

web

Useful for tracking the state of AI coding agent capabilities; relevant to discussions of AI autonomy, capability evaluations, and the pace of progress toward AI systems that can perform complex software engineering tasks independently.

Metadata

Importance: 45/100tool pagereference

Summary

SWE-bench Pro is a rigorous benchmark by Scale AI that evaluates AI agents on real-world software engineering tasks drawn from both public and private repositories. It addresses limitations of existing benchmarks by emphasizing realistic, challenging problem-solving scenarios. The leaderboard tracks and compares performance of leading AI coding agents.

Key Points

  • Evaluates AI agents on software engineering tasks sourced from public and private repositories, reducing benchmark contamination risks
  • Designed to address limitations of prior coding benchmarks by focusing on realistic, difficult problem-solving scenarios
  • Provides a public leaderboard ranking AI agent performance on software engineering tasks
  • Relevant for assessing current AI coding capabilities and tracking progress toward autonomous software development
  • Produced by Scale AI, a major player in AI data labeling and evaluation infrastructure

Review

SWE-Bench Pro represents a significant advancement in AI agent evaluation for software engineering tasks. By addressing critical limitations in existing benchmarks, such as data contamination, limited task diversity, and oversimplified problems, the benchmark offers a more authentic assessment of AI problem-solving capabilities. The methodology involves a sophisticated four-stage workflow that carefully sources, creates, and augments software engineering challenges from diverse repositories. The benchmark's key innovation lies in its rigorous design, which includes three distinct dataset subsets: a public set, a commercial set, and a held-out set. This approach allows for comprehensive testing across different coding environments and provides a more nuanced understanding of AI agents' generalization abilities. The results are striking, with top models like OpenAI GPT-5 and Claude Opus 4.1 scoring only around 23% on the public dataset, compared to 70%+ on previous benchmarks. This dramatic performance drop highlights the benchmark's increased complexity and its potential to drive meaningful improvements in AI software engineering capabilities.

Cited by 3 pages

PageTypeQuality
Autonomous CodingCapability63.0
Long-Horizon Autonomous TasksCapability65.0
Tool Use and Computer UseCapability67.0
Resource ID: 9dbe484d48b6787a | Stable ID: Yzc2OTk5OT