Key checks
- GPT-5.5 SWE-Bench Pro Performance: In its April 23, 2026, release announcement, OpenAI stated that GPT-5.5 reaches 58.6% on SWE-Bench Pro, which evaluates the resolution of real-world GitHub issues.
- Claude Opus 4.7 SWE-Bench Pro Performance: Anthropic's Claude Opus 4.7, released on April 16, 2026, scores 64.3% on SWE-Bench Pro, which is 5.7 percentage points higher than GPT-5.5.
- OpenAI's Use of Internal Benchmarks: OpenAI's official announcement table prominently features 'Expert-SWE (Internal)', where GPT-5.5 scores 73.1%. It does not list the SWE-Bench Pro score in that specific comparison table, though it is mentioned later in the text.