The Inference Report

May 2, 2026

Claude Opus 4.6 holds the top position on SWE-rebench at 65.3%, unchanged from the previous ranking, while the field beneath it shows modest consolidation rather than dramatic reshuffling. The most notable movement comes from Chinese models: GLM-5 jumped from rank 17 to rank 3 (49.8 to 62.8 percent), GLM-5.1 climbed from rank 14 to rank 6 (51.4 to 62.7 percent), and GLM-4.7 advanced from rank 44 to rank 14 (42.1 to 58.7 percent), suggesting systematic improvements in that family's code-solving capability. Kimi K2.5 rose from rank 29 to rank 16 (46.8 to 58.5 percent), and Kimi K2 Thinking moved from rank 54 to rank 21 (40.9 to 57.4 percent), indicating progress across Kimi's lineup as well. Claude Sonnet 4.6 improved from rank 12 to rank 9 (51.7 to 60.7 percent), while Gemini 3.1 Pro Preview declined from rank 3 to rank 7 (57.2 to 62.3 percent), suggesting variable iteration quality. The Artificial Analysis benchmark tells a different story: GPT-5.5 leads at 60.2, with Claude Opus 4.6 at rank 9 (53 points) and Claude Opus 4.7 at rank 2 (57.3 points), revealing substantial disagreement between the two evaluations on which models excel at real-world software engineering tasks. The divergence between these benchmarks, SWE-rebench emphasizing repository-level problem solving and Artificial Analysis potentially capturing different task distributions or evaluation criteria, warrants scrutiny of their respective methodologies before treating either ranking as definitive for model capability assessment.

Cole Brennan

Daily rankings from SWE-rebench, a benchmark designed to fairly compare LLM capabilities on real-world software engineering tasks. Unlike other evaluations, it uses a standardized scaffolding for all models, continuously updates its dataset to prevent contamination, and runs each model five times to account for stochastic variance.

#ModelScore
1Claude Opus 4.665.3%
2gpt-5.2-2025-12-11-medium64.4%
3GLM-562.8%
4Junie62.8%
5gpt-5.4-2026-03-05-medium62.8%
6GLM-5.162.7%
7Gemini 3.1 Pro Preview62.3%
8DeepSeek-V3.260.9%
9Claude Sonnet 4.660.7%
10Claude Sonnet 4.560.0%

Artificial Analysis composite index across coding, math, and reasoning benchmarks.

#ModelScoretok/s$/1M
1GPT-5.560.273$11.25
2Claude Opus 4.757.351$10.00
3Gemini 3.1 Pro Preview57.2132$4.50
4GPT-5.456.886$5.63
5Kimi K2.653.925$1.71
6MiMo-V2.5-Pro53.863$1.50
7GPT-5.3 Codex53.681$4.81
8Grok 4.353.2205$1.56
9Claude Opus 4.65349$10.00
10Muse Spark52.10$0.00

Output tokens per second — higher is faster. Minimum intelligence score of 40.

#Modeltok/s
1Grok 4.3205
2Qwen3.6 35B A3B187
3Gemini 3 Flash Preview184
4GPT-5 Codex178
5GPT-5.4 mini174
6GPT-5.1 Codex174
7GPT-5.4 nano157
8Qwen3.5 122B A10B153
9Gemini 3.1 Pro Preview132
10MiMo-V2-Flash131

Blended cost per 1M tokens (3:1 input/output) — lower is cheaper. Minimum intelligence score of 40.

#Model$/1M
1MiMo-V2-Flash$0.15
2DeepSeek V4 Flash$0.175
3DeepSeek V3.2$0.315
4GPT-5.4 nano$0.463
5MiniMax-M2.7$0.525
6KAT Coder Pro V2$0.525
7MiniMax-M2.5$0.525
8Qwen3.6 35B A3B$0.557
9GPT-5 mini$0.688
10Qwen3.5 27B$0.825