Model Rankings
Community-driven, updated daily
| Rank | Model | Provider | Overall | Code | Reasoning | Speed |
|---|---|---|---|---|---|---|
| Claude Opus 4 | Anthropic | 96.2 | 95.8 | 97.1 | 82.4 | |
| GPT-5.4 | OpenAI | 95.8 | 94.2 | 96.5 | 84.1 | |
| Gemini 2.5 Pro | 94.5 | 92.8 | 95.2 | 86.3 | ||
4 | o3 | OpenAI | 93.9 | 91.5 | 97.8 | 68.2 |
5 | Claude Sonnet 4 | Anthropic | 93.1 | 93.6 | 92.4 | 91.7 |
6 | GPT-5.2 Codex | OpenAI | 92.4 | 97.2 | 88.1 | 83.6 |
7 | Grok 4 | xAI | 91.8 | 89.4 | 93.2 | 85.5 |
8 | DeepSeek R1 | DeepSeek | 91.2 | 90.1 | 94.6 | 78.3 |
9 | Mistral Large | Mistral | 90.5 | 89.8 | 90.2 | 87.1 |
10 | Gemini Ultra | 90.1 | 88.3 | 93.8 | 71.2 |