AI BENCHY Category
Trivia Ranking
See which AI models perform best on Trivia, which ones stay reliable, and where the biggest gaps appear. Sort by: Tests Correct ↓.
Failure Reasons
| Rank | Model | Company | Trivia Score | Score | Tests Correct | Response Time (avg) |
|---|---|---|---|---|---|---|
| #131 | Mercury 2 none | Inception | 3.0 | 4.7 | 0/1 | 548ms |
| #132 | Qwen3.5-9B none | Qwen | 3.0 | 4.7 | 0/1 | 2.32s |
| #133 | HY3 Preview none | Tencent | 3.0 | 4.6 | 0/1 | 2.71s |
| #134 | Nemotron 3 Nano Omni 30b A3b Reasoning none | NVIDIA | 0.0 | 4.6 | 0/0 | 0ms |
| #135 | GPT-5.4 Nano none | OpenAI | 3.0 | 4.5 | 0/1 | 773ms |
| #136 | GLM 4.7 Flash medium | Z.ai | 3.0 | 4.5 | 0/1 | 11.1s |
| #137 | MiMo-V2-Flash none | Xiaomi | 3.0 | 4.5 | 0/1 | 1.82s |
| #138 | Ling-2.6-1T none | Inclusionai | 0.0 | 4.5 | 0/0 | 0ms |
| #139 | Grok 4.1 Fast none | X AI | 3.0 | 4.4 | 0/1 | 731ms |
| #140 | Qwen3.5-9B medium | Qwen | 3.0 | 4.3 | 0/1 | 177.0s |
| #142 | Granite 4.1 8B none | IBM Granite | 3.0 | 4.1 | 0/1 | 306ms |