AI BENCHY Category
Domain specific Ranking
See which AI models perform best on Domain specific, which ones stay reliable, and where the biggest gaps appear. Sort by: Tests Correct ↓.
| Rank | Model | Company | Domain specific Score | Score | Tests Correct | Response Time (avg) |
|---|---|---|---|---|---|---|
| #10 | Qwen3.5-27B medium | Qwen | 5.3 | 8.4 | 1/3 | 79.5s |
| #11 | Gemini 3.1 Flash Lite Preview high | 5.3 | 8.4 | 1/3 | 127.6s | |
| #12 | Gemini 3 PRO Preview medium | 5.3 | 8.4 | 1/3 | 7.01s | |
| #15 | Gemini 2.5 Flash medium | 5.9 | 8.2 | 1/3 | 37.3s | |
| #16 | GPT-5.4 medium | OpenAI | 5.3 | 8.2 | 1/3 | 74.3s |
| #22 | Gemini 3.1 Flash Lite Preview low | 5.3 | 8.1 | 1/3 | 2.36s | |
| #23 | MiMo-V2-Pro medium | Xiaomi | 5.3 | 8.1 | 1/3 | 6.00s |
| #25 | Grok 4.20 Beta medium | X AI | 5.3 | 8.0 | 1/3 | 21.3s |
| #27 | DeepSeek V3.2 medium | DeepSeek | 5.3 | 8.0 | 1/3 | 39.3s |
| #28 | GPT-5.2 Chat none | OpenAI | 5.3 | 7.9 | 1/3 | 17.8s |
| #29 | Gemini 3.1 Flash Lite Preview none | 5.3 | 7.9 | 1/3 | 942ms | |
| #30 | Step 3.5 Flash medium | Stepfun | 5.3 | 7.9 | 1/3 | 170.5s |
| #31 | GLM 5V Turbo medium | Z.ai | 5.3 | 7.8 | 1/3 | 38.1s |
| #32 | Qwen3.5-Flash medium | Qwen | 5.3 | 7.8 | 1/3 | 146.5s |
| #33 | GLM 5.1 medium | Z.ai | 5.3 | 7.8 | 1/3 | 29.8s |