AI BENCHY
Your ad here

AI BENCHY Category Failures

Domain specific: Wrong answer

Domain specific
Wrong answer

See which AI models are most likely to hit Wrong answer on Domain specific, so you can spot weak points faster. Sort by: Response Time (avg) ↑.

Models Shown

15

Total Failures

182

Most Affected Model

GLM 5 2
Rank Model Company Wrong answer Count Category Score Tests Correct Response Time (avg)
#56 Grok 4.20 Multi Agent Beta medium X AI 2 2.9 0/3 24.7s
#47 Grok 4.20 medium X AI 1 5.3 1/3 27.0s
#20 Qwen3.6 Plus medium Qwen 3 2.9 0/3 29.6s
#33 GLM 5.1 medium Z.ai 1 5.3 1/3 29.8s
#2 Gemini 3.1 Pro Preview medium Google 1 7.7 2/3 32.7s
#84 gpt-oss-120b none OpenAI 3 3.0 0/3 35.0s
#15 Gemini 2.5 Flash medium Google 2 5.9 1/3 37.3s
#31 GLM 5V Turbo medium Z.ai 2 5.3 1/3 38.1s
#38 GPT-5.4 Nano medium OpenAI 2 5.9 1/3 38.2s
#14 Gemma 4 31B medium Google 1 7.7 2/3 38.5s
#27 DeepSeek V3.2 medium DeepSeek 1 5.3 1/3 39.3s
#45 GPT-5 Mini medium OpenAI 2 3.6 0/3 44.6s
#68 gpt-oss-120b medium OpenAI 3 2.9 0/3 50.9s
#35 MiMo-V2-Omni medium Xiaomi 1 3.0 0/3 55.1s
#19 Qwen3.5-122B-A10B medium Qwen 3 2.9 0/3 63.4s

Top Models by Wrong answer Count

Wrong answer Count vs Score

Top Models by Response Time (avg)

Top Models by Estimated Wasted Cost