AI BENCHY
Compare Charts Methodology
❤️ Made by XCS
Your ad here

AI BENCHY Failures

Wrong answer Failures

See which AI models run into Wrong answer most often, so you can spot reliability risks before choosing one. Sort by: Avg Score ↑.

Models Shown

54

Total Failures

275

Most Affected Model

LFM2-24B-A2B 9
Rank Model Company Wrong answer Count Avg Score Tests Correct Response Time (avg)
#55 LFM2-24B-A2B none Liquid 9 2.6 1/16 811ms
#53 Grok 4.1 Fast none X AI 11 2.9 3/16 1.90s
#54 MiMo-V2-Flash none Xiaomi 10 2.9 3/16 2.97s
#52 GLM 4.7 Flash medium Z.ai 7 3.1 4/16 36.8s
#51 Mercury 2 none Inception 11 3.4 4/16 596ms
#50 Qwen3 Coder Next medium Qwen 8 3.5 3/16 12.5s
#49 GLM 4.7 Flash none Z.ai 9 3.9 4/16 2.99s
#47 GPT-4o-mini none OpenAI 11 4.0 4/16 2.07s
#48 Qwen3 Coder Next none Qwen 10 4.0 4/16 11.7s
#46 Kimi K2.5 none Moonshot AI 11 4.1 5/16 11.9s
#45 Trinity Large Preview none Arcee AI 9 4.2 5/16 3.15s
#44 GPT-5.4 none OpenAI 9 4.5 6/16 1.48s
#42 Qwen3.5-35B-A3B none Qwen 8 4.7 6/16 4.10s
#43 MiniMax M2.5 medium Minimax 5 4.7 5/16 43.0s
#41 Qwen3.5-27B none Qwen 9 4.9 5/16 1.75s
#40 Qwen3.5-122B-A10B none Qwen 9 5.0 6/16 3.72s
#39 gpt-oss-120b medium OpenAI 5 5.1 7/16 16.7s
#37 Qwen3.5-Flash none Qwen 8 5.2 7/16 3.54s
#38 Gemini 2.5 Flash none Google 9 5.2 6/16 923ms
#36 Mercury 2 medium Inception 5 5.3 7/16 2.36s
#33 DeepSeek V3.2 none DeepSeek 6 5.5 7/16 12.9s
#34 GPT-5 Nano medium OpenAI 5 5.5 7/16 47.9s
#35 Qwen3.5-35B-A3B medium Qwen 2 5.5 8/16 43.9s
#31 GLM 5 none Z.ai 7 6.0 9/16 4.03s
#32 GPT-5 Mini medium OpenAI 3 6.0 8/16 25.1s
#29 Qwen3.5 Plus 2026-02-15 none Qwen 7 6.2 9/16 2.65s
#30 Grok 4.1 Fast medium X AI 2 6.2 9/16 26.3s
#28 Kimi K2.5 medium Moonshot AI 3 6.4 9/16 69.8s
#27 GPT-5.2 medium OpenAI 1 6.5 10/16 15.3s
#26 Claude Opus 4.6 medium Anthropic 2 6.6 10/16 22.9s
#25 Claude Sonnet 4.6 none Anthropic 2 6.8 10/16 5.57s
#23 Seed-2.0-Mini medium Bytedance Seed 1 6.9 10/16 65.1s
#24 Qwen3.5-Flash medium Qwen 1 6.9 10/16 70.8s
#22 Gemini 3.1 Flash Lite Preview none Google 4 7.1 10/16 1.33s
#20 Gemini 3 Flash Preview none Google 5 7.2 11/16 1.75s
#21 MiMo-V2-Flash medium Xiaomi 3 7.2 11/16 25.3s
#17 Gemini 3.1 Flash Lite Preview low Google 4 7.3 11/16 3.36s
#18 DeepSeek V3.2 medium DeepSeek 3 7.3 11/16 39.5s
#19 GPT-5.3 Chat none OpenAI 4 7.3 10/16 5.96s
#13 Step 3.5 Flash medium Stepfun 3 7.4 10/16 29.1s
#14 GLM 5 medium Z.ai 2 7.4 11/16 16.2s
#15 GPT-5.2 Chat none OpenAI 4 7.4 11/16 7.03s
#16 Gemini 2.5 Flash medium Google 4 7.4 11/16 12.4s
#12 Gemini 3.1 Flash Lite Preview medium Google 4 7.5 11/16 3.83s
#10 Qwen3.5-122B-A10B medium Qwen 3 7.7 12/16 29.7s
#11 Claude Sonnet 4.6 medium Anthropic 1 7.7 12/16 11.2s
#9 GPT-5.4 medium OpenAI 2 8.0 12/16 20.1s
#5 Gemini 3 Flash Preview low Google 3 8.2 13/16 6.11s
#6 Gemini 3 Pro Preview medium Google 3 8.2 13/16 7.15s
#7 Qwen3.5-27B medium Qwen 1 8.2 12/16 52.1s
#8 Gemini 3.1 Flash Lite Preview high Google 3 8.2 12/16 68.8s
#4 Qwen3.5 Plus 2026-02-15 medium Qwen 1 8.3 13/16 34.5s
#3 GPT-5.3-Codex medium OpenAI 2 8.4 12/16 16.6s
#2 Gemini 3.1 Pro Preview medium Google 1 9.4 15/16 16.6s

Top Models by Wrong answer Count

Wrong answer Count vs Avg Score

Top Models by Response Time (avg)