AI BENCHY
Your ad here

#117

Laguna M.1

Poolside Release: 2026-04-28 Tested on: 2026-04-28 22:25 poolside/laguna-m.1::none
(medium) (none)

Score

5.1

Consistency

8.7

Total Cost

$0.000

Total Output Tokens

2,870

Input Price

$0.000 / 1M

Output Price

$0.000 / 1M

Tests Correct

Wrong Tests: 14

Attempt pass rate: 33.3%

Flaky tests

3

Flaky tests had mixed outcomes across runs (at least one pass and one fail).

Response Time (avg)

2.79s

Response Time (max): 15.42s

Response Time (total): 50.24s

Charts

Choose the first model, then click a second model to open a side-by-side page.

Total Output Tokens

Score vs Total Output Tokens

Quick Compare

Category Breakdown

Category Score Consistency Tests Correct
Anti-AI Tricks 3.4 7.9
Coding 7.5 3.8
Combined 3.0 10.0
Data parsing and extraction 10.0 10.0
Domain specific 3.6 7.2
General Intelligence 4.0 10.0
Instructions following 6.3 10.0
Puzzle Solving 3.2 10.0
Tool Calling 10.0 10.0

Compared models