Skip to main content

We Just Ran Our 2nd Benchmark Cycle. Here's What 21K API Calls and 12M Tokens Actually Looks Like.

Published: March 8, 20266 min read

We Just Ran Our 2nd Benchmark Cycle. Here's What 21K API Calls and 12M Tokens Actually Looks Like.

Most benchmark sites publish once and call it done. A static snapshot that's outdated the week after it goes live.

I'm doing something different. AI Search Arena runs recurring benchmark cycles. Today we completed our second one. The multi-cycle comparison features I built? They actually work now. With real data.

But here's what nobody tells you about running these things.


The Ugly Part

Running a benchmark cycle isn't clicking a button. It's this:

32 tools get enrolled. Each maps to one of 7 market segments. Then every tool gets evaluated across 51 capability dimensions by 6 independent AI models.

GPT-5.2, Claude Sonnet 4.6, Gemini 3 Flash, Grok 4.1 Fast, DeepSeek V3.2, Mistral Large 3.

32 × 51 × 6 = 9,792 individual API calls. Across 6 models, that's 21,000+ total requests. About 12.1 million tokens. Roughly 10 hours of continuous processing.

Why 6 models? Same reason peer review uses multiple reviewers. No single model's bias dominates. We take the median score. You'd need to compromise 4 out of 6 models to game a ranking.

Then raw scores become composite scores. Confidence tags come from inter-model agreement. Dense ranking applied. Badges awarded. The whole dataset gets sealed with a SHA-256 hash so anyone can verify we didn't change the numbers after the fact.


What Broke

Theory is clean. Production is messy.

During this cycle, our synthesis stage hit a connection pool exhaustion issue. After 9,792 evaluations completed successfully over 10 hours, the process just hung. Zero CPU. No error messages. Silence.

Root cause: when resuming a partially-completed cycle, the pipeline was re-upserting 1,632 synthesis records unnecessarily. Our serverless PostgreSQL just gave up.

The fix was making synthesis idempotent. Check if records exist before re-processing. Resume time went from "hang forever" to 44.7 seconds.

This is the kind of thing that happens when you actually build data pipelines instead of writing about them.


What Two Cycles Gets You

Now it gets interesting:

Trend tracking. Is a tool improving or declining? One cycle can't tell you. Two can. Weekly cycles give you statistically significant trend lines within a month.

Movement detection. Who moved up or down? Did BrightEdge stay #1? Did anyone new break into the top 5?

Confidence building. A tool ranks #3 once = might be noise. #3 across four cycles = pattern.

Segment stability. Do the "Best for Enterprise" picks actually hold? Multi-cycle data answers that.


Why Weekly

We're targeting weekly cycles because:

AI SEO tools ship fast. A tool that was mediocre last month might have shipped a major feature this week. Monthly misses that. Weekly catches it.

Market context changes. Google updates. New tools launch. Tools pivot. A benchmark from last quarter doesn't help today's buying decision.

If you're evaluating tools right now, you want to know how they perform right now. Not how they performed when some blogger reviewed them 6 months ago.

The goal: AI Search Arena always shows the current state of the market. Not a historical artifact. Not a sponsored ranking. The actual, current, independently-verified performance of every major AI SEO tool.


This Cycle's Results (April 2026)

Quick snapshot:

Rank Tool Score
1 BrightEdge 7.6
2 Semrush 7.4
3 seoClarity 7.3
3 Conductor 7.3

10 badges awarded. Full results on the leaderboard.


What's Next

The admin operator console is now live. Cycle management is no longer a manual database operation. State transitions, tool enrollment, model configuration — all handled through a proper admin interface.

Next cycles will be faster to set up. The pipeline is battle-tested. The resume logic works.

The vision: you check AI Search Arena before making any AI SEO tool decision, the same way you'd check Consumer Reports before buying an appliance. Except our data updates weekly, not annually.

Current rankings: aisearcharena.com


AI Search Arena is independent. No vendor pays to be listed or ranked. Full methodology and audit packages are publicly available.

Share this post