Home Models Coding Agents Compare Pricing Model Picker Source Data Local Models OpenClaw
← Back to all evals
What Is Chatbot Arena (LMSYS)? ELO Ratings Explained

What Is Chatbot Arena (LMSYS)? ELO Ratings Explained


Chatbot Arena is the largest crowdsourced AI model evaluation. Real humans chat with two anonymous models side-by-side, vote for the better one, and the results produce ELO ratings similar to chess rankings.

Unlike static benchmarks (SWE-Bench, GPQA), Chatbot Arena captures real user preferences across millions of conversations. Here’s how it works and what the ratings mean.

How Chatbot Arena Works

The Setup

  1. User enters a prompt — any question or task
  2. Two anonymous models respond — the user doesn’t know which models
  3. User votes — Model A wins, Model B wins, or Tie
  4. ELO updates — ratings adjust based on the outcome

This is “blind evaluation” — users can’t be biased by model names or reputations.

ELO Rating System

Chatbot Arena uses the Elo rating system (same as chess):

  • Higher ELO = stronger model
  • ELO difference predicts win probability
  • Beating a strong model gains more points

Example:

  • If Model A (1400 ELO) beats Model B (1300 ELO), A gains ~8 points, B loses ~8
  • If Model A (1400) beats Model C (1000), A gains ~2 points, C loses ~2

This makes the system self-correcting: models find their true level over time.

Key Properties

  • Scale: 4+ million votes collected
  • Coverage: 100+ models ranked
  • Categories: Overall, coding, hard prompts, language-specific
  • Updates: Continuous (new votes flow in daily)

Current Leaderboard (Feb 2026)

Overall ELO Rankings

RankModelELO95% CI
1Claude Opus 4.61417±4
2GPT-5.21408±3
3Gemini 2.5 Pro1392±5
4Claude Sonnet 4.51378±4
5GPT-4o1356±3
6DeepSeek V31341±6
7Gemini 2.5 Flash1329±4
8Llama 4 70B1315±5

Top of the board is tight: Claude Opus 4.6 leads by only 9 points over GPT-5.2. That’s within the margin of error — they’re effectively tied for everyday use.

Coding ELO Rankings

RankModelCoding ELO
1Claude Opus 4.61438
2GPT-5.21425
3Claude Sonnet 4.51401
4DeepSeek V31389
5Gemini 2.5 Pro1374

Claude’s coding lead is larger here (13 points) — consistent with SWE-Bench results.

Hard Prompts ELO

RankModelHard ELO
1Claude Opus 4.61392
2GPT-5.21378
3Gemini 2.5 Pro1361
4Claude Sonnet 4.51349

“Hard prompts” = multi-step reasoning, math, complex instruction following. Claude extends its lead here.

What ELO Scores Mean

ELO RangeTierInterpretation
1400+EliteTop-tier frontier models
1350-1400StrongCompetitive with leaders, slight gaps
1300-1350CapableGood for most tasks, struggles on hard problems
1250-1300SolidReliable for routine work
<1250DevelopingUseful for specific niches

ELO differences:

  • 50+ points: Clear, noticeable quality gap
  • 20-50 points: Consistent but small differences
  • <20 points: Within noise — effectively tied

Strengths of Chatbot Arena

Real Human Preferences

Unlike automated benchmarks, Arena captures what humans actually prefer:

  • Natural conversation flow
  • Nuanced instruction following
  • Helpfulness and tone
  • Error recovery

This makes it a better proxy for “how good does this feel to use?”

Broad Coverage

  • Diverse prompts: Users bring their own use cases
  • Adversarial testing: Users try to break models
  • Edge cases: Prompts that benchmarks don’t cover

Impossible to Game

Models can’t memorize Arena prompts — they’re generated fresh by users. This avoids the “data contamination” problem that plagues static benchmarks.

Limitations

User Bias

Arena users aren’t representative:

  • Skewed toward tech-savvy, English-speaking users
  • Over-index on certain prompt types (coding, trivia)
  • May prefer confident-sounding wrong answers over hedged correct ones

Prompt Distribution

The average Arena prompt isn’t your prompt. If you’re building:

  • Scientific tools: GPQA Diamond matters more
  • Coding assistants: SWE-Bench matters more
  • Customer support: Arena is highly relevant

Model Positioning

Models can be “optimized for Arena” — tuning tone, verbosity, and hedging to appeal to voters. This doesn’t always translate to production quality.

Statistical Noise

ELO ratings have confidence intervals. A 10-point difference might be real, or might be noise. Always check the CI.

How to Use Arena Data

For Model Selection

  1. Check overall ELO for general-purpose quality
  2. Check category ELO (coding, hard prompts) for your use case
  3. Look at CI — narrow intervals mean stable ratings
  4. Compare to benchmarks — consistency across Arena and static tests is a good sign

For Understanding AI Progress

Arena ELOs have risen ~200 points since 2023:

  • Early 2023: GPT-4 led at ~1200
  • Early 2024: GPT-4 Turbo at ~1280
  • Early 2025: Claude 3.5 Sonnet at ~1340
  • Early 2026: Claude Opus 4.6 at ~1417

This is a rough measure of AI capability improvement over time.

For Competitive Analysis

If you’re a model developer, Arena is a quick feedback loop:

  • Release a model
  • Watch ELO stabilize over 2-4 weeks
  • Compare to targets
  • Iterate

Arena vs. Static Benchmarks

AspectChatbot ArenaStatic Benchmarks
EvaluationHuman votesAutomated tests
CoverageBroad, unpredictableTargeted, known
GamingHardEasy (contamination)
SpeedSlow (weeks to stabilize)Fast (hours)
ObjectivitySubjective preferencesObjective scores
CostFree (crowdsourced)Expensive to run

Best practice: Use both. Arena for real-world feel, benchmarks for specific capabilities.

The Bottom Line

Chatbot Arena is the best measure of human preference for AI models. If you want to know which model feels best to use for real conversations, Arena ELO is the signal.

But it’s not the only signal. For engineering tasks, check SWE-Bench. For scientific reasoning, check GPQA Diamond. For your specific use case, run your own evals.

The current leader — Claude Opus 4.6 at 1417 ELO — is only 9 points ahead of GPT-5.2. That’s within the noise. The practical answer: both are excellent, choose based on cost, speed, and ecosystem.


Related: