April 2026 AI Model Rankings: Anthropic Tops LMArena, GPT-5.5 Rules AA Index

April 2026 AI Model Rankings: Anthropic Tops LMArena, GPT-5.5 Rules AA Index

Bottom Line

As of late April 2026, global AI model leaderboards show a “two-leaderboard, two-champions” pattern: Anthropic dominates LMArena (formerly Chatbot Arena) Elo rankings, while OpenAI’s GPT-5.5 series leads the Artificial Analysis Intelligence Index. Each reflects a different dimension of capability.

LMArena Elo: User Preference Rankings

Based on anonymous A/B voting, data as of April 24:

RankModelEloVendor
1Opus 4.7 (thinking)1503Anthropic
2Claude Opus 4.6 (thinking)1503Anthropic
3Claude Opus 4.61496Anthropic
4Opus 4.71494Anthropic
5Gemini 3.1 Pro Preview1493Google DeepMind
6Muse Spark1492Meta AI
7Gemini 3.0 Pro1486Google DeepMind
8grok-4.20-beta11482xAI
9gpt-5.4-high1481OpenAI
10grok-4.20-beta-reasoning1479xAI

Key signals: Anthropic holds 3 of top 4. Meta Muse Spark enters top 10 for the first time since early 2025.

AA Intelligence Index: Standardized Benchmark Rankings

Aggregating 10 standardized benchmarks (coding, math, science, reasoning, agents), data as of April 25:

RankModelScoreVendor
1GPT-5.5 (xhigh)60OpenAI
2GPT-5.5 (high)59OpenAI
3Opus 4.7 (max)57Anthropic
4Gemini 3.1 Pro Preview57Google DeepMind
5GPT-5.4 (xhigh)57OpenAI
6GPT-5.5 (medium)57OpenAI
7Kimi K2.654Moonshot AI
8MiMo-V2.5-Pro54Xiaomi
9GPT-5.3 Codex (xhigh)54OpenAI
10Muse Spark52Meta AI

Key signals: GPT-5.5 sweeps top 2, holds 4 of top 6. Kimi K2.6 (Moonshot AI) is the only Chinese model in top 10.

Selection Guide

  • Strongest overall benchmarks: GPT-5.5 (xhigh), AA score 60
  • Best user experience: Claude Opus 4.7 (thinking), LMArena 1503 Elo
  • Cost-effective: GPT-5.5 (medium), AA 57 at lower price
  • Chinese models: Kimi K2.6 at 54, highest-ranked domestic model
  • Open-source / semi-open: Muse Spark (Meta) at 52

Sources