Welcome to the April edition of our monthly LLM rankings. Each month, we pull real-world usage data from OpenRouter — a platform routing millions of API calls across every major LLM — and rank the models developers are actually paying for.
Not benchmarks. Not press releases. Real tokens, real money, real usage.
The Top 10 — April 2026
| Rank | Model | Provider | Move | Input / 1K tokens | Output / 1K tokens |
|---|---|---|---|---|---|
| 1 | Claude Sonnet 4.6 | Anthropic | ● | $0.003 | $0.015 |
| 2 | GPT-5.4 | OpenAI | ↑ | $0.002 | $0.01 |
| 3 | Gemini 3 Flash | ↑ | $0.0001 | $0.0004 | |
| 4 | DeepSeek V3 | DeepSeek | ↑ | $0.0003 | $0.0009 |
| 5 | Claude 3.5 Sonnet | Anthropic | ↓ | $0.003 | $0.015 |
| 6 | GPT-5.4 Mini | OpenAI | ↑ | $0.0003 | $0.0012 |
| 7 | Claude Opus 4.6 | Anthropic | ↑ | $0.015 | $0.075 |
| 8 | Gemini 3.1 Pro | ● | $0.00125 | $0.005 | |
| 9 | Qwen3.6 Plus | Alibaba | 🆕 | $0.0002 | $0.0008 |
| 10 | Gemma 4 31B | 🆕 | $0.0001 | $0.0003 |
New This Month
Qwen3.6 Plus (Mar 18) — Alibaba's latest general-purpose model brings serious multilingual capabilities at a price point that makes DeepSeek sweat. Strong adoption in Asia-Pacific markets, and surprisingly good English performance. The thinking mode is optional, which helps keep costs predictable.
Gemma 4 31B (Mar 22) — Google's open-weight contender re-enters the conversation. At 31B parameters, it's efficient enough to self-host on a single A100, and the quality rivals models 3x its size. Open-weight enthusiasts finally have a model worth deploying in production.
Claude Opus 4.6 Fast (Mar 30) — Anthropic's speed-optimized Opus variant launched at month's end. Same reasoning depth, 40% faster time-to-first-token. Too late to show up in April rankings, but expect it to appear in May as enterprise users migrate from standard Opus.
Dropped Out
DeepSeek R1 drops from #10 to #13. The dedicated reasoning model is losing volume to general-purpose models that now include built-in thinking modes (Qwen3 Max, GPT-5.4 with chain-of-thought). Dedicated reasoning models may be a dying category.
GPT-5.2 falls to #11 as GPT-5.4 absorbs its traffic. OpenAI's model ladder is working — users step up to the latest, and older models fade within two cycles.
The Story
Stability at the top. Claude Sonnet 4.6 holds #1 for the second consecutive month, and the gap is widening. Anthropic's model has become the default for coding agents, content generation, and complex instruction following. Developer surveys consistently rank it highest for "reliability" — the model does what you ask, consistently, without creative interpretations of your prompt.
GPT-5.4 climbs to #2, overtaking both Gemini 3 Flash and the legacy Claude 3.5 Sonnet. This is OpenAI's best position since October 2025. The 500K context window is a genuine differentiator — developers processing entire codebases or long legal documents are choosing GPT-5.4 over Claude for context-heavy tasks.
The budget tier is a bloodbath. GPT-5.4 Mini, DeepSeek V3, Gemini 3 Flash, Qwen3.6 Plus, and Gemma 4 are all fighting for price-sensitive volume. The quality gap between these models is shrinking every month, which means the competition is increasingly about throughput, latency, and ecosystem integration rather than raw capability.
Claude 3.5 Sonnet drops to #5 — its lowest position ever. This isn't a failure; it's a natural succession. Sonnet 4.6 is strictly better at the same price. The remaining 3.5 traffic is locked-in contracts, pinned model versions in production systems, and teams that haven't allocated engineering time to migrate. By June, it'll likely fall out of the top 10 entirely.
Market Share
| Provider | Share | Trend |
|---|---|---|
| Anthropic | ~31% | Stable |
| OpenAI | ~25% | Growing |
| ~23% | Growing | |
| DeepSeek | ~9% | Declining |
| Alibaba | ~5% | Growing |
| xAI | ~3% | Stable |
| Meta | ~2% | Declining |
| Mistral | ~1% | Declining |
| Others | ~1% | — |
Q1 2026 in Review
Since we now have four months of data, here's how the landscape shifted in Q1:
| Metric | January | April | Change |
|---|---|---|---|
| #1 Model | Claude 3.5 Sonnet | Claude Sonnet 4.6 | New champion |
| Models in Top 10 (Anthropic) | 2 | 3 | +1 |
| Models in Top 10 (OpenAI) | 2 | 2 | Stable (but newer) |
| Models in Top 10 (Google) | 2 | 3 | +1 |
| Cheapest Input (per 1K) | $0.0001 | $0.0001 | Floor held |
| Most Expensive Input (per 1K) | $0.015 | $0.015 | Ceiling held |
| Total Models Tracked | 80+ | 110+ | +37% |
The big takeaway: the number of competitive models is exploding, but usage is consolidating around fewer winners. The top 3 models command over 40% of all API traffic. The long tail of models — open-weight fine-tunes, regional specialists, niche reasoning models — is growing, but each individual model captures less volume.
What to Watch in May
- Claude Opus 4.6 Fast will enter the rankings — early testers report it's a genuine improvement for latency-sensitive enterprise use cases
- DeepSeek V4 rumors are getting louder; a Q2 launch seems likely
- Meta Llama 4 is overdue and losing open-weight mindshare to Gemma 4 and Qwen3
- Google I/O in May could bring Gemini 4 announcements that reshape the entire ranking
Monthly LLM rankings by Demand Signals, sourced from OpenRouter usage data. Subscribe to our blog for monthly updates.
Get a Free AI Demand Gen Audit
We'll analyze your current visibility across Google, AI assistants, and local directories — and show you exactly where the gaps are.