Anthropic: Claude Sonnet 4.5 vs xAI: Grok 3

Side-by-side specs, pricing, and benchmarks. Pick a winner for your team's use case.

Use it in a Space

Spin up a Switchy Space with either model — your whole team @-mentions it with shared context, pooled credits, one memory.

Pricing
Anthropic: Claude Sonnet 4.5xAI: Grok 3
Input $/Mtok$3.00 · $3.00
Output $/Mtok$15.00 · $15.00
Context window
Anthropic: Claude Sonnet 4.5xAI: Grok 3
Anthropic: Claude Sonnet 4.51000K tokens
xAI: Grok 3131K tokens

Bars use square-root scaling so a 1M-token window doesn't crush a 200K one.

Release timeline
Anthropic: Claude Sonnet 4.5xAI: Grok 3
2025-09-29
2025-06-10
2025-05-11today

Anthropic: Claude Sonnet 4.5

Provider
anthropic
Context
1000k
Input $/Mtok
$3.00
Output $/Mtok
$15.00
Max output
64000
Modalities
text, image, file

xAI: Grok 3

Provider
x-ai
Context
131k
Input $/Mtok
$3.00
Output $/Mtok
$15.00
Max output
Modalities
text

Price delta

Anthropic: Claude Sonnet 4.5 and xAI: Grok 3 charge the same on input tokens. Output: same price on both sides.

Which to pick

Pick **Claude Sonnet 4.5** for general-purpose team work — coding, writing, structured reasoning, multi-step planning. Sonnet leads Grok 3 on most reasoning and code evals, the 200k-token context comfortably exceeds Grok's 131k, and Anthropic's tool-use ergonomics are more predictable in long multi-turn chats. Pricing is identical at $3 in / $15 out per Mtok, so the choice rarely comes down to cost. Pick **Grok 3** when you specifically want X / xAI's tone, real-time signal access, or have a workflow that benefits from Grok's particular search integration. On pure model quality the gap favours Sonnet, but Grok's strengths are stylistic and ecosystem-shaped, not benchmark-shaped.
Data last verified 22 hours ago.Sources aggregated hourly to weekly. See docs/architecture/model-directory.md.