Anthropic Buys xAI Compute to Fix Claude Crunch
Theo - t3.gggo watch the original →
the gist
Anthropic partners with xAI/SpaceX for NVIDIA compute to alleviate severe capacity shortages from 80x demand growth, boosting Claude rate limits despite past animosity.
Anthropic's Explosive Growth Outpaces Compute Planning
Anthropic underestimated demand for Claude, hitting 80x annualized growth in revenue and usage in Q1, far beyond their planned 10x yearly expansion. CEO Dario Amodei admitted at Code with Claude conference: "this is the first year we've grown faster than the exponential... we tried to plan very well for a world of 10x growth per year um in the first quarter of this year we saw if you were to annualize it 80x growth per year in revenue and usage." Conservative GPU buys left them short, forcing compute splits across incompatible hardware: Amazon's Trainium (non-NVIDIA), Google TPUs, Microsoft Azure NVIDIA, and now xAI. Researchers prefer CUDA/NVIDIA for training, so inference gets shunted to non-optimal chips, but even that failed under load. Idle compute loses money every idle minute, while peak overloads throttle users—hence recent tier tweaks aimed at rationing, not revenue grabs.
SpaceX/xAI Deal Ends Beef for Survival
Despite history—Anthropic banned xAI from Claude in January—Elon Musk leased Colossus 1 (million H100-equivalent) to Anthropic after vetting their team: "I spent a lot of time last week with senior members of the anthropic team to understand what they do to ensure Claude is good for humanity and was impressed... no one set off my evil detector so long as they engage in critical self-examination Claude will probably be good." Musk, a vocal Anthropic critic (repeated "misanthropic" jabs), prioritizes compute sharing since SpaceX shifted training to Colossus 2. This underscores compute desperation: Anthropic hates xAI/SpaceX, but needs NVIDIA scale. xAI's Colossus provides what multi-cloud fragmentation can't—pure NVIDIA firepower.
User Wins: Higher Limits, But Weekly Caps Linger
Relief hits Claude subs and API: Pro/Max 5-hour limits double (e.g., Claude Code bursts), peak-hour throttling ends. API tiers explode—Tier 1 input TPM from 30k to millions, Tier 3 from 800k to 5M, Tier 4 to 10M—easing enterprise hosting pains. Charts show Amazon Trainium fastest (80-100 TPS), Google TPUs solid (50-70), Anthropic/Azure suspiciously mirrored (~50s TPS), hinting Azure proxies via Anthropic APIs, not true hosting. Multi-cloud strategy aids enterprise adoption (pick your cloud), but performance varies; NVIDIA edges for researchers. Bursty users benefit most; steady/heavy hit weekly resets unchanged, limiting parallel agents/autoreview gains.
xAI's Pivot: Compute Kingpins Eye Data via Cursor
xAI bids aggressively for Cursor ($60B acquisition option or $10B collab fee), pairing million-H100 Colossus with Cursor's dev distribution/data. Success needs research, data, compute—xAI nails compute but lags data. Twitter corpus is noisy (e.g., Reddit microwave noise spiked GPT-3 losses; Gemini loops on it). Labs shift to synthetic data: engineer fake products for histories, label human-AI chats. Transform user-model threads into one-shots for RLHF: isolate initial request, flip subsequent "user" corrections to "assistant" responses, creating ideal training pairs. Cursor's millions of dev sessions (build feature → part1 → fix → next) yield gold post-processing. xAI lacks this; Cursor deal grabs it, positioning for frontier catch-up.
Data Quality Trumps Scraping in Frontier Race
Raw web/Twitter poisons models—cleaning erases loss spikes. Post-web-scraping era demands manufactured data: full product sims, labeled interactions. Chat histories reinforce flaws unless recast as one-shots (user: build X; assistant: part1; user→assistant: next/correct). xAI's late start means buying data pipelines; Cursor provides dev-specific goldmine. OpenAI won early via novel research/data despite compute weakness; now compute leads, but idle talentless clusters waste (recall: excess compute pricey sans users).
Notable Quotes
- Dario Amodei: "we saw... 80x growth per year in revenue and usage" – on Q1 demand explosion crippling planning.
- Elon Musk: "no one set off my evil detector so long as they engage in critical self-examination Claude will probably be good" – post-Anthropic team meetings, greenlighting Colossus lease.
- Theo on subscriptions: "every minute your compute isn't running you are losing money... if you have more requests coming in than you have the compute to serve you're screwed."
- On xAI/Cursor: "the combination of Cursor's leading product and distribution... with SpaceX's million H100 equivalent Colossus... will allow us to build the world's most useful models."
Key Takeaways
- Monitor Claude limits: doubled 5h bursts help heavy coders; API TPM surges aid hosting (Tier 3: 5M input).
- Bet on compute leaders: xAI's Colossus scales via leasing; Anthropic's multi-cloud hedges but fragments.
- Clean data ruthlessly: scrapes poison (microwave loops); synthesize via fake products/chat recasts.
- Recast histories for RLHF: flip user corrections to assistant in one-shots for ideal training.
- Enterprise pick clouds wisely: Amazon Trainium fastest Claude inference (80-100 TPS).
- xAI eyes Cursor for dev data edge; watch $60B bid outcome.
- Idle compute bleeds cash—prioritize demand-matched scaling over hoarding.
- Researchers: CUDA/NVIDIA > Trainium/TPUs for training; inference offloads to free it.