Anthropic's negotiating a cloud deal with Google worth tens of billions while maintaining its AWS partnership. The structure reveals how foundation model companies are turning compute access into leverage—and why no single hyperscaler gets exclusivity anymore.
Meta cut 600 AI jobs from its established research units while protecting a new lab that barely exists. The $15 billion Scale AI deal brought new leadership who's now dismantling the structure that built Llama. Speed beats scale in the race to superintelligence.
Anthropic and Google circle a cloud pact measured in tens of billions
Anthropic's negotiating a cloud deal with Google worth tens of billions while maintaining its AWS partnership. The structure reveals how foundation model companies are turning compute access into leverage—and why no single hyperscaler gets exclusivity anymore.
A compute-for-TPUs deal would anchor Anthropic on Google Cloud while keeping AWS firmly in the frame.
A six-figure hourly compute bill meets a hyperscaler eager for flagship AI workloads. Anthropic is negotiating a cloud agreement with Google valued in the high tens of billions of dollars—an arrangement that would give the Claude maker priority access to Google’s tensor processing units and cement it as one of Google Cloud’s largest AI customers.
The discussions focus on services and silicon—not equity. Google already invested roughly $3 billion in Anthropic and, per court filings this spring, holds about 14 percent of the company. Both sides declined to comment, and the structure could change. What’s notable is the posture: Anthropic isn’t picking a side; it’s hedging at massive scale.
The Breakdown
• Anthropic negotiating Google cloud deal valued in high tens of billions for TPU access, not equity investment
• Google already holds 14% stake from $3B investment; Amazon committed $8B as primary cloud partner
• Multi-cloud strategy gives Anthropic leverage on pricing and chip access while limiting single-vendor lock-in
• Compute infrastructure emerging as binding constraint—pre-booked silicon access determines who competes at frontier
The infrastructure arithmetic
Training a frontier model consumes thousands of accelerators for weeks; serving it at scale is a permanent tax. That’s why the deal size matters more than any single chip spec. Anthropic needs guaranteed capacity over multiple cycles, not just a one-off tranche of GPUs or TPUs. A multi-year commitment also pre-allocates scarce power, networking, and floor space—the unglamorous constraints that increasingly decide who ships.
The economics favor partnerships. Building a proprietary footprint would take years and billions in capex before the first token moves. Renting from hyperscalers shifts risk, speeds deployment, and—crucially—locks pricing and supply in a market where demand still outruns wafer output. Optionality is a cost line, not a slogan. It keeps the models online.
What’s changed around Anthropic
The company just raised $13 billion at a $183 billion valuation, and it’s targeting an annual revenue run rate of about $9 billion by the end of 2025, according to people familiar and company statements. Those are the kinds of numbers that justify reserving entire data-center halls.
TPUs are the specific lure here. Google’s in-house accelerators are tightly integrated with its networking and software stack, which can reduce latency and increase throughput for certain training and inference patterns. For Anthropic, guaranteed access to that pipeline helps smooth production planning across model generations. For Google, landing a top-tier model lab concentrates usage on its proprietary ecosystem, which is the point.
Google’s foothold—without control
Despite the stake, Google has no board seat at Anthropic and no formal decision rights. That separation matters as the checks get larger. A supersized cloud contract deepens operational ties without altering governance. It also gives Google Cloud something it has chased for years: a banner AI tenant that makes its chips and fabric unavoidable for competitors watching from the sidelines.
The AWS complication
Amazon committed up to $8 billion to Anthropic and named it a primary cloud and training partner, with workloads spanning Nvidia GPUs and Amazon’s Trainium and Inferentia chips. AWS is also building “Project Rainier,” a Trainium2 UltraServer supercluster described as “hundreds of thousands” of chips across multiple U.S. data centers for Anthropic. A Google pact counted in tens of billions doesn’t evict AWS—but it does turn a primary into a peer, changing price dynamics and chip allocation on both sides.
The real contest: leverage, not loyalty
These deals aren’t just about compute; they’re about bargaining power when the next model cycle hits. Hyperscalers want durable, high-margin AI workloads and the prestige that attracts the next tier of customers. Model labs want predictable access to accelerators, power, and bandwidth at prices that don’t crater gross margin. Multi-cloud is therefore less about redundancy than about leverage. It lets Anthropic play silicon roadmaps, energy availability, and contract terms against each other, while insulating the business from any single vendor’s supply hiccup.
If finalized on the rumored scale, this agreement would also signal where AI economics are settling. The winners will be firms with pre-booked, preferential access to vertically integrated infrastructure—chips, fabric, storage, compilers, and ops teams—because that bundle, not a headline TFLOP figure, decides time-to-train and cost-to-serve. Everyone else will either accept worse unit economics or retreat to narrower product ambitions.
What could still go wrong
Early-stage talks can stall. Regulatory scrutiny of hyperscaler-model-lab entanglements is rising. And TPU specialization cuts both ways: if Anthropic’s next architectures favor a different accelerator or memory topology, the cost of migrating workload share could climb. None of that negates the logic. It only underscores why Anthropic is anchoring two providers at once.
Why this matters
Compute access—not clever prompts—is the binding constraint; firms with pre-booked silicon and power will set the pace in foundation models.
Multi-cloud at this scale forces hyperscalers to compete on price and chip access while limiting any one provider’s leverage over model cadence.
❓ Frequently Asked Questions
Q: What are TPUs and why does Anthropic want Google's specifically?
A: Tensor Processing Units are Google's custom chips designed for machine learning workloads. They're tightly integrated with Google's networking and software stack, which can reduce latency and increase throughput for certain training and inference patterns. For Anthropic, guaranteed TPU access means predictable performance across model generations without competing for scarce Nvidia GPUs on the open market.
Q: How does this deal compare to Microsoft's OpenAI partnership?
A: Microsoft invested $13 billion in OpenAI and secured exclusive cloud rights—OpenAI runs entirely on Azure. Anthropic's approach splits infrastructure: Amazon committed $8 billion as "primary cloud partner," Google invested $3 billion and now negotiates this cloud deal. Anthropic maintains leverage by avoiding single-vendor dependence, while OpenAI accepted exclusivity in exchange for deeper Microsoft integration and capital.
Q: What is AWS's Project Rainier mentioned in the article?
A: Project Rainier is a supercluster of AWS Trainium2 UltraServers—Amazon's custom AI chips—built specifically for Anthropic. It features "hundreds of thousands" of Trainium2 chips interconnected with low-latency networking across multiple U.S. data centers. It's Amazon's answer to keeping Anthropic workloads on AWS infrastructure rather than Nvidia GPUs or Google TPUs.
Q: Why doesn't Anthropic just build its own data centers?
A: Building proprietary data centers requires years and billions in upfront capital before deploying the first model. You need land, power infrastructure, cooling systems, networking, and ongoing operations teams. Renting from cloud providers shifts that risk, speeds deployment, and locks in pricing during supply shortages. For a company targeting $9 billion revenue by late 2025, capital flexibility matters more than infrastructure ownership.
Q: What does Anthropic's "$9 billion annual revenue run rate" target mean?
A: Annual revenue run rate projects yearly revenue based on current monthly or quarterly sales. If Anthropic hits $750 million in monthly revenue by December 2025, that's a $9 billion annual run rate ($750M × 12 months). It's not actual annual revenue—it's an extrapolation. The metric suggests Anthropic expects to roughly triple from its current estimated $3 billion run rate within 14 months.
Tech journalist. Lives in Marin County, north of San Francisco. Got his start writing for his high school newspaper. When not covering tech trends, he's swimming laps, gaming on PS4, or vibe coding through the night.
Meta cut 600 AI jobs from its established research units while protecting a new lab that barely exists. The $15 billion Scale AI deal brought new leadership who's now dismantling the structure that built Llama. Speed beats scale in the race to superintelligence.
Neolix just raised $500 million after deploying 10,000 robovans—the first autonomous vehicle maker to hit that milestone. The company breaks even at 1,000 units a month. China's delivery volume and cost structure prove the economics work, but can that advantage travel overseas?
OpenAI's Atlas browser doesn't just compete with Chrome—it restructures search economics. When conversational AI replaces link navigation, Google's $237 billion ad model faces its first credible technical challenge in 15 years. Investors noticed.
OpenAI's Atlas browser defaults to split-screen ChatGPT on every page—making the AI assistant permanent, not optional. Alphabet's stock dropped 3% as investors priced in Chrome's first credible challenger in the agent era.