Anthropic's SpaceX Compute Deal Caps Radical AI Partnership Resets
The $4 billion lease of Colossus 1 is only the most dramatic move in a spring that saw OpenAI break free of Microsoft, Meta sign with CoreWeave, and every major lab become a multi-cloud compute shopper.
Wired
On May 6, 2026, Anthropic announced it had signed a deal to run inference workloads on SpaceX's Colossus 1 supercomputer, a facility originally built to train xAI's Grok. The deal, valued at roughly $4 billion according to Forbes, gives the maker of Claude access to one of the largest single clusters of NVIDIA GPUs on the planet. The announcement landed less than five weeks after Anthropic had already struck a separate cloud deal with CoreWeave, and within a fortnight of Microsoft and OpenAI formally dismantling the exclusive hosting arrangement that had defined the lab-cloud relationship since 2019. In a single spring, the map of who provides compute to whom in frontier AI had been redrawn.
Colossus 1 houses 220,000 NVIDIA GPUs, TweakTown reported, and was built by xAI Holdings in 2024. Elon Musk had positioned the facility as the engine behind Grok, xAI's chatbot, which competes directly with Anthropic's Claude. By early 2026, however, Grok's monthly downloads had fallen 59 percent between January and April, according to data cited by MSN. The supercomputer's capacity was suddenly available, and Anthropic, facing what Wired described as a growth-driven compute shortage, was ready to pay for it.
Under the arrangement, Anthropic will lease the full capacity of the Colossus 1 data center, MSN reported. The immediate beneficiary is Claude's user base: the added compute would be used to raise Claude Code's session limits and improve API throughput, Anthropic told SiliconANGLE. The deal is structured as a straight capacity lease rather than a revenue-sharing or equity arrangement. That structure matters. It means Anthropic is treating compute as a commodity input it can shop for on price and availability, not as a strategic dependency that comes with governance strings.
The SpaceX deal did not arrive in isolation. On April 10, Reuters reported that CoreWeave had struck a deal to supply Anthropic with cloud computing capacity, sending the infrastructure provider's shares up. A day earlier, CoreWeave had signed a $21 billion deal with Meta, granting the social-media giant early access to Nvidia's next-generation Vera Rubin chips, Reuters reported separately. CoreWeave, once a niche Ethereum miner turned GPU cloud, had in the span of 48 hours become a compute supplier to two of the most capital-intensive AI labs in the world.
Then came the restructuring everyone had been expecting. On April 27, VentureBeat reported that Microsoft and OpenAI had dismantled the exclusive provisions of their partnership, ending Microsoft's sole right to sell OpenAI's models and freeing the company to offer its services through AWS, Google Cloud, and Oracle. Microsoft retained a 27 percent stake in OpenAI's for-profit entity and secured a cap on shared revenues at $38 billion, Yahoo Finance reported, citing Wedbush analysis. The AGI clause, a provision that had once threatened to cut Microsoft off from OpenAI's most advanced models, was scrapped entirely, WinBuzzer noted.
Taken together, the three deals describe a structural shift in how frontier AI labs procure compute. Eighteen months earlier, the market had a simple shape: OpenAI ran almost entirely on Azure, Anthropic leaned heavily on AWS, and Google DeepMind used Google Cloud. Each lab was tethered to a single hyperscaler, and that tethering was understood as a feature of the landscape, not a bug. The 2026 spring deals broke that pattern. Anthropic now draws compute from AWS, CoreWeave, and SpaceX. OpenAI is pushing aggressively into Amazon's cloud, CNBC reported, while still renting capacity from Microsoft. The labs have become multi-cloud by default.
The scale of the demand driving this diversification is difficult to overstate. Hyperscalers plan to spend between $635 billion and $665 billion on AI infrastructure in 2026 alone, a 67 percent jump from 2025, according to data cited by MSN. The Big Four cloud providers combined are approaching $725 billion in AI-related capital expenditure plans, the Irish Times and 24/7 Wall St. reported. Even at those sums, demand for the most advanced GPU clusters is outstripping supply. Labs that once relied on a single provider are discovering that no single provider, however well-capitalized, can build fast enough.
The supply constraint has a specific shape. Nvidia's newest chip generations, Blackwell and the forthcoming Vera Rubin, are being allocated through relationships that predate the current spending surge. New entrants and labs without legacy purchase agreements find themselves at the back of a queue that stretches into 2027. By leasing Colossus 1, Anthropic gained immediate access to 220,000 already-deployed H100-class GPUs without waiting for a new cluster to be built. The CoreWeave deal, meanwhile, gives Meta early access to Vera Rubin chips that would otherwise not be available until late 2026 or early 2027, Reuters reported. Speed to deployment has become a bargaining chip in its own right.
For SpaceX and its CEO Elon Musk, the deal represents a pivot that few would have predicted a year ago. Musk spent much of 2024 and early 2025 publicly criticizing Anthropic and its CEO Dario Amodei, calling the company a threat and framing xAI's Grok as the alternative. By leasing Colossus 1 to Anthropic, SpaceX is effectively monetizing capacity that Grok was not using at full scale. The decision turns a fixed-cost liability, a half-empty supercomputer, into a $4 billion revenue line. It also places Musk's company in the position of powering a product that competes directly with his own.
The reconfiguration of compute partnerships changes who holds leverage. Under exclusive arrangements, cloud providers could embed themselves into a lab's operations, offering credits and engineering support in exchange for preference on model releases and distribution. Multi-cloud procurement breaks that model. When a lab can credibly threaten to shift workloads from Azure to AWS, or from Google Cloud to CoreWeave, the hyperscaler's negotiating position weakens. The cheapest signal that this strategy is working for Anthropic is visible in the product itself: Claude Code's session limits doubled within days of the SpaceX deal being announced, TweakTown reported, a change that required no new model architecture, only more available compute.
The labs are not merely diversifying suppliers. They are becoming sophisticated compute shoppers, treating GPU hours as a market rather than a relationship. This has downstream consequences. Cloud providers are being forced to compete on price, latency, and chip generation rather than on bundled services. It also opens room for non-hyperscaler infrastructure firms like CoreWeave and, now, SpaceX to enter the market as credible alternatives to AWS, Azure, and Google Cloud. The barrier to entry for a compute provider is no longer 'build a full cloud platform,' but simply 'amass a large enough GPU cluster and offer competitive pricing.'
Microsoft's restructuring with OpenAI shows that the incumbent hyperscalers are not exiting the field. By securing a $38 billion revenue cap and retaining its equity stake, Microsoft traded exclusivity for certainty. It will no longer be the sole distributor of OpenAI's models, but it also will not be exposed to unlimited compute costs if OpenAI's usage continues to grow exponentially. The deal, eWeek reported, clears the way for a relationship that is commercial rather than quasi-parental, and in doing so may prove more durable than the arrangement it replaced.
OpenAI, freed from exclusivity, moved quickly. CNBC reported on April 29 that the company's drift toward Amazon had become "an aggressive move." The Stargate Norway data center project, originally conceived as an OpenAI-led infrastructure build, was being taken over by Microsoft while OpenAI began negotiating to rent capacity from other providers, CNBC reported. The image is of a lab unbundling itself from a decade-long dependency, one contract at a time.
The compute partnership story is not confined to GPU clusters. On April 29, WinBuzzer reported that MIT and IBM had launched the MIT-IBM Computing Research Lab, expanding a longstanding AI research partnership into quantum computing, hybrid algorithms, and next-generation chip architectures. The move signals that labs and research institutions are already looking past the current GPU supply crunch toward compute paradigms that could reset the playing field entirely. Quantum computing remains years from practical AI training workloads, but the partnership structure, a university and a legacy hardware firm pooling resources, echoes the multi-party arrangements that frontier AI labs are now pursuing for classical compute.
There is also a longer horizon. Forbes reported that Anthropic has expressed interest in partnering with SpaceX on the development of AI data centers in space. Orbital data centers, if they materialize, would sidestep the land, power, and permitting constraints that are slowing terrestrial data-center construction in Northern Virginia, Ireland, and Singapore. They would also introduce a new set of engineering challenges: cooling in a vacuum, radiation shielding for GPU memory, latency to ground stations. The fact that the idea is being discussed seriously between a frontier AI lab and a launch provider tells you something about how far the compute scramble has gone.
The season of partnership resets leaves three questions hanging. The first is whether multi-cloud procurement actually improves model performance, or simply spreads risk. Adding a new GPU cluster does not automatically make a model smarter; it makes inference faster and training runs larger, but only if the software stack can orchestrate across heterogeneous infrastructure. The second is whether the non-hyperscaler entrants can sustain their position. CoreWeave carried $35 billion in planned 2026 capital expenditure, Reuters reported, a number that will require continued access to debt markets that have grown wary of AI infrastructure exposure. The third is whether the labs, having broken free of single-provider dependency, will use their newfound bargaining power to drive down the cost of inference to the point where the economic model of frontier AI itself begins to change.
The Colossus 1 deal will serve as a test case. If Claude's performance and availability improve measurably over the next two quarters without a corresponding degradation in Anthropic's relationships with AWS and CoreWeave, the multi-provider model will have demonstrated its viability. If, on the other hand, managing three compute backends introduces latency, cost overruns, or coordination failures, the labs may discover that diversification has a complexity ceiling. The nameplate on the world's largest GPU cluster now reads, in effect, powered by SpaceX and rented by Anthropic. Whether that nameplate stays in place, and whether others like it appear, will tell us more about the future of AI compute than any earnings call or white paper.