Suprise Elon Anthropic Team Up Reshapes the AI Race

May 7, 2026 · Episode Links & Takeaways

HEADLINES

Managed Agents Get Memory, Oversight, and Orchestration

Anthropic's Developer Day, branded "Code with Claude," was a signal: models are no longer the main event. There was no Mythos reveal, no new Opus release — just a sharp focus on agents and the infrastructure around them. The central new releases were a major update to Claude Managed Agents, addressing what are essentially the three hardest challenges in agentic AI right now: memory, quality review, and coordination. The headline feature is something Anthropic is calling Dreaming — a scheduled memory review process that runs between sessions, extracts patterns from past work, and curates memory so agents improve over time. It's the same thing that drew people to open-source tools like Hermes, now baked into Anthropic's platform as a default.

Outcomes: Automated Quality Review for Non-Code Agents

Alongside Dreaming, Anthropic shipped Outcomes — a feature that lets users define a rubric for what success looks like, then scores agent output against that rubric using a separate grading agent. If the work doesn't pass, the task gets kicked back for another run. Anthropic claims it improved file generation quality by 8.4% for Word docs and 10.1% for PowerPoint slides in internal testing. External grading agents aren't a new idea in multi-agent system design, but they've mostly been used for coding tasks where rubrics are clear-cut. The interesting unlock here is extending the concept to subjective, non-code knowledge work — and making it the default rather than something you have to architect yourself. The managed agents platform also now supports full multi-agent orchestration, with a lead agent able to break jobs into pieces and delegate to specialists, all tracked and auditable via Claude Console.

Claude Finance: 10 Pre-Built Agents for Financial Services

The day before Dev Day, Anthropic dropped a suite of 10 predefined agents for financial services — pitch builder, meeting preparer, market researcher, valuation reviewer, month-end closer, and more. The idea is to hand financial services firms a starter pack rather than requiring custom builds from scratch, with a full cookbook included for teams that want to modify the internals. Early adopters are posting real numbers: AIG says they've compressed underwriting time fivefold while improving data accuracy from 75% to 90%. FIS says anti-money laundering investigations have gone from days to minutes. Nathaniel Lin, Anthropic's head of product for financial services, called finance "a great blueprint for the rest of knowledge work" — and said these use cases are only a few months behind coding. Worth noting: most of these agents aren't replacing high-skill finance work. They're going after the low-skill, repetitive task layer that was already semi-automated or outsourced.

"Infinite" Context Windows and What Comes Next for Models

The roadmap tease got real attention: Research Head of Product Diane Penn outlined three things Anthropic is actively working on — higher judgment and code taste, "infinite" context windows, and multi-agent coordination. The scare quotes around infinite were intentional; Penn said they're building "context windows that feel infinite," leaving open the question of whether this is better compaction or something more fundamental. Whatever it is, the direction is clear: all three of these features are aimed squarely at making agents more capable, not at chasing the next benchmark. Infinite context in particular is a necessary unlock for agents working across large codebases or running for extended periods.

Boris Cherny Is Done With "Vibe Coding"

Claude Code creator Boris Cherny used a panel at Dev Day to officially disavow the term. Less than 18 months after Andrej Karpathy coined it, "vibe coding" no longer describes how Anthropic — or most developers — actually uses AI. At Anthropic, there's no manually written code left; Claudes coordinate over Slack, code in loops, and resolve issues across the codebase, with copious automated testing before anything ships. The problem: Boris doesn't yet have a better term. Karpathy offered "agentic engineering," but that's not quite landing either. He says he's taking suggestions.

Dario: "We Planned for 10X. We Got 80X."

The clearest explanation yet for why Anthropic has been so compute-constrained came straight from Dario Amodei at Dev Day. In Q1 of this year, Anthropic saw 80x annualized growth in revenue and usage — against infrastructure that was built for 10x. That gap is the entire compute crisis in one number.

MAIN STORY

Elon and Anthropic: The Deal That Changes Everything

After noon Eastern on Wednesday, basically nothing else mattered. The Claude AI account posted that they'd agreed to a partnership with SpaceX — and 20 million views later, the AI conversation had shifted completely. This is not a deal that's merely interesting for tabloid reasons. It has the potential to genuinely reshape the AI race: fixing Anthropic's acute compute crisis while completing a pivot that repositions Elon from model builder to infrastructure czar.

COMPUTE ALLIANCE

What the Deal Actually Is
Anthropic gets the whole Colossus 1 data center — immediately.
Colossus 1 is the data center xAI built at record speed in Memphis in mid-2024, now scaled to 220,000 NVIDIA GPUs (mostly H100s) running at 300 megawatts. The deal covers that entire facility. Colossus 2 — the newer Blackwell cluster with ~550,000 GPUs — stays with SpaceX. Inference comes online within the month. For users, the immediate changes are: Claude Code's five-hour rate limit doubled for Pro, Max, Team, and Enterprise plans; peak hour throttling eliminated for Pro and Max; and API rate limits for Opus models up 2-10x depending on tier. Anthropic Head of Growth Amol Avasare noted that far more users hit the five-hour limit than weekly limits, so that's what they fixed first — with weekly limits to be revisited as more compute comes online.

Why It Had to Happen
Perfect asymmetry: Anthropic had the model, xAI had the GPUs.
As Derek Thompson put it, Musk had compute capacity but a meh model, and Anthropic had a fantastic model with weak capacity. Anthropic has been straining under a compute crunch all year, giving OpenAI an opening to reclaim ground it had lost. Meanwhile, xAI had the opposite problem: model improvement had stalled out with Grok 4.2 gathering no real buzz, the Cursor partnership wasn't going anywhere on the model side, co-founders had departed one by one, and the company had a warehouse full of GPUs with too little to do. Chamath Palihapitiya had essentially predicted this on the All-In Pod days earlier, arguing that Elon had massive leverage from excess compute capacity and suggesting he and Dario should do a deal. They did.

How Elon Explains Himself
"No one set off my evil detector."
In a tweet that now has tens of millions of views, Elon wrote that he spent time last week with senior Anthropic team members to understand how they ensure Claude is good for humanity, came away impressed, and was therefore comfortable leasing Colossus 1 — noting that SpaceX AI had already moved its own training to Colossus 2. Until this week, Elon had repeatedly called Anthropic "misanthropic" on Twitter. Apparently that's water under the bridge.

Elon's AI Play 3.0: From Model Builder to Compute Czar
The move that was always about compute, not Grok.
The deeper story here is about what Elon is actually trying to become. He's long wanted one company to rule them all in his empire — for a while that seemed like it would be Tesla, then briefly maybe xAI. But it's become increasingly clear that king-making in AI is going to be about compute infrastructure, not models. The SpaceX-xAI merger may have had less to do with giving SpaceX a model and more to do with giving SpaceX a footprint in terrestrial supercomputing that it could build on — from the earth, eventually, into orbit. The argument is that Elon realized his best path to shaping this industry wasn't being Sam Altman or Dario Amodei, but being Jensen Huang. If that's right, then Anthropic was the only call to make — the Sam feud rules out OpenAI, and Anthropic's disciplined, focused build approach maps more naturally to how Elon operates within a specific company. The AI play goes: 1.0 as OpenAI funder, 2.0 as model builder, 3.0 as compute czar. And Elon even tweeted the punctuation: xAI will be dissolved as a separate company. It'll just be SpaceXAI.

SpaceX as the Neocloud
"Elon fumbled the model but built a neocloud that works great for frontier labs."
That's Rohit's read, and Derek Thompson agreed: Musk has been world-class at compressing money, resources, and time to do known-but-hard things at scale — electric cars, cheaper rockets, now massive compute clusters. He's less suited to cracking open genuinely unknown breakthroughs, which is what frontier model research requires. So it makes sense that xAI lagged on models while building infrastructure that frontier labs need. Tom Brown, Anthropic's Chief Compute Officer, confirmed work starts immediately, posting that he's grateful to be partnering with SpaceX: "There's nobody better at quickly moving atoms, on or off planet Earth." Aaron Levie put it simply: "SpaceX as a vertically integrated AI compute company makes an insane amount of sense."

What It Means for Grok
"Certainly seems like a blow to the idea that Grok will remain a frontier model." — Ethan Mollick
Grok will stick around — X/Twitter needs something integrated, and there are options to keep open for Optimus and embodied robotics. But the direction is clear. Multiple observers noted that xAI renting out its flagship data center to a competitor is a signal about where the value in the business lies. The highest-return thing SpaceX AI can do with that compute isn't training Grok; it's renting to Anthropic. As one analyst put it: SpaceXAI will be more like AWS and less like OpenAI — and notably, one of those is wildly profitable.

What It Means for Users
A Claude that actually works — and that won't be known as "the Elon AI."
The compute crunch has been visibly degrading the product for months. The prospect of a Claude that doesn't throttle every few hours is the obvious upside. On the Elon association concern: you won't know whether you're being served from AWS or xAI. The infrastructure is invisible. Claude's brand stays Claude's — and the product may just be good enough that nobody cares about the plumbing.