Mira Murati's $3B Google Cloud Gamble: First Real Test of Post-OpenAI Strategy
Everyone's calling this a "strategic partnership." I call it desperation wrapped in press release poetry.
Mira Murati's Thinking Machines Lab just inked a multi-billion-dollar deal with Google Cloud—specifically valued in the single-digit billions, according to the April 22nd announcement at Cloud Next '26. The former OpenAI executive is betting her post-ChatGPT reputation on Google's AI Hypercomputer infrastructure and those shiny new NVIDIA GB300 NVL72 chips.
Here's what caught my attention: This isn't just another cloud migration story. TML became one of the first customers to get their hands on Google's A4X Max virtual machines powered by Blackwell architecture. Early testing showed a 2X improvement in training and serving speeds compared to previous-generation GPUs.
That's impressive. If true.
<> "Google Cloud got us running at record speed with the reliability we demand," said Myle Ott, TML's founding researcher./>
But let's dig deeper into what TML actually needs all this compute for. Their Tinker architecture runs on reinforcement learning—the same approach that gave us AlphaGo breakthroughs and ChatGPT's human-like responses. RL is notoriously compute-hungry, requiring massive parallel processing for continuous training alongside production serving.
The Infrastructure Reality Check
Google's selling TML on their integrated stack:
- Jupiter network for fast weight transfers
- GKE orchestration with automated remediation
- Spanner for metadata management
- Cloud Storage with custom caching for global workloads
Mark Lohmeyer, Google Cloud's VP of AI Infrastructure, emphasized the "seamless integration" angle. Translation: Please don't leave us for AWS like everyone else does.
The timing isn't coincidental. Google's scrambling to lock in frontier AI labs early, especially after watching Amazon and Microsoft carve up the market. Offering exclusive early access to GB300 chips? That's not partnership—that's cloud provider arms dealing.
The Elephant in the Room
Here's what nobody's talking about: This deal is non-exclusive.
TML can still run workloads on other cloud providers. Murati learned from OpenAI's Microsoft dependency—she's not putting all her chips (literally) in one basket. Smart move, but it also signals she's not fully convinced by Google's pitch.
The real question isn't whether GB300s are faster (they are) or whether Google's infrastructure is solid (it probably is). It's whether Murati can actually deliver on the reinforcement learning hype that's driving this massive spend.
TML's first cloud provider deal comes with enormous expectations. The company already secured NVIDIA investment earlier in 2026, so they've got the hardware vendor and cloud provider locked in. Now they need to prove their Tinker architecture can compete with whatever Sam Altman's cooking up at OpenAI.
Show Me the Models
I've seen too many "revolutionary" AI startups burn through billions in compute credits while producing glorified chatbots. Murati's betting that reinforcement learning will differentiate TML from the GPT copycats flooding the market.
Maybe she's right. Her OpenAI track record suggests she understands what it takes to ship frontier models. But $3+ billion buys a lot of AWS credits if this Google experiment doesn't pan out.
The cynical take? Google needed a marquee AI customer for Cloud Next '26, and TML needed compute credits. Both sides get what they want—until the first major outage or the quarterly burn rate review.
The optimistic take? Murati's building something genuinely different, and Google's infrastructure might actually accelerate breakthrough research instead of just expensive experiments.
We'll know which narrative is true when TML starts shipping products instead of press releases.

