- Roko's Basilisk
- Posts
- Who Controls AI Supply?
Who Controls AI Supply?
Plus: Positron vs Nvidia, Claude’s Super Bowl ad, Reddit–NASA drama.
Here’s what’s on our plate today:
🧪 The Laboratory: OpenAI–Amazon’s $10B chip pact decoded.
🧰 3 Things Worth Trying: hands-on tools for multi-cloud AI.
📰 Quick Bits, No Fluff: Positron chips, Claude ads, and Reddit drama.
🗳️ Thursday Poll: Is OpenAI–Amazon diversification smart or dangerous?
Let’s dive in. No floaties needed…

The AI Talent Bottleneck Ends Here
AI teams need PhD-level experts for post-training, evaluation, and reasoning data. But the U.S. talent pipeline can’t keep up.
Meet Athyna Intelligence: a vetted Latin American PhD & Master’s network for post-training, evaluation, and red-teaming.
Access vetted PhD experts, deep STEM knowledge, 40–60% savings, and U.S.-aligned collaboration.
*This is sponsored content

The Laboratory
Why the OpenAI-Amazon deal matters for the future of AI supply chains
The sheer scale of modern enterprise operations can often be difficult to comprehend. Take, for instance, the semiconductor supply chain. Chips are designed in California, printed in the Netherlands, fabricated in Taiwan and South Korea, and finally assembled in mainland China.
This model of working allows different companies to focus on a niche part of the production process, thereby enabling advancements at a scale that would otherwise be impossible.
Over the past couple of years, a similar model has evolved for the artificial intelligence industry. In this still-evolving model, AI labs like OpenAI, Anthropic, and Meta develop the algorithms; NVIDIA, AMD, Intel, and Amazon provide the chips and cloud infrastructure; and enterprises package the product. In this case, AI is then integrated into workflows and chatbots that reach users.
The evolution of this model depends on participants scaling their operations while strengthening their symbiotic relationship. From this perspective, companies operating in this space have signed interfirm agreements that link their growth to one another.
The interconnected future of AI
The latest deal between the companies that make modern AI systems possible is expected to be signed between OpenAI and Amazon. According to a report from The Information, the companies are negotiating a $10B investment that would value OpenAI at more than $500B. The deal is expected to include commitments from OpenAI to adopt Amazon Web Services’ proprietary Trainium and Inferentia AI chips alongside its existing NVIDIA-dominated infrastructure.
This represents the most significant strategic realignment in AI infrastructure since Microsoft’s initial $13B investment in OpenAI in 2019.
Why is OpenAI turning to Amazon?
The deal between OpenAI and Amazon makes visceral sense when examined through the lens of scarcity rather than strategy.
At the time the two companies are strengthening their ties, NVIDIA is reportedly reducing production of the GeForce RTX 50 series by 30-40% due to memory shortages. The chipmaker, in the meantime, is prioritizing AI data center chips that generate revenue orders of magnitude higher.
In this scenario, OpenAI faces a brutal calculation: it cannot scale on NVIDIA alone because NVIDIA cannot manufacture enough chips to meet demand, even if OpenAI could afford them.
Similarly, Microsoft’s Azure infrastructure is heavily committed to existing enterprise workloads. Oracle’s Stargate project won’t deliver meaningful capacity until late 2026. In the meantime, OpenAI needs compute, and Amazon has it, even if they are in the form of chips that aren’t NVIDIA.
Charles Fitzgerald, the cloud infrastructure investor, frames it plainly: OpenAI doesn’t have the cash to honor even its $38B existing AWS commitment, let alone a fraction of the $1.4T in aggregate infrastructure commitments it has signed.
The $10B investment from Amazon may be less about capital injection and more about Amazon prefunding infrastructure that OpenAI will rent back, creating a closed-loop accounting structure that makes the burn rate appear more manageable to investors.
For OpenAI, the choice is simple: either wait for NVIDIA chips, which would mean missing growth targets and jeopardizing the company's future, or seek an alternative chip source. Which, in this case, would mean Amazon.
And for OpenAI, the timing is spot on.
Why Amazon’s chips are a viable alternative to NVIDIA
For years, AWS promoted Trainium and Inferentia with a straightforward value proposition: lower cost. The promise was comparable inference performance at 40-70% lower cost than NVIDIA. The market response was consistent and dismissive: “But does it run CUDA?”
However, that objection is steadily losing force. The first reason is structural. The economics of AI computing have shifted. More than 80% of AI compute consumption today is devoted to inference rather than training. Inference workloads account for the majority of recurring operational expenditure. As a result, cost reductions at the inference layer compound over time. At a sufficient scale, they become decisive.
The second reason is evidence. Amazon’s Project Rainier is not an experimental deployment. It involves approximately 400K to 500K Trainium2 chips training Anthropic’s Claude models exclusively. This represents roughly five times the compute Anthropic used for prior Claude generations, all operating on non-NVIDIA silicon. This is frontier-scale AI training in production, without CUDA as the underlying standard. That marks a meaningful shift.
Amazon’s roadmap further reinforces this direction. Trainium3 delivers approximately four times the performance of its predecessor with a 40% improvement in energy efficiency. While NVIDIA retains an absolute performance lead, the gap is narrowing, and Amazon’s cost advantage persists.
More strategically, Trainium4 is expected to support NVIDIA’s NVLink Fusion interconnect. This signals recognition of CUDA’s entrenched position while enabling hybrid clusters that combine Amazon and NVIDIA silicon. For enterprises, this creates a credible path toward diversification without requiring a full departure from existing NVIDIA infrastructure.
By signing a deal with Amazon, OpenAI will have successfully diversified its chip procurement structure. However, it is not just OpenAI that stands to gain from the deal. Amazon’s investment in OpenAI means more than a customer for its chips.
What Amazon really gains
For Amazon, the value proposition is credibility rather than revenue. Despite being the world’s largest cloud provider, AWS has struggled to position itself as first-tier in generative AI.
Microsoft locked in OpenAI early with its investment and exclusive Azure partnership. Google built Gemini entirely on its own TPU infrastructure, demonstrating that custom silicon could power frontier models from training through deployment. Amazon, meanwhile, spent years pitching Trainium to a market unconvinced it could handle serious AI workloads.
Landing OpenAI changes that narrative overnight. As Anshel Sag, principal analyst at Moor Insights & Strategy, observes: “ChatGPT is still seen as the Kleenex of AI. If OpenAI uses your hardware at any scale, that’s a huge validation.” When the company behind the most recognized AI product in history publicly treats Amazon chips as production-grade infrastructure, it signals to every enterprise customer that Trainium is safe, viable, and future-proof.
However, the story does not end here. Some analysts caution that Amazon may be paying for the wrong kind of exposure. If OpenAI’s messaging treats Trainium as good enough for some workloads rather than optimal, it validates AWS silicon as second-tier infrastructure rather than a genuine alternative.
Amazon needs OpenAI to treat Trainium the way Anthropic does: as the primary training and inference platform, not a cost-cutting fallback for non-critical tasks.
What the deal means for single-vendor AI
While the deal between Amazon and OpenAI is bound to attract attention, a step back from its specifics reveals a broader pattern. The single-vendor AI infrastructure era has definitively ended.
Google trains Gemini 3 exclusively on TPUs, delivering 67% better performance per watt than H100 GPUs. Meta reportedly secured multibillion-dollar commitments for TPU starting in mid-2026, despite being NVIDIA’s largest customer. Microsoft deploys Maia 100 chips internally while continuing to deploy massive NVIDIA GPU clusters for Azure customers. These numbers indicate that even NVIDIA’s closest partners are hedging.
And this fragmentation reflects economic reality more than technical preference. Custom silicon delivers 15-25% lower production inference costs, and inference represents two-thirds of AI compute spending moving forward. The training market remains NVIDIA-dominated due to advantages in the software ecosystem and raw performance requirements, but the inference market is fragmenting rapidly, with ASICs optimized for specific workloads.
For enterprise infrastructure teams, the OpenAI-Amazon deal signals a strategic shift: multi-vendor AI infrastructure is no longer just acceptable but necessary.
NVIDIA supply constraints mean relying exclusively on H100s or GB200s creates deployment risk. Rising energy costs make efficiency paramount, favoring purpose-built ASICs over general-purpose GPUs for production workloads. The question shifts from ‘can we trust non-NVIDIA silicon?’ to ‘can we afford not to diversify?’
The era when NVIDIA’s CUDA moat guaranteed single-vendor dominance is ending. Not because CUDA is less important, but because economic pressures and supply scarcity compel the industry to build around it. The companies that recognize this early gain a competitive advantage. Those who wait until supply constraints force their hand may find themselves scrambling for capacity in a market where the winners have already locked in long-term commitments.
Interdependence replaces dominance
The modern AI industry is beginning to resemble the semiconductor supply chain it once admired from a distance. Just as chipmaking fragmented across geographies and specializations to unlock scale, AI is fragmenting across models, chips, clouds, and capital providers. What once appeared to be vertical integration is giving way to enforced interdependence.
No single company can design the model, supply the silicon, fund the infrastructure, and absorb the financial risk alone. The OpenAI Amazon deal is not an anomaly but a symptom of this maturation, where scarcity, cost pressure, and capital intensity reshape alliances that once appeared purely strategic.
What ultimately emerges from this shift is a more fragile but more realistic AI ecosystem. Growth is no longer constrained by ideas or ambition but by physical supply chains, power availability, and balance sheets.
The next phase of AI will be defined less by architectural breakthroughs and more by who can secure compute, finance it sustainably, and deploy it efficiently at scale. In that environment, dominance will not rest with a single model builder or chipmaker, but with those who recognize that modern AI, like modern semiconductors, advances only when the entire system moves forward together.


Quick Bits, No Fluff
Blurry Space Drama: Reddit mods briefly deleted NASA astronaut Don Pettit’s stunning ISS photo of a jet as ‘blurry,’ then restored it after users mocked the pedantic rule-lawyering.
Positron vs Nvidia: Chip startup Positron raised a $230M Series B to scale its Atlas memory chips, claiming H100-level inference performance at under one-third the power, as hyperscalers hunt for non-Nvidia options.
Claude Skips Super Bowl: Anthropic is running a meta Super Bowl campaign touting Claude as the chatbot that doesn’t show ads, positioning quiet, ad-free AI against flashier, ad-driven rivals.

Outperform the competition.
Business is hard. And sometimes you don’t really have the necessary tools to be great in your job. Well, Open Source CEO is here to change that.
Tools & resources, ranging from playbooks, databases, courses, and more.
Deep dives on famous visionary leaders.
Interviews with entrepreneurs and playbook breakdowns.
Are you ready to see what’s all about?
*This is sponsored content

Thursday Poll
🗳️ How would you hedge your AI infra bet today? |

3 Things Worth Trying
AWS Trainium: Spin up a small Trainium instance on AWS and port one inference or fine-tuning workflow from GPUs.
CoreWeave Free Tier: Test a CoreWeave GPU cluster with a real training or heavy inference job (not just a toy notebook).
ProsperOps: Plug the cloud cost platform into your stack and build a by-provider, by-model-family, and by-workload.
Rate This Edition
What did you think of today's email? |




