OpenAI just locked in $38B in compute. Here's what that means for your unit economics.
OpenAI just locked in $38B in compute. Here's what that means for your unit economics.
Issue #6 ·  · 5 min read
← All editions

OpenAI's $38B AWS Deal Tells You Everything About AI Unit Economics

Ground Model — March 7, 2026


Lead Story

OpenAI and AWS announced a multi-year strategic partnership worth $38 billion, giving OpenAI access to hundreds of thousands of NVIDIA GPUs (GB200s and GB300s) via Amazon EC2 UltraServers, with the ability to scale to tens of millions of CPUs. All capacity is targeted for deployment before the end of 2026, with expansion into 2027 and beyond.

This deal has already been expanded—OpenAI and Amazon are growing the existing $38B agreement by an additional $100 billion over 8 years, including approximately 2 gigawatts of Trainium capacity through AWS infrastructure.

The Builder's Math

Let's be blunt about what this means for anyone building on OpenAI's APIs or competing with products built on them.

$138 billion in compute commitments is not a number that gets amortized quietly. That capital has to flow somewhere, and it flows into the price of every token you consume. OpenAI is making an enormous bet that demand will scale faster than cost—that inference volume will grow exponentially while per-token costs decline. If they're right, prices drop and your margins improve. If they're wrong, you're subsidizing their infrastructure buildout through your API bill.

Here's what most builders miss: this deal is a moat play disguised as an infrastructure deal. By locking in this much capacity on AWS, OpenAI is signaling that they expect to be the inference layer for a massive chunk of enterprise software. The clustering of GPUs via EC2 UltraServers—optimized for low-latency interconnected systems—is specifically designed to make ChatGPT inference and next-gen model training faster and cheaper at scale. That scale advantage compounds. Smaller model providers can't negotiate $138B infrastructure deals. They pay retail.

What This Actually Means for Your Product

1. Token prices will likely decline, but not as fast as you think. OpenAI has to recoup $138B+ in compute costs. They'll pass savings along selectively—probably through volume discounts and enterprise tiers, not across-the-board API price cuts. If you're a startup doing $10K/month in API spend, you're not the customer this deal is optimized for.

2. The inference moat is real and growing. We've said it before: cloud providers are weaponizing capital to embed AI deeply into their platforms. This deal confirms it. AWS gets OpenAI workloads locked into their cloud. OpenAI gets infrastructure nobody else can match. The independent model provider running on a patchwork of GPU clouds just became even less competitive.

3. Your defensibility can't be at the model layer. If OpenAI is spending $138B on infrastructure alone, you cannot compete on model quality. Full stop. Your defensibility has to come from proprietary data, workflow integration, or distribution. The model is the commodity. The glue code is the product—and we said exactly this last edition.

4. Plan for multi-cloud AI. OpenAI running on AWS doesn't mean they'll leave Azure. It means they're hedging. You should too. Build abstraction layers that let you swap model providers. Your unit economics depend on it.

The Uncomfortable Question

If OpenAI needs $138 billion in infrastructure to serve its workloads profitably, what does that tell you about the actual margin structure of AI-native products? Every AI product built on third-party inference is effectively a gross-margin-constrained business until inference costs hit near-zero. We're not there yet. Build accordingly.


Quick Hits

  • OpenAI is retiring GPT-4o, GPT-4.1, GPT-4.1 mini, and o4-mini from ChatGPT. Model deprecation is the hidden tax on AI products. If you've fine-tuned or prompt-engineered around a specific model's quirks, retirement means rework. Budget 10-15% of eng time for model migration. OpenAI

  • Reasoning models can't fully control their chains of thought—and OpenAI says that's a feature, not a bug. If you're building products that expose chain-of-thought to users, this is a transparency win. But it also means you can't fully steer reasoning for cost optimization. Longer chains = more tokens = higher costs. OpenAI

  • The Pentagon officially labeled Anthropic a supply-chain risk after the company refused to allow military use of its AI for mass surveillance or autonomous weapons. If you're selling to defense or federal customers, your model provider choice is now a compliance decision, not just a technical one. TechCrunch

  • BNY Mellon is building "AI for everyone, everywhere" with OpenAI. Another data point in OpenAI's enterprise embedding strategy. Financial services data flowing through OpenAI systems creates deep lock-in. OpenAI

  • Plex Coffee uses ChatGPT Business for customer service. The interesting signal is OpenAI marketing to SMBs—millions of small businesses at $30-100/month is a better margin structure than heavy enterprise customization. OpenAI


Company Watch: Anthropic's Costly Principles

The Pentagon's designation of Anthropic as a supply-chain risk is unprecedented for an AI company. The formal declaration requires defense vendors and contractors to certify that they don't use Anthropic products, effectively cutting the company off from the entire U.S. government contracting ecosystem.

The U.S. defense and intelligence community is arguably the highest-value AI customer segment on Earth—unlimited budget, mission-critical use cases, long contract cycles, minimal price sensitivity. Anthropic just lost access to all of it.

The market bifurcation is now explicit:

  • OpenAI has been systematically embedding into military defense systems, national security operations, and federal infrastructure. They're optimizing for total addressable market.
  • Anthropic drew a line on ethical use. That line now has a dollar sign attached to it.

For builders: your model provider's policy positions are now part of your go-to-market strategy. Choose accordingly.


Tool of the Day: Build an Inference Cost Dashboard

With OpenAI retiring older models and reasoning models consuming unpredictable token volumes, you need real-time visibility into your inference costs per feature, per user, per workflow.

What to track:
- Tokens consumed per API call (input + output, separately)
- Cost per user action (not per API call—per actual product interaction)
- Gross margin per customer tier
- Model version → cost mapping
- Reasoning chain length distribution

Use your cloud provider's cost allocation tags, pipe API usage logs into your data warehouse, and build alerts for per-user cost spikes. Table stakes for 2026.


Stat of the Day

$138 billion — The combined infrastructure commitment between OpenAI and AWS ($38B initial + $100B expansion over 8 years). That's what it actually costs to run frontier AI at scale. Source: OpenAI and Amazon partnership announcements.

Get Ground Model in your inbox

Daily AI briefings for builders. No hype. Just signal. Free, Mon-Fri.

Subscribe free