Skip to main content
MindStudio
Pricing
Blog About
My Workspace

AWS Free Cash Flow Collapsed from $26B to $1.2B in One Year — Here's Where Every Dollar Is Going

Amazon's free cash flow dropped from $26B to $1.2B in a single year. It's not a problem — it's a signal. Here's the full picture of AWS's AI buildout bet.

MindStudio Team RSS
AWS Free Cash Flow Collapsed from $26B to $1.2B in One Year — Here's Where Every Dollar Is Going

Amazon Just Spent Its Way to $1.2B in Free Cash Flow — Down From $26B

AWS free cash flow collapsed from $26 billion to $1.2 billion in a single year. Not a typo. Not an accounting anomaly. Amazon’s dominant cloud unit generated roughly 95% less free cash in Q1 2026 than it did in Q1 2025, and CEO Andy Jassy’s response on the earnings call was essentially: good, keep going.

If you build on AWS, deploy on AWS, or compete with anyone who does, this number tells you something important about the next two to three years of the infrastructure landscape. The question isn’t whether Amazon is in trouble. It’s what they’re betting on, and whether that bet changes the environment you’re building in.


The Numbers Behind the Collapse

Amazon’s Q1 2026 CapEx came in at $43.2 billion — a 60% jump from the same quarter last year, and on pace to hit their stated $200 billion annual target. That’s the largest absolute CapEx number among the four major hyperscalers this quarter.

The free cash flow math is simple: when you spend $43 billion in a quarter and your business generates roughly that much, you have no cash left over. Amazon is not in distress. They are choosing to convert every dollar of operating profit into physical infrastructure as fast as the supply chain allows.

Hire a contractor. Not another power tool.

Cursor, Bolt, Lovable, v0 are tools. You still run the project.
With Remy, the project runs itself.

AWS revenue itself was up 28% year-over-year — its fastest growth rate since climbing out of a trough in 2021. The business is accelerating. The cash is just going straight back into the ground in the form of data centers, networking, and chips.

Jassy was explicit about why. “We have such demand right now for Trainium from various companies who will consume as much as we make,” he said on the call. He added that most of the new capacity being built is already spoken for before it comes online. This is not speculative construction. It’s backfill.


What Amazon Is Actually Building

The CapEx isn’t going into generic compute. A meaningful portion is going into Amazon’s in-house silicon program — specifically Trainium, their custom AI training chip.

Jassy made a claim on the earnings call that deserves more attention than it got: if Amazon’s custom silicon business were a standalone company booking its own revenue rather than subsidizing AWS pricing, it would be sitting at $50 billion ARR. He called it “one of the top three data center chip businesses in the world” and noted the speed of getting there as extraordinary.

That’s a significant statement. Nvidia has dominated AI chip revenue so thoroughly that most coverage treats the chip market as a one-player game. Jassy is asserting that Amazon has quietly built the second or third largest chip business on the planet, measured by the compute it deploys, even if that compute doesn’t show up as chip revenue on their income statement.

The strategic logic is straightforward. If you’re going to spend $200 billion a year on infrastructure, and a large fraction of that is chips, you have enormous incentive to make those chips yourself. Every dollar of margin you’d otherwise pay to a chip vendor is a dollar you can reinvest in capacity or pass through to customers as lower prices.


Why This Matters If You’re Building on Cloud

You might be wondering what a hyperscaler’s free cash flow has to do with your architecture decisions. The connection is more direct than it looks.

First, the supply constraint is real and it affects you now. OpenAI CFO Sarah Fryer described the current situation as “a vertical wall of demand with compute being the bottleneck.” The Wall Street Journal reported that demand for AI tools is causing outages and surging GPU rental prices — up 40% over the last six months. When Amazon says they’re building as fast as they can, they’re also implicitly saying that if you need capacity today, you may not get as much as you want at the price you expect.

Cursor
ChatGPT
Figma
Linear
GitHub
Vercel
Supabase
remy.msagent.ai

Seven tools to build an app. Or just Remy.

Editor, preview, AI agents, deploy — all in one tab. Nothing to install.

Second, the OpenAI-AWS partnership changes the competitive dynamics on Bedrock. Amazon’s cloud unit now hosts both Anthropic’s Claude and OpenAI’s models. As one analyst noted, many companies defaulted to Anthropic and Claude simply because they were already on Bedrock and it was the path of least resistance. That same gravitational pull now works for OpenAI models too. If you’re building multi-model workflows — routing different tasks to different models based on cost and capability — the fact that both frontier model families now live in the same cloud matters for your architecture. MindStudio handles this kind of orchestration across 200+ models and 1,000+ integrations, which becomes more relevant as the “which model for which task” question gets more complex; having a single visual builder that spans the full model landscape matters more when your cloud provider is consolidating that landscape under one roof.

Third, the free cash flow collapse signals something about pricing stability. Amazon is not in a position to compete on price right now. They’re spending everything on capacity. When capacity catches up to demand — and eventually it will — you’ll likely see more aggressive pricing. Building cost-sensitive workflows today that can take advantage of lower prices later is a reasonable hedge. If you’re already thinking about cost reduction at the model layer, the approaches covered in how to use Open Router free models with Claude Code to cut AI costs by 99% apply equally well to workflows running on AWS Bedrock.


The Non-Obvious Detail: Everyone Is Underestimating Demand

The most interesting thing buried in these earnings reports isn’t Amazon’s number. It’s the consistency of the surprise across every major player.

Meta CFO Susan Li said on their earnings call: “Our experience so far has been that we have underestimated our compute needs, even as we have been ramping capacity significantly.” Meta raised their CapEx forecast from $135 billion to $145 billion for the year, and the increase wasn’t because they added new projects — it was because existing projects cost more than expected and they need more capacity than they planned for.

Microsoft raised CapEx guidance by $25 billion to $190 billion for the year, attributing the entire increase to higher component prices. Azure grew 40% year-over-year, right in line with expectations, but the infrastructure cost to support that growth keeps going up.

Google Cloud grew 63% year-over-year and reported a $460 billion backlog in new orders — up from $240 billion at the end of Q4 2025. Analyst Joseph Carlson posted the backlog chart and wrote that it “literally looks fake” because the curve is so steep. Google’s CEO noted that cloud revenue would have been even higher if they could have met demand. They were compute-constrained. For a deeper look at how Google is structuring the pricing side of that demand, the breakdown of Google Flow pricing, credits, and tiers shows how hyperscalers are beginning to tier access as supply remains tight.

The pattern across all four companies is the same: they thought they knew how much compute they needed, they built for that, and they were wrong. Not by a little. By enough that they’re each spending tens of billions more than they planned.

This is the actual signal. It’s not that one company made a bold bet. It’s that every company with visibility into real enterprise AI workloads is discovering the same thing: the demand for tokens is growing faster than anyone’s models predicted.


The Chip Business Nobody Is Talking About

Amazon’s Trainium claim deserves a standalone moment because it reframes how you should think about the infrastructure stack.

The conventional narrative is: Nvidia makes chips, hyperscalers buy chips, hyperscalers rent compute to you. In that model, Nvidia captures the margin and the hyperscalers compete on price and services.

Remy is new. The platform isn't.

Remy
Product Manager Agent
THE PLATFORM
200+ models 1,000+ integrations Managed DB Auth Payments Deploy
BUILT BY MINDSTUDIO
Shipping agent infrastructure since 2021

Remy is the latest expression of years of platform work. Not a hastily wrapped LLM.

What Amazon is building is a different model: make your own chips, deploy them internally, price your compute against the market rate for Nvidia-based compute, and capture the chip margin yourself. If Trainium is genuinely competitive with Nvidia’s H100 and H200 for training workloads — and Jassy’s $50B ARR claim suggests it is, at least for the workloads Amazon runs — then Amazon has structurally lower costs than any cloud provider that buys all its chips externally.

Jassy said he expects to start selling Trainium racks externally over the coming years. That would be a significant shift: Amazon not just using custom silicon internally, but competing with Nvidia in the merchant chip market. The $200 billion annual CapEx target starts to look different if a meaningful fraction of it is building chip manufacturing capacity rather than just buying chips from someone else.

For builders, this matters because it affects which cloud has the most room to lower prices as the current supply crunch eases. The provider with the lowest underlying cost structure has the most flexibility. Right now, that appears to be Amazon.


The Search Revenue Subplot

One more data point worth pulling out, because it complicates the standard narrative about AI destroying incumbent businesses.

Google search revenue grew 19% year-over-year in Q1 2026. Queries hit an all-time high. The prevailing prediction for the last two years was that AI chatbots would cannibalize Google search — people would get answers from Claude or ChatGPT and stop Googling. The opposite is happening.

This matters for how you think about building AI-powered products. The assumption that AI is a zero-sum replacement for existing tools has been wrong repeatedly. Google turned what looked like an existential threat into a growth accelerator. The same dynamic may apply to whatever category you’re building in.

The corollary for builders: don’t assume that adding AI to a workflow destroys the existing workflow. Often it augments it. The products that win are usually the ones that figure out how to make the existing behavior more valuable, not the ones that try to replace it entirely.


What the Infrastructure Bet Means for What You Build

The infrastructure spending at this scale has a direct implication for the tools and abstractions available to you as a builder.

When Amazon spends $200 billion on infrastructure, a large fraction of that goes into the harness layer — the agent runtimes, sandboxing, orchestration, and tooling that sit between raw compute and the applications you build. AWS Bedrock, managed agents, and the broader agent infrastructure stack all benefit from that investment. The same is true at Google and Microsoft.

This is why the harness engineering conversation has accelerated so quickly. The underlying models are good enough that the bottleneck has shifted to the environment in which they operate. If you’re building production agents today, the quality of your harness — persistent memory, tool dispatch, error handling, context management — matters as much as which model you pick. For teams building full-stack applications on top of these agents, Remy takes a different approach: you write a spec in annotated markdown, and it compiles into a complete TypeScript backend, database, auth, and deployment. The spec is the source of truth; the infrastructure is derived output. That abstraction makes sense when the underlying infrastructure is as reliable and commodity-like as Amazon is trying to make it.

Plans first. Then code.

PROJECTYOUR APP
SCREENS12
DB TABLES6
BUILT BYREMY
1280 px · TYP.
yourapp.msagent.ai
A · UI · FRONT END

Remy writes the spec, manages the build, and ships the app.

The $1.2 billion free cash flow number is Amazon’s down payment on that future. They’re betting that if they build enough capacity, fast enough, the demand will be there to fill it. Every signal from their own customers, from Google’s backlog, from Meta’s compute shortfall, and from the secondary market valuation of Anthropic — reportedly trading at $1 trillion in some transactions, above OpenAI’s last official $825 billion round — suggests that bet is well-founded.


What to Watch

Three things worth tracking over the next two quarters:

Whether the capacity catches up. Amazon said most new capacity is already spoken for. If that’s true, the supply crunch continues through at least 2026. If demand softens or new capacity comes online faster than expected, you’ll see it in GPU rental prices first.

Trainium’s external availability. Jassy’s comment about selling racks externally was a signal, not an announcement. If Amazon starts offering Trainium-based compute as a distinct product line, it changes the competitive dynamics for every AI workload that doesn’t require Nvidia-specific software.

Free cash flow recovery. Amazon’s investors are tolerating the current cash burn because Jassy has credibility from the original AWS buildout. But the market will eventually want to see cash flow recover. Watch for the quarter where CapEx growth starts to decelerate — that’s when the infrastructure bet starts paying out.

The $26 billion to $1.2 billion drop is not a warning sign. It’s a commitment. Understanding what Amazon is committing to — and why every other hyperscaler is making the same bet simultaneously — is the context you need to make good decisions about what to build and where to build it.

The open-source agent era of building by hand, wiring every component yourself, is giving way to infrastructure that’s been pre-built and pre-scaled by companies spending $200 billion a year to make it reliable. That changes what’s possible for builders who aren’t Amazon. The question is whether you’re positioned to take advantage of it.

Presented by MindStudio

No spam. Unsubscribe anytime.