← Research
Ralph Capital Research March 2026

AI Infrastructure:
The $690 Billion Sprint

The five largest U.S. hyperscalers will spend $660-690 billion on capex in 2026 — nearly double the $443 billion deployed in 2025. Roughly 75% is directed at AI. This report maps the capital flows, bottlenecks, and investment implications of the largest infrastructure buildout since the transcontinental railroad.

01 Scale of the Buildout

AI infrastructure spending has entered a phase with no historical parallel. The combined 2026 capex of Amazon, Microsoft, Google, Meta, and Oracle — approximately $690 billion — exceeds the GDP of Switzerland. It represents a 56% increase over 2025 and a 4x increase from 2022 levels.

This is not speculative R&D spending. It is industrial-scale deployment of compute, power, cooling, and networking infrastructure to serve inference workloads that are already generating revenue. The shift from model training (concentrated, episodic) to inference serving (distributed, continuous) has fundamentally changed the economics — and the capital intensity is permanent.

$690B
2026 hyperscaler capex
Top 5 U.S. hyperscalers combined. Nearly 2x the $443B spent in 2025.
75%
Allocated to AI
~$450B directed at AI compute, networking, and supporting infrastructure.
45-57%
Capex/revenue ratio
Historically unprecedented capital intensity. Traditional tech ran at 10-15%.
Hyperscaler capex acceleration ($B)
Combined spend of Amazon, Microsoft, Google, Meta, Oracle
2022
2023
2024
2025
2026

02 Who Is Spending What

The spending is concentrated but the strategies diverge. Each hyperscaler is making distinct bets on compute architecture, power sourcing, and market positioning.

Company2025 capex2026 capexYoY growthPrimary focus
Amazon / AWS$125B$200B+60%AWS data centers, custom Trainium chips, global edge
Alphabet / Google$91B$175-185B+95%TPU v6, Gemini inference, 60% servers / 40% DC+network
Meta$72B$115-135B+75%Superintelligence Labs, Llama training clusters
Microsoft$80B+$120B++50%Azure AI, OpenAI partnership, $80B Azure backlog
Oracle~$30B~$50B+67%OCI Gen2, enterprise AI infrastructure

A critical detail: Microsoft has disclosed an $80 billion Azure backlog that it cannot fulfill due to power constraints. GPUs are sitting in warehouses waiting for data centers to be energized. This is not a demand problem — it is a supply problem, and it defines the investment opportunity.

Financing the buildout

Hyperscalers raised $108 billion in debt during 2025 alone, with projections suggesting $1.5 trillion in cumulative debt issuance will be needed over coming years. Capital intensity has reached 45-57% of revenue — historically unprecedented for technology companies that traditionally ran at 10-15%. This creates a multi-decade infrastructure financing opportunity similar to telecom in the 1990s, but at 5-10x the scale.

03 The Power Bottleneck

Power has replaced GPUs as the binding constraint on AI infrastructure deployment. The grid cannot keep up with demand, and the gap is widening.

1,000+
TWh by 2026
Global data center electricity consumption. More than Japan's entire national consumption.
10 GW
NVIDIA AI capacity
AI data center capacity deployed through NVIDIA partnerships in 2025 alone.
128 wk
Transformer lead time
Power transformer procurement lead times. Over 2.5 years to get the basic grid equipment.

The numbers tell a stark story. AI racks require 30-60 kW each — 10-20x the power density of traditional server racks. A single large AI training cluster consumes 100+ MW, equivalent to powering a small city. The IEA projects data center electricity consumption will surge from 460 TWh in 2024 to over 1,000 TWh by 2026 and 1,300 TWh by 2035.

Grid connection crisis

  • Virginia (Loudoun County): The world's largest data center market. Power connection waiting lists stretch to 2028-2029. New major connections are effectively closed.
  • Texas: Some developers are required to build substations costing $100M+ just to connect to the grid.
  • Ireland: EirGrid has imposed a moratorium on new data center grid connections in Dublin until 2028.
  • Power transformer lead times: 128 weeks and growing. The bottleneck isn't just generation — it's the physical grid infrastructure to deliver power.
Investment implication: Power access is now the primary competitive moat in data center development. Operators with contracted power, grid priority, or behind-the-meter generation have structural advantages that are nearly impossible to replicate on short timescales.

04 GPU Economics & Supply

NVIDIA dominates the AI accelerator market with approximately 80-90% market share in data center GPUs. The H100/H200 and B100/B200 series are the de facto standard for both training and inference workloads.

GPUList priceMarket priceMemoryUse case
NVIDIA H100 SXM$25,000$30-40K80GB HBM3Training + inference (current gen)
NVIDIA H200$30,000$35-45K141GB HBM3eLarge model inference, memory-bound tasks
NVIDIA B200$30-37K$40-50K192GB HBM3eNext-gen training + inference
NVIDIA GB200 NVL72~$2-3M/rackLimited avail.72×192GBFull-rack training supercompute
AMD MI300X$10-15K$12-18K192GB HBM3Inference alternative, price-competitive
Google TPU v5e/v6InternalCloud onlyVariesGoogle Cloud workloads, Gemini training
AWS Trainium2InternalCloud onlyVariesAWS-native training, cost optimization

The inference economics shift

The market is transitioning from training-dominated (large, concentrated clusters, episodic usage) to inference-dominated (distributed, continuous, cost-sensitive). This changes the hardware calculus:

  • Training: Performance is paramount. Customers pay premium for the fastest chips. NVIDIA's moat is widest here.
  • Inference: Cost-per-token matters most. AMD, custom ASICs (Google TPU, AWS Trainium), and even older GPUs become competitive. This opens the market.

By 2027, inference is projected to represent 70-80% of AI compute spending, up from ~50% today. This structural shift creates opportunities for alternative hardware providers and for operators who can optimize inference workloads on mixed hardware fleets.

05 Data Center Market Structure

The global data center market has bifurcated into two distinct segments with different economics and competitive dynamics.

Hyperscale (self-built)

Amazon, Google, Meta, and Microsoft are increasingly building their own data centers rather than leasing from third parties. This vertical integration is driven by the need for custom power, cooling, and networking architectures optimized for AI workloads. Hyperscale self-build now represents the majority of new capacity additions.

Colocation / third-party

Companies like Equinix, Digital Realty, CyrusOne, and QTS serve enterprises that lack the scale or expertise to self-build. This segment is consolidating rapidly, with three major dynamics:

  • Power scarcity premium: Colo operators with secured power capacity are seeing record-high lease rates — up 20-40% in power-constrained markets.
  • AI-ready requirements: Customers increasingly demand liquid cooling, high-density racks (30+ kW), and 100G+ networking. Retrofitting existing facilities is expensive.
  • Long-term contracts: Average lease terms have extended from 3-5 years to 7-10+ years as enterprises lock in capacity. This improves cash flow visibility but reduces flexibility.
Global data center capacity growth (GW of IT load)
Combined hyperscale + colocation, AI is driving non-linear expansion
2022
2024
2026
2028
2030

06 The Nuclear Pivot

With grid connections constrained and renewable intermittency limiting AI workloads that require 24/7 baseload power, hyperscalers are making aggressive moves into nuclear energy — both existing plants and new technologies.

CompanyNuclear dealCapacityTimelineStructure
MicrosoftThree Mile Island Unit 1 restart835 MW2028$1.6B 20-year PPA with Constellation
AmazonSusquehanna nuclear campus2.5 GW siteOperatingDirect data center connection to existing plant
GoogleKairos Power SMRs75-300 MW each2030+Multiple small modular reactors, behind-the-meter
OracleSMR-powered data centers1 GW+ planned2030+Co-located generation + compute

The logic is straightforward: a 1 GW nuclear plant provides 8,760 GWh/year of baseload power at a marginal cost of $20-30/MWh — enough to power 10-15 large AI data center campuses indefinitely. Compare this to grid power at $50-120/MWh with zero availability guarantees. Nuclear is the only technology that can provide gigawatt-scale, 24/7, carbon-free power for AI.

Investment angle: Nuclear-adjacent companies (uranium miners, fuel fabricators, reactor services, power plant developers) are one of the most asymmetric AI trades available. The demand is locked in by hyperscaler commitments, but the supply chain is still priced at pre-AI levels.

07 Where Value Accrues

Not all AI infrastructure spending creates equal value. The $690B is distributed across a supply chain with very different margin profiles and competitive dynamics.

LayerShare of spendGross marginCompetitive moatKey players
Silicon / GPUs~30%70-75%Extremely highNVIDIA, AMD, Google, AWS
Networking~10%55-65%HighArista, Broadcom, Cisco
Memory (HBM)~12%50-60%High (3 suppliers)SK Hynix, Samsung, Micron
Power infrastructure~15%25-40%Medium-HighEaton, Schneider, Vertiv
Cooling~8%30-45%MediumVertiv, CoolIT, Iceotope
Construction / EPC~15%8-15%Low-MediumQuanta, MasTec, Jacobs
Land + permitting~5%VariesHigh (scarcity)REITs, utilities, landowners
Energy generation~5%30-50%Very highUtilities, nuclear, IPPs

The value chain shows a clear pattern: the higher the margin, the harder the moat. NVIDIA's 75% gross margin reflects a near-monopoly in AI training silicon. Power infrastructure margins are lower but access is increasingly scarce. Construction is commoditized but volume is enormous.

08 Revenue & Return Framework

The AI infrastructure buildout creates investable opportunities across multiple time horizons and risk profiles. Below is a framework for evaluating returns across the value chain.

Investment thesisEntry point5yr revenue CAGRMargin profileRisk
GPU / silicon suppliersPublic equities25-35%Gross 65-75%Concentration, custom chip threat
Power infrastructurePublic + PE15-25%Gross 30-45%Execution, permitting
Data center REITsPublic equities12-18%NOI 50-60%Interest rates, power costs
Nuclear / uraniumPublic equities20-30%Gross 40-55%Regulatory, timeline
Cooling technologyPE / venture30-50%Gross 40-55%Technology risk, competition
AI cloud servicesPublic equities30-40%Gross 55-70%Capex intensity, execution
Fiber / connectivityPE / infra funds10-15%EBITDA 45-55%Commoditization
AI cloud services revenue trajectory ($B)
Combined AWS, Azure, GCP AI-specific revenue estimates
2024
2025
2026
2028
2030

09 Risk Analysis

RiskDescriptionSeverityMitigation
OverbuildingCapex outpaces revenue growth; utilization falls below 60%MediumLong-term contracts, diverse workload mix, cloud marketplace
Power grid failureGrid constraints delay deployments by 12-24 monthsHighBehind-the-meter generation, nuclear PPAs, distributed deployment
GPU commoditizationCustom chips (TPU, Trainium) erode NVIDIA's pricing powerMediumCUDA ecosystem lock-in, continued performance leadership
Regulatory backlashEnergy consumption, water usage, or export controls constrain growthMediumRenewable PPAs, cooling innovation, geographic diversification
Debt sustainability$1.5T projected debt issuance in a rising-rate environmentMedium-HighStrong cash flows, investment-grade ratings, long maturities
AI demand plateauEnterprise AI adoption slower than projectedLow-MediumConsumer inference already generating revenue at scale (ChatGPT, Gemini)

10 Outlook & Positioning

The AI infrastructure buildout is a multi-decade trend, not a cycle. Even if AI model improvement slows, the existing trajectory of inference workload growth, enterprise migration, and applications deployment will sustain infrastructure demand well into the 2030s.

12-month view

Most conviction
Power infrastructure

128-week lead times, grid constraints, and nuclear deals create a multi-year demand runway with limited new supply. Picks-and-shovels with pricing power.

High conviction
Networking (Arista, Broadcom)

AI clusters require 100G-800G networking at scale. Each new GPU deployed requires proportional network investment. Moats are deep.

Selective
GPU / silicon

NVIDIA remains dominant but priced for perfection. Custom chip threat from Google/Amazon creates long-term margin risk. Position on pullbacks.

The fundamental insight: $690 billion in annual spending creates a supply chain where the constraining resources — power, cooling, land with grid access — appreciate faster than the compute itself. The investment opportunity is not in buying GPUs; it is in owning the infrastructure that GPUs depend on to operate.

Reference Sources

Futurum Group, "AI Capex 2026: The $690B Infrastructure Sprint."

CNBC, "Tech AI spending approaches $700 billion in 2026," Feb 2026.

Introl, "Hyperscaler CapEx Hits $690B in 2026."

IEA, "Energy supply for AI — Energy and AI," 2025-2026.

Enki AI, "NVIDIA's AI Data Center Energy Demand: A 10 Gigawatt Challenge."

Enki AI, "AI Power Demand 2026: A Trillion-Dollar Power Crisis."

Deloitte, "Data Center Sustainability Insights," 2025.

Company filings and earnings calls: Amazon, Microsoft, Alphabet, Meta, Oracle (Q4 2025, Q1 2026).

This report is produced by Ralph Capital for informational purposes only. It does not constitute investment advice, an offer to buy or sell any security, or a solicitation. Data and analysis reflect publicly available sources as of March 2026; accuracy is not guaranteed. Scenario projections are analytical estimates, not forecasts. Ralph Capital may hold positions in assets discussed in this report.