01 Scale of the Buildout
AI infrastructure spending has entered a phase with no historical parallel. The combined 2026 capex of Amazon, Microsoft, Google, Meta, and Oracle — approximately $690 billion — exceeds the GDP of Switzerland. It represents a 56% increase over 2025 and a 4x increase from 2022 levels.
This is not speculative R&D spending. It is industrial-scale deployment of compute, power, cooling, and networking infrastructure to serve inference workloads that are already generating revenue. The shift from model training (concentrated, episodic) to inference serving (distributed, continuous) has fundamentally changed the economics — and the capital intensity is permanent.
02 Who Is Spending What
The spending is concentrated but the strategies diverge. Each hyperscaler is making distinct bets on compute architecture, power sourcing, and market positioning.
| Company | 2025 capex | 2026 capex | YoY growth | Primary focus |
|---|---|---|---|---|
| Amazon / AWS | $125B | $200B | +60% | AWS data centers, custom Trainium chips, global edge |
| Alphabet / Google | $91B | $175-185B | +95% | TPU v6, Gemini inference, 60% servers / 40% DC+network |
| Meta | $72B | $115-135B | +75% | Superintelligence Labs, Llama training clusters |
| Microsoft | $80B+ | $120B+ | +50% | Azure AI, OpenAI partnership, $80B Azure backlog |
| Oracle | ~$30B | ~$50B | +67% | OCI Gen2, enterprise AI infrastructure |
A critical detail: Microsoft has disclosed an $80 billion Azure backlog that it cannot fulfill due to power constraints. GPUs are sitting in warehouses waiting for data centers to be energized. This is not a demand problem — it is a supply problem, and it defines the investment opportunity.
Financing the buildout
Hyperscalers raised $108 billion in debt during 2025 alone, with projections suggesting $1.5 trillion in cumulative debt issuance will be needed over coming years. Capital intensity has reached 45-57% of revenue — historically unprecedented for technology companies that traditionally ran at 10-15%. This creates a multi-decade infrastructure financing opportunity similar to telecom in the 1990s, but at 5-10x the scale.
03 The Power Bottleneck
Power has replaced GPUs as the binding constraint on AI infrastructure deployment. The grid cannot keep up with demand, and the gap is widening.
The numbers tell a stark story. AI racks require 30-60 kW each — 10-20x the power density of traditional server racks. A single large AI training cluster consumes 100+ MW, equivalent to powering a small city. The IEA projects data center electricity consumption will surge from 460 TWh in 2024 to over 1,000 TWh by 2026 and 1,300 TWh by 2035.
Grid connection crisis
- Virginia (Loudoun County): The world's largest data center market. Power connection waiting lists stretch to 2028-2029. New major connections are effectively closed.
- Texas: Some developers are required to build substations costing $100M+ just to connect to the grid.
- Ireland: EirGrid has imposed a moratorium on new data center grid connections in Dublin until 2028.
- Power transformer lead times: 128 weeks and growing. The bottleneck isn't just generation — it's the physical grid infrastructure to deliver power.
04 GPU Economics & Supply
NVIDIA dominates the AI accelerator market with approximately 80-90% market share in data center GPUs. The H100/H200 and B100/B200 series are the de facto standard for both training and inference workloads.
| GPU | List price | Market price | Memory | Use case |
|---|---|---|---|---|
| NVIDIA H100 SXM | $25,000 | $30-40K | 80GB HBM3 | Training + inference (current gen) |
| NVIDIA H200 | $30,000 | $35-45K | 141GB HBM3e | Large model inference, memory-bound tasks |
| NVIDIA B200 | $30-37K | $40-50K | 192GB HBM3e | Next-gen training + inference |
| NVIDIA GB200 NVL72 | ~$2-3M/rack | Limited avail. | 72×192GB | Full-rack training supercompute |
| AMD MI300X | $10-15K | $12-18K | 192GB HBM3 | Inference alternative, price-competitive |
| Google TPU v5e/v6 | Internal | Cloud only | Varies | Google Cloud workloads, Gemini training |
| AWS Trainium2 | Internal | Cloud only | Varies | AWS-native training, cost optimization |
The inference economics shift
The market is transitioning from training-dominated (large, concentrated clusters, episodic usage) to inference-dominated (distributed, continuous, cost-sensitive). This changes the hardware calculus:
- Training: Performance is paramount. Customers pay premium for the fastest chips. NVIDIA's moat is widest here.
- Inference: Cost-per-token matters most. AMD, custom ASICs (Google TPU, AWS Trainium), and even older GPUs become competitive. This opens the market.
By 2027, inference is projected to represent 70-80% of AI compute spending, up from ~50% today. This structural shift creates opportunities for alternative hardware providers and for operators who can optimize inference workloads on mixed hardware fleets.
05 Data Center Market Structure
The global data center market has bifurcated into two distinct segments with different economics and competitive dynamics.
Hyperscale (self-built)
Amazon, Google, Meta, and Microsoft are increasingly building their own data centers rather than leasing from third parties. This vertical integration is driven by the need for custom power, cooling, and networking architectures optimized for AI workloads. Hyperscale self-build now represents the majority of new capacity additions.
Colocation / third-party
Companies like Equinix, Digital Realty, CyrusOne, and QTS serve enterprises that lack the scale or expertise to self-build. This segment is consolidating rapidly, with three major dynamics:
- Power scarcity premium: Colo operators with secured power capacity are seeing record-high lease rates — up 20-40% in power-constrained markets.
- AI-ready requirements: Customers increasingly demand liquid cooling, high-density racks (30+ kW), and 100G+ networking. Retrofitting existing facilities is expensive.
- Long-term contracts: Average lease terms have extended from 3-5 years to 7-10+ years as enterprises lock in capacity. This improves cash flow visibility but reduces flexibility.
06 The Nuclear Pivot
With grid connections constrained and renewable intermittency limiting AI workloads that require 24/7 baseload power, hyperscalers are making aggressive moves into nuclear energy — both existing plants and new technologies.
| Company | Nuclear deal | Capacity | Timeline | Structure |
|---|---|---|---|---|
| Microsoft | Three Mile Island Unit 1 restart | 835 MW | 2028 | $1.6B 20-year PPA with Constellation |
| Amazon | Susquehanna nuclear campus | 2.5 GW site | Operating | Direct data center connection to existing plant |
| Kairos Power SMRs | 75-300 MW each | 2030+ | Multiple small modular reactors, behind-the-meter | |
| Oracle | SMR-powered data centers | 1 GW+ planned | 2030+ | Co-located generation + compute |
The logic is straightforward: a 1 GW nuclear plant provides 8,760 GWh/year of baseload power at a marginal cost of $20-30/MWh — enough to power 10-15 large AI data center campuses indefinitely. Compare this to grid power at $50-120/MWh with zero availability guarantees. Nuclear is the only technology that can provide gigawatt-scale, 24/7, carbon-free power for AI.
07 Where Value Accrues
Not all AI infrastructure spending creates equal value. The $690B is distributed across a supply chain with very different margin profiles and competitive dynamics.
| Layer | Share of spend | Gross margin | Competitive moat | Key players |
|---|---|---|---|---|
| Silicon / GPUs | ~30% | 70-75% | Extremely high | NVIDIA, AMD, Google, AWS |
| Networking | ~10% | 55-65% | High | Arista, Broadcom, Cisco |
| Memory (HBM) | ~12% | 50-60% | High (3 suppliers) | SK Hynix, Samsung, Micron |
| Power infrastructure | ~15% | 25-40% | Medium-High | Eaton, Schneider, Vertiv |
| Cooling | ~8% | 30-45% | Medium | Vertiv, CoolIT, Iceotope |
| Construction / EPC | ~15% | 8-15% | Low-Medium | Quanta, MasTec, Jacobs |
| Land + permitting | ~5% | Varies | High (scarcity) | REITs, utilities, landowners |
| Energy generation | ~5% | 30-50% | Very high | Utilities, nuclear, IPPs |
The value chain shows a clear pattern: the higher the margin, the harder the moat. NVIDIA's 75% gross margin reflects a near-monopoly in AI training silicon. Power infrastructure margins are lower but access is increasingly scarce. Construction is commoditized but volume is enormous.
08 Revenue & Return Framework
The AI infrastructure buildout creates investable opportunities across multiple time horizons and risk profiles. Below is a framework for evaluating returns across the value chain.
| Investment thesis | Entry point | 5yr revenue CAGR | Margin profile | Risk |
|---|---|---|---|---|
| GPU / silicon suppliers | Public equities | 25-35% | Gross 65-75% | Concentration, custom chip threat |
| Power infrastructure | Public + PE | 15-25% | Gross 30-45% | Execution, permitting |
| Data center REITs | Public equities | 12-18% | NOI 50-60% | Interest rates, power costs |
| Nuclear / uranium | Public equities | 20-30% | Gross 40-55% | Regulatory, timeline |
| Cooling technology | PE / venture | 30-50% | Gross 40-55% | Technology risk, competition |
| AI cloud services | Public equities | 30-40% | Gross 55-70% | Capex intensity, execution |
| Fiber / connectivity | PE / infra funds | 10-15% | EBITDA 45-55% | Commoditization |
09 Risk Analysis
| Risk | Description | Severity | Mitigation |
|---|---|---|---|
| Overbuilding | Capex outpaces revenue growth; utilization falls below 60% | Medium | Long-term contracts, diverse workload mix, cloud marketplace |
| Power grid failure | Grid constraints delay deployments by 12-24 months | High | Behind-the-meter generation, nuclear PPAs, distributed deployment |
| GPU commoditization | Custom chips (TPU, Trainium) erode NVIDIA's pricing power | Medium | CUDA ecosystem lock-in, continued performance leadership |
| Regulatory backlash | Energy consumption, water usage, or export controls constrain growth | Medium | Renewable PPAs, cooling innovation, geographic diversification |
| Debt sustainability | $1.5T projected debt issuance in a rising-rate environment | Medium-High | Strong cash flows, investment-grade ratings, long maturities |
| AI demand plateau | Enterprise AI adoption slower than projected | Low-Medium | Consumer inference already generating revenue at scale (ChatGPT, Gemini) |
10 Outlook & Positioning
The AI infrastructure buildout is a multi-decade trend, not a cycle. Even if AI model improvement slows, the existing trajectory of inference workload growth, enterprise migration, and applications deployment will sustain infrastructure demand well into the 2030s.
12-month view
128-week lead times, grid constraints, and nuclear deals create a multi-year demand runway with limited new supply. Picks-and-shovels with pricing power.
AI clusters require 100G-800G networking at scale. Each new GPU deployed requires proportional network investment. Moats are deep.
NVIDIA remains dominant but priced for perfection. Custom chip threat from Google/Amazon creates long-term margin risk. Position on pullbacks.
The fundamental insight: $690 billion in annual spending creates a supply chain where the constraining resources — power, cooling, land with grid access — appreciate faster than the compute itself. The investment opportunity is not in buying GPUs; it is in owning the infrastructure that GPUs depend on to operate.
Reference Sources
Futurum Group, "AI Capex 2026: The $690B Infrastructure Sprint."
CNBC, "Tech AI spending approaches $700 billion in 2026," Feb 2026.
Introl, "Hyperscaler CapEx Hits $690B in 2026."
IEA, "Energy supply for AI — Energy and AI," 2025-2026.
Enki AI, "NVIDIA's AI Data Center Energy Demand: A 10 Gigawatt Challenge."
Enki AI, "AI Power Demand 2026: A Trillion-Dollar Power Crisis."
Deloitte, "Data Center Sustainability Insights," 2025.
Company filings and earnings calls: Amazon, Microsoft, Alphabet, Meta, Oracle (Q4 2025, Q1 2026).
This report is produced by Ralph Capital for informational purposes only. It does not constitute investment advice, an offer to buy or sell any security, or a solicitation. Data and analysis reflect publicly available sources as of March 2026; accuracy is not guaranteed. Scenario projections are analytical estimates, not forecasts. Ralph Capital may hold positions in assets discussed in this report.