A Transition in Progress
For years, the AI industry operated under a fundamental assumption: computational power was the primary constraint. Build a faster chip, win the race.
That assumption is evolving. While chip innovation remains critical, a parallel competition has emerged—one centred on infrastructure orchestration rather than silicon alone. We are witnessing the early stages of a shift from a chip race to an infrastructure race.
Four Shifts Reshaping the Landscape
Inference Economics Are Rising in Importance
Training a large AI model is a discrete, one-time event. Serving that model to users is a continuous operation that scales with adoption.
We are still in the early stages of AI deployment. Most organizations are experimenting, integrating LLMs into existing platforms rather than rebuilding from first principles. But as AI moves from experimentation to production—and software platforms fundamentally rewrite their platforms around AI capabilities—the economics will shift decisively. The cost of delivering intelligence will rival, then exceed, the cost of building it.
This transition has begun, but it is far from complete.
Memory: Physics and Supply Chains Converge
Contemporary AI systems increasingly struggle not with calculation, but with data movement. Memory bandwidth and capacity are becoming limiting factors.
This constraint has two dimensions:
Physics: Moving data between memory and processors consumes time and energy. Context windows are ultimately bounded by these physical realities.
Supply chains: High Bandwidth Memory (HBM) production is concentrated and constrained. Shortages are not merely inconveniences—they are strategic bottlenecks that determine who can scale and who cannot. We know that trend will not resolve itself in 2026.
Understanding the memory constraint requires acknowledging both dimensions.
Thermal Management: The Hidden Constraint
Next-generation AI processors generate thermal densities that traditional cooling methods cannot accommodate. Rack power densities have increased from the historical standard of 10kW to well over 100kW in advanced AI deployments. Air cooling has reached its physical limits.
Immersion cooling—submerging servers in thermally conductive dielectric fluid—is emerging as a compelling solution. Such systems can reduce cooling-related energy consumption by 50-60%, whilst enabling the density that AI workloads require.
Cooling infrastructure is becoming as strategically important as power infrastructure.
A Multi-Winner Market Is Forming
The demand for AI infrastructure has expanded to such a degree that no single approach can satisfy it. Multiple architectures, vendors, and approaches will coexist.
This is not a winner-take-all market. It is a market where the winners will be those who can orchestrate complexity—coordinating power, compute, memory, and thermal management as integrated systems.
What Follows
The chip race rewards innovation in silicon. The infrastructure race rewards orchestration—the capacity to coordinate multiple systems at scale.
We are in the early stages of this transition. Most organisations are still integrating AI into existing workflows rather than rebuilding around it. The full shift will take years, not months.
But the direction is clear. The organisations that recognise this shift early—and position accordingly—will have significant advantages as the market matures.
The question is not whether this transition will occur. It is who adapts first.