Tali Rosenwaks, Partner at Next Gear Ventures.
Opinion

Physical AI is breaking the hyperscale model - not because of compute, but because of where compute can exist

Physical AI is shifting computation toward the real world. Robotics, autonomous logistics, manufacturing systems, and sensor-dense environments require inference at the point of action.

A structural shift is underway in the AI infrastructure economy. It is no longer driven by compute scarcity, but by a deeper system constraint: where compute can physically exist at scale.
On one side is Physical AI - robotics, autonomous systems, and industrial intelligence - where software is embedded directly into machines operating in the real world. On the other are two reinforcing constraints reshaping infrastructure design. Hyperscale AI is increasingly limited by grid delivery capacity, while Physical AI introduces a latency requirement: intelligence must sit close to the systems it controls. Centralized compute becomes less viable for real-time operation.
1 View gallery
Tali Rosenwaks Next Gear Ventures
Tali Rosenwaks Next Gear Ventures
Tali Rosenwaks, Partner at Next Gear Ventures.
(Courtesy)
This tension is already visible in deployment patterns. Crusoe’s Edge Zones place compute at energy-adjacent sites, including stranded or underutilized power assets, rather than waiting for hyperscale interconnection. In parallel, early microgrid-based AI facilities in Ireland and Texas
integrate on-site generation and storage into the data center, collapsing the separation between energy production and compute consumption.
For more than a decade, hyperscale data centers defined the cloud era, built on centralization, network effects, and the assumption that energy and connectivity could be extended to support concentrated compute regions. That assumption is now breaking under physical constraints.
Physical AI is shifting computation toward the real world. Robotics, autonomous logistics, manufacturing systems, and sensor-dense environments require inference at the point of action. In these systems, distance is a hard constraint that defines feasibility.
At the same time, energy has become the binding constraint on hyperscale expansion. Grid capacity is uneven, slow to expand, and increasingly saturated. In leading hubs such as Northern Virginia and parts of Europe, interconnection queues already span multiple years. Even where capital is abundant, electricity delivery is now the limiting factor.
This is amplified by scale. Frontier AI campuses require power comparable to medium-sized cities. Yet transmission, permitting, and generation operate on decade-long cycles. AI demand operates on product cycles measured in months. The result is a structural mismatch between compute velocity and energy velocity.
It is in this gap that a new architecture is emerging.
Instead of extending grids toward compute, compute is being relocated toward energy. This includes curtailed renewables, stranded transmission nodes, remote natural gas resources, and industrial sites with excess power. Energy availability becomes the primary organizing principle of compute geography.
This marks a reversal of hyperscale logic. Location is no longer determined by network density or demand proximity, but by energy physics: where electrons can be converted into computation at scale.
Distributed systems address this through modular deployment. On-site generation, hybrid renewables, and storage allow capacity to scale alongside demand. Energy and compute become co-designed rather than sequentially provisioned.
This does not displace hyperscale infrastructure. Large centralized campuses remain essential for frontier model training and global workload aggregation. But hyperscale alone is no longer sufficient to define the compute map.
The emerging architecture is bifurcated: centralized systems dominate training, while distributed systems handle physical-world intelligence and real-time inference at the edge.
The implications are structural rather than incremental. The binding constraint is no longer capital or chips, but the coordination of energy availability, physical proximity, and real-time compute demand. Alignment across these variables is becoming the core design problem of the industry.
As a result, the data center is being redefined. It is no longer a passive compute factory but an active node within an energy-industrial system, co-located with generation, storage, and grid constraints.
This creates a new competitive axis. Hyperscalers remain central orchestration layers, but advantage is shifting toward operators who control distributed, energy-coupled infrastructure. The challenge is no longer only scaling compute, but allocating it across fragmented and energy-variable environments.
This introduces a new requirement: workloads must be dynamically routed based on real-time electricity availability. Power is no longer a fixed input but a continuously optimized constraint. The question shifts from where to build compute to where compute should run at any moment.
A new infrastructure stack is emerging. At the physical layer: modular generation, high-density cooling, and storage systems designed for constrained environments. Above it: a software layer that matches workloads to energy conditions and converts stranded power into usable compute capacity in real time.
However, distributed systems introduce risks including fragmentation and coordination overhead. Their viability depends on whether orchestration software can abstract this complexity at scale.
Ultimately, as AI extends into robotics, logistics, manufacturing, and defense-adjacent systems, it becomes governed again by physical constraints: energy, geography, and throughput.
The cloud era abstracted these constraints away.
The Physical AI era reinstates them as first principles.
In doing so, it reshapes not only data center architecture, but the geography of intelligence itself.
Tali Rosenwaks is a Partner at Next Gear Ventures.