Training a frontier language model consumes electrical power equivalent to tens of thousands of households running continuously for weeks or months. Inference — the ongoing computation required to serve user queries — consumes even more in aggregate. The International Energy Agency has flagged AI data centers as a significant and growing source of global electricity demand. The current trajectory, if maintained, would require power generation capacity that does not yet exist and may not exist on the timeline the scaling laws demand.
The semiconductor industry encountered the same wall, and its response was instructive. When power consumption became the binding constraint on chip performance, the industry developed techniques for reducing power per computation: lower voltage operation, dynamic frequency scaling, specialized low-power architectures for mobile devices. These techniques did not eliminate the power constraint — they changed the terms under which it operated, buying time for the trajectory to continue. AI will require analogous techniques: more efficient model architectures, hardware optimized for inference rather than training, on-device processing that reduces the load on centralized data centers, and new approaches to cooling and power management.
Some of these are already in development. Others will emerge from the pressure the wall itself creates — in Moore's experience, walls are the most reliable source of engineering innovation. The problems that matter most are the ones that threaten the trajectory. In several regions, data center construction has been delayed or blocked by insufficient grid capacity, meaning the magic is already bumping against the physics of power generation and distribution.
The energy wall differs from Moore's thermal wall in one structurally significant way: it is not primarily a problem of heat dissipation on a chip but of power generation at civilizational scale. Managing it requires not just engineering innovation but infrastructure investment — new power plants, expanded grids, regulatory approvals — on timelines that move slower than compute demand is growing. This introduces coordination failures the semiconductor industry did not have to solve: the scaling of power generation is not controlled by the AI industry and does not respond to AI industry economics.
Concerns about AI energy consumption entered mainstream policy discussion through International Energy Agency reports beginning around 2023, with subsequent analysis by researchers including de Vries, Masanet, and others quantifying data center power requirements. The framing as a structural analog to the semiconductor thermal wall — and as a precursor to dimensional rotation — is articulated in this volume.
Inference dominates aggregate consumption. Training is expensive but one-time; inference is continuous and, in aggregate, exceeds training power consumption.
The grid, not the chip, is the bottleneck. Unlike semiconductor thermal walls, the energy wall is a problem of civilizational infrastructure, not on-die engineering.
Rotation responses are emerging. Lower-voltage hardware, on-device inference, specialized inference architectures, and novel cooling approaches parallel the semiconductor industry's responses to thermal limits.
Coordination failures compound. Power generation scales on timelines controlled by regulators and utilities, not by AI industry demand.
Walls drive innovation. The pressure the energy wall creates will produce engineering responses, but the timing of those responses relative to continued compute growth is uncertain.