Bush developed this distinction in reaction to the automation discourse of the 1940s, when industrial automation was displacing factory workers at scale. He saw clearly that automation's logic (replace the worker, capture the savings) and augmentation's logic (enhance the worker, expand the possible) led to incompatible futures. Bush advocated for augmentation not from technophobia but from engineering realism: the problems he cared about—scientific synthesis, cross-disciplinary connection, navigation of accumulated knowledge—required human judgment that no conceivable machine could replicate. The memex was designed to handle what machines could do (storage, retrieval, display) so humans could focus on what machines couldn't (asking which connections matter, evaluating which trails are worth following).
The distinction has been rediscovered repeatedly—Licklider's man-computer symbiosis (1960), Engelbart's augmentation framework (1962), the user-centered design movement of the 1980s, and now You On AI's amplifier metaphor all restate Bush's original insight. Yet each generation of technology builders has had to relearn the lesson because the economic incentives consistently favor automation over augmentation. Automation reduces headcount; augmentation expands what existing workers can accomplish. Automation produces immediate margin improvement; augmentation requires patience while workers develop new capabilities. The market rewards quarters, not decades, which explains why automation becomes the default and augmentation requires deliberate institutional commitment.
The simulation in this volume argues that large language models represent augmentation's apotheosis and its crisis. Apotheosis because the natural-language interface finally removes the translation barrier Bush identified as the bottleneck—users can now direct machines in their own associative vocabulary. Crisis because the same systems that augment can replace, and the economic logic determining which path prevails operates at speeds that outrun the institutional responses Bush spent his post-war career building. The National Science Foundation, the peer review system, the research university—all designed to ensure augmentation served broad human flourishing—now face adaptive challenges they weren't built to handle.
Bush's augmentation principle emerged from his World War II experience managing the coordination of 6,000 scientists across dozens of institutions. He observed that the bottleneck wasn't individual researcher capability but coordination—getting the right knowledge to the right person at the right time. The memex was designed to solve a coordination problem by giving each individual researcher augmented navigation capability. This institutional insight (that individual augmentation solves collective coordination challenges) distinguishes Bush's framework from purely individualist technology enthusiasm.
The philosophical ground was William James's pragmatism, particularly the thesis that consciousness is fundamentally selective—attention as a spotlight that illuminates some features while leaving others in darkness. Bush translated this into design principle: build tools that extend selective attention's reach without replacing the selection itself. The machine should retrieve; the human should choose what to retrieve. The machine should display; the human should interpret what is displayed. Maintaining this division required deliberate architectural choices that resisted the temptation to automate judgment alongside execution.
The irreplaceable human contribution. Bush identified associative judgment—deciding which connections matter—as the cognitive operation machines should support but never perform autonomously.
Economic vs. ethical logics. Automation optimizes for cost reduction, augmentation for capability expansion—incompatible objectives that produce divergent technological trajectories and require different institutional frameworks.
The coordination insight. Individual augmentation solves collective coordination problems more effectively than centralized control—a principle that anticipated distributed computing, open source, and now AI-democratized capability.
Judgment as the ascending challenge. When retrieval becomes trivial, evaluation becomes decisive—the pattern You On AI calls ascending friction, visible in Bush's framework seventy years before the language models arrived.