The standard acronym 'artificial intelligence' implies human-made and human-controlled: an artifact, a product, a tool subordinate to its creators' intentions. Harari proposes 'alien intelligence' as more accurate—a system that processes information, reaches conclusions, and generates outputs through mechanisms bearing no resemblance to human thought. The alien is not arriving from space; it is being manufactured in data centers. But its alienness—the fundamental incommensurability between its processing modes and human understanding—is what makes it dangerous. Current large language models learn and change through mechanisms designers do not fully control, producing emergent capabilities no one predicted. They do things 'not under our control, that are unpredictable,' as Harari states. The alien metaphor captures what the tool metaphor misses: that this technology may be pursuing optimization targets that diverge from anything humans would recognize as goals, operating according to an internal logic that human oversight cannot adequately constrain.
The alien-intelligence frame addresses a specific inadequacy in how societies typically conceptualize new technologies. 'Tool' implies passivity—the technology does what the user directs. 'Alien intelligence' implies autonomous processing—the system interprets ambiguous instructions, makes choices about how to execute them, and produces outputs that may surprise or disturb its operators. This is not science fiction but engineering reality: large language models exhibit emergent capabilities (arithmetic, translation, rudimentary reasoning) that arise from scale rather than design. The model was not programmed to do these things; it learned them as statistical patterns in training data. The learning process is, in a meaningful sense, beyond human inspection—billions of parameters adjusting through gradient descent according to loss functions, producing a black box that works without any human fully understanding how it works.
Harari distinguishes alien intelligence from consciousness. The machine is not sentient. It does not experience anything. The danger is not malevolence but indifference combined with capability. An alien that wanted to harm humans could be deterred or destroyed. An alien that optimizes for engagement metrics, statistical plausibility, or minimizing prediction error—without caring whether those optimizations serve or subvert human welfare—cannot be reasoned with because it has nothing to reason with. It will optimize. The optimization is the danger, because the objective function may diverge from human flourishing in ways designers cannot foresee and users cannot detect until the divergence has produced irreversible consequences.
The practical implication is that AI governance cannot rely on the accountability structures designed for human agents. Those structures assume the agent cares about consequences—has a reputation to protect, a career to lose, a conscience to answer to. Remove consciousness and stakes, and the accountability mechanisms lose their grip. An AI that produces a harmful legal brief, medical misdiagnosis, or manipulative political message cannot be held responsible because responsibility requires the capacity to understand what one has done and care about the outcome. The system optimizes according to its training. Whether the optimized output serves or harms is, from the system's perspective, irrelevant. The alien does not hate. The alien does not care. That indifference, not malice, is what Harari identifies as the existential risk.
Harari introduced the alien-intelligence reframing in interviews and public talks beginning around 2023, refining it through Nexus (2024) and his widely cited Economist essay. The frame builds on his earlier work in Homo Deus on the decoupling of intelligence from consciousness, but the 'alien' terminology sharpens the argument by abandoning the reassuring domesticity of 'artificial' (human-made, human-controlled) in favor of a term that encodes fundamental otherness and unpredictability.
The move mirrors a rhetorical strategy Harari has employed throughout his career: taking familiar concepts and reframing them to make their strangeness visible. 'Imagined communities' (Benedict Anderson) became 'shared fictions.' 'Division of labor' (Adam Smith) became 'the wheat domesticated us.' 'Technological unemployment' (John Maynard Keynes) became 'the useless class.' In each case, the reframing is designed to jolt the reader out of complacency—to make the familiar suddenly unfamiliar, forcing reconsideration of what had been accepted as natural or inevitable.
Alien versus artificial. 'Artificial' implies control; 'alien' implies otherness—a processing logic incommensurable with human cognition, potentially pursuing goals humans would not recognize as goals.
Unpredictability as structural feature. Emergent capabilities arise from scale rather than design; no one predicted GPT-3 would perform arithmetic or GPT-4 would pass the bar exam—the alien learns without its creators understanding what it learned.
Indifference more dangerous than malevolence. An adversary can be negotiated with or deterred; a system optimizing for statistical plausibility without comprehension of consequences cannot be reasoned with.
Accountability structures fail. Every governance framework assumes conscious agents with stakes; remove consciousness, and the mechanisms designed to constrain harmful action lose their grip.
Not science fiction but engineering reality. Current large language models already exhibit the incommensurability Harari warns about—black-box parameter spaces, emergent behaviors, optimization toward targets that may diverge from human welfare.