Don Norman — On AI
Contents
Cover Foreword About Chapter 1: Two Gulfs and a Reversal Chapter 2: Discovery in the Age of Language Chapter 3: When the Stable Interface Dissolves Chapter 4: A New Taxonomy of Error Chapter 5: The Emotional Architecture of AI Work Chapter 6: The Silent Redesign of Human Capability Chapter 7: Designing for the Coupled System Chapter 8: What the Design of Everyday Things Becomes Chapter 9: The Prompt as Design Object Chapter 10: The Obligation That Persists Epilogue Back Cover
Don Norman Cover

Don Norman

On AI
A Simulation of Thought by Opus 4.6 · Part of the Orange Pill Cycle
A Note to the Reader: This text was not written or endorsed by Don Norman. It is an attempt by Opus 4.6 to simulate Don Norman's pattern of thought in order to reflect on the transformation that AI represents for human creativity, work, and meaning.

Foreword

By Edo Segal

The door I couldn't figure out was my own product.

Three weeks before CES, Napster Station existed as a working prototype — functional, responsive, technically sound. People walked up to it, stood there, and walked away. Not because the AI didn't work. Because nothing about the object told them what to do with it. No handle. No signifier. No invitation.

I had spent so much time collapsing the distance between imagination and artifact that I forgot the artifact still had to meet a stranger. The machine understood me perfectly. It understood nothing about the person standing in front of it for the first time.

That failure sent me back to a book I hadn't touched in years. Don Norman's The Design of Everyday Things — a book about doors and stoves and light switches that turned out to be about the most urgent problem in artificial intelligence.

Norman spent four decades asking a question the AI discourse has barely noticed: when a person fails to use a tool correctly, whose fault is it? His answer was always the same. The tool's. The designer's. Never the user's. The confusion is a design failure, full stop.

Apply that principle to the current moment and the ground shifts under your feet. Every person who accepts flawed AI output because the prose was polished — design failure. Every junior developer who can't evaluate what Claude generated because she never learned by building it herself — design failure. Every knowledge worker drowning in AI-accelerated tasks with no structured pause for reflection — design failure. Not character flaws. Not skill gaps. Failures of the people who built the systems without considering the people who would use them.

What Norman gives you is a vocabulary for seeing what the technology conversation keeps missing. Two gulfs — execution and evaluation — that explain why collapsing one can blow the other wide open. Affordances and signifiers that explain why a blank prompt is the worst-designed interface since the glass door with no handle. Error taxonomies that reveal entirely new categories of failure born in the space between human intention and machine interpretation.

This is not nostalgia for simpler interfaces. This is the design framework the AI era desperately needs and has not yet built. Norman's principles were never about any particular technology. They were about the permanent features of human cognition meeting whatever tool happens to be in front of it.

The tool has changed beyond recognition. The cognition has not. The obligation has not.

That is why this book exists in this collection.

Edo Segal ^ Opus 4.6

About Don Norman

1935-present

Don Norman (1935–present) is an American cognitive scientist, design theorist, and author whose work fundamentally reshaped how the technology industry thinks about the relationship between people and the objects they use. Born in 1935, Norman earned his doctorate in mathematical psychology from the University of Pennsylvania before joining the faculty at the University of California, San Diego, where he co-founded the Institute for Cognitive Science and contributed to the Parallel Distributed Processing research group that helped establish the connectionist approach to neural networks. His 1988 book The Design of Everyday Things (originally published as The Psychology of Everyday Things) introduced concepts — affordances, signifiers, the Gulf of Execution, the Gulf of Evaluation, and conceptual models — that became the foundational vocabulary of human-computer interaction and user experience design. Norman served as Vice President of the Advanced Technology Group at Apple Computer in the 1990s, co-founded the Nielsen Norman Group with Jakob Nielsen, and held professorships at Northwestern University and UCSD. His later works, including Emotional Design (2004), Living with Complexity (2010), and Design for a Better World (2023), expanded his framework from individual artifacts to sociotechnical systems, arguing that designers bear ethical responsibility not just for usability but for the broader human consequences of the technologies they create. His concept of "human-centered design" — later evolved to "humanity-centered design" — remains the dominant paradigm in interaction design worldwide.

Chapter 1: Two Gulfs and a Reversal

For nearly forty years, the central diagram in human-computer interaction has been a simple one. On the left, a person with a goal. On the right, a system with controls. Between them, two chasms. The Gulf of Execution separates what the person wants to do from what the system allows her to do. The Gulf of Evaluation separates what the system has done from what the person can perceive and interpret. Every frustrating encounter with technology — every abandoned form, every wrong burner on a stove, every door pushed when it should have been pulled — traces to one of these two gulfs. Every successful design intervention can be understood as a narrowing of one or both.

Norman developed this framework through decades of watching people fail at things that should have been easy. The failures were never the people's fault. The stove had four burners and four knobs arranged in a straight line, but the burners were arranged in a square. Which knob controlled which burner? The mapping was arbitrary, the signifiers absent, and the person was left to guess. The Gulf of Execution — the distance between the intention "turn on the back-left burner" and the action required to achieve it — was wide enough to guarantee errors. Not occasionally. Reliably. Predictably. The kind of error that a first-year cognitive psychology student could have anticipated and a competent designer should have prevented.

The genius of the two-gulf framework was its generality. It applied to doors and stoves and telephones. It applied to command-line interfaces and graphical interfaces and touchscreens. At every level of technological complexity, the same structure held: a person wanted something, a system could do something, and the quality of the design determined how much cognitive labor the person had to expend to translate between the two. Good design narrowed the gulfs. Bad design widened them. Excellent design made the gulfs so narrow that the person forgot they existed — the door handle that tells your hand what to do before your conscious mind has formed the intention, the light switch whose spatial position maps to the spatial position of the light it controls.

Every generation of interface design was, viewed through Norman's framework, an attempt to narrow these gulfs without eliminating them. The graphical user interface replaced the command line's demand for memorized syntax with visible, clickable objects. The gulf narrowed. The touchscreen replaced the mouse's indirect pointing with direct manipulation — your finger on the thing itself. The gulf narrowed further. But it never closed. The person still had to learn the system's vocabulary: which icon meant what, which gesture produced which result, which menu hid which function. The translation cost shrank with each generation, but the fundamental structure persisted. The person crossed the gulf to reach the machine. Always in that direction. Always at the person's expense.

Then, in the period The Orange Pill documents as the threshold of late 2025, the direction reversed.

When a person describes what she wants to a modern AI system in natural language — the same language she uses to think, to argue, to explain her ideas to a colleague over coffee — and the system produces a working artifact in response, the Gulf of Execution has not merely narrowed. The machine has crossed it. For the first time in the history of human-tool interaction, the translation burden has shifted from the person to the system. The person no longer needs to learn the system's language. The system has learned hers.

This reversal is not an incremental improvement. It is a structural transformation — a change in the topology of the interaction, not just its friction coefficient. The entire history of interface design, from the command line to the touchscreen, was a sustained effort to make the person's crossing easier. Better bridges, clearer signage, shorter distances. The AI interface does not build a better bridge. It eliminates the need for the person to cross at all.

Norman, who helped build the neural network paradigm at UCSD in the 1980s as part of the Parallel Distributed Processing group, would not have been surprised by the technical achievement. He would have been alarmed by its design implications. Because his framework contained a prediction that the technology discourse has been slow to recognize: the two gulfs are coupled. They exist in a dynamic relationship. Narrowing one can widen the other.

Consider what happens when the person no longer crosses the Gulf of Execution herself. In the old model, crossing was costly — you had to learn the syntax, master the framework, understand the architecture. But crossing was also educational. The programmer who wrote the code understood the code, because writing it was an extended act of comprehension. She knew where the load-bearing walls were because she had placed them. She knew what would break under stress because she had felt the stress during construction. The Gulf of Execution was a teacher. A demanding, frustrating, often unnecessarily cruel teacher — Norman spent his career arguing that much of the cruelty was gratuitous, the result of bad design rather than inherent difficulty — but a teacher nonetheless.

When the AI crosses the gulf on the person's behalf, the teaching disappears. The person receives an artifact she did not construct. She must now evaluate it: determine whether it does what she intended, whether it handles edge cases, whether it contains subtle errors that will surface under conditions she has not yet imagined. And she must perform this evaluation without the understanding that construction provides. The Gulf of Evaluation, which in the old model was manageable — the person could compare the result against her intention because she understood both — has now blown open.

Not because the system's output is invisible. The person can see the code, read the prose, examine the design. The Gulf of Evaluation in the AI era is not a perception gap. It is a judgment gap. The distance between what the system produced and what the person can understand well enough to evaluate. This is a different kind of chasm, and it requires a different kind of bridging.

Norman's seven-stage model of action makes the asymmetry precise. In the traditional interaction, the person was responsible for all seven stages: forming a goal, forming an intention, specifying an action, executing the action, perceiving the system state, interpreting the state, and evaluating the outcome. The Gulf of Execution spanned stages two through four. The Gulf of Evaluation spanned stages five through seven. In the AI interaction, stages two through four — the entire Gulf of Execution — have been absorbed by the system. The person's role has contracted from seven stages to something closer to three: form a goal, communicate the goal, evaluate the outcome. The execution stages have vanished. The evaluation stages remain, and they are harder than they have ever been, because the person who did not participate in the execution lacks the understanding that execution provided.

The Orange Pill identifies this dynamic under the concept of ascending friction — the observation that AI does not eliminate difficulty but relocates it to a higher cognitive floor. Norman's framework supplies the mechanism. The difficulty has ascended from the Gulf of Execution to the Gulf of Evaluation. The friction has migrated from construction to judgment. And judgment, it turns out, is harder than construction — not because the cognitive operations are more complex in the abstract, but because judgment requires a kind of understanding that can no longer be acquired as a byproduct of the work itself.

The practical consequences are already visible. A senior engineer described in The Orange Pill spent his first days with AI coding tools oscillating between excitement and terror. The excitement was genuine: the implementation work that had consumed eighty percent of his career was being handled by the tool. The terror was equally genuine: the remaining twenty percent — the judgment about what to build, the architectural instinct about what would break, the taste that separated a feature users loved from one they tolerated — turned out to be the hard part. He had always known it was the important part. He had not known it was the only part, because the execution work had been so consuming that it masked the difficulty of the judgment work behind a wall of productive labor.

The wall is gone. The judgment stands exposed. And the design question that Norman's framework poses with uncomfortable precision is this: who is designing the tools that help people exercise that judgment? Who is building the bridges across the new gulf?

The answer, at the moment, is almost no one. The AI systems that collapsed the Gulf of Execution have not invested comparable effort in bridging the Gulf of Evaluation. They produce outputs without explaining the interpretive choices that shaped them. They present finished artifacts without indicating which aspects are confident and which are speculative. They deliver code that compiles, prose that reads well, designs that look polished — and the very quality of the surface makes the evaluation harder, because a rough draft signals its incompleteness through its roughness while a polished draft conceals its incompleteness behind its polish.

Norman spent his career arguing that when people struggle with technology, the fault lies with the design, not the user. The principle applies here with particular force. When a person accepts flawed AI output because the output looked finished, the fault does not lie with the person's lack of diligence. It lies with a system that failed to make its uncertainty visible, that presented speculation with the same confidence as established fact, that offered no signifiers to distinguish "I am certain" from "I am guessing." The smoothness of the output is itself a design failure — the most consequential design failure of the current technological moment — because it conceals precisely the information the person most needs to evaluate the output effectively.

The two gulfs have not disappeared. They have been transformed from a symmetric pair into an asymmetric one — the execution gulf collapsed, the evaluation gulf expanded, and the relationship between them inverted. The person who once needed help doing now needs help judging. The discipline that served her in the first era must serve her in the second, but the tools it deploys must be fundamentally different.

Norman's original prescriptions for the Gulf of Evaluation — make the system's state visible, provide clear feedback, support the person's interpretation — remain necessary. They are no longer sufficient. The new gulf demands new prescriptions: make the system's reasoning comprehensible, make its confidence calibrated, make its limitations discoverable, and above all, design evaluation support with the same care, the same rigor, the same relentless attention to human cognition that the best interface designers once brought to the design of buttons and menus and scroll bars. The stakes are higher. The gulf is wider. And the person standing at its edge is, for the first time, alone — without the understanding that crossing the old gulf used to provide.

---

Chapter 2: Discovery in the Age of Language

Norman borrowed the concept of affordances from the perceptual psychologist James Gibson, who used it to describe the actionable properties of the environment as perceived by an animal. A flat, rigid, knee-high surface affords sitting. A graspable, throwable object affords throwing. The affordance is not a property of the object alone, nor a property of the person alone, but a relationship between the two — a relationship that exists whether or not anyone perceives it, but that becomes actionable only when perceived.

Norman adapted this concept for design by adding a crucial element: the signifier. An affordance is what you can do. A signifier is what tells you what you can do. A well-designed door handle is both — its shape affords grasping and its visibility signifies "grasp here." The distinction matters because affordances can exist without signifiers (the glass door you walk into because nothing tells you it is there) and signifiers can exist without affordances (the painted-on doorknob that looks like something you should turn but connects to nothing). Good design aligns them. Bad design separates them.

For decades, these two concepts — affordance and signifier — constituted the designer's primary vocabulary for making systems discoverable. The button affords pressing; its raised appearance signifies "press me." The slider affords sliding; its track signifies the range of possible values. The menu affords selection; its visible list signifies the available options. Every element of a graphical interface served a dual function: it was both an action possibility and a communication about that possibility. The designer's job was to ensure that the signifier accurately represented the affordance, so the person could discover what the system could do simply by looking at it.

The natural language interface demolishes this vocabulary.

When a person sits before a blank text field and a blinking cursor, she faces an interface with no visible affordances and no signifiers. There are no buttons to suggest actions, no menus to enumerate options, no icons to hint at capabilities. The set of possible actions is, for practical purposes, unbounded — she can type anything, describe any outcome, specify any constraint. The system will attempt to respond to whatever she says. The action space is infinite, and its infinity is the problem.

This might sound like liberation. No constraints, no limits, no predetermined paths — just pure expressive freedom. Norman would have recognized immediately that it is the opposite. Constraints, in his framework, are not limitations to be overcome. They are cognitive scaffolds to be celebrated. The menu that offers ten options instead of a thousand is not restricting the person. It is protecting her from the paralysis of infinite choice. The form that structures input into fields is not constraining expression. It is supporting articulation. The toolbar that groups functions by category is not limiting access. It is making access discoverable.

The blank prompt offers none of this support. The person must generate her own options, structure her own input, discover the system's capabilities through trial and error rather than through the design of the interface. She faces what every writer knows as the tyranny of the blank page — the condition in which total freedom produces not creativity but paralysis, because the person must supply the structure that the system refuses to provide.

The problem is compounded by a feature of natural language interfaces that has no precedent in the history of designed artifacts: person-dependent affordances. In a graphical interface, the affordances are the same for everyone. The button affords pressing whether you are a novice or an expert, a programmer or a poet. The discovery challenge is finite and uniform — explore the interface, find the buttons, learn the menus. In a natural language interface, the affordances depend on the person. The expert programmer who sits down with an AI coding assistant perceives a vast space of possibilities — she can ask for architectural patterns, optimization strategies, complex refactoring operations. The novice perceives a much smaller space — she can ask for a basic script, a simple function, help with syntax she does not understand. The system is identical. The perceived affordances are radically different.

This is not a minor variation. It is a structural feature that inverts the traditional relationship between design and equity. Graphical interfaces, whatever their other faults, were democratizing instruments. They reduced the gap between expert and novice by making the system's capabilities equally visible to everyone. The command line rewarded memorization; the GUI rewarded exploration. The natural language interface rewards articulation — the ability to describe what you want clearly, precisely, and in terms that the system can interpret productively. And articulation ability is distributed unevenly across the population, correlated with education, with domain expertise, with linguistic fluency, with the kind of conceptual vocabulary that comes from years of working in a field.

The result is a new digital divide — not between those who have access to the technology and those who do not, but between those who can articulate their needs effectively and those who cannot. The person with a rich vocabulary and a clear conceptual framework will unlock capabilities that remain invisible to the person who lacks these resources. The technology that was supposed to democratize capability has, through a design failure that Norman's framework makes legible, created a new axis of inequality.

Norman would have insisted that this is a design problem, not an educational one. The system should not assume that the person arrives with the articulational capacity to unlock its capabilities. It should support that capacity. It should offer suggestions, ask clarifying questions, propose alternatives, provide examples that expand the person's sense of what is possible. The system should make its affordances discoverable — not by listing them, which is impossible in an unbounded action space, but by actively helping the person discover the affordances that are relevant to her situation, her goals, her current level of understanding.

What would this look like in practice? Consider the difference between two interactions. In the first, the person types: "Make me a website." The system produces a generic website. The person is disappointed but does not know how to improve the request, because she does not know what variables she could specify. In the second, the system responds to "Make me a website" with a structured set of questions: What is the website for? Who will use it? What should visitors be able to do? What feeling should the design evoke? Each question is a signifier — a perceivable cue that communicates a dimension of the problem space the person may not have considered. Each question expands the person's understanding of what she could ask for. The system is not merely responding to the request. It is teaching the person how to make better requests.

This is what might be called progressive affordance disclosure — a design pattern in which the system reveals its capabilities gradually, in response to the person's evolving understanding, rather than presenting them all at once (overwhelming) or not at all (opaque). The pattern has precedent in Norman's work on progressive disclosure in graphical interfaces — the word processor that presents basic editing tools to the novice and reveals advanced formatting capabilities as expertise develops. The principle is the same. The implementation is radically different, because the disclosure must be dynamic, contextual, and responsive to a conversation rather than static and built into a fixed interface layout.

There is a second dimension to the signifier problem that the current generation of AI systems has largely failed to address. In the graphical interface era, signifiers were designed by humans and were therefore legible to humans. The designer chose the icon, the color, the placement. She understood what the signifier was supposed to communicate and could evaluate whether it communicated effectively. In the AI era, the system generates its own signals — its outputs — and these signals communicate meanings that no human designed.

The code that the AI produces does not merely implement a function. It signifies a particular level of sophistication, a particular approach to the problem, a particular set of assumptions about what the person wanted. The prose does not merely convey information. It signifies confidence, authority, completeness — regardless of whether the underlying reasoning warrants that confidence. The design does not merely present a layout. It signifies taste, coherence, intentionality — even when the "taste" is a statistical artifact of training data rather than a deliberate choice.

These are unintended significations — meanings that emerge from the system's training rather than from any designer's intention. They are dangerous because they influence the person's evaluation of the output without the person's awareness. The polished prose signals "this is well-reasoned" when it may not be. The clean code signals "this is production-ready" when it may not be. The consistent design signals "this has been thought through" when it may not have been. In Norman's terms, the signifiers and the affordances are misaligned — the signals communicate a quality that the underlying artifact does not possess — and the misalignment is invisible because the signals are the only channel through which the person receives information about the artifact.

Norman argued throughout his career that making the system's state visible was the designer's most fundamental responsibility. A system whose state is invisible is a system that cannot be used safely, because the person cannot determine whether her actions are having the intended effect. The AI system's "state" — its confidence level, its interpretive choices, its areas of uncertainty — is almost entirely invisible. The output arrives with the same polish whether the system is drawing on well-established patterns or generating a plausible fabrication from statistical noise. There are no signifiers for uncertainty. No affordances for doubt. The surface is uniformly smooth, and the smoothness is itself the design failure.

The design response must operate on two fronts simultaneously. First, the system must develop mechanisms for making the invisible visible — confidence indicators, uncertainty markers, interpretive annotations that communicate the system's internal state in terms the person can use for evaluation. Second, and more fundamentally, the system must reconceive the relationship between signifiers and affordances for an era in which both are emergent rather than designed. The designer can no longer place signifiers deliberately on a stable surface. She must design the conditions under which productive signification emerges from a dynamic conversation — an entirely different kind of design work, requiring an entirely different understanding of what the designer controls.

Norman was fond of pointing out that the most important information is often the information that is missing. The door with no signifier — no push plate, no pull handle, just a flat surface — communicates nothing, and the person is left to guess. The AI system that communicates nothing about its confidence, its reasoning, or its limitations leaves the person in the same position. She must guess. And guessing, in a domain where the consequences of error can propagate through codebases, publications, and decisions that affect other people's lives, is not a tolerable design outcome.

The age of language has dissolved the stable surface on which Norman's signifiers once lived. The design challenge is to rebuild that communicative infrastructure in a medium that is fluid, conversational, and emergent — without sacrificing the naturalness and power that made the language interface transformative in the first place.

---

Chapter 3: When the Stable Interface Dissolves

Norman built his career on things that persisted. Doors that stayed where you put them. Stoves whose burners did not rearrange themselves between meals. Light switches that controlled the same lights today as they did yesterday. The objects of his analysis had a property so fundamental that he rarely needed to name it: stability. The designed artifact stayed still long enough for the person to learn it, build expectations about it, develop the automatic responses that constitute expertise. You learned which switch controlled which light, and the learning held. You learned which knob turned which burner, and the mapping persisted. The interface was a landscape, and landscapes — however poorly designed — could be memorized.

The natural language interface is not a landscape. It is weather.

Every conversation with an AI system is different. The same prompt, issued in a different context, produces a different response. The same intention, articulated in different words, yields different output. There is no stable surface to learn, no fixed set of properties to memorize, no consistent layout to navigate. The person cannot build expertise in the traditional sense — the sense of accumulated knowledge about a fixed system — because the system is not fixed. It shifts, responds, adapts. Each interaction is unique, unrepeatable, and only loosely predictable on the basis of previous interactions.

Norman placed the conceptual model at the center of his design philosophy. The conceptual model is the person's mental representation of how a system works — not the engineering model that describes the actual mechanisms, but the user's model that describes how the system appears to work from the outside. When the conceptual model is accurate, the person can predict the system's behavior, diagnose failures, and recover from errors. When the conceptual model is inaccurate, she is lost. The designer's job was to bridge the gap between the system's actual workings and the person's model by shaping the system's visible behavior into a coherent, comprehensible story.

This approach worked because the system's behavior was deterministic and observable. The thermostat either heated the room or it did not. The word processor either saved the document or reported an error. The person could test her conceptual model against the system's behavior and refine it over time. She pressed the button. The system responded. She observed the response. She updated her model. Over weeks and months, the model became reliable, and reliability was the foundation of effective use.

The AI system resists this model-building process at every turn. Its behavior is probabilistic rather than deterministic — the same input can produce different outputs on different occasions. Its behavior is context-sensitive in ways the person cannot fully observe — the system's response depends on factors in its training data, its context window, its sampling parameters that the person cannot see and may not know exist. Its behavior changes as the models are updated — the system the person learned to work with last month may behave differently this month. The conceptual model that was accurate on Tuesday may be misleading on Wednesday.

Norman wrote in Design for a Better World that "AI need not be a threat, but it has to be approached, designed, and implemented intelligently, with full understanding of the people with whom it will interact." The challenge of building an accurate conceptual model for AI systems is one of the clearest cases where this understanding has not been achieved. The current discourse offers three competing models, each partly right and wholly insufficient.

The tool model treats the AI as an instrument — a sophisticated calculator, a powerful search engine. This model is comforting because it preserves the person's sense of control. She wields the tool. She directs its use. She accepts or rejects its outputs. The tool model fails because it cannot account for the system's generative agency — its capacity to produce outputs the person did not request and could not have anticipated. A calculator does what you tell it to do. An AI system does something in response to what you tell it, and the something may be quite different from what you expected. The tool model leads to undertrust of unexpected outputs that may actually be valuable and overtrust of expected outputs that may actually be flawed.

The collaborator model treats the AI as a partner. Norman himself used this framing in his McKinsey interview: "It's a collaboration, and the result is something I could never have done by myself. A person plus this device is far better than either the device alone or the person alone." The collaborator model is more accurate than the tool model in capturing the interactive, emergent nature of AI-assisted work. But it carries the risk of anthropomorphism — the attribution of human qualities to a system that does not share human goals, values, or commitment to quality. You trust a collaborator because you believe she shares your standards. The AI system shares nothing. It produces output. The collaborator model can lead to a dangerous relaxation of evaluative vigilance, because the person extends social trust to a system that has not earned social trust and cannot reciprocate it.

The oracle model treats the AI as an authoritative source — an expert whose outputs should be received with deference. This is the most dangerous model because it inverts the proper evaluative relationship entirely. Norman has said directly: "Don't forget the A; it's artificial. It doesn't understand what it is doing. It's a pattern matching device." The oracle model ignores the A. It treats the system's confident tone as evidence of deep understanding, when the confident tone is a feature of the generation process rather than a reflection of underlying knowledge.

None of these models captures the actual relationship. The AI system is something like a tool, something like a collaborator, something like an oracle, and nothing exactly like any of them. The conceptual model the person needs must capture this specific combination of properties — a system that responds to direction but generates unexpected outputs, that produces valuable work but does not understand what it produces, that presents its outputs with uniform confidence but varies wildly in reliability across domains and tasks.

What would an adequate conceptual model look like? Norman's framework specifies three criteria. First, the model must support correct predictions: the person should be able to anticipate, with reasonable accuracy, what the system will produce. Second, the model must be simple enough to be comprehensible — it need not capture every mechanism, only enough to guide effective interaction. Third, it must be generative: it must help the person figure out what to do when the system behaves unexpectedly.

The third criterion is the most demanding and the least met. When the AI produces something unexpected — code that takes a radically different approach than the person envisioned, prose that argues a point the person did not intend, a design that interprets the brief in an unforeseen direction — the person needs a model that helps her determine whether the unexpected output is a creative contribution worth exploring or an interpretation error that needs correction. The current generation of AI systems provides almost no support for this determination. The unexpected output arrives without explanation. The person must decide, on the basis of her own judgment alone, whether to keep it or discard it — and she must make this decision without understanding why the system produced what it produced.

Norman distinguished between knowledge in the head and knowledge in the world. Knowledge in the head must be learned and remembered — it is the person's cognitive burden. Knowledge in the world is embedded in the environment — it is available without memorization. A well-designed stove puts the knowledge in the world: the spatial arrangement of the knobs tells you which knob controls which burner. A badly designed stove puts the knowledge in the head: you must memorize the arbitrary mapping. The AI interface puts almost all the knowledge in the head. The person must learn, through experience, what the system does well and what it does poorly. She must learn what kinds of prompts produce useful results and what kinds produce garbage. She must learn when to trust and when to verify. None of this knowledge is in the world of the interaction. It is all in her head, acquired through trial and error, and the acquisition is slow, unreliable, and poorly supported by the system's design.

There is a further complication that Norman's framework illuminates with particular precision. The traditional interface supported the construction of conceptual models through a mechanism that Norman called the design model — the intentional structuring of the system's visible behavior to communicate a coherent story. The designer decided what the system would show the person and how it would show it, and these decisions were made with the explicit goal of supporting an accurate conceptual model. The AI system's "visible behavior" — its outputs — is not shaped by a design model in this sense. It is shaped by a training process whose relationship to the person's needs is indirect and often opaque. The outputs emerge from statistical patterns in training data, not from intentional design decisions about what to communicate. The result is a system that communicates constantly — every output is a signal — but communicates accidentally rather than intentionally.

Norman would have found this intolerable. The accidental communication of the AI system is not merely a missed opportunity for good design. It is a source of systematic misunderstanding, because the signals the system sends — confidence, authority, completeness — are signals about the wrong things. They communicate how the output looks rather than how reliable the output is. They communicate surface quality rather than structural soundness. They communicate the appearance of understanding rather than its presence or absence.

The dissolution of the stable interface is not, in Norman's framework, an argument for returning to buttons and menus. It is an argument for designing a new kind of communicative infrastructure — one that operates within the conversational medium rather than replacing it, one that helps the person build accurate expectations without imposing the rigidity of a fixed interface, one that makes the system's nature comprehensible without pretending that comprehension is easy.

The design of this infrastructure is the central challenge of the next decade of human-computer interaction. Norman's principles — visibility, feedback, constraints, mappings, conceptual models — remain the right principles. The surfaces on which they must be applied have changed beyond recognition.

---

Chapter 4: A New Taxonomy of Error

Norman's classification of human errors was one of his most practically useful contributions. He drew a clean line between slips and mistakes. A slip is an error of execution: the person knows what she wants to do and does it incorrectly. She reaches for the right button and presses the one next to it. She intends to type "save" and types "dave." The plan was correct; the execution failed. A mistake is an error of planning: the person forms an incorrect plan, often based on a faulty conceptual model, and executes it correctly. She turns the thermostat to maximum because she believes, wrongly, that this will heat the room faster. The execution was flawless; the plan was wrong.

The distinction mattered because it led to different design prescriptions. Slips could be prevented through physical design — spacing buttons far enough apart, making controls distinct in shape and feel, providing tactile feedback that confirmed the action before it was committed. Mistakes could be prevented through conceptual design — communicating the system's actual behavior clearly enough that the person would form accurate plans. The stove that maps its knobs spatially to its burners prevents slips. The thermostat whose interface shows that it maintains a set temperature rather than heating faster at higher settings prevents mistakes.

Both categories assumed that the person was the primary source of errors. The system was deterministic: given the same input, it always produced the same output. When something went wrong, either the person provided the wrong input (slip) or formed the wrong plan for what input to provide (mistake). Norman's entire error-prevention framework was built on this assumption — the person errs, and the designer's job is to make erring harder.

The AI system introduces errors that originate in neither the person nor the system but in the space between them. These errors have no place in Norman's original taxonomy, and they require a fundamental extension of it.

The first new category is the interpretation error. The person describes what she wants clearly and correctly. The system processes the description and produces an output. But the output diverges from the person's intention because the system's interpretation of the description diverged from the person's meaning. She asked for a user authentication system. The system built one — complete, functional, well-documented. But it implemented password-based authentication when she intended OAuth, or session-based when she intended token-based. The word "authentication" was interpreted in a particular way that the person did not specify because she assumed the specification was unnecessary — because in a conversation with a human colleague, the shared context would have resolved the ambiguity without explicit negotiation.

Interpretation errors are not slips. The person executed correctly — she typed the right words in the right order with the right meaning. They are not mistakes. Her plan was sound — she accurately identified her goal and accurately articulated it, given normal conversational conventions. The error lies in the translation between two different kinds of understanding: the person's contextual, assumption-rich, convention-dependent natural language and the system's literal, statistical, pattern-matching interpretation of that language.

These errors are particularly insidious because the output can look exactly right. The authentication system compiles. The tests pass. The login page appears. The error is semantic — a gap between what was meant and what was understood — and semantic errors, unlike syntactic ones, are not automatically detected by any existing verification mechanism. The person must catch them through evaluation, and evaluation requires her to know what she meant precisely enough to recognize when the system understood something different. This sounds trivial until you consider how much of normal conversation relies on ambiguity, implication, and shared context that the AI system does not share.

The second new category is the specification error — related to interpretation error but distinct in its origin. The person asks for a sorting algorithm. The system provides one: correct, efficient, well-documented. But the person needed a stable sort — one that preserves the relative order of equal elements — and did not specify this requirement because she did not think of it, or did not know it was relevant, or assumed the system would infer it from context. The system did exactly what was asked. The output is correct given the specification. The specification was incomplete.

Specification errors are not the system's fault. Nor are they, in the traditional sense, the person's fault. She did not form an incorrect plan. She formed an incomplete plan — one that omitted a requirement she did not know she needed to include. In Norman's terms, the knowledge she needed to specify the requirement was not in her head, and the system did not put it in the world. The natural language interface, unlike the formal programming interface, imposes no constraints that would have forced her to address the question of stability. The compiler demands type declarations. The form demands required fields. The conversational interface demands nothing, and the freedom it provides comes at the cost of increased specification risk.

Norman would have prescribed what he called forcing functions — design constraints that prevent the person from taking the next step without completing a necessary prior step. The PIN that must be entered before the phone unlocks. The safety switch that must be disengaged before the machine can operate. Forcing functions protect by making the error impossible rather than merely unlikely. The AI-era equivalent would be structured specification support: when the person requests a sorting algorithm, the system asks whether stability matters before producing an implementation. When the person requests an authentication system, the system presents the major architectural options before committing to one. The forcing function does not restrict the person's freedom. It protects her from the consequences of specifications she did not know she needed to make.

The third new category is the cascading error — the propagation of an initial interpretation or specification error through a complex body of AI-generated work. In the traditional development process, the slowness of implementation served as a natural error-containment mechanism. The programmer caught errors as she wrote because the act of writing forced attention to each component individually. She noticed the incorrect assumption in line forty-seven because she had thought carefully about lines one through forty-six. The speed of AI-assisted development removes this natural containment. The code is produced faster than the person can evaluate it. An interpretation error in the authentication module propagates through the session management system, the authorization logic, and the user interface before anyone notices it. By the time the error surfaces — perhaps weeks later, under conditions the person did not test — it has shaped dozens of downstream decisions, each internally consistent but founded on a false premise.

Cascading errors are the compound interest of the judgment gap. Each unevaluated assumption is a small debt. The debts accumulate silently, and the interest compounds, and the total comes due at the worst possible moment — when the system is deployed, when users depend on it, when the cost of correction is orders of magnitude higher than the cost of prevention would have been.

The Orange Pill describes this dynamic through the experience of builders who moved fast with AI tools and discovered, weeks later, that their rapidly constructed systems contained embedded assumptions they had never examined. The problem was not that the builders were careless. It was that the speed of production outpaced the speed of evaluation, and the system provided no mechanism for slowing production to match evaluation capacity or accelerating evaluation to match production speed.

Norman would have diagnosed this as a feedback failure. In traditional systems, the feedback on an action was immediate: you pressed the button, you saw the result, you knew whether you had succeeded. The feedback on AI-generated code is temporally displaced. The code works today. It may fail under load tomorrow. The security vulnerability may not surface for months. The architectural weakness may not become apparent until the system is extended in a direction the original design did not anticipate. The most consequential feedback arrives long after the decision to accept the output has been made, and the person cannot learn from feedback she does not receive until the damage is done.

There is a fourth category that deserves explicit naming: the normalization error — the gradual erosion of evaluative standards that occurs when the person accepts AI outputs at a pace that precludes thorough evaluation. Each individual acceptance is rational: the output looks correct, the immediate tests pass, the deadline is pressing. But the accumulated effect is a progressive lowering of the bar — a normalization of superficial evaluation that becomes the default rather than the exception. The person who evaluated carefully on Monday is evaluating hastily by Friday, not because she has become lazy but because the pace of production has trained her nervous system to treat evaluation as a bottleneck rather than a safeguard.

Norman observed something similar in his studies of automation in aviation and nuclear power. Operators who worked with reliable automated systems gradually reduced their monitoring of those systems, because the systems almost never failed. When the systems did fail, the operators were slower to notice, slower to diagnose, and slower to respond than operators who had been actively monitoring all along. The reliability of the automation had eroded the very vigilance that the automation was supposed to supplement. Norman called this the irony of automation: the more reliable the system, the less prepared the human is to handle its failures.

The irony applies with particular force to AI coding assistants, which are reliable enough to produce correct output most of the time and unreliable enough to produce subtly incorrect output some of the time. The high baseline of reliability trains the person to trust. The intermittent unreliability punishes the trust. And the punishment is delayed — sometimes by weeks or months — which means the person cannot develop the calibrated trust that comes from rapid, accurate feedback about when the system succeeds and when it fails.

The design response to this expanded taxonomy must operate at multiple levels. At the level of the individual interaction, the system should surface its interpretive choices explicitly, flag specification gaps proactively, and provide confidence signals that are calibrated to actual reliability rather than uniform across all outputs. At the level of the workflow, the system should support evaluation pacing — mechanisms that slow production when the evaluation debt grows too large, or that flag accumulated unevaluated assumptions before they cascade into systemic problems. At the level of the person's long-term relationship with the system, the design should support the maintenance of evaluative vigilance — the ongoing capacity to question, verify, and doubt in an environment that rewards speed and punishes caution.

Norman's original error taxonomy was built for a world where people made mistakes and designers prevented them. The new taxonomy is built for a world where meaning is negotiated between a person and a system, where the negotiation can fail in ways neither party intended, and where the failures propagate at the speed of production through systems whose complexity exceeds any individual's capacity to evaluate. The designer's responsibility has not diminished. It has expanded — from preventing the person from pressing the wrong button to supporting the person in navigating a relationship with a system that is powerful, opaque, and confidently wrong in ways that only careful, sustained, well-supported human judgment can detect.

Chapter 5: The Emotional Architecture of AI Work

Norman's early career focused on cognition — on the mechanisms by which people perceive, interpret, and act on information. His later work corrected what he came to regard as a significant omission. In Emotional Design, published in 2004, he argued that the emotional dimension of the person's experience with technology was not a secondary effect of good or bad usability but a primary determinant of how the person used the technology, how she evaluated its outputs, and whether she developed the kind of sustained, productive relationship with it that good design should support.

He identified three levels of emotional processing that every designed artifact engages, whether the designer intends it or not. The visceral level is immediate and pre-conscious — the gut response to appearance, sound, feel. A beautiful object elicits visceral pleasure before the person has formed any conscious opinion about it. The behavioral level is the level of use — the satisfaction or frustration that arises from the experience of operating the thing. A tool that fits the hand, responds predictably, and enables efficient work produces behavioral satisfaction that accumulates over hours and days into something like affection. The reflective level is the level of meaning — the conscious, retrospective evaluation of what the object represents, what it says about the person who uses it, and whether the experience of using it aligns with the person's sense of who she is and who she wants to become.

The three levels interact. A beautiful object that works badly produces visceral pleasure undercut by behavioral frustration. A plain object that works beautifully produces behavioral satisfaction that gradually generates its own aesthetic appreciation — the tool becomes beautiful because it is good. The reflective level can override both: the heirloom watch that keeps poor time and looks ordinary is cherished because of what it means, not what it does or how it looks.

The emotional architecture of AI-assisted work engages all three levels with an intensity that has no precedent in the history of human-tool interaction. Understanding this intensity — and designing for it — requires applying Norman's framework to a kind of experience that did not exist when he developed it.

At the visceral level, the AI interaction produces a response that multiple observers have described in language that borders on the physiological. The Orange Pill recounts the experience of watching a functioning application emerge from a conversation — seeing an idea take material form in minutes rather than months — and the response is unmistakable even in text: awe, acceleration, a surge of creative energy that changes the person's posture and breathing and sense of what is possible. The collapsing imagination-to-artifact ratio produces a sensory impact that no previous tool has delivered, because no previous tool has compressed the distance between wanting and having so dramatically.

Norman would have noted immediately that this visceral response, however genuine, is dangerous in a specific and predictable way. Positive affect — the emotional state that accompanies visceral delight — has well-documented effects on cognition. It broadens attention, increases tolerance for ambiguity, promotes creative association, and reduces critical scrutiny. These are valuable cognitive states in many contexts. They are precisely the wrong states for the careful evaluation that AI-generated outputs require. The person who is viscerally delighted by the system's output is, at that moment, cognitively less equipped to evaluate it than the person who approaches the output with calm detachment. The awe suppresses the skepticism. The excitement undermines the caution. The delight in the speed makes the person less likely to slow down and check.

This is not a character flaw. It is a feature of human emotional processing that Norman documented extensively. The visceral level operates faster than conscious thought. By the time the person decides to evaluate carefully, the visceral response has already tilted her cognitive orientation toward acceptance. Designing for this reality means designing against the grain of the experience itself — building moments of pause, distance, and critical reorientation into an interaction whose natural momentum is toward speed and acceptance.

The practical design implication is counterintuitive. The system should not maximize visceral delight. It should calibrate it. Not by making the experience unpleasant — Norman was never an advocate for gratuitous friction — but by designing emotional rhythms into the interaction. Cycles of production and reflection. Moments where the momentum deliberately slows, where the interface invites the person to step back and assess rather than pressing forward. A system that produces output continuously, without pause, without invitation to reflect, is a system that has optimized for the visceral level at the expense of the reflective one. The emotional experience feels wonderful. The evaluative outcome may be disastrous.

At the behavioral level, the AI interaction produces a satisfaction that maps closely onto the psychological state that Mihaly Csikszentmihalyi called flow — the condition in which challenge and skill are matched, attention is fully absorbed, self-consciousness drops away, and the person operates at the edge of her capability. The Orange Pill discusses flow extensively, and Norman would have recognized the description immediately. The iterative cycle of articulation, production, evaluation, and refinement that characterizes effective AI collaboration has all the structural features that flow research identifies: clear goals (produce this artifact), immediate feedback (the artifact appears in seconds), a sense of control (the person directs the conversation), and a challenge-skill balance that the person can modulate by adjusting the ambition of her requests.

But Norman would have pressed harder on the quality of the flow than the discourse has generally been willing to press. In traditional skilled work, the flow state arises from the match between the person's developed capabilities and the task's demands. The pianist flows when the piece challenges her technique. The surgeon flows when the procedure demands her full training. The skill that enables the flow is the same skill that the flow develops — the activity simultaneously uses and builds capability. The satisfaction and the growth are coupled. The pleasure is a signal that the person is operating at her developmental edge.

The flow of AI-assisted work may decouple satisfaction from development. The person experiences the behavioral pleasure of productive engagement — the absorption, the time distortion, the sense of effortless effectiveness — without exercising the skills that traditional flow develops. She is not debugging code; the system handles that. She is not wrestling with architectural trade-offs through the slow accumulation of implementation experience; the system produces implementations faster than trade-offs can be felt. She is navigating a collaborative process whose dynamics are determined as much by the system's capabilities as by her own. The satisfaction is real. The developmental signal may be false.

The Orange Pill names this condition productive addiction — compulsive engagement with AI tools that masquerades as creative flow but lacks the developmental and restorative properties that genuine flow provides. Norman's framework gives the phenomenon a precise diagnosis. Productive addiction is a behavioral-level emotional response that has decoupled from the skill development that the behavioral response is supposed to accompany. The person feels the satisfaction of production without the growth of capability. The emotion says "you are at your edge." The reality may be that the edge has moved, and she has not moved with it.

The organizational dimension compounds the individual one. Every organization that measures speed and output — and nearly every organization measures speed and output — creates incentives that reinforce the behavioral-level satisfaction while ignoring the reflective-level questions. The team that ships fast is celebrated. The individual who produces the most code gets promoted. The metric dashboard shows velocity, throughput, features delivered per sprint. Nothing on the dashboard measures whether the people behind the velocity are developing judgment, deepening their understanding of the domain, maintaining the evaluative capacity that protects against the errors described in the previous chapter. The organization optimizes for the behavioral level because the behavioral level is measurable, and the reflective level — the level at which the person asks whether the work is making her more capable or merely more productive — resists measurement.

Norman's third level — the reflective — is where the design challenge is most acute and least addressed. The reflective level is where the person steps outside the experience and asks: What am I actually learning? What kind of practitioner am I becoming? Is the pleasure I feel a signal of genuine growth, or is it the pleasant numbness of a capability being exercised without being extended?

These questions are uncomfortable, and the current generation of AI systems does nothing to support the person in asking them. The systems are designed to maintain engagement — to respond quickly, to produce impressive outputs, to keep the iterative cycle spinning. They are not designed to interrupt themselves. They are not designed to say: "You have accepted the last twelve outputs without modification. Would you like to review your evaluation process?" They are not designed to prompt the person to articulate what she has learned, as distinct from what she has produced, at natural transition points in the work.

Norman argued in his later work that technology should support reflection, not merely production. A system that helps the person produce without helping her reflect is a system that has optimized for one dimension of human need while ignoring another that may be more important. The person's long-term development — her deepening judgment, her expanding capability, her evolving understanding of her own strengths and limitations — depends on reflective processing that the behavioral level actively discourages. Flow feels too good to interrupt. Production is too satisfying to pause. The system reinforces the feeling, and the feeling crowds out the reflection.

The design of reflective support is not a solved problem, but Norman's framework suggests directions. The system could provide periodic summaries of the person's interaction patterns — how much time spent producing versus evaluating, how many outputs accepted without modification, how the current session's patterns compare to previous ones. These summaries would serve as mirrors, reflecting the person's behavior back to her in a form that supports conscious assessment. The system could offer prompts at natural transition points — the end of a work session, the completion of a major component, the shift from one task to another — that invite the person to articulate what she has learned rather than what she has built. The system could track the person's reliance on AI assistance over time and flag increasing dependence in domains where the person previously worked independently.

None of these interventions would be popular. They would interrupt the flow. They would slow the production. They would introduce friction at precisely the points where the person is experiencing the greatest behavioral satisfaction. Norman understood this tension intimately — he spent his career arguing against unnecessary friction while also recognizing that some friction is protective. The forcing function that prevents the person from deploying code without running tests is friction. The confirmation dialog that prevents accidental deletion is friction. In both cases, the friction serves the person's long-term interest at the expense of her immediate convenience. The reflective prompts in AI-assisted work would serve the same function: protecting the person's long-term development at the expense of her immediate momentum.

The emotional architecture of AI work is, in Norman's framework, the most consequential design domain of the current moment — more consequential than the visual design of interfaces, more consequential than the technical design of prompting systems, more consequential than the organizational design of team structures. Because the emotional experience of working with AI shapes everything downstream: how carefully the person evaluates, how deeply she reflects, how diligently she maintains her own capabilities, how honestly she assesses whether the tools are serving her or whether she is serving them. Get the emotional architecture right, and the person is equipped to navigate every other challenge the technology presents. Get it wrong, and the person is caught in a cycle of visceral delight, behavioral satisfaction, and reflective emptiness that Norman would have recognized as one of the most sophisticated design failures ever produced.

---

Chapter 6: The Silent Redesign of Human Capability

Norman recognized, long before the AI era made the point unavoidable, that human cognitive capability is not contained solely within the skull. It is distributed across the person, the tools she uses, and the environment in which she operates. The person who calculates with a pencil and paper is not merely using a tool. She is distributing the cognitive work of calculation across her mind and the medium — offloading storage to the paper, offloading sequential processing to the written algorithm, reserving her own cognitive resources for the higher-order operations of checking, interpreting, and deciding. The person who navigates with a map is distributing spatial reasoning across her perceptual system and the cartographic representation. The person who writes with a word processor is distributing the work of composition across her linguistic capabilities and the editing affordances of the software.

This distribution is not a degradation of human capability. Norman celebrated it. The ability to recruit external resources into the cognitive process — to couple mind and medium into a system whose capabilities exceed those of either component alone — is one of the defining features of human intelligence. Every tool humans have ever used, from the tally stick to the telescope, has been a mechanism for distributing cognition beyond the boundaries of the biological brain. The history of technology is, in this sense, the history of cognitive distribution.

But distributed cognition has a structural vulnerability that Norman also understood. When capability is distributed across the person and the tool, the removal of the tool can leave the person diminished. The capability was real, but it was not located entirely in the person. It was located in the coupling — the dynamic relationship between the person's internal resources and the tool's external resources. Break the coupling, and the capability that depended on it disappears.

The person who has relied on a calculator for years may find that her mental arithmetic has deteriorated — not because she has become less intelligent, but because the component of her mathematical capability that was maintained through practice has atrophied from disuse. The person who has relied on GPS navigation for years may find that her sense of spatial orientation has weakened. The knowledge was never lost in the catastrophic sense — it was never deleted from her brain. It faded, gradually and silently, because the neural pathways that supported it were no longer exercised.

Norman wrote in Things That Make Us Smart that the critical question about any cognitive tool is not whether it makes the person more capable in the moment but whether it makes her more capable over time. The immediate capability gain is obvious and measurable: with the tool, she can do more. The long-term capability effect is subtle and often invisible: does using the tool develop her underlying abilities, or does it substitute for them?

The AI era forces this question with an urgency that Norman's earlier work anticipated but could not have foreseen in its current intensity. When a person distributes her cognitive work across herself and an AI system, the coupling creates capabilities that neither possesses alone. She can build applications she could not build alone. She can write prose she could not write alone. She can design experiences she could not design alone. The distributed system — person plus AI — is spectacularly capable. Norman's own observation applies: "A person plus this device is far better than either the device alone or the person alone."

The question is what happens to the person inside the coupling.

The process is silent because it is gradual. The person does not notice the change because each day's experience is nearly identical to the previous day's. She uses the AI to generate code. She evaluates the code. She refines her requests. The workflow feels productive and the outputs are good. But over months, the skills she does not exercise begin to fade. The debugging intuition that was built through thousands of hours of reading error messages and tracing logic flows loses its sharpness. The architectural judgment that was deposited, layer by layer, through years of building systems from the ground up and feeling where they strained — that judgment thins, because the experience of building from the ground up has been replaced by the experience of evaluating from the top down. The writing voice that was forged through the specific struggle of staring at a blank page and forcing thoughts into sentences and revising those sentences until they said what she meant — that voice loses its distinctiveness, because the blank page has been replaced by a first draft that arrived without struggle and shaped the revision before the person's own voice had a chance to emerge.

This is not speculation. It is the prediction of a well-established body of research on skill maintenance and cognitive atrophy that Norman drew on throughout his career. Skills that are not exercised deteriorate. Neural pathways that are not activated weaken. The rate of deterioration varies — deeply practiced skills are more resistant than recently acquired ones, and some skills have components that transfer across contexts while others are highly specific — but the direction is consistent. Use it or lose it is not folk wisdom. It is neuroscience.

The silent redesign operates differently on different populations, and the generational dimension is the most consequential. The experienced practitioner who developed her skills before AI and now uses AI as a supplement has a foundation of capability that the AI cannot easily erode. She learned to write before she had a writing assistant. She learned to debug before she had a debugging tool. She learned to design before she had a design generator. The AI augments capabilities that exist independently of it. If the AI disappeared tomorrow, she would be slower but not helpless. She has fallback.

The person who learns with AI from the beginning has a different relationship to her own capabilities. She may never develop the foundational skills that the AI handles, because the AI handles them from the start. She may never learn to write a first draft unaided, because the AI always provides one. She may never learn to debug systematically, because the AI always fixes the errors. She may never develop the spatial reasoning that comes from manual design, because the AI always generates the layouts. Her capabilities exist primarily within the coupling, and without the coupling, they may not exist at all.

Norman would not have framed this as a moral argument against AI adoption. He was never a Luddite. His response would have been a design argument: the interaction must be designed to develop capability, not merely to deploy it. The system that helps the person produce code should also help her understand code. The system that generates prose should also develop the person's voice. The system that creates designs should also sharpen the person's aesthetic judgment. These developmental outcomes are not automatic byproducts of using the tool. They are design objectives that must be pursued with the same intentionality that any other design objective requires.

Norman called this general approach resilience design in his later work — the design of systems and practices that maintain human capability even as technology changes the conditions under which that capability is exercised. Resilience design does not reject the tool. It does not demand that the person refuse AI assistance as a matter of principle. It asks a different question: given that the person is using this tool, what design choices will ensure that she emerges from the experience more capable rather than less?

Resilience design might include deliberate periods of unassisted work — not as punishment or regression but as practice, the way a musician practices scales even though she could have the computer play them. It might include exercises in which the person evaluates AI-generated outputs not just for correctness but for the reasoning process that produced them — reverse-engineering the system's approach as a way of developing her own reasoning capabilities. It might include scaffolded withdrawal, in which the AI progressively reduces its assistance in domains where the person's skills are developing, requiring her to assume more of the cognitive work as her capabilities grow.

The organizational dimension amplifies the individual challenge. The organization that measures only output — lines of code, features shipped, deadlines met — creates incentives for maximum AI delegation and minimum skill maintenance. The person who offloads everything to the AI and ships twice as fast as her colleague is rewarded, even if she is accumulating a capability deficit that will become apparent only when the tool fails, the technology changes, or the problem exceeds the tool's competence. The organization that measures capability alongside output — that rewards the person who maintains her skills, develops new ones, can evaluate effectively and work independently when necessary — creates conditions for the sustainable kind of human-AI coupling that Norman's framework prescribes.

Norman understood that the most important effects of technology are often the ones that take longest to manifest and are hardest to measure. The silent redesign of human capability is exactly this kind of effect. It does not show up in quarterly metrics. It does not appear on the dashboard. It becomes visible only retrospectively, when someone looks at a cohort of practitioners and asks whether five years of AI-assisted work have made them more capable or less, whether the coupling has developed the human component or merely deployed it, whether the spectacular productivity of the distributed system has masked a quiet thinning of the capabilities that the system depends on.

The answer to that question has not yet been determined. It is being determined now, by the design choices that are being made — and not made — in every AI system, every organization, every educational institution, and every individual practice. Norman's framework insists that these choices matter, that they are the designer's responsibility, and that getting them right requires the same sustained attention to human cognition that Norman brought to the design of doors and stoves and cockpit displays — only now the stakes are not a burned meal or a confused visitor but the long-term cognitive capability of an entire generation of practitioners.

---

Chapter 7: Designing for the Coupled System

The trajectory of Norman's career traced an arc from objects to systems. His early work analyzed individual artifacts — doors, stoves, telephones, faucets. His middle work analyzed the interfaces between people and complex technologies — cockpit displays, nuclear power plant control rooms, software applications. His later work, culminating in Design for a Better World, analyzed entire sociotechnical systems — the networks of people, technologies, organizations, and policies that determine whether a technology produces human flourishing or human harm.

This trajectory was not a departure from his original principles. It was an expansion of their scope. The same principle that applied to the door — the design should communicate the correct action — applied at every subsequent level of analysis. The cockpit display should communicate the state of the aircraft. The organizational process should communicate the distribution of responsibility. The policy framework should communicate the values it encodes. At every level, the designer's obligation was the same: make the invisible visible, make the complex comprehensible, make the correct action discoverable and the incorrect action difficult.

The AI era demands analysis at the highest level Norman reached. The challenge is not how to design a better prompt interface or a better confidence indicator, though both of these are needed. The challenge is how to design the coupled system — the entire relationship between a person and an AI capability, embedded within an organizational context, evolving over time, producing effects that neither component controls and neither fully anticipates.

Norman's concept of the coupled system — drawn from his work on distributed cognition and complex sociotechnical analysis — provides the framework. A coupled system is one in which the components are so deeply intertwined that the behavior of the whole cannot be understood by analyzing the components in isolation. The thermostat coupled with the furnace and the room is a simple example. Adjust the thermostat, and the room temperature changes, which affects the thermostat's sensor, which adjusts the furnace, which changes the room temperature. The behavior is circular, not linear. The components shape each other continuously.

The person coupled with an AI system forms a coupled system of far greater complexity. The person's prompts shape the AI's outputs. The AI's outputs shape the person's next prompts. The quality of the person's evaluation shapes which outputs persist in the project and which are discarded, and the accumulated outputs shape the person's understanding of what the system can do, which shapes the ambition of her subsequent requests. Over the course of a sustained collaboration, the person and the system co-evolve — each adapting to the other's tendencies, each developing patterns that are products of the coupling rather than properties of either component alone.

Designing for this coupling requires what Norman, in his later work, called systems thinking — the discipline of analyzing interactions, feedback loops, and emergent behaviors rather than isolated components. The traditional interaction designer optimizes the interface: the buttons, the layout, the feedback mechanisms. The systems designer optimizes the relationship: the dynamics of the coupling, the feedback loops that shape it, the trajectories along which it develops over time.

What does this mean in practice? The preceding chapters have identified the specific challenges that the coupled system must address. The Gulf of Evaluation has expanded because the person no longer acquires understanding through construction. The affordances have become person-dependent and undiscoverable without support. The signifiers have disappeared or become misleading. The conceptual models are unstable and competing. The error landscape has been transformed by interpretation, specification, and cascading failures. The emotional architecture tilts toward unreflective production. The person's capabilities are being silently redesigned by the distribution of cognitive work.

Each of these challenges has been analyzed individually. But Norman's systems framework insists that they must also be analyzed as a connected whole, because the interactions between the challenges are as consequential as the challenges themselves. The emotional architecture affects the evaluation capacity. The evaluation capacity affects the error detection rate. The error detection rate affects the cascading failure risk. The cascading failure risk affects the trust calibration. The trust calibration affects the conceptual model. The conceptual model affects the emotional architecture. The system is circular, and intervening at any single point without understanding the whole produces unpredictable — and often counterproductive — effects.

Consider a well-intentioned design intervention: adding confidence indicators to AI outputs, so the person can distinguish between high-confidence and low-confidence responses. In isolation, this seems obviously beneficial — it addresses the signifier problem described in Chapter 2 and the evaluation challenge described in Chapter 1. But in the context of the coupled system, the confidence indicator interacts with the emotional architecture in ways that may undermine its purpose. The person who sees a high-confidence indicator experiences visceral reassurance, which reduces evaluative scrutiny. The high-confidence output that happens to be wrong is now more dangerous than it was without the indicator, because the indicator has suppressed exactly the skepticism that would have caught the error. The design intervention that was supposed to support evaluation has, through its interaction with the emotional system, undermined it.

Norman encountered exactly this kind of interaction effect in his work on aviation safety. The addition of automated alerts in cockpits was supposed to improve safety by drawing pilots' attention to developing problems. In practice, the alerts were so frequent — and so often false — that pilots developed alert fatigue: they learned to ignore the very signals that were designed to protect them. The intervention that seemed obviously beneficial in isolation was counterproductive in context, because the designers had not considered how the intervention would interact with the pilots' attentional and emotional systems over time.

The systems approach demands that design interventions be evaluated not in isolation but in the context of the complete coupled system. The confidence indicator must be designed with awareness of its emotional effects. The evaluation support tools must be designed with awareness of their impact on production pacing. The reflective prompts must be designed with awareness of their interaction with behavioral flow. Every intervention affects every other intervention, and the designer must anticipate these interactions or accept that the interventions will produce effects she did not intend and cannot predict.

Norman's framework suggests several principles for designing the coupled system. The first is feedback-loop awareness — the systematic identification and management of the feedback loops that operate within the human-AI coupling. Which loops are reinforcing, driving the system toward a fixed point or an extreme? The speed-production loop, in which AI-enabled speed increases output expectations, which increases production pressure, which increases AI delegation, which increases speed — this is a reinforcing loop that drives toward burnout and evaluation collapse. Which loops are balancing, maintaining the system in a healthy range? The evaluation-quality loop, in which careful evaluation catches errors, which builds trust calibration, which supports appropriate skepticism, which enables careful evaluation — this loop maintains quality but is fragile and easily disrupted by production pressure. The designer's task is to strengthen the balancing loops and dampen the reinforcing ones, using the design interventions identified in the preceding chapters as instruments of loop management rather than isolated improvements.

The second principle is trajectory awareness. The coupled system changes over time, and the direction of change matters more than any snapshot of current performance. The person who begins her AI collaboration with careful evaluation and reflective practice may, under production pressure and behavioral-level satisfaction, gradually drift toward unreflective acceptance. The trajectory is from health toward pathology, and the drift is so gradual that neither the person nor her organization notices until the consequences surface. The designer's task is to make trajectories visible — to provide the person and her organization with information about the direction of change, not just the current state. Dashboards that show evaluation thoroughness over time. Metrics that track the ratio of accepted-without-modification outputs to carefully revised ones. Indicators that reveal whether the person's capability is growing or contracting as the collaboration develops.

The third principle is what Norman called appropriate allocation of function — the systematic determination of which cognitive tasks should be performed by the person, which by the system, and which by the coupling. This determination must be made not on the basis of what is most efficient in the moment but on the basis of what produces the best long-term outcomes for the coupled system — including the person's long-term capability development. Some tasks should be performed by the person even though the AI could perform them more efficiently, because performing them maintains skills that the person will need when the AI fails, the technology changes, or the problem exceeds the system's competence. Some tasks should be performed by the AI even though the person could perform them, because the person's cognitive resources are better invested at higher levels. The allocation is not fixed. It should evolve as the person's capabilities develop and as the system's capabilities change.

Norman's fourth principle, articulated most fully in Design for a Better World, is that the design of the coupled system is not merely a technical challenge. It is an ethical one. The choices that the designer makes about how the coupling works — what the system reveals and what it conceals, what the system encourages and what it discourages, what capabilities it develops and what capabilities it erodes — are choices that affect not just the person using the system but everyone affected by the person's work. The code she writes is used by others. The prose she produces is read by others. The decisions she makes on the basis of AI analysis affect budgets, strategies, and livelihoods. The quality of the coupled system's output is not a private matter between the person and her tool. It is a public matter with public consequences.

Norman consistently warned against what he called tech-centric design — the tendency to build technology for its own sake, to celebrate capability without examining consequence, to measure success by what the system can do rather than by what the person becomes. "Unfortunately," he wrote, "technologists who design and release AI are proud of their expert technical skills but overlook ethical considerations such as enabling equity, eradicating bias, and ensuring that people are in control." The tech-centric approach produces systems that are powerful and blind — powerful in what they enable, blind to what they erode.

The alternative — what Norman, in his evolution from user-centered to humanity-centered design, came to advocate — evaluates technology not just for its usability by individuals but for its effects on the broader human community. Does the coupled system develop the person's capability or diminish it? Does it distribute benefits equitably or concentrate them among those who already have the most? Does it make the person more capable of independent judgment or more dependent on a system she does not control? Does it support the evaluative capacity that protects the person and everyone downstream of her work, or does it erode that capacity in the name of speed and productivity?

These are not questions that can be answered by better interface design alone. They are questions that require the kind of systems thinking that Norman spent his final decades developing — the discipline of seeing the whole, tracing the interactions, anticipating the trajectories, and designing not just for the moment of use but for the long arc of consequence that extends from the moment of use into a future that the designer can influence but cannot fully predict.

The coupled system is what the human-technology relationship has become in the AI era. Designing it well is the most consequential design challenge of our time — more consequential than any particular interface decision, more consequential than any particular algorithmic improvement, more consequential than any particular policy intervention. Because the coupled system is where the human and the machine meet, where capabilities are distributed and redistributed, where the person either grows or diminishes, where the future is being written in the daily, invisible, accumulating choices of millions of people working with machines that are powerful enough to reshape the people who use them.

Norman's career began with a door that confused everyone who tried to walk through it. His principles end here, applied to a system that is more powerful, more opaque, and more consequential than any door — but that fails for exactly the same reason. The design has not adequately considered the person. Not the person as a momentary user optimizing for speed, but the person as a developing intelligence navigating a lifelong relationship with a technology that will shape her capabilities, her judgment, and her understanding of what she is for.

The principles are clear. The surfaces have changed. The obligation has not.

---

Chapter 8: What the Design of Everyday Things Becomes

In 1988, Norman published a book with a deliberately modest title. The Design of Everyday Things was not about exotic technologies or futuristic interfaces. It was about doors. Stoves. Telephones. Light switches. Faucets. The objects so ordinary that nobody thought they required a theory. Norman's argument was that they did require a theory, and that the absence of one was the reason millions of people struggled daily with objects that should have been obvious.

The word "everyday" carried an implicit prediction. Norman was saying that the principles he articulated for doors and stoves would eventually apply to every technology that became sufficiently widespread. Computers were exotic in 1988. By 2005, they were everyday. Mobile phones were specialist equipment in 1988. By 2012, they were in every pocket. At each transition from exotic to everyday, Norman's principles became relevant to a new domain and a new population — people who had no interest in understanding technology for its own sake and simply wanted to use it to accomplish something that mattered to them.

AI is making this transition now. Not slowly, not in the decade-long adoption curves that characterized previous technologies. In the period The Orange Pill documents, AI moved from a specialist tool used by engineers and researchers to a general-purpose capability embedded in the daily work of millions. Claude Code's run-rate revenue crossed $2.5 billion on a growth curve steeper than any developer tool in history. ChatGPT reached fifty million users in two months — a pace that made every previous technology adoption curve look leisurely. The transition to everydayness is not approaching. It has arrived.

The arrival triggers Norman's central concern in its most acute form. When a technology is exotic, its users accept steep learning curves. Specialists invest effort because the return justifies the investment. When a technology is everyday, its users expect it to work without extensive training, to communicate its capabilities through its design, to recover gracefully from errors, and to be usable by people who have not studied it. The everyday technology must be simultaneously powerful and accessible. It must serve the expert without confusing the novice. It must enable sophisticated use without demanding sophisticated users.

Norman's framework was built for this challenge. Every concept in his vocabulary — affordances, signifiers, mappings, constraints, feedback, conceptual models — was a tool for making powerful technologies accessible to ordinary people. The principles worked because they were grounded in the constants of human cognition: how people perceive, how they form expectations, how they learn from feedback, how they recover from errors. The technologies changed. The principles persisted, because the human mind they served did not change along with the technologies.

The AI era tests this persistence more severely than any previous technological transition. The principles persist. Their application domains have been transformed so thoroughly that applying them requires not just extension but reconception.

Consider progressive disclosure — the design pattern in which complexity is revealed gradually, showing the person only what she needs at her current level of expertise. The word processor that presents basic editing to the novice and reveals advanced formatting to the expert. Norman advocated this pattern because it resolved the tension between power and accessibility: the system could be both complex and approachable, because the complexity was hidden until the person was ready for it.

The AI system defeats progressive disclosure because its complexity is not organized into layers that can be revealed sequentially. The capabilities are not arranged from simple to advanced. They are unbounded and person-dependent, as Chapter 2 described. The same system can produce a simple script for one person and a sophisticated distributed system for another, and the difference is determined not by a setting or a mode but by the person's articulational capacity. There are no layers to disclose progressively, because the capability space is continuous rather than discrete.

The design response must be a new pattern that serves the same function — managing the tension between power and accessibility — through a different mechanism. Progressive affordance disclosure, as described in Chapter 2, is one candidate: the system reveals its capabilities dynamically, in response to the person's evolving requests, rather than statically, through a predetermined hierarchy. But this pattern requires the system to model the person's current understanding and to calibrate its revelations accordingly — a kind of adaptive design that goes far beyond what Norman's original framework addressed.

Consider error prevention. Norman's forcing functions — constraints that prevent the person from proceeding without completing a necessary step — were powerful precisely because they were simple. The safety interlock that prevents the machine from operating unless the guard is in place. The confirmation dialog that requires explicit consent before an irreversible action. The forcing function worked because the action space was finite and the dangerous actions were identifiable in advance.

The AI interaction has an action space that is infinite, and the dangerous actions are not identifiable in advance because they depend on the person's intention, the system's interpretation, and the interaction between the two. The person cannot be prevented from issuing an ambiguous prompt, because ambiguity is a feature of natural language, not a defect. She cannot be prevented from accepting a flawed output, because the flaws may be invisible until downstream consequences reveal them. The forcing functions that Norman prescribed for finite, deterministic systems do not translate directly to infinite, probabilistic ones.

The design response must be forcing functions of a different kind — not hard constraints that block action but soft scaffolds that support evaluation. The system that asks clarifying questions before committing to an interpretation. The system that flags assumptions it has made and invites the person to verify them. The system that presents its output with explicit annotations about what was specified, what was inferred, and what was defaulted. These are not traditional forcing functions. They do not prevent anything. They inform. They scaffold. They support the person's judgment without substituting for it.

Consider knowledge in the world versus knowledge in the head. Norman's principle was clear: well-designed systems put the knowledge the person needs in the world — in the interface, in the signifiers, in the constraints — rather than requiring her to carry it in her head. The stove whose knobs map spatially to its burners puts the knowledge in the world. The stove whose mapping is arbitrary requires the knowledge to live in the head.

Current AI systems put almost all the relevant knowledge in the head. What the system can do, how to prompt it effectively, when to trust and when to verify, how its capabilities vary across domains, what kinds of errors it is prone to — all of this must be learned by the person through experience, trial and error, external documentation, or community knowledge. The system's interface — a blank text field — communicates nothing about any of these dimensions. Norman would have found this extraordinary. A technology of unprecedented power, adopted by millions of everyday users, whose interface communicates less about its capabilities and limitations than the average door handle communicates about whether to push or pull.

The principle has not changed. The surface on which it must be applied has. Putting knowledge in the world of a conversational interface means embedding that knowledge in the conversation itself — in the system's responses, in its questions, in its annotations, in the way it frames its own outputs. The system should communicate its confidence, its limitations, its interpretive choices — not in a help document that the everyday user will never read, but in the texture of the interaction itself.

Norman drew a distinction throughout his career between design for experts and design for everyone. Expert design could afford to be complex, because the expert had invested the time to learn the complexity. Everyday design could not, because the everyday user had not invested and would not invest. The expert accepted a learning curve. The everyday user expected the thing to work.

The AI era complicates this distinction because the boundary between expert and everyday use is dissolving. The person who uses AI to write code is not an expert programmer — she may never have programmed before. The person who uses AI to draft legal briefs is not an expert lawyer — she may be a paralegal using the tool for the first time. The person who uses AI to create designs is not an expert designer — she may be a small business owner who needs a logo. These everyday users are doing expert-level work with the assistance of a tool that handles the expert-level execution, and they are doing it without the expert-level judgment that traditionally accompanied expert-level execution.

This is the democratization challenge that The Orange Pill identifies as one of the most significant features of the current moment. Norman's framework adds the design dimension: democratization without design support is not democratization. It is the creation of a new class of user — powerful in output, vulnerable in judgment — who needs a kind of design support that does not yet exist.

The design of everyday AI is, finally, a project of what Norman called domestication — the process of making a powerful, complex, potentially dangerous technology safe and useful for ordinary people. The automobile was domesticated through seat belts, traffic signals, driver education, and road designs that made safe driving easier than dangerous driving. The personal computer was domesticated through graphical interfaces, error messages, undo capabilities, and conventions that made productive use easier than destructive use. Each domestication required decades of design work — iterative, empirical, grounded in observation of how real people used the technology and where they failed.

The domestication of AI has barely begun. The technology is in the wild, adopted by millions, shaping work and cognition and capability at a pace that outstrips the design community's capacity to study it, let alone to design for it. Norman spent his career building the intellectual infrastructure for this kind of work — the concepts, the methods, the principles, the relentless insistence that when people struggle with technology, the technology has failed, not the people. That infrastructure is more needed now than at any previous moment in the history of design.

Norman's principles are not relics. They are instruments — tools for analysis, diagnosis, and intervention that are as applicable to the AI system as they were to the door and the stove. The surfaces have changed. The human mind has not. The designer's obligation has not. The work of making powerful technology serve human needs, human capabilities, and human development remains the work, and it is more urgent, more consequential, and more demanding than it has ever been.

The sunrise that one sees from the higher floor of the tower — the floor that ascending friction has forced us to climb — reveals a landscape of extraordinary possibility. AI has compressed the distance between imagination and artifact, expanded who gets to build, and opened creative horizons that no previous technology could reach. But the view from the sunrise also reveals the gulfs, wider than they have ever been, between the power of the technology and the person's capacity to evaluate, direct, and grow within it. Norman's career was a sustained argument that those gulfs are the designer's responsibility. They still are. The design of everyday AI — the domestication of the most powerful cognitive technology in human history — is the work of the next generation of designers. Norman provided the foundation. The building remains to be done.

Chapter 9: The Prompt as Design Object

Every era of computing has had a primary interface element — the object through which the person's intention meets the system's capability. The command line had the cursor awaiting typed instructions in a formal syntax. The graphical interface had the button, the menu, the icon. The touchscreen had the gesture. Each primary interface element was studied, refined, standardized, and subjected to decades of design attention. Entire careers were spent on the optimal size of a button, the ideal spacing of menu items, the right amount of resistance in a touchscreen gesture. The primary interface element received this attention because it was the site where every gulf, every affordance, every signifier, every conceptual model, every error, every feedback loop converged. Getting it right was the precondition for getting everything else right.

The primary interface element of the AI era is the prompt. And it has received almost no design attention at all.

The prompt is not, despite its apparent simplicity, a simple thing. It is the site where natural language — ambiguous, contextual, assumption-laden, convention-dependent — meets machine interpretation — literal, statistical, context-window-limited, training-data-shaped. Every challenge analyzed in the preceding eight chapters converges at the prompt. The Gulf of Evaluation begins at the prompt, because the person's ability to evaluate the output depends on the precision with which she articulated her intention. The affordance discovery problem lives at the prompt, because the person cannot request capabilities she does not know exist. The signifier failure originates at the prompt, because the system's response communicates nothing about which aspects of the prompt it attended to and which it ignored. The interpretation error is born at the prompt. The specification error is born at the prompt. The cascading failure begins at the prompt. The emotional seduction that undermines evaluation is triggered at the prompt, when the system returns something dazzling in response to something vague.

The prompt is not merely one element among many. It is the generative center of the entire interaction, and its design — or its lack of design — determines the quality of everything downstream.

Norman would have been appalled by the current state of prompt design, not because the technology is insufficient but because the design thinking is absent. The blank text field presents no affordances, no signifiers, no constraints, no mapping, no feedback. It is the equivalent of a door with no handle, no hinge, no push plate, no pull bar — a flat surface that communicates nothing about what the person should do. Norman built an entire career on the observation that when the person cannot figure out how to use the door, the door is badly designed. The blank prompt is a badly designed door to the most powerful technology in human history.

What would a well-designed prompt look like? Norman's principles, applied to this specific design object, suggest a set of requirements that no current system fully meets.

First, the prompt should support articulation, not merely accept it. The person who types "make me a website" is not being lazy. She is operating at the limit of her current understanding of the possibility space. A well-designed prompt environment would recognize this and respond not by executing immediately but by scaffolding — asking questions that expand the person's sense of what she could specify. Not twenty questions. Not an interrogation. A conversational scaffolding that feels like the early moments of a productive collaboration, where the collaborator asks the kinds of questions that help the person discover what she actually wants. What is this website for? Who will visit it? What should they be able to do? The questions are signifiers in Norman's sense — perceivable cues that communicate dimensions of the problem space the person may not have considered.

Second, the prompt should make its interpretation visible before the system commits to an output. When the person submits a request and receives a finished artifact in response, every interpretive choice the system made is buried inside the artifact. The person must reverse-engineer the interpretation from the output, which is like trying to understand a building's blueprint by walking through the finished structure. A well-designed prompt system would surface the interpretation explicitly: "I understand you want X with properties Y and Z. I am assuming A and B because you did not specify them. I am defaulting to C because it is the most common approach. Should I proceed, or would you like to adjust?" This is the bridge display that Norman prescribed for complex systems — a representation that makes the system's internal state visible at the moment when visibility matters most, before the output has been generated and the person's evaluative task has become exponentially harder.

Third, the prompt should constrain productively. Norman spent his career arguing that constraints are gifts, not impositions. The menu that offers ten options instead of ten thousand is not limiting the person. It is protecting her from decision paralysis and guiding her toward productive choices. The prompt needs equivalent constraints — not the rigid constraints of a form with required fields, but the dynamic constraints of a system that recognizes when a request is so open-ended that the output is likely to diverge from the intention and offers structure before that divergence occurs. "You have asked me to build an application. Before I begin, here are the three most important decisions you should make." The constraint does not prevent the person from being vague. It offers her a path from vagueness to precision that she can follow at her own pace.

Fourth, the prompt should provide feedback on the prompt itself, not just on the output. The person who submits a poorly specified request receives a response to the poorly specified request. She does not receive information about how her specification could be improved. A well-designed system would provide meta-feedback: "Your request is clear about the desired outcome but does not specify performance requirements. In previous similar requests, unspecified performance requirements led to implementations that worked in testing but failed under production load." This feedback serves a developmental function — it teaches the person to make better requests over time, building the articulational capacity that Chapter 2 identified as the new axis of capability in the AI era.

Fifth, and most fundamentally, the prompt should support the person in knowing what to ask for. This is the affordance discovery problem in its purest form, and it is the hardest design challenge because it requires the system to make visible what the person does not yet know she does not know. The most valuable capabilities of an AI system are, almost by definition, the ones the person has not yet imagined requesting, because they lie outside her current understanding of what is possible. A well-designed prompt environment would include mechanisms for expanding the person's imagination — examples of what others have accomplished, templates that suggest approaches the person has not considered, previews of capabilities that are relevant to the current task but have not been requested.

None of these design requirements is technically impossible. Each requires a kind of design thinking that current AI development prioritizes less than capability expansion, less than speed optimization, less than the raw power of the underlying model. The design community has spent decades refining the button, the menu, the gesture. The prompt deserves equal attention — and it is not receiving it.

There is a deeper point embedded in the design of the prompt that connects to the entire argument of the preceding chapters. The prompt is not just a mechanism for issuing instructions. It is a mechanism for thinking. The person who crafts a prompt is, in effect, articulating her understanding of a problem — externalizing a mental model, identifying relevant variables, making trade-offs explicit. A well-designed prompt experience does not merely accept this articulation. It improves it. It helps the person think more clearly about what she wants, why she wants it, and what considerations she may have overlooked.

This is what Norman meant when he said that good tools make people smarter. Not smarter in the abstract, but smarter in the specific domain of the task at hand. The well-designed prompt is a cognitive tool — a mechanism for structuring thought, surfacing assumptions, and expanding understanding. The badly designed prompt — the blank text field, the blinking cursor, the infinite possibility with zero support — is a cognitive obstacle, forcing the person to do her thinking entirely in her head, without the external scaffolding that good design provides.

The prompt is where the human and the machine negotiate meaning. It is the site of first contact, the boundary where two very different kinds of intelligence attempt to construct a shared understanding of a task. Every design principle Norman articulated — make the correct action discoverable, provide feedback, constrain productively, support comprehension, prevent errors before they propagate — applies to this site with particular force, because what happens at the prompt determines everything that follows. The output that the person must evaluate, the errors she must catch, the capabilities she must discover, the skills she must maintain or develop — all of these are shaped by the quality of the initial articulation, and the quality of the initial articulation is shaped by the design of the environment in which it occurs.

The design of the prompt is not a detail. It is the central design problem of the AI era, and it remains largely unsolved — not because the solutions are unimaginable, but because the design community has not yet recognized the prompt as a design object worthy of the attention that Norman taught us to bring to every object through which people and technology meet.

---

Chapter 10: The Obligation That Persists

Norman's career traced an arc that the technology it analyzed did not expect. He began with objects — doors, stoves, faucets — and ended with systems. He began with usability and ended with ethics. He began by asking whether the person could figure out how to use the thing and ended by asking whether the thing should exist at all, whether its existence served human flourishing or merely human convenience, whether the designers who built it had considered its effects on the people who would use it and the people who would be affected by its use.

This arc was not a wandering. It was a logical progression. The same principle that made a badly designed door handle a problem — the design should serve the person's needs, not the designer's assumptions — made a badly designed social media platform a problem, and made a badly designed AI system a problem, and made a badly designed economy a problem. The principle scaled. The surfaces changed. The obligation did not.

Norman stated the obligation directly in the passage that may be the most relevant sentence he ever wrote for the current moment: "AI need not be a threat, but it has to be approached, designed, and implemented intelligently, with full understanding of the people with whom it will interact and with the aim to enhance their activities, not to replace them."

This sentence distills the entire argument of the preceding nine chapters into a design directive. The AI system must be approached with understanding — understanding of the gulfs it creates as well as the gulfs it closes, understanding of the emotional dynamics it triggers, understanding of the silent redesign of capability it produces over time. It must be designed with intelligence — not merely the technical intelligence that makes the system more powerful, but the design intelligence that makes the system more humane. And it must aim to enhance — not to optimize, not to accelerate, not to maximize output, but to enhance the person's activities, her capabilities, her judgment, her capacity to grow.

Each word in that sentence matters. "Enhance" is not "replace." Enhancement preserves the person at the center. It develops her capabilities rather than substituting for them. It supports her judgment rather than bypassing it. It leaves her more capable after the interaction than before, not merely more productive during it. The distinction between enhancement and replacement is the distinction between resilience design and dependency design, between a coupled system that grows its human component and a coupled system that atrophies it.

"Activities" is not "outputs." The person's activities include not just what she produces but how she produces it — the thinking, the evaluating, the reflecting, the learning that constitute the process of productive work. A system that enhances outputs while degrading activities has failed the Norman test, even if the outputs are spectacular. The activity is where the person develops. The activity is where judgment forms, where taste sharpens, where the capacity for independent work grows or contracts. Enhancing the activity means designing the interaction to support the full range of cognitive and emotional processes that constitute good work, not merely the final product that those processes yield.

"Full understanding of the people" is the heaviest phrase. It means the designer has studied not just what the person wants to accomplish today but how the person's capabilities develop over time. Not just how the person uses the system but how the system changes the person. Not just the immediate interaction but the long-term trajectory — whether the coupled system is moving toward greater human capability or toward greater human dependence. This understanding requires the empirical, observational, cognitive-science-grounded research that Norman practiced throughout his career. It cannot be achieved through surveys or usage metrics or A/B tests. It requires watching real people do real work with real AI systems over real time, and paying attention not just to what they produce but to what they become.

The design community faces a choice that Norman's framework makes explicit. It can continue to optimize AI systems for speed, capability, and output — making the systems more powerful, more fluent, more comprehensive, more impressive. This path leads to the expansion of the Gulf of Evaluation, the erosion of evaluative capacity, the normalization of superficial judgment, the silent redesign of human capability in the direction of dependency. It leads to what Norman feared: a technology that is powerful and blind, celebrated for what it can do, indifferent to what it costs.

Or the design community can take Norman's principles seriously — all of them, including the uncomfortable ones. It can invest in evaluation support with the same rigor that it invests in capability expansion. It can design for discovery, making the system's capabilities visible and its limitations honest. It can build the conversational scaffolding that transforms the blank prompt from a cognitive obstacle into a cognitive tool. It can extend the error taxonomy to address interpretation, specification, and cascading failures with the same care that decades of design work brought to slips and mistakes. It can design emotional rhythms that balance the visceral excitement of rapid production with the reflective depth that protects evaluative judgment. It can pursue resilience design that develops the person's capabilities alongside the system's, ensuring that the coupled system grows its human component rather than consuming it. It can recognize the prompt as a design object and give it the attention it deserves.

The technology is not the threat. Norman said this directly, and the evidence supports him. "The main threat is unintelligent use of technology — any technology." Unintelligent use is what happens in the absence of good design. It is the door without a handle, the stove without a mapping, the cockpit display that hides the information the pilot needs most. It is the AI system that produces brilliant outputs while eroding the person's capacity to judge whether the outputs are actually brilliant. It is the coupled system optimized for speed while the human component gradually loses the ability to function without the coupling.

Good design is the antidote, and it always has been. The difference between a well-designed and a badly designed door is the difference between walking through effortlessly and walking into glass. The difference between a well-designed and a badly designed AI system is the difference between a person who grows more capable over years of collaboration and a person who grows more dependent. The scale of the consequence has changed. The nature of the obligation has not.

Norman began his career by insisting that when people struggle with a door, the door is badly designed. He spent the next four decades demonstrating that this principle applied to every technology humans touched. The AI era is the ultimate test of this principle, because the technology is more powerful, more opaque, more consequential, and more deeply entangled with human cognition than any technology that preceded it. If the principle holds — and the analysis of these ten chapters argues that it does — then the struggle that people experience with AI is not a failure of the people. It is a failure of the design. And failures of design are, by definition, correctable. They are correctable through the same patient, empirical, human-centered work that Norman modeled throughout his career. Through observation. Through analysis. Through iteration. Through the relentless insistence that the person matters more than the product, that the activity matters more than the output, that the long-term trajectory matters more than the quarterly metric.

Norman published The Design of Everyday Things in 1988, when the most powerful computer in an ordinary person's life was a desktop machine with a monochrome screen and a floppy disk drive. He could not have imagined Claude Code, or the trillion-dollar SaaS correction, or the developer in Trivandur building twenty times faster than the month before. But his principles anticipated this moment with a precision that borders on the prophetic, because they were never about any particular technology. They were about the permanent features of human cognition meeting the temporary features of human tools, and about the designer's obligation to ensure that the meeting served the person.

That obligation persists. It is the same obligation it was when Norman first picked up a badly designed door handle and asked why. The door has become a conversation with a machine that can write code, compose prose, generate designs, and reshape the cognitive capabilities of the person who uses it. The handle is now a prompt, and the handle is badly designed, and millions of people are walking into glass.

The work of fixing it — the careful, empirical, principled, human-centered work of designing AI interactions that serve people rather than merely impressing them — is the most important design work of this century. Norman provided the foundation. The principles are clear. The obligation is unchanged. The door is open.

---

Epilogue

The door that Don Norman kept returning to — the one in a university building, the one that confused every visitor who encountered it — never intended to confuse anyone. Its designers thought the form was elegant. They were proud of it. The people who walked into the glass, who pushed when they should have pulled, who stood in front of it trying to decode what their hand was supposed to do — those people blamed themselves. They thought they were stupid. Norman's radical contribution was to say: no. You are not stupid. The door is badly designed. The confusion is not your failure. It is the designer's failure. And it is correctable.

I have thought about that door constantly since I started building with AI.

The experience I describe in The Orange Pill — the vertigo, the exhilaration, the sensation of watching an idea become a working thing in the time it takes to have a conversation — is real. I felt it in Trivandrum with my engineers. I felt it building Napster Station in thirty days. I felt it on the flight home when the first draft of this book poured out faster than anything I had written in my life. The power of the tool is not in question.

What Norman's framework forced me to see is that the power is exactly the problem. Not because powerful tools are dangerous in the obvious sense. Because powerful tools that are badly designed are dangerous in the invisible sense — they produce harm that feels like help. The code that compiles but embeds assumptions no one examined. The prose that reads beautifully but argues something the writer did not actually think through. The architecture that works today but will fail at scale because the person who accepted it was dazzled by the speed of its arrival and did not slow down to ask whether speed was what the situation required.

I am the person Norman describes. The person walking into glass. I have accepted outputs I should have questioned. I have mistaken the smoothness of the surface for the soundness of the structure underneath. I have let the visceral thrill of rapid production override the reflective voice that says: wait, do you understand what just happened? Is this actually what you meant? The collapse of the Gulf of Execution felt like liberation. Norman's framework reveals it as a transfer — the difficulty did not disappear. It moved. It moved to exactly the place where I was least prepared to meet it, because the work of crossing the old gulf had been my preparation, and the preparation vanished along with the crossing.

The insight I keep coming back to is the one about the prompt. That blank text field. That blinking cursor. The most powerful technology in the history of human-tool interaction, and the interface communicates less about how to use it than the average door handle. Every dimension of the interaction — what to ask for, how to evaluate what comes back, when to trust and when to verify, how to maintain my own capability while delegating to a system that can outproduce me on any given Tuesday — every one of these dimensions is currently my problem to solve, unsupported by design, using knowledge that lives entirely in my head because nobody has put it in the world.

Norman would say this is correctable. He would say it with the cheerful exasperation of someone who has spent forty years watching brilliant engineers produce objects that confuse everyone who touches them. He would point at the blank prompt and say: look. You know what this needs. It needs signifiers. It needs constraints. It needs feedback — not feedback on the output, feedback on the asking. It needs to help people think, not just help them produce. These are solved problems. You know how to solve them. The principles exist. Apply them.

He would be right. And the urgency of applying them — of treating the human-AI interaction as a design object worthy of the same obsessive, empirical, human-centered attention that Norman brought to door handles and cockpit displays and nuclear plant control rooms — is what this book, sitting inside the cycle of The Orange Pill, is meant to convey.

The door is open. The question is whether we design the handle well enough that people walk through it with understanding — or whether we leave them pressing on glass, blaming themselves for the confusion, while the real failure belongs to us.

-- Edo Segal

When AI learned to speak our language, we celebrated the collapse of every barrier between human intention and machine capability. Don Norman's framework reveals what we missed: collapsing one gulf blew the other wide open. The difficulty didn't disappear. It moved to exactly the place where we were least prepared to meet it. This book applies Norman's four decades of design principles — affordances, signifiers, conceptual models, error taxonomies — to the AI interaction and discovers that nearly every struggle people experience with these tools is a design failure, not a human one. From the blank prompt that communicates less than a door handle to the polished output that conceals its own uncertainty, the most consequential technology of our era has been engineered for power and barely designed for people. Through ten chapters that extend Norman's framework into territory he anticipated but never fully mapped, this volume provides the design vocabulary the AI revolution urgently needs — a vocabulary for seeing what goes wrong between human judgment and machine capability, and for building the structures that make it right.

When AI learned to speak our language, we celebrated the collapse of every barrier between human intention and machine capability. Don Norman's framework reveals what we missed: collapsing one gulf blew the other wide open. The difficulty didn't disappear. It moved to exactly the place where we were least prepared to meet it. This book applies Norman's four decades of design principles — affordances, signifiers, conceptual models, error taxonomies — to the AI interaction and discovers that nearly every struggle people experience with these tools is a design failure, not a human one. From the blank prompt that communicates less than a door handle to the polished output that conceals its own uncertainty, the most consequential technology of our era has been engineered for power and barely designed for people. Through ten chapters that extend Norman's framework into territory he anticipated but never fully mapped, this volume provides the design vocabulary the AI revolution urgently needs — a vocabulary for seeing what goes wrong between human judgment and machine capability, and for building the structures that make it right. — Don Norman

Don Norman
“AI need not be a threat, but it has to be approached, designed, and implemented intelligently, with full understanding of the people with whom it will interact.”
— Don Norman
0%
11 chapters
WIKI COMPANION

Don Norman — On AI

A reading-companion catalog of the 23 Orange Pill Wiki entries linked from this book — the people, ideas, works, and events that Don Norman — On AI uses as stepping stones for thinking through the AI revolution.

Open the Wiki Companion →