The mechanism is straightforward: the developer describes what the software should do in English — her own language, with all its ambiguity and implication — and the model produces implementation across the stack. Database schema, backend logic, API design, frontend rendering, deployment configuration. The gap between the abstraction level (human intention) and the underlying layer (executing code) is not one step. It is every step, collapsed into a single conversational interface. This is what Edo Segal calls the orange pill moment and what Spolsky's framework identifies as the outer limit of concealment's reach.
The power is real and transformative. For routine operations — CRUD endpoints, standard authentication flows, data validation, UI rendering, configuration management — AI-generated code is not merely adequate but often better than what the median developer would produce by hand, because the training data encodes millions of well-reviewed examples. The engineer in Trivandrum who built a complete frontend feature in two days without prior frontend experience was operating within this reliable domain. The imagination-to-artifact ratio has collapsed, and the liberation is genuine.
But AI-generated code differs from previous abstractions on three structural dimensions that make its leak profile uniquely severe. It expands the scope of what is concealed from implementation details to architectural decisions. It conceals not only how but why — the generated code lacks recoverable intent because no mind with reasons produced it. And it conceals the interactions between components, because components generated in separate conversations may carry mismatched assumptions that no human negotiated. The first dimension expands the territory of the leak; the second prevents diagnostic archaeology from reconstructing design intent; the third introduces integration leaks as a distinct and often undiagnosable class of failure.
The result is a technology that is transformative within its reliable domain and uniquely fragile at its boundary. The boundary is defined by three characteristics: pattern density (how many examples of this problem exist in the training data), specification precision (how well the developer can describe what she wants), and isolation (how independent the component is from others). When all three are high, the abstraction holds brilliantly. When one or more is low, the abstraction leaks in ways that previous frameworks for understanding software reliability cannot fully address.
The phrase 'AI-generated code' is older than the current moment — it was used for earlier template engines and code generators — but its current meaning crystallized with the release of GPT-4, Claude, and competitors capable of producing working full-stack implementations from natural language. The Spolsky-lens reading of the category emerged in 2024–2026 as practitioners began encountering leaks that their AI-era tooling could not explain, prompting a rediscovery of the 2002 law as the most adequate framework for naming what they were experiencing.
Full-stack concealment. Unlike prior abstractions, AI-generated code hides every layer simultaneously — database, backend, API, frontend, deployment.
Architectural decisions are generated, not designed. The schema, the authentication scheme, the session model — all chosen by the system, not by the developer.
No recoverable intent. The code's specific implementation choices have no 'why' in the human sense, only statistical pattern-matching from training data.
Integration leaks. Components generated separately may carry mismatched assumptions that no human negotiated, producing failures that live in the space between components.
The reliable domain is real. High pattern density, precise specification, and component isolation together produce a zone where the abstraction holds brilliantly.