The traditional theory began with the retinal image because that image was what ophthalmologists could measure and laboratory researchers could control. The image was flat, two-dimensional, and frequently ambiguous about the three-dimensional world — which led, naturally, to the conclusion that perception must involve inferential supplementation. Gibson's move was to point out that no organism actually sees through a static retinal image. The eye moves. The head turns. The body walks. What the organism samples across these movements is not a sequence of images but structured transformations of the ambient array, and the transformations themselves specify what the static image could not.
The richness of the ambient array is not uniform. A dense forest presents an array of extraordinary structural complexity — overlapping surfaces at multiple distances, texture gradients from every trunk and leaf, occlusion patterns specifying depth relationships among hundreds of objects, color gradients specifying light direction. A bare room with white walls presents an impoverished array: undifferentiated surfaces, minimal texture, simple occlusion. The organism in the first environment has more to detect. The organism in the second has less to work with.
The distinction between content richness and structural richness — a distinction Gibson's framework makes precise — is the key to understanding what AI-augmented environments do to perception. A library contains enormous information content but presents a structurally impoverished ambient array (book spines on shelves). The same information encountered through active engagement with the systems it describes presents a structurally rich array: textured surfaces, cascading consequences, events that reveal dynamic properties through their unfolding.
The AI-augmented builder's environment is the library-shaped version of what was once the forest-shaped environment of software construction. Content richness has increased enormously. Structural richness — the degree to which information is arranged for perceptual pickup through active exploration — has decreased. The smoothing of surfaces, the prevention of events through error-avoidance, the delivery of pre-processed explanations all reduce the ambient array's structural complexity while increasing what is nominally available.
Gibson developed the concept across the 1950s and 1960s, arriving at the mature formulation in The Senses Considered as Perceptual Systems (1966) and refining it in The Ecological Approach (1979). The wartime studies of pilot perception were formative: pilots navigating real terrain could not be treated as passive recipients of retinal images, because the information they relied on — optic flow, texture gradients on approach surfaces — only existed in the full ambient array that their movement continuously resampled.
The totality claim. Perception uses the full 360-degree ambient array, not a sampled retinal fragment.
Structural vs content richness. An environment can contain enormous information content while presenting a structurally impoverished array for perceptual pickup.
Texture specifies layout. Gradients of texture density, grain size, and contrast carry mathematically precise information about surface orientation and distance.
Transformations reveal invariants. The changes in the ambient array as the observer moves specify properties that no static snapshot could encode.
The AI paradox. Contemporary computing environments have maximized content richness while reducing structural richness — a configuration Gibson's framework predicts will produce organisms with access to vast information and diminished perceptual differentiation.