Large language models are, by their architecture, opaque. Their internal processing is not transparent to the user, and the relationship between input features and output quality is not deducible from observation of the input-output relationship. The user who prompts a model cannot see which features of the prompt the model attended to, which features of its training produced its response, or why a particular variation produced a particular change in output quality. This opacity is a permanent feature of the interaction, not a correctable deficiency, and it creates the conditions under which organisms reliably develop superstitious behavior — reinforcement delivery insensitive to specific response features combined with natural behavioral variability producing spurious correlations between prompt features and response quality.
The opacity of large language models has multiple sources. The models themselves are large neural networks with billions of parameters, whose internal computations are not accessible through direct inspection. The training procedures — pretraining on large corpora, reinforcement learning from human feedback — produce emergent behaviors that were not explicitly programmed and are not always predictable from the training setup. The interfaces through which users interact with the models further abstract away the model's operations, presenting the response without the derivation.
The research program of mechanistic interpretability is the attempt to reduce this opacity by reverse-engineering what is actually happening inside neural networks. The program has made significant progress but has not eliminated the opacity faced by users in ordinary interaction — mechanistic interpretability remains a specialized research activity that does not propagate to the user experience.
The Skinner volume treats opacity as the structural condition that makes AI-user interactions inescapably conducive to superstitious conditioning. The conditions for superstitious conditioning — reinforcement insensitive to specific response features, combined with natural behavioral variability — are permanently present. The question is not whether superstitious behaviors will develop. They will. The question is whether the development will be recognized, measured, and managed through systematic methods, or allowed to accumulate into an elaborate culture of prompting ritual.
The diagnosis of AI opacity as the structural condition for superstitious behavior development is a 2026 contribution of the Skinner volume, synthesizing the technical fact of neural network opacity with Skinner's 1948 analysis of superstitious conditioning.
Opacity is structural, not incidental. The neural network architecture produces opacity that cannot be eliminated without abandoning the architecture.
Users rely on temporal contiguity to infer causation. In the absence of visible mechanism, the only inference available is from coincidence to cause.
Superstitious conditioning is therefore structural. The conditions for its development are permanent features of the interaction.
The remedy is methodological. Controlled variation can distinguish genuine effects from superstitious rituals, but requires explicit application of the method.