Interrogative vigilance names the metacognitive discipline that AI collaboration demands: the active, effortful investigation of outputs that appear competent, the generation of internal critical signals when the environment provides none, the maintenance of skepticism toward confident assertions whose confidence the medium automatically produces. The Dweck volume identifies this as the cognitive practice that distinguishes genuine growth from the false growth of passive AI consumption. It is not suspicion, which rejects; it is vigilance, which investigates. And it extends Dweck's established framework — which focused on responses to visible failure — into the new territory of smooth failure, where the failure conceals itself beneath the aesthetics of polished output.
Dweck's original research assumed that failure was visible. The math problem produces the wrong answer. The experiment fails to replicate. The project misses its deadline. These failures arrive with markers that tell the individual where to direct attention. The growth mindset's characteristic response — engage with failure, extract the learning signal, adjust approach — depends on the failure being detectable.
AI-generated output disrupts this assumption at its foundation. The machine produces failures that do not announce themselves: confident wrongness dressed in good prose, plausible assertions backed by fabricated evidence, coherent arguments constructed on foundations that do not exist. The surface quality actively conceals the substrate failures that growth-mindset engagement would normally detect and learn from.
Daniel Kahneman's research on cognitive fluency demonstrates why this concealment is so effective. Information presented in smooth, easy-to-process format is judged more credible than information presented with friction, regardless of actual accuracy. The fluency itself becomes a proxy for truth — an efficient heuristic under normal conditions but a catastrophic vulnerability when a machine can produce polished presentation of fabricated content with equal facility.
Interrogative vigilance is the disciplined countermeasure: slowing down when the tool is urging speed, generating one's own signal of potential failure when nothing in the environment suggests one, asking "is something wrong here?" when everything in the output says nothing is wrong. The practice is psychologically expensive because it runs counter to the cognitive ease that smooth output is designed to produce. It depends on domain knowledge deep enough to catch the machine's errors — creating the paradox that the expertise being displaced is also the foundation on which vigilance depends.
The concept synthesizes Dweck's growth-mindset framework with the broader literature on metacognitive awareness (Flavell, 1979), cognitive fluency bias (Alter & Oppenheimer, 2009), and the emerging research on AI-assisted cognitive work. The 2025 Zain and Habib study in Research Journal for Social Affairs found that researchers who used AI as a "cognitive co-worker" — actively interrogating outputs — showed growth-mindset hallmarks, while those who used AI passively showed none. The finding mapped precisely onto the distinction between interrogative vigilance and passive consumption.
The Dweck volume's extension names the practice and grounds it in the specific cognitive and institutional challenges the AI moment creates.
Vigilance is active investigation. Not suspicion that rejects but disciplined questioning that investigates — a dual orientation that holds openness and skepticism simultaneously.
The signal must be generated internally. Unlike visible failure, smooth failure provides no external prompt; the practitioner must generate her own critical signal from domain knowledge and metacognitive discipline.
Fluency bias is the adversary. The brain uses processing ease as a heuristic for reliability; AI output exploits this heuristic by producing fluent presentation regardless of underlying accuracy.
Domain knowledge enables vigilance. You cannot catch the error if you lack the knowledge to recognize what belongs and what does not — creating the paradox that the displaced expertise is the foundation on which vigilance depends.
It is the AI-age growth-mindset discipline. Responding to failure that announces itself is the traditional growth response; responding to failure that conceals itself is the AI-age extension.
An unresolved tension: interrogative vigilance depends on deep domain knowledge, but the AI efficiency that makes vigilance necessary also tempts practitioners to bypass the friction-rich learning that builds that knowledge. The false growth mindset version of AI collaboration — passive consumption of fluent output — erodes the very capacity that genuine collaboration requires. The prescription is that educational and professional environments must protect spaces of unassisted practice not as nostalgic exercises but as deliberate investments in the domain knowledge that makes vigilance possible.