The expertise paradox names a temporal contradiction at the heart of AI deployment. AI systems depend on human expertise for reliable oversight — the capacity to detect errors, evaluate outputs, and intervene when situations fall outside the system's training distribution. But this oversight capacity was built, for the current generation of practitioners, through direct engagement with the domain before AI arrived. The next generation, trained in AI-augmented environments, will lack the experiential foundation that makes effective oversight possible. The oversight capacity is a non-renewable resource under current deployment patterns — built under conditions the technology is eliminating, with no mechanism within the technology itself to regenerate it. The paradox has a temporal dimension that makes it especially dangerous: the degradation does not manifest immediately, but gradually, as the current generation retires and is replaced by practitioners who lack the pattern libraries that effective oversight requires.
Klein encountered the paradox through his work on DARPA's Explainable AI program, where DARPA assembled eleven teams of AI researchers to build explainable systems alongside a separate team of cognitive psychologists led by Klein. The AI teams focused on technical transparency; Klein's team asked what humans actually need to know in order to form accurate mental models of how the systems work. The research revealed that effective oversight depends on experiential foundations the AI cannot provide.
The paradox manifests in the practice framework organizations need but rarely adopt. The military's manual reversion training — pilots required to fly without automation at regular intervals — exists because the military learned catastrophically that automation erodes the manual skills it depends upon for backup. The analogy to AI-augmented knowledge work is direct: organizations deploying AI coding assistants need developers who still write code by hand, organizations deploying AI diagnostic tools need clinicians who still examine patients, not because manual practice is more efficient but because it is the only way to build and maintain the pattern libraries that effective review requires.
The temporal invisibility of the degradation is its most dangerous feature. In the first months and years of AI deployment, the workflow appears to work well — outputs are reviewed, errors are caught, the system performs as expected. Organizational leaders observe smooth functioning and conclude the system is reliable. The degradation manifests later, gradually, as the generation of experts who built their libraries pre-AI retires and the replacements bring thinner libraries to the task. The errors that would have been caught are not caught, but the misses do not show up in organizational metrics because the metrics were designed to capture errors the system itself detects.
Klein's career-long adversarial collaboration with Daniel Kahneman established that expert intuition is reliable under two conditions: the environment provides valid cues, and the expert has had sufficient learning opportunity. AI-augmented work disrupts both conditions — mediating the cues through processing that may strip contextual richness, and automating the tasks that provide the learning.
Klein formulated the paradox through observation of how automation historically affected expertise across aviation, medicine, and military operations. The pattern was consistent: practitioners who developed skills before automation could oversee automated systems effectively; practitioners trained within automated environments could not, because the experiences that built oversight competence had been eliminated by the automation.
The AI era intensifies the paradox because the capabilities being automated — reasoning, judgment, synthesis — are precisely the capabilities that oversight requires, and the pace of deployment is faster than any prior automation wave.
Temporal mismatch. The first generation of AI reviewers has pre-AI expertise; the next generation will not.
Invisible degradation. The loss of oversight capacity is gradual and undetectable by standard metrics until a failure makes it visible.
Non-renewable resource. Current deployment patterns consume expertise without providing the conditions for its regeneration.
Counterfactual invisibility. Errors that would have been caught by expert oversight do not show up in organizational data — the counterfactual is unmeasurable.
Design intervention required. The paradox cannot be resolved by better AI; it requires deliberate organizational design that preserves the conditions under which expertise develops.
A central debate concerns whether AI training programs can substitute for direct domain experience in building oversight capacity. Klein's position is that training can supplement but not replace direct engagement, because the pattern libraries that support anomaly detection are built through encountering unexpected situations — and training programs by their nature curate the situations practitioners encounter. The paradox therefore requires structural work redesign, not merely curriculum reform.