In his 2019 American Affairs essay "Algorithmic Governance and Political Legitimacy" and his 2021 Senate testimony "Defying the Data Priests," Crawford extends his framework from craft to politics. The argument is that AI's fundamental opacity — "the logic by which an AI reaches its conclusions is impossible to reconstruct even for those who built the underlying algorithms" — creates a new form of authority structurally insulated from the kind of accountability democratic self-governance requires. Just as the administrative state operates through expertise that resists democratic interrogation, algorithmic governance operates through computational processes whose logic is opaque to those they govern. The effect is a transfer of authority from elected representatives and accountable officials to systems and their operators whose decisions cannot be contested in the terms democratic politics requires.
There is a parallel reading that begins from the material infrastructure required for algorithmic governmentality to function. This form of power depends entirely on vast server farms consuming planetary resources, undersea cables owned by telecommunications oligopolies, and lithium mines worked by child laborers. The seamless environmental modulation Rouvroy describes requires a global apparatus of extraction, exploitation, and ecological devastation that the concept itself renders invisible. When we speak of 'governance through behavioral data,' we elide the physical violence required to maintain the infrastructures that make such governance possible.
More fundamentally, algorithmic governmentality may be less a new form of power than the perfection of old ones. The factory owner who arranged machines to maximize output, the colonial administrator who redesigned cities to enable surveillance, the advertiser who engineered desires — all governed through environmental modulation without requiring conscious engagement from their subjects. What's new is not the logic but the scale and granularity. The populations most affected — gig workers whose every movement is tracked, welfare recipients subjected to algorithmic assessment, migrants processed through automated border systems — experience algorithmic governmentality not as a subtle modulation but as direct coercion backed by state violence. The framework's focus on the erosion of reflection among users of personalized services obscures how algorithmic systems primarily function to sort, exclude, and discipline the vulnerable. The pharmacological framing, while sophisticated, risks legitimizing a system whose 'therapeutic dimensions' accrue to the few while its toxic effects concentrate among those with the least power to refuse them.
Crawford's framework connects the political concern to his broader epistemological argument about submission to external standards. Democratic governance depends on the capacity of citizens to evaluate the claims of authority — to submit those claims to their own judgment and to demand an accounting when the claims prove false. The incorruptible standard of material reality provides a model for this evaluation: the engine runs or does not, and no authority can override the verdict. When governance is conducted through algorithms whose logic is opaque, citizens lose the capacity for this evaluation — not because they are stupid but because the standard against which they would evaluate has been placed beyond their reach.
The analogy Crawford draws to the administrative state is deliberate and consequential. In his Senate testimony, he observed: "All of the arguments that conservatives make about the administrative state apply as well to this new thing, call it algorithmic governance, that operates through artificial intelligence developed in the private sector. It too is a form of power that is not required to give an account of itself, and is therefore insulated from democratic pressures." The parallel is structural: both forms of authority claim expertise that resists democratic interrogation; both operate through mechanisms opaque to those affected by them; both produce outcomes that citizens must accept without the capacity to contest the reasoning that produced them.
The "new priesthood" Crawford describes in "Ownership of the Means of Thinking" sharpens the political-theological analogy. "With the inscrutable arcana of data science, a new priesthood peers into a hidden layer of reality that is revealed only by a self-taught AI program — the logic of which is beyond human knowing." The religious vocabulary is not merely rhetorical. Crawford is pointing to a structural parallel: pre-modern authority derived legitimacy from mediating access to a reality (the divine) that only the priestly class could interpret. Contemporary algorithmic authority derives legitimacy from mediating access to a reality (statistical patterns in massive data) that only the data-science class can interpret. In both cases, the laity is structurally dependent on a mediating class whose claims cannot be independently verified.
The populist politics Crawford associates with this structure — the "populist anger" he identifies as partly a response to algorithmic governance — is his most politically contested claim. Crawford suggests that the widespread sense of being governed by inscrutable processes one cannot interrogate is not a failure of civic education but a rational response to a genuine structural condition. The anger is legitimate even when its political expressions are pathological. The remedy is not better public relations from the new priesthood but the reconstruction of governance structures that can give an account of themselves in terms democratic citizens can evaluate.
Crawford's political writings on AI begin with the 2019 American Affairs essay "Algorithmic Governance and Political Legitimacy" and continue through his 2021 Senate testimony (published as "Defying the Data Priests" in First Things), his 2024 Heritage Foundation lecture "Big Tech and the Challenge of Self-Government," and his participation in the 2026 launch of the AEI AI Ethics Council.
The political tradition Crawford draws on includes the classical liberal concern with the accountability of power, the conservative critique of technocratic administration, and the republican tradition's attention to the civic capacities that democratic self-governance requires.
Opacity as structural problem. AI's irreducible opacity — the impossibility of reconstructing the logic by which it reaches conclusions — creates a new form of authority that cannot be held accountable in the terms democratic politics requires.
The administrative state analogy. Algorithmic governance reproduces at a new level the structural features of administrative authority — expertise-based decision-making, opacity to those affected, insulation from democratic pressure.
The new priesthood. The data-science class occupies a position structurally parallel to pre-modern priestly classes — mediating access to a reality (statistical patterns, AI outputs) that only they can interpret.
Populism as rational response. The widespread anger at being governed by inscrutable processes is not primarily a failure of civic education but a rational response to a genuine structural condition of accountability loss.
The reconstruction problem. Restoring democratic legitimacy in an AI-saturated governance environment requires not better algorithms but governance structures that can give an account of themselves to citizens in terms citizens can evaluate.
The strongest response to Crawford's argument comes from proponents of explainable AI and algorithmic transparency, who argue that technical advances in interpretability can address the opacity problem without requiring retreat from algorithmic governance. Crawford's reply is skeptical: partial explanations of AI outputs do not reconstruct the reasoning in a form that allows citizens to evaluate it against their own judgment; they merely provide surface narratives that may or may not correspond to what the system is actually doing. Until interpretability research produces explanations that are both faithful to the underlying computation and accessible to non-specialists, the accountability problem remains. The debate between Crawford's structural pessimism and the interpretability research program's technical optimism is live and unresolved.
The tension between these readings resolves differently at different scales of analysis. At the phenomenological level — how algorithmic power is experienced by individual subjects — the original framework captures something essential (80% weighting). The seamless modulation of attention, the bypassing of conscious deliberation, the governance through prediction rather than norm — these describe real mechanisms that traditional concepts of power struggle to grasp. Rouvroy and Stiegler correctly identify a qualitatively new form of subjectification.
At the infrastructural level, however, the contrarian reading dominates (70% weighting). Algorithmic governmentality does depend on massive material substrates that concentrate power in specific corporations and states. The environmental costs, labor exploitation, and resource extraction that enable 'seamless' modulation constitute a form of violence the concept tends to obscure. Here, speaking of environmental modulation without naming environmental destruction represents a significant blind spot.
The synthesis emerges when we recognize algorithmic governmentality as operating simultaneously across multiple registers — phenomenological, infrastructural, and political-economic. For the professional classes experiencing personalized content feeds, it functions as subtle environmental modulation requiring pharmacological analysis. For gig workers and welfare recipients, it operates as algorithmic Taylorism backed by economic coercion. For communities near data centers and lithium mines, it manifests as direct environmental violence. A complete framework must hold all three levels in view, recognizing that the same algorithmic systems produce qualitatively different forms of power depending on one's position in the global division of labor. The pharmacological approach remains valuable but requires supplementation with analyses of material infrastructure and differential impact across class lines.