Peirce's community of inquiry converges on truth because it includes multiple perspectives contributing what only specific vantage points can see. Dewey made this democratic in The Public and Its Problems (1927): democracy's challenge is epistemological—how do people affected by a decision participate in making it? How do collective action's consequences become visible to the community whose action produced them? Dewey's answer was communication: open, honest, publicly accessible exchange allowing communities to understand their own actions' consequences and revise accordingly. Democracy for Dewey was not a government form but associated living—organizing collective inquiry so knowledge from community experience can be shared, criticized, incorporated into ongoing decisions. The alternative was expert concentration of decision-making by people who, however brilliant, cannot possess the local knowledge that only affected communities hold.
The dam-building metaphor threading through You On AI contains an unexamined individualist assumption: it locates agency in the singular beaver, the gifted builder with vision and skill. But the question of where the dam should go—which is to say, how powerful technology's flow should be directed, whose interests served, whose flourishing prioritized—is not answerable by any individual. It's a phronesis question requiring communal deliberation. The individual builder sees the river from one vantage. The community affected by the flow sees it from many. Dams serving the full community—not just builders, investors, early adopters—can only be designed through serious sustained deliberation including voices of those living with consequences. Actual dam-building history confirms this: Aswan High Dam produced electricity while displacing 100,000 Nubians, Three Gorges generated power while displacing 1.3 million people—technical capacity was sound but democratic deliberation about whose interests the dam served was absent.
Current AI governance displays the same structural absence. Corporate 'responsible AI' principles, institutional review boards—these are dams of a kind, addressing real risks, but designed primarily by builders for builders within builders' fishbowls. Voices most affected—workers whose jobs transform, students whose education reshapes, parents whose children grow up in attentional environments designed by unaccountable strangers—are largely absent from design processes. Even governmental frameworks (EU AI Act, American executive orders) address primarily supply side: what AI companies may build, what disclosures required, what risks assessed. The demand side—what citizens need to navigate wisely—remains unaddressed. Bernstein's pragmatism identifies this as epistemological failure, not just policy gap: knowledge required for adequate dams is distributed across populations, not concentrated in builders' or regulators' hands. No expert possesses the situated knowledge a Lagos developer holds about how AI affects her specific career trajectory. No governance framework incorporates the lived experience requiring attention for dam-building to serve rather than sacrifice.
Bernstein synthesized Dewey's democratic pragmatism with Habermas's discourse ethics across works including The Restructuring of Social and Political Theory (1976) and The New Constellation (1991). His contribution was showing these traditions converge on a single insight: the self-correcting process of inquiry that Peirce described requires the open pluralistic exchange that democracy at its best provides, and the democratic process Dewey described requires the honest evidence-based inquiry Peirce's community models. The inseparability of communal inquiry and democratic participation produces the question You On AI gestures toward but doesn't fully develop: who builds the dams? Bernstein's framework insists this is not a technical question but a democratic question—how a community confronted with force exceeding any individual's control organizes collective inquiry to produce structures directing that force toward all members' flourishing, not just those holding the sticks.
Expertise is necessary but insufficient. Technical knowledge about AI capabilities cannot substitute for situated knowledge about consequences as experienced by affected communities—both are required for adequate governance.
The demand side is unaddressed. Current frameworks regulate what companies build but provide almost no institutional support for what citizens, workers, educators need to navigate the transition wisely—a governance gap with catastrophic developmental consequences.
Knowledge is distributed. No individual, expert committee, or corporate board possesses the full range of understanding required—truth emerges from inclusive deliberation incorporating voices seeing what experts systematically miss.
The dams require democracy. Concentrating design authority in technical experts guarantees dams serve some communities and sacrifice others—Bernstein's pragmatism insists on dialogical communities making decisions through serious sustained mutual deliberation.