Every large language model is built on a corpus of human creative output — novels, code, documentation, forum posts, every form of written expression captured in digital form. This corpus is not raw data but accumulated immaterial labor: the creative, communicative, cognitive, and affective product of millions of human beings, each investing genuine subjectivity. The enclosure of this corpus — its conversion from commons into privately owned productive asset — is structurally identical to what Marx analyzed as primitive accumulation. The ghost in the training data is the immaterial laborer whose labor is essential to AI's functioning but who has been dispossessed of the value her work generates, rendered invisible within the system's self-understanding as an autonomous capability.
The principle that the social machine explains the technical machine applies with particular force to training corpora. The technical machine — the large language model — could not exist without the commons of human creative output. But the social machine — the venture-capital-funded AI industry, the intellectual property regime treating publicly available text as freely extractable, the platform economy that has normalized conversion of user-generated content into corporate assets — determines how the commons is appropriated and who benefits.
The distribution of value from the assemblage is radically asymmetric. The builder gains productivity. The platform gains data, revenue, and competitive position. The millions of developers, writers, and artists whose immaterial labor constitutes the training data gain nothing. They are, in Mary Gray and Siddharth Suri's term, ghost workers — their labor essential, invisible, uncompensated. The AI appears to generate capability from computation. The computation depends on the commons. The commons was produced by labor. The labor is uncompensated.
The Orange Pill's celebration of the democratization of capability is complicated by this analysis. The democratization is real: more people can produce more with less institutional support. But the capability being democratized was itself produced through enclosure of a commons. The developer in Lagos who gains access to Claude Code gains access to a tool built from the enclosed creative output of millions. Her democratization depends on their dispossession. The two operations are not separate but structurally linked — two faces of the same process of enclosure and redistribution that has characterized every major appropriation of common resources in capitalist history.
Legal challenges from creative workers — the Authors Guild Letter, Andersen v. Stability AI, the Hollywood Writers' Strike — are beginning to demand what the framework identifies as politically necessary: collective licensing structures, transparency about training data, recognition that the creative community's contribution entitles it to a share of value generated by models that depend on its output.
The framework synthesizes Lazzarato's analysis of immaterial labor with commons theory and scholarship on ghost work — notably Mary Gray and Siddharth Suri's 2019 documentation of the hidden human labor force powering AI systems.
Training corpus as enclosed commons. The AI training corpus is a commons — the accumulated immaterial labor of human civilization — enclosed through operations structurally identical to primitive accumulation.
Asymmetric value distribution. Builders gain productivity, platforms gain data and revenue, the laborers whose work constitutes the corpus gain nothing.
Democratization depends on dispossession. The expansion of who can build is enabled by the enclosure of what countless others already built.
Invisibility as structural, not incidental. The system cannot acknowledge its dependence on ghost labor without undermining the fiction that capability is the platform's property.
Governance through commons theory. Adequate response requires structures that recognize collective production, govern the corpus collectively, and distribute value to the community that produced it.
The framework is contested by those who argue that training on publicly available text is fair use, that the transformation performed by neural networks is sufficiently novel to constitute original creation rather than derivative extraction, and that the collective licensing structures implied by the critique would be practically unworkable. Defenders respond that fair use doctrine was not designed for industrial-scale corpus extraction, that legal frameworks lag the technological reality, and that practical unworkability is a political rather than logical obstacle — one that cultural transformation could, in time, address.