Data dignity is Lanier's prescription, arrived at after fifteen years of diagnosing the extractive architecture of the digital economy. The principle is elementary: labor that creates value deserves compensation. Its application to AI is radical: if large language models derive their capability from the accumulated labor of millions of human contributors, those contributors are owed both acknowledgment and a share of the value their data generates. Data dignity rejects the framing of training data as a free resource to be harvested. It insists that data is the product of specific human effort and that the people performing that effort retain moral and economic claims on what is built from it. The principle has technical, economic, and institutional dimensions — each of them feasible, each of them politically difficult, each of them currently unbuilt because the beneficiaries of the status quo have no reason to build them.
Lanier introduced data dignity in a 2018 Harvard Business Review article co-authored with E. Glen Weyl, though the underlying concept had been developing across his work for nearly a decade. The 2018 framing responded to a specific moment: the post-Cambridge Analytica reckoning with social media's data practices, the rising concern about AI's trajectory, and the growing awareness that the 'users are the product' critique of Web 2.0 platforms needed to be translated into constructive institutional design.
The framework became urgent rather than merely interesting with the rise of large language models. If the siren server diagnosis applied to Facebook and Google — companies whose data use was at least partially transparent and whose users had some awareness of the exchange — it applied with vastly greater force to AI training, where the contributors rarely knew their work was being used and had no meaningful mechanism to opt out. Data dignity moved from speculative prescription to urgent agenda.
The principle has three technical layers, each addressing a specific failure of the current architecture. The provenance layer requires AI systems to maintain at least approximate records of which training data influenced which outputs — not perfect attribution but statistical attribution sufficient to identify major contributors. The compensation layer requires payments to flow back to contributors when their data generates value, analogous to the royalty structures that exist (however imperfectly) in music and publishing. The institutional layer requires organizations — Lanier and Weyl called them Mediators of Individual Data — that aggregate the bargaining power of individual contributors into collective force capable of negotiating with platforms.
Data dignity is not utopian. Every technical component exists in some form in existing industries. Provenance tracking exists in publishing, academic citation, and music sampling. Micro-compensation exists in streaming royalties, app store revenue sharing, and advertising networks. Collective bargaining exists in labor unions, collection societies, and professional associations. The innovation Lanier proposes is not new institutions but the application of existing institutional forms to the specific context of AI training data. What is missing is not technology but political will.
The roots of data dignity trace to Lanier's 2010 You Are Not a Gadget and its 2013 successor Who Owns the Future?, which argued that the dominant ideology of Silicon Valley treated human beings as components in a computational system rather than as irreducible individuals. The 2013 book proposed that digital networks should preserve the connection between a person's contribution and the value it generated — the embryonic form of what would become data dignity.
The 2018 co-authored HBR article with E. Glen Weyl gave the concept its name and its institutional structure. Weyl, an economist working on plurality and radical markets, brought technical economic framing that complemented Lanier's critique. The partnership extended through subsequent publications and policy advocacy, with data dignity becoming one of several related proposals — alongside Mediators of Individual Data and broader plurality frameworks — for restructuring the relationship between individuals and digital systems.
Labor creates value; value creators deserve compensation. The foundational claim is elementary and universal: people whose work contributes to a valuable output have moral and economic claims on that output. Data dignity applies this universal principle to the specific case of AI training data.
Provenance is a design choice. The current architecture's inability to track contributions is not a technical limitation but an engineering decision. Systems could be built to maintain provenance. They are not built that way because the current business model does not require it.
Compensation must be automatic and granular. Individual payments will be tiny — fractions of a cent per interaction — but collectively significant, because the same contributor's work may influence millions of interactions. The infrastructure for micro-payments exists; applying it to AI training data is a matter of will.
Collective organization is indispensable. Individual contributors have no leverage over AI companies. Collectively, the millions of contributors whose labor powers the models represent an indispensable resource. Data dignity requires institutions that aggregate individual bargaining power into collective force.
Sustainability joins justice as motivation. Data dignity is not merely about fairness to individual creators. It is about the sustainability of the AI ecosystem itself, which depends on continued production of high-quality human-generated training data. Without compensation flowing back to creators, the pipeline degrades and the models eventually suffer.
Critics argue data dignity is technically infeasible at scale, economically prohibitive, and politically unachievable. Each objection has been addressed by Lanier and allied researchers. Technical infeasibility: attribution methods exist and are improving; approximate attribution is vastly better than none. Economic prohibition: payments would be small relative to AI industry revenues, and the cost of not compensating contributors includes the eventual degradation of training data. Political unachievability: every labor protection in industrial history was declared politically impossible until it became politically inevitable. The deeper debate concerns whether AI companies will build data dignity voluntarily or whether it will require regulatory compulsion. The current evidence suggests that voluntary action is not coming, and the initiative is shifting to courtrooms (the Authors Guild and Andersen v. Stability AI litigation) and legislative bodies (the EU AI Act, emerging US proposals).