Shannon's source coding theorem establishes that any source with entropy rate H can be compressed to H bits per symbol without loss, but compression below H inevitably destroys information. Natural language, as a compression format for human intention, is near-optimal for a specific class of information: propositional content, logical relationships, functional specifications. Its semantic bandwidth — carrying denotation, connotation, implication, context simultaneously — vastly exceeds the statistical entropy of the character sequence. But natural language cannot carry the full entropy of every dimension of human knowledge. Embodied intuition, aesthetic judgment, contextual expertise — these reside in patterns of experience that resist verbalization, with entropy rates exceeding what language can encode. The AI interface is therefore a highly efficient compressor for the compressible component of knowledge and a lossy compressor for the incompressible component.
Shannon's 1948 and 1951 experiments estimated the entropy of printed English at roughly one bit per character — reflecting the high redundancy and predictability of letter sequences in context. But this measures statistical entropy of the character stream, not semantic entropy of the meaning carried. The semantic bandwidth of natural language is far higher, because it exploits context, shared knowledge, and pragmatic inference.
When Segal describes a product vision in a few paragraphs and receives a working prototype, the compression is remarkably efficient for functional requirements. The paragraphs carry enough information — enough constraint on the space of valid implementations — that the model can reconstruct a working artifact. The reconstruction is not perfect, but the imperfection is the imperfection of a single compression stage, not cumulative imperfection of five.
For another class of information, the compression is lossy below the entropy rate. The senior engineer's architectural intuition, the designer's aesthetic judgment, the craftsperson's feel for materials — these have entropy rates that exceed what natural language can carry. The information is lost not through careless encoding but through the mathematical impossibility of fitting a high-entropy source into a low-capacity code.
The implication is asymmetric quality. AI-assisted products excel on explicit dimensions (functionality, features, logic) and fall short on tacit ones (feel, craft, specificity). They work correctly but feel generic. The asymmetry is not a failure of the language model — it is a property of the channel. Natural language is a verbal medium, and verbal media carry verbal information more faithfully than non-verbal information.
The analysis synthesizes Shannon's 1948 source coding theorem with the philosophical tradition — from Polanyi through Dreyfus — that distinguishes tacit from explicit knowledge. The synthesis becomes operationally consequential when natural language becomes a machine interface, because the channel's compression characteristics now determine what can and cannot be built through conversation with machines.
Semantic bandwidth exceeds character entropy. Natural language carries far more meaning per symbol than its statistical entropy would suggest, through context and pragmatic inference.
Propositional content compresses well. Functional requirements, logical constraints, and explicit specifications can be encoded in language with high fidelity.
Tacit knowledge resists encoding. Embodied intuition, aesthetic judgment, and contextual expertise have entropy rates exceeding what language can carry.
Lossy compression is mathematical, not technical. Shannon's theorem establishes that no encoder can preserve information exceeding the channel's capacity.
Asymmetric product quality. AI-assisted work succeeds on explicit dimensions and falls short on tacit ones, because the channel transmits the former faithfully and the latter not at all.