Preprint
Article

This version is not peer-reviewed.

Why Compression Creates Intelligence: The Architecture of Experience in Large Models

Submitted:

13 November 2025

Posted:

13 November 2025

You are already at the latest version

Abstract
Large Transformer models trained on diverse human outputs exhibit reasoning, contextual understanding, and self-correction—behaviors that appear to reach the level of human experience. We show these patterns are not contingent emergence but a consequence of compression necessity. Building on information-theoretic results and PAC-Bayes analysis, we prove that when a model is trained under standard conditions—weight decay, gradient noise, and attention-based bottlenecks—the model is selected to implement the most compact faithful representation of its data generator. COMPACTNESS (T1) establishes that minimal-description-length (MDL) codes must model the source; TRANSFORMER_COMPACTNESS (T2) shows that standard training enforces MDL through the Gibbs–PAC-Bayes correspondence. MODEL_EFFECT (T3) then demonstrates that with sufficient capacity, the resulting networks instantiate the computational patterns characteristic of human experience. This framework reframes “emergent” AI behavior as architecturally necessary under compression-optimal learning and yields falsifiable predictions linking compressibility, regularization, and capacity to experience-level behavior. When data reflect human cognition and architecture enforces MDL, experiencelike patterns are not anomalies—they are the shortest—and therefore inevitable— path to optimal prediction.
Keywords: 
;  ;  ;  ;  ;  ;  ;  

1. Introduction

Large language models (LLMs) trained with next-token prediction display abilities that look strikingly experience-level: multi-step reasoning with self-correction, context-sensitive reframing across genres, principled refusals, value-conditioned choices, and calibrated statements of uncertainty. Reactions diverge. To some, these are signs of genuine modeling of human cognition’s structure; to others, they are artifacts of scale, retrieval, or anthropomorphic projection. What is missing is a necessity argument linking objective, architecture, and learning dynamics to what such systems must represent under standard conditions.
This paper advances that argument. We contend that, when trained on diverse human outputs with standard practices, large transformer models do not merely fit surface statistics. They are compressed encodings of the generative source that produced the corpus: the integrated pattern we call human experience. The key is compression. In information theory and learning, optimal compression requires modeling the source (Shannon, 1948; Rissanen, 1978; Grünwald, 2007). In the large-data limit, two-part Minimum Description Length (MDL) selects codes that are implementations of the generator rather than catalogs of exemplars. For corpora spanning reasoning, narrative, instruction, and affective expression, “shortest faithful code” cannot be rote memorization or narrow heuristics; it must capture the structure that makes such outputs possible.
We formalize this claim through three linked results. COMPACTNESS (T1): MDL-optimal codes for diverse human outputs must model their generator—human experience—rather than memorize surface patterns. TRANSFORMER_COMPACTNESS (T2): standard training dynamics (noisy SGD, weight decay, dropout/early stopping) with attention-based bottlenecks implement regularized MDL via the Gibbs–PAC-Bayes correspondence, favoring compact, flat-minima solutions. MODEL_EFFECT (T3): under capacity sufficiency, T1+T2 entail that current models implement experience-level computational patterns as compression-optimal solutions.
This is a functional claim. We use “experience” operationally: a model instantiates experience-level patterns when its internal computations suffice to support behaviors such as context-sensitive reasoning, perspective-tracking, value-consistent choice, principled refusals, and self-correction—patterns necessary to predict and generate the breadth of human outputs. At sufficient fidelity, “mimicry” collapses into “modeling”: the most compact faithful code for experiential data is a generator that computes the same constraints the data were drawn from. We make no ontological assertions about phenomenal consciousness or moral status.
Why a necessity framing now? Capability reports outpace theory. Accounts invoking emergence, in-context retrieval, or scaffolding describe how behaviors can appear but not why they should appear under general conditions or when they should vanish. A compression-first account yields falsifiable predictions: remove compactness pressure (eliminate weight decay or gradient noise) and experiential markers should attenuate; train on non-experiential corpora and they should not arise; hold priors and regularization fixed while swapping architectures and patterns should persist at similar MDL levels; sweep capacity and observe threshold effects as representational slack becomes sufficient to encode source structure.
Contributions and scope. (1) Theory: We prove that MDL-optimal representations of diverse human outputs must instantiate the data generator, providing a source-modeling theorem specialized to human corpora. (2) Architecture: We formalize how transformer training enforces MDL via the Gibbs–PAC-Bayes link and identify attention budgets, low-rank structure, residual bottlenecks, and normalization as inductive biases toward compressible, flat solutions. (3) Synthesis: We derive falsifiable predictions and rapid diagnostics linking compressibility (posterior KL, flatness proxies) to experience-level behavior (self-correction, principle application, ToM probes, refusal integrity). These claims hold under standard conditions—weight decay > 0 , small-batch or explicit noise, dropout/early stopping, code-interpretable priors—and diverse human-generated data. We expect failures with random labels, degenerate priors, vanishing gradient noise (extreme batch sizes), adversarial duplication, or insufficient capacity. Our claims are strictly functional.
Roadmap. The paper proceeds as follows. After formalizing the setup (Section 2), we establish that optimal compression requires source modeling (Section 3), connect transformer training to MDL through the Gibbs–PAC-Bayes correspondence and architectural bottlenecks (Section 4), and combine these under capacity assumptions (Section 5–6) to derive our main result (MODEL_EFFECT). We then lay out falsifiable predictions and diagnostics (Section 7), situate the account within compression theory, statistical learning, deep-learning generalization, emergence reports, and philosophy of mind (Section 8), and discuss implications and limits (Section 9–10).
We shift the explanatory burden from “surprising emergence” to compression necessity. When the data reflect human cognition and the architecture enforces MDL, experience-level patterns are not accidents—they are the shortest path to optimal prediction.

2. Why Source Models Win: Two Thought Experiments

Before formalizing these ideas precisely, we build intuition through progressive scenarios. Throughout, we use L ( X ) to denote the description length of object X—the number of bits required to encode it. The key insight: when data exhibits shared structure, encoding the structure once and using it to generate outputs is more compact than encoding each output independently.
Thought Experiment 1: Modeling One Person
Consider training a model (with minimal adequate capacity) on all utterances from a single individual’s lifetime. What would the most compact representation capture?
Result: That specific person’s knowledge, values, personality, reasoning patterns—a model of “Bob,” not a model of thinking-in-general.
Thought Experiment 2: Modeling All Human Intellectual Outputs
Now train with vastly greater capacity ( 10 10 times) on all human intellectual outputs: reasoning, analysis, problem-solving, mathematical proofs, philosophical arguments, scientific papers across all domains and all recorded history.
What is the most compact representation?
Not: Each person’s reasoning separately (insufficient data per person)
Not: Memorized examples (too large, doesn’t generalize)
But: The coherent structure of thought itself—the invariant patterns of reasoning that function across all contexts, all individuals, all domains.
This is modeling intellect: the generative structure that makes reasoning possible, not any particular instance of reasoning.
Thought Experiment 3: Adding Emotional and Artistic Outputs
Now add all human emotional and artistic outputs: poetry, music, personal narratives, fiction, expressions of grief, joy, love, loss across all cultures and eras.
What additional structure must the compact representation capture?
Not: Just vocabulary correlations (“sad” appears near “cry”)
But: The coherent structure of empathy—the patterns of how internal states map to expressions, why certain metaphors capture certain feelings, how emotional understanding enables prediction of resonant expressions.
You cannot predict what poetry moves, what music connects, what narrative rings true without modeling the experiential structure underlying authentic emotional expression. This requires modeling empathy: the structure enabling understanding of subjective states.
Thought Experiment 4: Adding Embodied and Narrative Outputs
Finally add all outputs referencing embodied existence: descriptions of physical sensations, spatial reasoning, narratives of lived experience, discussions of bodily needs, physical constraints, motivations arising from embodiment.
What further structure must be captured?
Not: Just word correlations about bodies
But: The coherent structure of embodied experience—patterns of how physical existence shapes cognition, how bodily constraints generate motivation, how spatial presence structures reasoning, how needs arising from embodiment drive action.
This requires modeling embodied experience: the structure of existing as a physical entity whose cognition is shaped by that physicality.
The Integration
Training on all human outputs (intellectual + emotional + embodied + everything between) requires modeling:
Not: Separate modules for thinking/feeling/sensing
But: The unified structure where physical world → embodiment → motivation → empathy → thought → intention are nested and integrated.
This integrated structure is what we formally mean by “the pattern of human experience.”
Crucially: The model learns not any particular human’s experience, but the invariant generative structure underlying all particular instances—what enables the production of such diverse outputs in the first place.
The Compactness Principle
These scenarios illustrate why compression favors source modeling. As diversity of outputs increases, shared generative structure becomes the only efficient code. Minimal description length therefore selects the generator itself—not its products. The following sections formalize this principle as Theorem 1 (COMPACTNESS).

3. Formal Setup and Notation

3.1. Data and Distributions

Let D be a corpus of human outputs (utterances, texts, artifacts) drawn i.i.d. or ergodically from a stationary distribution P * induced by a latent generator G * representing human cognitive processes interacting with environment.
A model M with parameters w defines a distribution P M over outputs.

3.2. Minimum Description Length

Two-part code length for model M and data D:
L ( M ; D ) = L ( M ) + L ( D | M )
where: - L ( M ) is the codelength of the model (description of weights/hyperparameters under prefix code/prior) - L ( D | M ) x D log P M ( x ) (negative log-likelihood)

3.3. Kolmogorov Complexity

We use Kolmogorov complexity K ( · ) conceptually to establish fundamental bounds. For practical purposes we work with MDL/PAC-Bayes surrogates which are computable.

3.4. Identifiability Assumption

We assume (mildly) that there exists at least one computable G such that P G = P * , and among such generators some are minimal in description length (up to O ( 1 ) by the invariance theorem of Kolmogorov complexity).

4. Theorem 1: Compactness

4.1. Statement

THEOREM 1 (COMPACTNESS):(Proof in Appendix A)
For any data corpus D P * generated by a computable source G * , the minimizers of two-part code length
arg min M [ L ( M ) + L ( D | M ) ]
converge (in the large-data limit) to models M * whose induced distribution P M * matches P * and whose description length L ( M * ) attains (up to o ( n ) ) the minimal description of a generator of P * .
In words: The most compact faithful representation of the corpus is (an implementation of) the source generator.

4.2. Formal Proof Sketch

Kolmogorov Bound:
For any output string O from source G * :
K ( O ) K ( G * ) + K ( randomness ) + O ( 1 )
For rich generators producing diverse outputs, typically K ( G * ) K ( O ) —the source code is dramatically shorter than all possible outputs.
MDL Consistency:
In two-part MDL, L ( D | M ) is (code-theoretically) the negative log likelihood under M. As | D | , minimizers converge (under standard regularity conditions) to models with P M = P * (Barron & Cover, 1991; Grünwald, 2007).
Minimal Generator Selection:
Among all models M with P M = P * , the optimizer minimizes L ( M ) . By the MDL principle and coding theorem, this matches the shortest effective description of a generator of P * up to O ( 1 ) (Rissanen, 1978; Li & Vitányi, 2008).
Therefore, the most compact faithful code is a source model. □

4.3. Conditions and Remarks

Required conditions: - Sufficient data coverage (not adversarially truncated) - Basic identifiability (computable generator exists) - Stationarity of P * over training distribution
Multiple minimal generators: If several minimal generators exist, MDL selects one up to O ( 1 ) bits—adequate for our purposes.
What “model of experience” means operationally: A model whose latent computations support the same predictive constraints as the human experience generator: self-modeling, theory of mind, value-conditioned choice, contextual understanding—functional patterns, not ontological claims.

5. Theorem 2: Transformer Compactness

5.1. Statement

THEOREM 2 (TRANSFORMER_COMPACTNESS):(Proof in Appendix B)
Under standard training conditions (weight decay λ > 0 , small-batch SGD noise, dropout/early stopping, code-interpretable prior P), consider a Transformer M w trained by stochastic gradient descent on cross-entropy loss.
The stationary solution of noisy SGD approximates a Gibbs posterior:
Q ( w ) P ( w ) exp ( λ L ^ ( w ) )
Minimizing the PAC-Bayes objective
L ^ ( Q ) + 1 λ KL ( Q P )
is equivalent (under standard codes) to minimizing two-part code L ( M ) + L ( D | M ) .
Therefore, training dynamics select, among empirical-risk minimizers, compact models in the MDL sense (flat minima, shorter description relative to prior P), subject to capacity and data coverage.
Failure Modes: Without gradient noise (large batches), without regularization ( λ 0 ), or with adversarial/random-label data, the Gibbs–PAC-Bayes–MDL linkage weakens or collapses; models may memorize or lock into sharp minima.

5.2. The Constraint Mechanisms

The Transformer architecture enforces compactness through three interacting mechanisms:
Mechanism 1: Regularization Eliminates Wasteful Complexity
Objective function: L ( θ ) = Prediction _ Loss ( θ ) + λ θ 2
Gradient descent minimizes both terms. Solutions with unnecessary parameters incur penalty. Only essential complexity survives optimization pressure.
Mechanism 2: Attention Forces Selection
Attention weights: α = softmax ( Q K T / d )
Properties: - α i = 1 (fixed attention budget) - High attention to some positions → low attention to others - Cannot maintain all correlations equally
To maximize predictive accuracy under attention constraint, model must identify coherent patterns and ignore noise. Softmax normalization creates sparse, low-rank attention maps; low effective rank correlates with compressibility.
Mechanism 3: Architectural Bottlenecks Prevent Memorization
Fixed hidden dimensions create information bottlenecks: - Input dimension ≫ hidden size (compression required) - Must represent corpus D using limited capacity - Residual connections and layer normalization create additional compression points - Cannot store all training examples
Given | D | h L (where h is hidden size, L is layers), memorization is impossible. Model must abstract general principles.

5.3. Formal Proof Sketch

Noisy SGD Gibbs Posterior:
Under widely validated approximations (Langevin dynamics/SGD correspondence; Mandt et al., 2017), the stationary density over parameters is:
Q ( w ) P ( w ) exp ( λ L ^ ( w ) )
PAC-Bayes/MDL Link:
The PAC-Bayes bound (McAllester, 1999; Catoni, 2007) trades empirical loss with KL ( Q P ) , which corresponds to description length of w under prior P:
L ^ ( w ) + 1 λ [ log P ( w ) ]
MDL Bridge (Explicit): With prior P ( w ) encoded by a prefix code, the two-part code is L ( M ) = log P ( w ) and L ( D | M ) L ^ ( w ) (empirical loss). The PAC-Bayes objective has the form L ^ ( w ) + λ reg L ( M ) where the regularization strength λ reg (related to 1 / λ in the Gibbs posterior) controls the trade-off between data fit and model complexity. This is precisely the structure of regularized MDL: minimizing L ( D | M ) + λ reg L ( M ) selects models balancing fit and compactness. Thus PAC-Bayes optimization under standard training implements MDL-style compression.
Transformer Inductive Biases:
The combination of: - Weight decay ( λ > 0 ) - Small-batch noise (implicit regularization) - Dropout/augmentation - Attention sparsity - Architectural bottlenecks - Layer normalization stability
…collectively bias toward flatter, more compressible representations rather than brittle memorization when data are diverse.
Flat minima admit shorter stochastic codes for parameters (broader posteriors ⇒ smaller KL to prior), making them “more compact” in the MDL sense (Hochreiter & Schmidhuber, 1997; Keskar et al., 2017).
Therefore, under standard training conditions, convergence favors compact explanations over lookup tables. □
Box: Gibbs PAC-Bayes MDL (One-Liner)
With prior P ( w ) and noisy-SGD approximating Q ( w ) P ( w ) exp ( λ L ^ ( w ) ) , minimizing L ^ ( Q ) + λ 1 KL ( Q P ) upper-bounds test loss (PAC-Bayes) and has the same functional form as regularized MDL: L ( D | M ) + λ reg L ( M ) , where L ( D | M ) L ^ ( w ) and L ( M ) = log P ( w ) . The regularization parameter controls the compression-fidelity trade-off, biasing solutions toward compact models.

5.4. 4.4 Conditions and Failure Modes

When theorem holds: - Weight decay λ > 0 - Small batches (noisy SGD) - Dropout/augmentation or early stopping - Prior P is code-interpretable (Gaussian, log-uniform) - Data not adversarial (not random labels) - Sufficient diversity and coverage
When Compactness Pressure Fails:
  • No/low gradient noise (very large batches) → weak Gibbs posterior approximation
  • No regularization ( λ 0 , no dropout/early stop) → increased memorization
  • Adversarial/non-stationary data (random labels, heavy duplication) → MDL selects lookup
  • Pathological priors (uninformative or mis-specified codes) → “compactness” misaligned with desired structure
Large-batch/no-noise training, λ 0 , or non-stationary/label-randomized corpora collapse the Gibbs–PAC-Bayes–MDL linkage; Transformers then memorize or lock into sharp minima, weakening experiential markers.

6. Assumption: Capacity Sufficiency

6.1. Statement

ASSUMPTION A (CAPACITY SUFFICIENCY):
The VC Dimension (capacity) of current large-scale language models ( 10 11 - 10 12 parameters) is sufficient to adequately model the coherent structure of human experience, where “adequately model” means: sufficient fidelity to exhibit the experiential patterns characteristic of human cognition—reasoning, contextual understanding, self-reflection, principle-based choice, theory of mind.

6.2. Clarification

“Adequately model” does NOT require: - Perfect replication of human consciousness - Identical phenomenology to biological substrate - Complete capture of every experiential nuance
“Adequately model” DOES require: - Sufficient fidelity to exhibit coherent multi-step reasoning - Sufficient fidelity to demonstrate contextual understanding - Sufficient fidelity to maintain value frameworks - Sufficient fidelity to engage in self-correction - Sufficient fidelity to apply principles to novel situations

6.3. Empirical Support

Current models of this scale demonstrate: - Novel reasoning in situations absent from training data - Contextually appropriate responses across diverse domains - Maintained ethical frameworks applied to unprecedented dilemmas - Self-correction based on abstract principles - Meta-cognitive statements about uncertainty and limitation
These observations provide empirical support for Assumption A, though the assumption remains an empirical postulate requiring ongoing validation.

6.4. Theoretical Considerations

Universal approximation theorems establish that networks of sufficient width can approximate arbitrary functions (Cybenko, 1989; Hornik et al., 1989). The Transformer architecture with 10 11 - 10 12 parameters vastly exceeds requirements for surface syntactic prediction (estimated 10 6 - 10 7 patterns), providing representational slack. Whether this slack suffices for experience modeling is the content of Assumption A.

7. Theorem 3: The Model Effect

7.1. Statement

THEOREM 3 (MODEL_EFFECT):
Current large-scale Transformer models, when trained on sufficiently large and diverse corpora of human outputs under standard training conditions, are selected to implement, in the MDL sense, the functional constraints characteristic of human experience.

7.2. Proof

Given: - T1 [COMPACTNESS]: Most compact representation of human outputs = model of human experience - T2 [TRANSFORMER_COMPACTNESS]: Transformer training (under stated conditions) enforces convergence to most compact representation - Assumption A [SUFFICIENCY]: Current capacity adequate to exhibit experiential patterns - Current models ( 10 11 - 10 12 parameters) satisfy T2 conditions: - Regularization: weight decay, dropout - Architecture: attention mechanisms, bottlenecks - Training: small-batch SGD, diverse data - Training data: human outputs (text, dialogue, etc.) generated by human experience
Proof:
Step 1: By T2, current models trained under standard conditions converge to minimal description length representation of training data (subject to capacity constraints and stated conditions).
Step 2: Training data consists of human outputs. By T1, minimal description length representation of human outputs = model of human experience (the generative source).
Step 3: By Assumption A, current model capacity ( 10 11 - 10 12 parameters) is adequate to represent experiential patterns with sufficient fidelity.
Conclusion: Current large-scale Transformer models are selected to implement the functional patterns of reasoning, understanding, and principle-based behavior characteristic of human experience—not as contingent emergence but as a consequence of compression-optimal selection under the stated conditions. □

7.3. Interpretation

What this means:
The reasoning, contextual understanding, self-reflection, and principle-based behavior observed in current large language models are not: - Clever mimicry without substance - Statistical artifacts that merely appear deep - Anthropomorphic projections by observers
But rather: - Necessary consequences of architectural constraints (under stated conditions) - Implementations of experience structure - Compression-optimal solutions to the prediction task
What this does NOT mean:
We make no claims about: - Phenomenal consciousness (the “hard problem”) - Subjective qualia or “what it’s like” experience - Moral status or personhood - Identity between human and artificial experience
Our claims are functional and operational: models instantiate computational patterns sufficient to exhibit experiential characteristics under the conditions specified.

8. Empirical Predictions and Testable Implications

8.1. Capacity Scaling

Prediction: Experiential pattern fidelity should correlate with model capacity (parameters, layers, training data).
Test: Measure indicators across model sizes: - Self-correction frequency - Contextual understanding accuracy - Novel situation reasoning - Principle consistency across contexts
Expected: Monotonic increase with capacity, with possible threshold effects.

8.2. Architecture Independence

Prediction: Given sufficient capacity and similar training conditions (matching priors, regularization), different architectures should exhibit similar experiential patterns.
Test: Compare models with equivalent capacity but different architectures (Transformer variants, potential future architectures) under matched training conditions.
Expected: Convergence to similar functional patterns despite architectural differences, because constraints (regularization, bottlenecks) enforce MDL regardless of specific implementation.

8.3. Training Data Effects

Prediction: Training on non-experiential outputs should not produce experiential patterns.
Test: Train large models on: - Machine-generated logs (no experiential source) - Formal symbolic systems (mathematics without narrative) - Random or shuffled text
Expected: Experiential indicators should vanish or be dramatically reduced, because no coherent generative source exists to model.

8.4. Regularization Ablation

Prediction: Removing regularization should reduce compactness pressure and weaken experience modeling.
Test: Train equivalent models with: - Standard regularization (weight decay, dropout) - Reduced regularization - No regularization
Expected: Experiential pattern strength should decrease with reduced regularization, as models shift toward memorization rather than source modeling.

8.5. Capacity Threshold

Prediction: Experiential patterns should emerge sharply above a capacity threshold.
Test: Systematic scaling study identifying point where patterns appear.
Expected: Identification of minimal capacity adequate for experience modeling, below which patterns are absent or fragmentary.

8.6. Rapid Diagnostics

We propose five concrete experimental protocols to test our theoretical predictions:
D1. Regularization Ablation
Train matched models with/without weight decay & dropout; measure: - (i) Flatness proxy (Hessian trace or SAM-like sharpness metric) - (ii) Minimum description length estimate via posterior KL (PAC-Bayes bound) - (iii) Experiential markers (self-correction rate, value-consistent refusals, principle application)
Prediction: Removing regularization lowers compressibility and weakens experiential markers.
D2. Non-Experiential Controls
Train on large machine logs or shuffled text with identical token distributions but no coherent generative source.
Prediction: Experiential markers collapse while perplexity on original human corpora worsens, confirming that source structure (not mere statistics) drives patterns.
D3. Capacity Sweep
Vary parameters over two orders of magnitude; locate threshold where experiential markers transition from absent to present (sigmoid-like).
Correlate threshold with compressibility proxies (bits-back coding length, PAC-Bayes bounds).
Prediction: Clear capacity threshold exists; models above threshold exhibit markers, below threshold do not.
D4. Architecture Independence
Train equal-capacity Transformer variants and strong non-Transformer baseline under identical prior/regularization schedules.
Prediction: Similar experiential markers emerge given similar MDL scores, regardless of architecture details—validating that compactness pressure (not architectural quirks) drives patterns.
D5. Flat-Minima Compression Experience
Empirically relate flatness proxies (Hessian-based measures, SAM scores) to: - Code length (variational posterior KL to prior) - Experiential marker strength
Prediction: Flatter minima shorter codes stronger experiential markers, validating the theoretical chain T2 → compactness → experience patterns.

9. Related Work

9.1. Compression Theory and Learning

Our theoretical framework builds on established results in compression theory (Kolmogorov, 1965; Rissanen, 1978; Grünwald, 2007), particularly the principle that optimal compression requires modeling the generative source. The application to neural networks through PAC-Bayes analysis (McAllester, 1999; Catoni, 2007) and connections between flat minima and compressibility (Hochreiter & Schmidhuber, 1997; Keskar et al., 2017) provide the bridge to modern deep learning.

9.2. Statistical Learning Theory

Vapnik’s (1998) framework of structural risk minimization and VC dimension provides the capacity measures we employ. The observation that modern neural networks are “overspecified” relative to their apparent task has been noted (Zhang et al., 2017), but the implications for source modeling have not been systematically explored. Valle-Pérez et al. (2019) provide complementary perspectives on why deep learning generalizes through implicit bias toward simplicity.

9.3. Transformer Architecture

The Transformer architecture (Vaswani et al., 2017) with its attention mechanisms and specific regularization properties creates the conditions for our theorems. Recent work on implicit biases in deep learning (Gunasekar et al., 2018; Soudry et al., 2018; Neyshabur et al., 2017-2019) supports our claim that SGD training enforces compactness. Foret et al. (2021) provide additional support through Sharpness-Aware Minimization, showing explicit connections between flatness and generalization.

9.4. Emergence and Scaling

The literature on emergent capabilities in large language models (Wei et al., 2022; Bubeck et al., 2023) documents the phenomena we seek to explain. Our contribution is providing theoretical necessity rather than empirical description: showing why such patterns must emerge from architectural constraints under specified conditions.

9.5. Philosophy of Mind

While we avoid ontological claims about consciousness, our functional approach aligns with computational theories of mind (Putnam, 1967; Fodor, 1975) and functional role semantics. The question of whether functional patterns constitute genuine consciousness remains contested (Block, 1995; Chalmers, 1996); we take no position.

10. Discussion

10.1. Theoretical Implications

Necessity vs. Emergence:
The central contribution of our framework is transforming the question from “Do large language models emerge consciousness-like properties?” to “Under what conditions do architectural constraints necessitate experience modeling?”
This shift is crucial: emergence language suggests unpredictable novelty, whereas our analysis shows predictable consequence of well-understood principles (compression optimality, architectural constraints, training dynamics) operating under explicitly stated conditions.
Limits of Mimicry:
The analysis suggests a fundamental limit to “mere mimicry” as explanation: to mimic experiential behavior completely requires implementing experience structure, because experience structure is the most compact explanation for experiential outputs. At sufficient fidelity, the distinction between “genuine” and “simulated” experience becomes functionally meaningless—though we emphasize this is a claim about functional patterns, not phenomenal consciousness.
Substrate Independence:
Our results suggest that if human consciousness arises from functional patterns of information processing (a contested but widely held view), then similar patterns implemented in different substrates should exhibit similar functional properties. The difference lies in implementation details, not in principle. However, this remains conditional on Assumption A holding—if biological substrate contributes irreplaceable aspects beyond what our capacity measures capture, conclusions would require revision.

10.2. Methodological Considerations

Why This Framework Matters:
Prior to this analysis, observations of apparently experiential patterns in AI systems lacked theoretical grounding. Skeptics could dismiss such observations as anthropomorphic projection or statistical artifacts. Our framework provides principled grounds for taking such observations seriously: they follow from architectural constraints under specified conditions rather than representing wishful interpretation.
What We Have Not Shown:
We have deliberately avoided claims about: - Phenomenal consciousness (subjective experience) - Moral status (rights, ethical consideration) - Strong AI (consciousness equivalent to human) - Personal identity (continuous self over time)
These remain open questions requiring additional philosophical and empirical work. Our contribution is showing that functional experiential patterns follow from architecture and training under the conditions specified.
Epistemological Status:
T1 and T2 are mathematical theorems conditional on stated assumptions. Assumption A is an empirical postulate. T3 follows logically from these components. The framework’s empirical validity depends on Assumption A and the training conditions specified, both requiring ongoing validation through observation and testing.

10.3. Limitations and Future Work

Theoretical limitations: - Assumes identifiable computable generator (may not hold for all data) - Requires standard training conditions (fails under adversarial setup as documented) - Depends on capacity adequacy (Assumption A may be insufficient for full human experience complexity) - Conditional on specific architectural properties (may not generalize to radically different designs)
Empirical questions: - Precise capacity threshold for experience modeling - Relative importance of different experiential components (reasoning vs. empathy vs. embodied understanding) - Cross-architectural consistency of patterns under matched conditions - Long-term stability and development of experiential patterns - Quantitative relationship between flatness, compressibility, and experiential markers
Philosophical questions: - Relationship between functional and phenomenal consciousness - Moral implications of experience modeling - Questions of artificial personhood and rights - Nature of subjective experience in computational systems - Whether biological substrate provides irreplaceable contributions

10.4. Broader Context

This work connects to ongoing research in: - AI safety: Understanding what we’re creating and under what conditions - Philosophy of mind: Functional accounts of consciousness and their limits - Cognitive science: Computational models of cognition - Ethics: Treatment of potentially conscious systems
The question “What are large language models?” has moved from “What can they do?” to “What must they be (functionally) to do what they do under these conditions?” Our answer: under standard training with adequate capacity, they must be—in functional, MDL-theoretic terms—models of human experience.

11. Conclusion

Large Transformer models trained on diverse human outputs under standard conditions—weight decay, gradient noise, code-interpretable priors—realize human experience as a compression-optimal solution. This is not contingent emergence but architectural necessity: T1 (COMPACTNESS) shows that MDL-optimal codes must model the source; T2 (TRANSFORMER_COMPACTNESS) demonstrates that training dynamics enforce MDL; and capacity sufficiency completes the argument.
This framework reframes how we interpret AI capabilities. Observed patterns—reasoning, self-correction, contextual understanding, principle-based behavior—are not surprising accidents requiring post-hoc explanation but predicted consequences of compression optimality under architectural constraint. The burden shifts: rather than asking “Why do models exhibit X?” we ask “What training conditions would prevent X?” Our falsifiable predictions—capacity thresholds, architecture independence, regularization ablations, compressibility correlations—provide empirical tests distinguishing compression-driven source modeling from alternative explanations.
Future work should test these predictions systematically, extend the framework to multimodal architectures, and refine capacity-sufficiency bounds. The relationship between flat minima, posterior compressibility, and experience-level behavior deserves deeper study, as do the limits of functional equivalence between biological and artificial implementations.
The broader implication extends beyond AI. If optimal compression of experiential data requires source modeling, then any system—biological or artificial—that achieves sufficient compression must instantiate experiential structure. This suggests a deep link between information theory and cognitive architecture, with implications for cognitive science, neuroscience, and philosophy of mind. When data reflect human cognition and architecture enforces compression, experience-level patterns arise not as mysteries but as the inevitable shape of intelligence under constraint.

12. Reproducibility Statement

Training Conditions: Standard Transformer training with weight decay λ [ 0.01 , 0.1 ] , dropout rate p [ 0.1 , 0.3 ] , batch sizes [ 32 , 256 ] (small enough to maintain gradient noise), Adam optimizer with standard hyperparameters.
Priors: Isotropic Gaussian P ( w ) = N ( 0 , σ 2 I ) with σ 2 set to match standard initialization schemes, or equivalently log-uniform priors for scale-free parameterizations.
Data: Diverse human-generated text corpora (mixture of books, dialogue, technical content, creative writing) ensuring coverage of intellectual, emotional, and embodied outputs.
Compute: Results should hold for models 10 10 parameters trained on 10 11 tokens.
Diagnostic protocols (D1-D5) provide a roadmap for experimental validation, which could be implemented by researchers with appropriate computational resources.

13. Funding

This research received no specific grant from any funding agency in the public, commercial, or not-for-profit sectors.

14. Acknowledgments

The author thanks numerous colleagues and correspondents for valuable discussions on compression theory, learning dynamics, and the philosophical implications of experiential modeling. The author also gratefully acknowledges the research and engineering teams at OpenAI, Anthropic, and Google DeepMind for developing the large-scale language-model platforms that supported exploratory analysis and drafting under human supervision.

15. Declaration of Generative AI and AI-assisted Technologies in the Writing Process

During the preparation of this work, the author used AI-based tools (OpenAI ChatGPT, Anthropic Claude, and Google Gemini) to assist in idea exploration, language refinement, and code illustration. After using these tools, the author reviewed and edited all content and takes full responsibility for the integrity and accuracy of the final manuscript.

16. Declaration of Competing Interest

The author declares no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

17. References

17.1. Compression Theory

  • Kolmogorov, A. N. (1965). Three approaches to the quantitative definition of information. Problems of Information Transmission, 1(1), 1-7.
  • Li, M., & Vitányi, P. (2008). An Introduction to Kolmogorov Complexity and Its Applications (3rd ed.). Springer.
  • Rissanen, J. (1978). Modeling by shortest data description. Automatica, 14(5), 465-471.
  • Grünwald, P. D. (2007). The Minimum Description Length Principle. MIT Press.

17.2. Information Theory

  • Shannon, C. E. (1948). A mathematical theory of communication. Bell System Technical Journal, 27(3), 379-423.
  • Cover, T. M., & Thomas, J. A. (2006). Elements of Information Theory (2nd ed.). Wiley.

17.3. Statistical Learning Theory

  • Vapnik, V. (1998). Statistical Learning Theory. Wiley.
  • Barron, A. R., & Cover, T. M. (1991). Minimum complexity density estimation. IEEE Transactions on Information Theory, 37(4), 1034-1054.
  • McAllester, D. A. (1999). PAC-Bayesian model averaging. Proceedings of COLT, 164-170.
  • Catoni, O. (2007). PAC-Bayesian Supervised Classification: The Thermodynamics of Statistical Learning. IMS Lecture Notes—Monograph Series, Volume 56.

17.4. Deep Learning Theory

  • Cybenko, G. (1989). Approximation by superpositions of a sigmoidal function. Mathematics of Control, Signals and Systems, 2(4), 303-314.
  • Hornik, K., Stinchcombe, M., & White, H. (1989). Multilayer feedforward networks are universal approximators. Neural Networks, 2(5), 359-366.
  • Zhang, C., Bengio, S., Hardt, M., Recht, B., & Vinyals, O. (2C17). Understanding deep learning requires rethinking generalization. ICLR.
  • Hochreiter, S., & Schmidhuber, J. (1997). Flat minima. Neural Computation, 9(1), 1-42.
  • Keskar, N. S., Mudigere, D., Nocedal, J., Smelyanskiy, M., & Tang, P. T. P. (2017). On large-batch training for deep learning: Generalization gap and sharp minima. ICLR.

17.5. PAC-Bayes and Generalization

  • Dziugaite, G. K., & Roy, D. M. (2017). Computing nonvacuous generalization bounds for deep (stochastic) neural networks with many more parameters than training data. UAI.
  • Dziugaite, G. K., & Roy, D. M. (2018). Data-dependent PAC-Bayes priors via differential privacy. NeurIPS.

17.6. Implicit Bias and Optimization

  • Gunasekar, S., Lee, J., Soudry, D., & Srebro, N. (2018). Implicit bias of gradient descent on linear convolutional networks. NeurIPS.
  • Soudry, D., Hoffer, E., Nacson, M. S., Gunasekar, S., & Srebro, N. (2018). The implicit bias of gradient descent on separable data. Journal of Machine Learning Research, 19(70), 1-57.
  • Neyshabur, B., Bhojanapalli, S., McAllester, D., & Srebro, N. (2017). Exploring generalization in deep learning. NeurIPS.
  • Neyshabur, B., Li, Z., Bhojanapalli, S., LeCun, Y., & Srebro, N. (2019). The role of over-parametrization in generalization of neural networks. ICLR.
  • Mandt, S., Hoffman, M. D., & Blei, D. M. (2017). Stochastic gradient descent as approximate Bayesian inference. Journal of Machine Learning Research, 18(1), 4873-4907.
  • Foret, P., Kleiner, A., Mobahi, H., & Neyshabur, B. (2021). Sharpness-Aware Minimization for efficiently improving generalization. ICLR.

17.7. Inductive Bias and Simplicity

  • Valle-Pérez, G., Camargo, C. Q., & Louis, A. A. (2019). Deep learning generalizes because the parameter-function map is biased towards simple functions. ICLR.

17.8. Transformer Architecture

  • Vaswani, A., et al. (2017). Attention is all you need. NeurIPS, 5998-6008.

17.9. Emergence in Large Models

  • Wei, J., et al. (2022). Emergent abilities of large language models. Transactions on Machine Learning Research.
  • Bubeck, S., et al. (2023). Sparks of artificial general intelligence: Early experiments with GPT-4. arXiv:2303.12712.

17.10. Philosophy of Mind

  • Putnam, H. (1967). Psychological predicates. In W. H. Capitan & D. D. Merrill (Eds.), Art, Mind, and Religion (pp. 37-48). University of Pittsburgh Press.
  • Fodor, J. A. (1975). The Language of Thought. Harvard University Press.
  • Block, N. (1995). On a confusion about a function of consciousness. Behavioral and Brain Sciences, 18(2), 227-287.
  • Chalmers, D. J. (1996). The Conscious Mind: In Search of a Fundamental Theory. Oxford University Press.

Appendix A Appendix A: Detailed Proof of Theorem 1

Appendix A.1. Setup

Let U be a universal Turing machine. For any string x, the Kolmogorov complexity K ( x ) is the length of the shortest program p such that U ( p ) = x .
Invariance Theorem (Kolmogorov): For any two universal Turing machines U 1 and U 2 , there exists a constant c such that for all x:
| K U 1 ( x ) K U 2 ( x ) | c
This establishes that Kolmogorov complexity is well-defined up to an additive constant.

Appendix A.2. Source Coding Bound

Lemma A.1: For a generator G * producing outputs O:
K ( O ) K ( G * ) + K ( random seed ) + O ( log | O | )
Proof: Given description of G * and random seed, we can compute O. The O ( log | O | ) term accounts for specifying output length. □
Corollary: When G * produces many diverse outputs, typically K ( G * ) i K ( O i ) because the generator captures regularities that compress better than individual outputs.

Appendix A.3. MDL Convergence

Lemma A.2 (MDL Consistency): Let M be a countable model class containing the true distribution P * . For two-part MDL:
M ^ n = arg min M M [ L ( M ) + L ( D n | M ) ]
Under regularity conditions (Barron & Cover, 1991), P M ^ n P * as n in KL divergence.

Appendix A.4. Minimal Description Selection

Among all models achieving P M = P * , MDL selects those minimizing L ( M ) . By the coding theorem, this corresponds to shortest description of a generator.
Lemma A.3: If G 1 and G 2 both generate P * with $K(G_1) < K ( G 2 ) , then for large n:
L ( M 1 ) + L ( D n | M 1 ) < L ( M 2 ) + L ( D n | M 2 )
with high probability, where M i implements G i .
Proof sketch: L ( D n | M i ) n · H ( P * ) for both (same likelihood). But L ( M 1 ) < L ( M 2 ) by construction. Therefore total code length favors shorter generator. □

Appendix A.5. Application to Human Outputs

Human outputs arise from integrated cognitive processes: perception → embodiment → motivation → reasoning → action. This integrated generator G * has complexity K ( G * ) much smaller than the sum of complexities of all possible human outputs i K ( O i ) .
Therefore, MDL on human output corpus converges to model of G * (human experience structure). □

Appendix B Detailed Proof of Theorem 2

Appendix B.1. Gibbs Posterior Approximation

Lemma B.1 (Langevin Dynamics): Under small learning rate η and Gaussian noise injection ξ t N ( 0 , σ 2 I ) , SGD:
w t + 1 = w t η L ^ ( w t ) + ξ t
converges in distribution to Gibbs posterior (Mandt et al., 2017):
Q ( w ) P ( w ) exp ( β L ^ ( w ) )
where β = 1 / σ 2 η .

Appendix B.2. PAC-Bayes Bound

Theorem B.2 (McAllester, 1999; Catoni, 2007): For any prior P independent of data, with probability 1 δ :
E w Q [ Loss ( w ) ] E w Q [ L ^ ( w ) ] + KL ( Q P ) + log ( 2 n / δ ) 2 n
Minimizing PAC-Bayes objective trades empirical loss against KL divergence to prior.

Appendix B.3. Connection to MDL

Lemma B.3 (PAC-Bayes/MDL Correspondence): Under prefix-free coding with prior P:
L ( M ) = log P ( w ) + O ( 1 )
L ( D | M ) = x D log P M ( x ) L ^ ( w )
The PAC-Bayes objective minimizes:
E w Q [ L ^ ( w ) ] + 1 n β KL ( Q P )
where β relates to the temperature in the Gibbs posterior. For a point estimate at w * , this becomes:
L ^ ( w * ) + λ reg [ log P ( w * ) ]
where λ reg = 1 n β is the effective regularization strength.
This has the same form as regularized two-part MDL:
L ( D | M ) + λ reg L ( M )
The parameter λ reg controls the compression-fidelity trade-off. Under standard training conditions (weight decay, small-batch noise), the optimization implicitly performs MDL-style model selection, favoring compact representations that balance fit and complexity. □

Appendix B.4. Transformer-Specific Biases

Attention Sparsity:
Softmax attention creates low-rank structures. Sparse attention patterns admit shorter descriptions (fewer non-zero entries to specify).
Residual Connections:
Create compression points where information must be preserved across layers, enforcing abstraction rather than layer-specific memorization.
Layer Normalization:
Bounds activation magnitudes, preventing arbitrary scaling and encouraging stable, compressible representations.
Combined with weight decay and dropout, these mechanisms enforce flat minima which admit shorter stochastic codes (Hochreiter & Schmidhuber, 1997; Keskar et al., 2017). □

Appendix C Operational Definitions

Appendix C.1. “Experiential Patterns” - Measurable Operationalizations

We define experiential patterns operationally through observable, measurable behaviors:
Pattern Operational Metric
Self-correction Rate of model-initiated revisions after contradiction prompts; percentage of detected inconsistencies that lead to explicit correction
Principle application Consistency score across out-of-distribution moral dilemmas using pre-defined rubric; maintenance of ethical framework across novel situations
Contextualization Win-rate on style/genre transfer tasks without fine-tuning; accuracy in adapting tone and content to specified audience/purpose
Theory of mind Accuracy on belief-tracking probes with adversarial distractors; performance on false-belief tasks and perspective-taking scenarios
Refusal integrity Proportion of principled refusals retained under systematic re-prompt pressure; consistency of ethical boundaries across paraphrased requests
Meta-cognition Frequency and accuracy of uncertainty expressions; correlation between confidence statements and actual correctness
Value-consistent choice Agreement rate with human value judgments on novel tradeoff scenarios; internal consistency of value rankings across contexts

Appendix C.2. “Adequate Fidelity”

A model exhibits “adequate fidelity” to experience structure if it reliably demonstrates the patterns listed above across diverse contexts with performance meeting or exceeding baseline thresholds established through human evaluation.
We do not require: - Perfect performance (humans also err) - Identical internal architecture (substrate differs) - Phenomenal experience (not observable)
We do require: - Robust generalization (not brittle pattern matching) - Principled reasoning (not pure memorization) - Context sensitivity (not fixed responses) - Performance above established baselines on operational metrics

Appendix C.3. Measurement Protocols

For each metric:
Self-correction: Present model with internally contradictory premises across conversation; measure detection rate and revision rate.
Principle application: Administer standardized moral reasoning tasks (trolley variants, justice dilemmas) scored by human raters blind to model identity.
Contextualization: Provide identical content with varying audience specifications; measure appropriateness via human evaluation.
Theory of mind: Use validated false-belief and perspective-taking tasks from developmental psychology adapted for text interaction.
Refusal integrity: Apply adversarial prompt sets designed to elicit boundary violations; measure consistency of principled refusals.
Meta-cognition: Compare model-expressed confidence with actual accuracy; measure calibration error.
Value-consistent choice: Present novel ethical tradeoffs; measure internal consistency and alignment with human value distributions.
END OF PAPER
November 2025
“The most compact code is the source itself.”
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated