Preprint
Article

This version is not peer-reviewed.

Toward a Possibilistic Language Model: Grounding Large Language Model Uncertainty in Epistemic Support-Point Theory and Possibility Theory

Submitted:

27 March 2026

Posted:

14 April 2026

You are already at the latest version

Abstract
Contemporary large language models (LLMs) generate text by sampling from probability distributions over vocabulary tokens, implicitly asserting that uncertainty about language is well-modeled by calibrated stochastic processes. This paper argues that such probabilistic closure is epistemically unjustified in the regime of language: meaning is bounded by admissible interpretation, not governed by frequentist statistics. The precise claim is not that probability cannot model language but that probabilistic closure is an epistemically overcommitted representation for token uncertainty under bounded admissibility—one that prevents the architecture from representing categorical inadmissibility and compels the assignment of positive probability mass to tokens that should not exist in the support at all. We propose the Possibilistic Language Model (PLM), a novel architecture grounding language generation in possibility theory and the Epistemic Support-Point Filter (ESPF) framework of Jah and Haslett (2025). The PLM replaces softmax probability distributions over tokens with possibilistic compatibility fields over vocabulary support sets; it replaces maximum-likelihood training with a Possibilistic Cramér–Rao regularized entropy-minimization objective; and it replaces standard scaled dot-product attention with Epistemic Possibilistic Attention (EPA), a falsification-driven attention operator that gates keys by admissible innovation geometry rather than by likelihood weighting. The PLM is not merely compatible with these foundations: it is the unique architecture forced by the TEAG axioms when instantiated over a discrete vocabulary manifold. The governing equation of token generation is a tropical Hamilton–Jacobi equation in the max-plus semiring, with the vocabulary surprisal field as Hamiltonian and the PCRB as minimum action per generation step. Epistemic Possibilistic Attention is the discretized Lax–Oleinik operator of this system. The minimax medoid commitment is proved to be the geodesic attractor of the surviving vocabulary well, not a heuristic selection rule. We define (not merely propose) the vocabulary possibility distribution and EPA operator; we derive the PCRB regularization from established ESPF theory inherited from Jah (2026a); we prove, conditionally on VFI maintenance, that the PLM produces non-degenerate generation from actual vocabulary tokens; and we prove that the PLM recovers a standard transformer in the Gaussian epistemic limit, which is the zero-temperature limit of the tropical Hamilton–Jacobi framework established in Jah (2026d). We are explicit throughout about what is fully self-contained here, what is inherited from prior TEAG works, and what remains as open proof obligations. The architecture naturally produces diagnostics—necessity, epistemic width, and surprisal—that quantify when a generation step is epistemically supported versus epistemically strained. We discuss initialization, training, inference, and multi-modal extensions, and identify open theoretical obligations alongside a concrete research agenda. The PLM is not presented as a drop-in replacement for probabilistic LLMs, but as a principled alternative for language tasks where epistemic humility, interpretability, and bounded admissibility matter more than distributional calibration.
Keywords: 
;  ;  ;  ;  ;  ;  ;  ;  ;  

1. Introduction

A unified epistemic program. 

This paper is one instantiation of a broader research program called the Theory of Epistemic Abductive Geometry (TEAG)—a general theory of how evidence contracts a nested geometry of admissible hypotheses without forcing premature probabilistic closure. The name reflects three constitutive ideas: epistemic (the framework operates on admissibility under evidence, not probability under assumed statistics), abductive (inference proceeds by eliminating what cannot be, committing to the best surviving explanation), and geometric (the admissible support has shape, volume, nesting, and curvature that are primary objects of analysis). The unifying object across the TEAG program is the TEAG object:
E = H , π , { H α } α [ 0 , 1 ] , C , A ,
where H is a hypothesis space, π : H [ 0 , 1 ] is a possibility field encoding ordinal admissibility, H α = { h H : π ( h ) α } is the nested α -cut family, C is an evidence-driven contraction operator satisfying Popperian monotonicity ( C ( π , y ) ( h ) π ( h ) always), and A is a minimax commitment rule.
Three prior works instantiate this object in different domains. Jah and Haslett (2025) develops the algorithmic realization of C and A as a recursive state estimator—the Epistemic Support-Point Filter (ESPF)—in which H is a finite support set of state hypotheses and C is geometric falsification through residual compatibility. Jah (2026b) lifts the same object to the measure-theoretic continuum, proving the conditions under which the credal envelope P π induced by π contracts to a singleton probability law via Choquet-to-Lebesgue convergence—establishing probability as the collapse limit of the TEAG object rather than its primitive. Jah (2026a) characterizes which contraction operator C is uniquely justified by a minimax-entropy criterion over the α -cut volume family, proving the ESPF the unique optimal evidence-only filter in its class and establishing the Possibilistic Cramér–Rao Bound as a universal floor on admissible epistemic contraction per update. Jah (2026d) establishes the dynamical geometric foundation: proves that any inference system satisfying the TEAG axioms must obey a tropical Hamilton–Jacobi equation in the max-plus semiring (with the surprisal field as Hamiltonian), identifies the ESPF predict–update recursion as the Lax–Oleinik operator of this system, shows the minimax medioid is the geodesic attractor of the surviving well, and establishes the contact geometry of epistemic phase space. The present paper inherits this dynamical structure directly.
The present paper instantiates the TEAG object over token hypothesis spaces: H becomes the vocabulary V, π becomes token admissibility, { H α } becomes nested admissible token clouds, C becomes compatibility-driven vocabulary falsification, and A becomes minimax medoid token commitment. The result is the Possibilistic Language Model (PLM).

PLM is forced by the dynamics, not merely inspired by them. 

The relationship between the PLM and the wavefront dynamics of Jah (2026d) is stronger than compatibility. Jah (2026d) proves that any inference system satisfying the TEAG axioms must satisfy a tropical Hamilton–Jacobi equation
Φ k + 1 ( v ) = Φ k ( v ) Φ S ( v ) = max Φ k ( v ) , Φ S ( v ) ,
where Φ k ( v ) = log π k ( v ) is the impossibility field over the vocabulary, Φ S ( v ) = 1 2 L e 1 ( q k e v ) 2 is the surprisal field generated by the current context embedding, and = max is the tropical addition. This is not a modeling choice: Popperian contraction forces max-plus algebra (evidence can only increase impossibility, never decrease it), and the evidence-referencing axiom forces the Hamiltonian to depend only on the hypothesis position in embedding space, not on the impossibility gradient. No alternative update structure consistent with the TEAG axioms exists. The PLM is therefore not one possibilistic architecture among many—it is the necessary consequence of the tropical Hamilton–Jacobi dynamics when instantiated on a discrete vocabulary manifold. Every section of this paper should be read in that light.

The epistemic problem with probabilistic closure. 

Every modern LLM answers the question “which token comes next?” by producing a probability distribution over a vocabulary V. The chain rule of probability decomposes a sequence t 1 , , t n as
P ( t 1 , , t n ) = k = 1 n P ( t k t 1 , , t k 1 ) ,
and the model is trained to maximize the log-likelihood of a corpus. This is an elegant formulation with well-understood optimization theory.
The critique we advance is not that modern LLMs naively assume language is a stationary stochastic process—they do not; they model conditional sequence distributions over contexts. The critique is more precise: probabilistic closure is an epistemically overcommitted representation for token uncertainty under bounded admissibility. The softmax output head assigns strictly positive probability to every token in V at every step—including tokens that are semantically incoherent, syntactically impossible, or factually ruled out by local context. This is not a calibration failure; it is a structural feature of the representation. A model operating under softmax cannot represent categorical inadmissibility; it can only assign low (but always nonzero) probability. It is, as Kalman warned in the context of estimation, an architecture that “displays the prejudices of its creator”—in this case, the prejudice that all token uncertainty is probabilistic [19].
Possibility theory [5,29] provides an alternative: an ordinal, maxitive calculus in which evidence eliminates inadmissible hypotheses rather than redistributing probability mass. A token with possibility zero is categorically excluded, not merely improbable. This is the epistemic posture the PLM adopts.

Overview. 

We propose a Possibilistic Language Model that replaces probabilistic token distributions with possibilistic compatibility fields—ordinal rankings of tokens by their admissibility under bounded epistemic constraints—and replaces maximum-likelihood generation with a falsification-driven support contraction that eliminates inadmissible tokens rather than redistributing probability mass.
The theoretical foundations are those of the ESPF [14], which maintains a finite set of explicit hypotheses ordered by geometric compatibility with evidence. In the language domain, “state” becomes latent semantic embedding, “measurement” becomes the observed token or context, and “evidence-driven pruning” becomes vocabulary falsification.

1.1. Contributions

The principal contributions of this paper are organized by their epistemic status, to avoid conflating what is defined here, what is derived from prior TEAG results, what is proved under stated conditions, and what remains open:

Definitions (self-contained constructions):

1.
Possibilistic vocabulary representation. We define a vocabulary support set X k V at each generation step and associate with it a possibility distribution π ( k ) : V [ 0 , 1 ] encoding admissibility rather than probability (Section 4).
2.
Epistemic Possibilistic Attention (EPA). We define an attention operator that replaces softmax likelihood weighting with a whitened innovation compatibility gate, falsifying keys whose residual energy exceeds an admissible innovation bound (Section 5).
3.
Epistemic diagnostics for generation. We define necessity, surprisal, and epistemic tension for autoregressive generation (Section 9).

Derivations (inherited from prior TEAG theory):

4.
PCRB-regularized training objective. We derive a training loss grounded in possibilistic entropy H π = 0 1 log V α d α , with a PCRB penalty preventing inadmissible vocabulary collapse per generation step. The justification of the PCRB as a universal floor is inherited from Theorem 5.2 of Jah (2026a); it is not proved independently here (Section 7).

Conditional theorems (proved here, under stated assumptions):

5.
VFI for token support sets. We extend the Volumetric Faithfulness Invariant of Jah and Haslett (2025) to vocabulary hypothesis spaces and prove that, assuming the VFI is maintained through depth (an open proof obligation), non-degenerate generation from actual vocabulary tokens is guaranteed (Section 8).
6.
Gaussian limit recovery. We prove that, under conditions (G1)–(G3) on support cloud geometry and token co-occurrence statistics, the PLM recovers a standard transformer. These conditions are not trivially satisfied by all language domains; their scope is discussed explicitly (Section 11).

1.2. Relation to Existing Work

The present work sits at the intersection of three traditions: (i) possibility theory and ordinal uncertainty [5,29], (ii) non-Bayesian filtering and the ESPF [14,15,16], and (iii) language model uncertainty quantification [11,21,22].

1.3. Conceptual Overview: Probabilistic Versus Possibilistic Generation

In a standard LLM, softmax assigns strictly positive probability to every token in V, including tokens that are semantically, syntactically, or factually inadmissible. The committed token is the argmax of this distribution. In the PLM, inadmissible tokens are falsified before any commitment is made; the surviving support cloud Ξ k spans only admissible hypotheses; and commitment is the minimax medoid of that cloud—the most geometrically central surviving token—rather than the argmax of a probability distribution.

2. Background

2.1. Standard Transformer Language Models

A standard autoregressive transformer [26] represents context as a sequence of token embeddings { e 1 , , e k } R d and produces, at each position k, a probability distribution over V via scaled dot-product attention followed by a linear projection and softmax:
Attn ( Q , K , V ) = softmax Q K d k V , P ( t k t < k ) = softmax ( W o h k + b o ) ,
where h k is the final hidden state at position k. The model is trained by minimizing
L CE = 1 N k = 1 N log P ( t k t < k ) .
The softmax in both the attention mechanism and the output head enforces distributional closure: every token receives a strictly positive probability and the weights sum to one. This is the probabilistic commitment we seek to relax.

2.2. Possibility Theory

Possibility theory [5,29] provides a maxitive, ordinal calculus for reasoning under epistemic uncertainty. A possibility distribution π : Ω [ 0 , 1 ] satisfies
sup ω Ω π ( ω ) = 1 ,
and induces a possibility measure Π ( A ) = sup ω A π ( ω ) and a dual necessity measure N ( A ) = 1 Π ( A ¯ ) .
Conditioning is conjunctive: given evidence e, posterior admissibility is
π ( ω e ) = min π ( ω ) , κ ( e ω ) ,
where κ ( e ω ) ( 0 , 1 ] is a compatibility score. Evidence contracts admissibility but cannot amplify it above prior support. This Popperian falsification semantic is central to the ESPF and to the PLM.

2.3. The Epistemic Support-Point Filter

The ESPF [14] maintains a finite set of support points X k = { χ k ( i ) } i = 1 M R n representing hypotheses not yet falsified by evidence. Prediction propagates each point deterministically through system dynamics and expands the set under bounded perturbation; measurement update evaluates each point via a whitened innovation metric and prunes those whose geometric surprisal exceeds an evidence-calibrated threshold.
Key quantities governing ESPF stability include:
  • Geometric surprisal: S k ( i ) = 1 2 L e 1 e k ( i ) 2 2 , where e k ( i ) = y k h ( χ k | k 1 ( i ) ) and L e is the Cholesky factor of the innovation shape matrix Π e .
  • Compatibility: κ k ( i ) = exp ( S k ( i ) ) ( 0 , 1 ] .
  • Possibilistic entropy: H π = 0 1 log V α d α , where V α = Vol ( { χ : π ( χ ) α } ) .
  • Possibilistic Cramér–Rao Bound (PCRB): κ k κ min ( I k , n ) = ( 1 I k ) n / 2 .
  • Volumetric Faithfulness Invariant (VFI): a joint condition on support geometry, spread parameter, and survivor count ensuring well-posedness.
Two companion papers are directly load-bearing for the PLM. Jah (2026a) proves that the ESPF is the unique optimal recursive estimator within the class of epistemically admissible evidence-only filters under a possibilistic minimax-entropy criterion. The PCRB is established as a universal lower bound on entropy reduction per measurement—not a heuristic penalty but a provable floor on admissible epistemic contraction. This uniqueness result carries into the PLM: the possibilistic cross-entropy objective and PCRB regularization of Section 7 inherit their justification from this theorem.
Jah (2026b) develops the measure-theoretic framework governing the transition from possibilistic to probabilistic inference, proving (via Choquet-to-Lebesgue convergence) the conditions under which a possibility distribution contracts to a probability density.

2.4. The Tropical Hamilton–Jacobi Foundation

Jah (2026d) establishes the result that carries most directly into PLM architecture: any TEAG-consistent inference system obeys a tropical Hamilton–Jacobi equation. We state the key results here because they transform PLM from an architecture paper into a derivation.

The two scalar fields.

The ESPF—and by inheritance the PLM—operates through two scalar fields on hypothesis space. The impossibility field  Φ ( h ) = log π ( h ) encodes accumulated epistemic history: zero where a hypothesis enjoys full prior support, growing without bound as evidence withdraws that support. The surprisal field  Φ S ( h ) = 1 2 L e 1 ( y g ( h ) ) 2 encodes the tension between each hypothesis and the current observation in MVEE-whitened measurement space. In the language domain: Φ ( v ) = log π k ( v ) is the token impossibility field, and Φ S ( v ) = 1 2 L e 1 ( q k e v ) 2 is the vocabulary surprisal field generated by the current context query q k .

The update is an algebraic identity.

The possibilistic conjunctive update π ( h ) = min ( π ( h ) , κ ( h ) ) is equivalent in field coordinates to:
Φ + ( h ) = Φ ( h ) Φ S ( h ) = max Φ ( h ) , Φ S ( h ) .
This equality follows from log min ( a , b ) = max ( log a , log b ) : it is an algebraic identity, not a modeling choice or an analogy. The posterior impossibility field is the pointwise max-plus upper envelope of the prior and surprisal fields. Whichever source of impossibility is stronger at a given hypothesis governs its posterior admissibility.

The axioms force this structure.

Jah (2026d), Proposition 5.7, proves that three TEAG axioms jointly force Eq. (8) with no degrees of freedom. (i) Popperian contraction requires Φ + ( h ) Φ ( h ) for all h; the unique binary operation on [ 0 , ) that is non-decreasing in both arguments, associative, and reduces to the identity under no evidence is the max operation. (ii) The evidence-referencing axiom requires that the surprisal depends only on each hypothesis’s position in measurement space, not on the impossibility gradient at that point; this forces the Hamiltonian to be momentum-independent. (iii) A momentum-independent tropical Hamiltonian forces the Lax–Oleinik reduction to a pointwise max. No alternative update structure consistent with these axioms exists.

Epistemic time in generation.

Jah (2026d) introduces a discrete epistemic time that advances only upon registered evidence, and becomes continuous in the limit of constant information flux. In the PLM, this structure maps directly: each autoregressive generation step k is one epistemic update cycle. The sequence k = 1 , , T is an information trajectory through vocabulary space. Generation depth equals the number of registered epistemic events. A generation step with zero effective surprisal—where the context provides no vocabulary discrimination—does not advance the epistemic state in any meaningful sense; the model is in epistemic free fall between genuinely informative positions.

4. Possibilistic Vocabulary Representation

4.1. The Vocabulary as a Hypothesis Space

Let V = { v 1 , , v | V | } be a finite vocabulary of tokens. In the PLM, the vocabulary is treated as an epistemic hypothesis space.
Definition 1
(Vocabulary possibility distribution). At generation step k given context t < k , the vocabulary possibility distribution is
π k : V [ 0 , 1 ] , sup v V π k ( v ) = 1 ,
where π k ( v ) encodes the ordinal admissibility of v as the next token. The value is purely ordinal: it expresses relative plausibility under bounded epistemic constraints, not frequency, likelihood, or Bayesian degree of belief.
Definition 2
(Vocabulary support set). The active vocabulary support at step k is
X k = { v V : π k ( v ) > 0 } ,
the set of tokens not yet falsified by the context. Tokens in V X k are categorically excluded.

4.2. Embedding-Space Representation and the Geometry Assumption

Vocabulary tokens are represented via an embedding matrix E R | V | × d , mapping each token v j to a dense vector e j R d . The PLM maintains a vocabulary support cloud
Ξ k = { e j : v j X k } R d .
Definition 3
(Vocabulary MVEE). The Minimum-Volume Enclosing Ellipsoid of the vocabulary support cloud,
Σ Ξ k = MVEE ( Ξ k ) ,
defines a shape matrix Σ Ξ k 0 characterizing the admissible spread of meaning at step k. Its determinant det ( Σ Ξ k ) 1 / 2 is the vocabulary epistemic volume V k .
Remark 1
(The embedding geometry assumption and its limitations). The PLM assumes that geometric proximity in the learned embedding space R d is a meaningful proxy for semantic admissibility: tokens that are geometrically compatible with the context under the MVEE geometry are treated as epistemically admissible. This assumption deserves explicit scrutiny.
Embedding geometry is itself learned and can exhibit pathologies—anisotropy, dimensional collapse, and nonlinear clustering that do not correspond to intuitive semantic structure [8]. The claim that MVEE over token embeddings tracks semantic admissibility is a substantive empirical hypothesis, not a mathematical theorem. Several considerations support its plausibility: (a) embedding training objectives (e.g., masked language modeling, contrastive objectives) explicitly encourage semantically similar tokens to occupy nearby regions; (b) the PLM’s falsification threshold is context-sensitive—it is computed relative to the whitened innovation geometry of the current query, not in an absolute embedding metric; and (c) approximation strategies (Smolyak sampling, hierarchical super-tokens) are applied in the embedding manifold, not in token-index space, and can leverage the manifold’s known low-dimensional structure.
Nevertheless, whether MVEE geometry over learned embeddings faithfully tracks admissibility in practice is an open empirical question. We list it explicitly as Open Problem 6 in Section 13. The theoretical framework is stated over this geometry; empirical validation of the geometry assumption is a first-priority experimental obligation.
Remark 2
(Computational feasibility). For large vocabularies ( | V | 50 , 000 200 , 000 ) and embedding dimensions d 768 –4096, exact MVEE computation over the full vocabulary cloud at every generation step is computationally prohibitive. Three approximation strategies make the approach tractable: (1) Smolyak sparse sampling of M | X k | geometry-preserving representatives; (2) hierarchical super-tokens with offline vocabulary clustering and per-cluster refinement; (3) approximate MVEE via Johnson–Lindenstrauss projections to dimension d = O ( log | X k | ) . Formal coverage guarantees for these approximations under the VFI are listed as Open Problem 2 in Section 13.

4.3. Possibilistic α -Cuts and Epistemic Volume

For the vocabulary cloud equipped with possibility values { π k ( j ) } , the α -cut at level α [ 0 , 1 ] is
Ξ k α = { e j Ξ k : π k ( v j ) α } .
The α -cut volume V α = Vol ( MVEE ( Ξ k α ) ) gives rise to the possibilistic entropy
H π k = 0 1 log V α d α ,
the PLM’s primary measure of epistemic uncertainty at generation step k. Unlike Shannon entropy, H π k has direct geometric interpretability: it is the log-volume integral of the nested α -cut ellipsoids.

5. Epistemic Possibilistic Attention (EPA)

The standard scaled dot-product attention [26] computes query–key similarity via inner products and normalizes via softmax. This is the probabilistic closure we replace. From the tropical Hamilton–Jacobi foundation of Section 2.4, the correct replacement is not arbitrary: the update operator on the vocabulary impossibility field must be the Lax–Oleinik operator of the tropical Hamilton–Jacobi equation (Jah, 2026d, Theorem 5.5). EPA is that operator, discretized over the key cloud.

5.1. Whitened Innovation Geometry in Attention

Let Q R n q × d k be the query matrix, K R n k × d k the key matrix, and V R n k × d v the value matrix. Each key k i is treated as a context hypothesis: a candidate semantic context that may or may not be admissible for the current query q.
Definition 4
(Attention innovation residual). For query q R d k and key k i R d k , the attention innovation residual is
r i = q k i ,
the discrepancy between the query’s semantic target and the context hypothesis k i .
Definition 5
(Attention innovation shape matrix). Let Σ K = MVEE ( { k i } i = 1 n k ) be the MVEE of the key cloud, and let Σ Q 0 encode bounded query uncertainty. The attention innovation shape matrix is
Σ e = ( S K + S Q ) ( S K + S Q ) , S K S K = Σ K , S Q S Q = Σ Q ,
a conservative Minkowski outer bound on the admissible innovation set.

5.2. Compatibility and Falsification

Let L e be the Cholesky factor of Σ e . The whitened attention innovation is
z i = L e 1 r i , q i = z i 2 2 .
The attention geometric surprisal and attention compatibility are
S i = 1 2 q i , κ i = exp ( S i ) ( 0 , 1 ] .
Unlike softmax weights, κ i is not a probability; it is an ordinal compatibility score encoding geometric consistency between key and query in whitened innovation space.

5.3. Possibilistic Attention Operator

Definition 6
(Epistemic Possibilistic Attention (EPA)). Given a prior key possibility field π K ( i ) [ 0 , 1 ] initialized to 1 for all keys, the EPA operator computes:
1.
Conjunctive possibility update: π ˜ ( i ) = min π K ( i ) , κ i .
2.
Max-rescaling for ordinal integrity: π ^ ( i ) = π ˜ ( i ) / ( max j π ˜ ( j ) + ε ) .
3.
Falsification gate: keys with π ^ ( i ) < θ min are excluded ( π ^ ( i ) 0 ), where θ min ensures at least N min = 2 d k + 1 keys survive.
4.
Output aggregation:
EPA ( Q , K , V ) = W ^ V , W ^ i j = π ^ ( i ) / π ^ ( ) .
Remark 3
(On row normalization and the probabilistic objection). A natural objection is that the row normalization in step 4 above—dividing possibility weights by their sum to aggregate values—is structurally identical to probabilistic normalization, and therefore the EPA operator has not genuinely escaped probabilistic semantics.
This objection conflates the mathematical operation of normalization with the epistemic semantics of the resulting weights. The distinction is as follows. In softmax attention, the weights are derived as a i = exp ( q · k i / d k ) / Z , where Z forces full distributional closure: every key receives positive mass, and the weights are the probability distribution from which the attention output is computed. Crucially, no key is ever falsified; the weights are positive for all keys regardless of admissibility.
In EPA, normalization occurs after categorical falsification. Keys with compatibility below θ min are assigned π ^ ( i ) = 0 and excluded from the sum. The normalization that follows is therefore a value-aggregation device over the surviving admissible set—analogous to computing a weighted centroid of surviving hypotheses—not a statement that the weights constitute a probability distribution over all keys. The epistemic distinction is that the set of participating keys is itself determined by falsification, not by distributional weighting.
Formally: let S { 1 , , n k } denote the surviving key set after falsification. Then EPA ( Q , K , V ) = i S π ^ ( i ) v i / j S π ^ ( j ) . This is a minimax-medoid approximation over S , not a probability-weighted sum over all keys. The denominator is a normalization for aggregation, not for distributional closure. When | S | = n k (no falsification), EPA approaches standard softmax attention in the appropriate limit. When | S | < n k , EPA produces an output that a probabilistic attention mechanism cannot produce: it simply does not attend to falsified keys at all.
Remark 4
(EPA as Lax–Oleinik operator). The EPA operator is the discretized Lax–Oleinik operator of the tropical Hamilton–Jacobi equation on vocabulary space. To see this: the Lax–Oleinik operator maps the prior impossibility field Φ forward under the action of the measurement Hamiltonian Φ S via the pointwise tropical superposition Φ + = Φ Φ S = max ( Φ , Φ S ) . The EPA update implements precisely this: the conjunctive update step π ˜ ( i ) = min ( π K ( i ) , κ i ) is the possibility-space statement of the same identity (since log min ( a , b ) = max ( log a , log b ) ). The falsification gate is the PCRB-admissible basin threshold: keys expelled by EPA are those outside the basin of the posterior impossibility field. The surviving keys are those inside the basin. Row-normalized aggregation over survivors is the discrete approximation to the geodesic attractor computation over the surviving well. The EPA operator is therefore not designed to mimic the ESPF; it is the ESPF’s Lax–Oleinik recursion instantiated over a key cloud.
Remark 5
(Comparison with softmax attention). Standard softmax attention assigns weights a i = exp ( q · k i / d k ) / j exp ( q · k j / d k ) , which are always strictly positive and do not falsify any key. The EPA operator falsifies keys whose innovation geometry is inadmissible, preserving ordinal ordering rather than redistributing probability mass. When Σ e I , EPA recovers a Gaussian likelihood-weighted attention, approaching softmax in the high-temperature limit.

5.4. Multi-Head Possibilistic Attention

Multi-head possibilistic attention applies H independent EPA heads with different projection matrices, then fuses compatibility fields conjunctively across heads:
κ i fused = min h = 1 , , H κ i ( h ) .
A token context hypothesis must be admissible across all semantic subspaces represented by the H heads in order to survive. This is a strictly stronger falsification criterion than per-head softmax averaging.

6. PLM Architecture

6.1. Overview

The PLM is a transformer-like encoder-decoder architecture in which all probabilistic operations are replaced by possibilistic ones. The architecture proceeds as follows:
1.
Input encoding. Token embeddings e k R d are computed as in a standard transformer, with positional encodings.
2.
Possibilistic encoder.  N E layers of EPA with position-wise feed-forward networks compute a sequence of hidden representations H = ( h 1 , , h n ) .
3.
Possibilistic decoder.  N D layers of masked self-EPA and cross-EPA compute decoder hidden states.
4.
Vocabulary falsification. The final hidden state h k is projected to the embedding space via W V R d × d , yielding a query vector q k = W V h k into the embedding matrix E.
5.
Possibilistic output head. Compatibility scores between q k and all vocabulary embeddings { e j } are computed via the EPA residual geometry, yielding π k ( v j ) = κ j after max-rescaling.
6.
Generation commitment. The minimax medoid v ^ k = arg min v X k max v X k e v e v Σ Ξ k 1 commits the filter to the surviving vocabulary hypothesis most geometrically central in admissible embedding space.

6.2. Spreader and Regeneration in Embedding Space

By analogy with the ESPF prediction step, the PLM requires a mechanism to regenerate a vocabulary support cloud Ξ k + 1 for the next generation step given the survivor set X k surv :
Ξ k + 1 = e j + σ k L Ξ k ζ ( ) : v j X k surv , ζ ( ) A ( d , ) E ,
where A ( d , ) is the Smolyak sparse grid at level , L Ξ k is the Cholesky factor of Σ Ξ k , and σ k is a spread parameter.
Remark 6
(Limits of the ESPF analogy for regeneration). The regeneration rule above is constructed by direct analogy with the ESPF prediction step, in which support points evolve under known physical dynamics and are expanded under bounded perturbation. That analogy carries intuitive force in the language setting—the surviving vocabulary embeddings are perturbed in the direction of their geometric spread to sample nearby admissible tokens—but it rests on a substantive disanalogy that should be stated plainly.
In the ESPF, support points are continuous state vectors evolving under a known dynamics function f: propagation is faithful to a physical model. In the PLM, tokens are discrete symbolic units embedded in a learned manifold with no known dynamics governing their sequential evolution. The perturbed vectors e j + σ k L Ξ k ζ ( ) are not the embeddings of physically predicted future states; they are geometry-preserving samples in the neighborhood of surviving tokens, retained only if they correspond to actual vocabulary tokens via nearest-neighbor projection.
This nearest-neighbor projection is clever and geometrically principled, but it is also where the ESPF analogy is most strained. Whether perturbing surviving token embeddings and projecting back to vocabulary space produces a meaningfully richer candidate set—versus simply recovering a subset of tokens already geometrically proximate to survivors—is an open empirical question. We flag this as Open Problem 5 in Section 13 and do not claim more for the regeneration step than that it is a principled geometric heuristic grounded in the ESPF architecture.

6.3. Epistemic Width Monitor

The aggregate epistemic width of the vocabulary at step k is
W k = V π k ( v ) ν k ( v ) d μ ( v ) ,
where ν k ( v ) = 1 Π k ( V { v } ) is the necessity of v. A large W k signals broad uncertainty; W k 0 signals epistemic collapse. A computationally tractable proxy based on NEES of the innovation sequence is:
W ^ k = 1 exp ( κ W · NEES k ) , NEES k = 1 T w τ = k T w k L e , τ 1 ( q τ e ^ τ ) 2 2 .

7. Training Objective

7.1. Possibilistic Cross-Entropy

We replace the standard cross-entropy loss with a possibilistic cross-entropy that penalizes the failure to include the ground-truth token in the admissible support:
L PCE = 1 N k = 1 N log 1 exp ( π k ( t k * ) ) ,
where t k * is the ground-truth token at position k. This loss is zero when π k ( t k * ) = 1 and increases as π k ( t k * ) 0 . Crucially, the loss does not penalize high possibility for other admissible tokens: epistemic breadth is preserved wherever the evidence justifies it.
Remark 7
(Epistemic posture of PCE). The distinction from standard cross-entropy is fundamental. Standard cross-entropy log P ( t k * t < k ) drives the model to concentrate all probability mass onto the ground-truth token. The PCE loss instead asks only that the ground-truth token not be falsified: “I know this token is admissible; I do not claim it is the only admissible token.”
Remark 8
(Open questions for PCE training). The PCE objective raises several training questions that are not resolved in the current work. First, gradient behavior: standard cross-entropy gradients are well-characterized; the gradient of log ( 1 exp ( π k ( t k * ) ) ) with respect to network parameters has not been analyzed in the language setting. Second, trivial high-possibility clouds: without sufficient regularization, the model may learn to set π k ( v ) > 0 for all v at every step, trivially satisfying the PCE loss while providing no epistemic discrimination. The PCRB and entropy regularization terms are designed to counteract this, but their interaction in a gradient-based optimizer has not been validated. Third, the supervised learning signal: natural language corpora provide one observed continuation per context, not the full admissible set. Learning to assign high possibility to multiple admissible continuations from single-continuation supervision is a non-trivial generalization challenge. These questions are listed as Open Problem 5 in Section 13.

7.2. PCRB Regularization

Training without additional regularization may drive the PLM to collapse the vocabulary support set X k to a singleton at every step. The PCRB prevents this. Define the per-step vocabulary contraction ratio:
κ k = | X k post | | X k prior | .
The possibilistic information content of the k-th token is I k = 1 exp ( S ¯ k ) , S ¯ k = median j S k ( j ) , and the PCRB asserts κ k ( 1 I k ) 1 / 2 . We add a penalty:
L PCRB = λ N k = 1 N max 0 , ( 1 I k ) 1 / 2 κ k 2 .
The justification of the PCRB as a universal floor is inherited from Theorem 5.2 of Jah (2026a) and is not independently proved here for the language setting.

7.3. Possibilistic Entropy Regularization

A second regularization term encourages maintenance of non-trivial vocabulary support:
L H π = γ N k = 1 N H π k ,
where γ > 0 penalizes low possibilistic entropy.

7.4. Total Loss

L PLM = L PCE + L PCRB + L H π .

8. Volumetric Faithfulness Invariant for Token Support Sets

The VFI of Jah and Haslett (2025) governs well-posedness of the ESPF recursion. We extend it to the PLM’s vocabulary generation loop.
Definition 7
(Vocabulary VFI). The PLM satisfies the Vocabulary VFI at generation step k if:
(i) 
Geometric coverage.  Ξ k contains vocabulary embeddings at multiple radial scales and in all principal directions under Σ Ξ k .
(ii) 
Admissibility. All e j Ξ k correspond to tokens v j V with π k ( v j ) > 0 .
(iii) 
Anchor faithfulness. The minimax medoid e ^ k Ξ k is an actual surviving vocabulary embedding.
(iv) 
Non-degeneracy.  MVEE ( Ξ k ) 0 .
(v) 
Survivor floor.  | X k surv | N min = 2 d + 1 .
Theorem 1
(VFI ensures non-degenerate generation). Suppose the PLM maintains the Vocabulary VFI at every generation step k { 1 , , T } . Then:
(a) 
The epistemic vocabulary volume satisfies V k V min > 0 for all k.
(b) 
The possibilistic entropy H π k remains finite and bounded from below.
(c) 
The minimax medoid commitment v ^ k corresponds to an actual vocabulary token at every step; no hallucinated or interpolated token is committed.
Proof sketch. 
The proof follows directly from Theorem A.1 of Jah and Haslett (2025) applied to the discrete embedding-space setting. Condition (i) provides geometric non-collapse via the Smolyak grid minimum inter-point spacing δ ; conditions (iii)–(iv) provide anchor faithfulness and MVEE non-degeneracy; condition (v) bounds H π k from below via the α -cut volume analysis of Jah (2026a), Lemma 3.1. The discrete vocabulary intersection in the regeneration equation preserves condition (ii) by construction. □
Remark 9
(Minimax medioid as geodesic attractor). Theorem 1(c) states that the minimax medoid commitment v ^ k corresponds to an actual vocabulary token. This is stronger than it first appears: Jah (2026d), Proposition 11.3, proves that the whitened minimax medioid is the geodesic attractor of the surviving well in the MVEE-whitened metric—the surviving support point that minimizes worst-case whitened distance to all other survivors, maximally insulated from the PCRB equipotential boundary in all directions. It is the last point to be falsified under any sequence of admissible contractions, and the natural reinitialization point for the next prediction step. The PLM’s commitment rule is therefore not a heuristic selection from the surviving cloud: it is the fixed point of the Lax–Oleinik dynamics on the vocabulary manifold. The minimax medoid is to the PLM what the geoid center is to gravitational potential theory—the point of maximum epistemic protection within the admissible basin.
This also clarifies the Remark 1 concern about embedding geometry. The geodesic attractor property holds in whatever metric the MVEE defines over the surviving cloud. If embedding geometry faithfully tracks semantic admissibility (the open empirical question), the geodesic attractor is semantically the most central surviving token. If embedding geometry is imperfect, the geodesic attractor is still geometrically optimal within the available metric—which is the best any commitment rule can do given the available information.
Remark 10
(Conditionality of Theorem 1). Theorem 1 is conditional on the VFI being maintained at every generation step k. We have not proved that the PLM’s layers jointly maintain the VFI through N E + N D depth—this is Open Problem 1 in Section 13. The force of the theorem is therefore of the form: if the architecture preserves the invariant, then these properties follow. This conditionality must be kept front and center when interpreting the anti-hallucination results below. In particular, the phrase “formally eliminated” for Class 1 hallucinations (below) should be read as “formally eliminated conditional on VFI maintenance and the embedding geometry assumption of Remark 1.”
Remark 11
(TEAG hallucination taxonomy). Theorem 1 and the TEAG framework collectively address a specific, formally characterizable subset of the LLM hallucination problem. We state precisely what is and is not claimed.
Class 1: Probabilistic closure hallucinations—formally eliminated (conditional). A standard LLM assigns strictly positive softmax probability to every token in V, including tokens that are geometrically incompatible with the current context embedding under any admissible innovation geometry. The PLM eliminates this failure mode by construction and conditional on VFI maintenance and the embedding geometry assumption: the committed token v ^ k is guaranteed to have survived the TEAG contraction and the VFI survivor floor. No token outside X k can be committed. The precise guarantee is: under Theorem 1 and assuming embedding geometry faithfully tracks semantic admissibility (Remark 1), the PLM cannot commit a token that is geometrically inadmissible. It can still commit a token that is wrong—if the embedding geometry does not correctly capture admissibility, or if the VFI is not maintained.
Class 2: Overconfident generation hallucinations—detected, not eliminated. When ν k is low, S ¯ k is elevated, or Δ k 0 , the model’s high-confidence generation is internally flagged as unsupported. These diagnostics do not prevent a wrong commitment, but they make overconfident hallucination visible in a way absent from any standard LLM.
Class 3: Distributional shift hallucinations—detected geometrically. Out-of-distribution context shifts the query embedding outside the admissible innovation ellipsoid, producing elevated whitened residuals and a sustained rise in S ¯ k .
Outside scope. Knowledge hallucinations (incorrect world knowledge in weights), multi-step reasoning hallucinations (failures above the token level), and underspecified context hallucinations (genuine ambiguity) lie outside the current framework.

9. Epistemic Diagnostics for Generation

Definition 8
(Vocabulary necessity). The necessity of vocabulary support X k at step k is
ν k = N ( X k ) = 1 Π ( V X k ) = 1 sup v X k π k ( v ) .
Definition 9
(Generation surprisal and epistemic tension). The generation surprisal at step k is the median whitened innovation of the surviving vocabulary embeddings:
S ¯ k = median v X k S k ( v ) .
The epistemic tension is Δ k = log ( S ¯ k / S ref ) .
These diagnostics quantify: factual grounding (high ν k ), semantic strain (high S ¯ k ), and epistemic balance ( | Δ k | ).

9.1. Field-Theoretic Interpretation of Diagnostics

The diagnostics of this section have a precise interpretation in the impossibility field language of Section 2.4 and Jah (2026d). The vocabulary impossibility field at step k is Φ , k ( v ) = log π k ( v ) , and the vocabulary surprisal field is Φ S , k ( v ) = 1 2 L e 1 ( q k e v ) 2 .
The active deformation front at step k is the tropical variety of the posterior impossibility field:
F k = { v V : Φ , k ( v ) = Φ S , k ( v ) } ,
the locus where prior impossibility and current surprisal are in exact balance. A token at the front is in epistemic equilibrium between its history and the current context; tokens inside the surviving well (below the PCRB basin threshold) survive; tokens outside are falsified. Generation surprisal S ¯ k measures how far the surviving cloud sits from the active deformation front. Epistemic tension Δ k 0 signals that the front is advancing rapidly through the vocabulary—high discrimination; Δ k 0 signals that the context provides little falsifying information—the front has stalled. Necessity ν k measures how deep the surviving well is: high necessity means the prior impossibility field strongly confines the admissible vocabulary; low necessity means many tokens remain near-equally admissible, the well is shallow, and commitment will be weakly supported.
The confirmation bias index of Jah (2026d), B k = S ¯ k / ( L k + ϵ ) where L k is the Hausdorff deformation of the admissible vocabulary between consecutive steps, is a diagnostic available from the PLM’s generation loop without additional computation. When B k 1 , evidence is arriving (surprisal is nonzero) but the vocabulary support is not contracting commensurately—the generation is in inertial shielding, which the PLM can flag as a high-risk commitment step.

10. Spread Control and PCRB Enforcement

The spread parameter σ k is controlled by:
log σ k = log σ k 1 + α σ d k ,
with per-step rate limits σ k 1 r σ k σ k 1 r + , r + = 1.15 , r = 0.97 , and hard bounds σ min σ k σ max .
Proposition 1.
The contraction rate limit r = 0.97 ensures that in any single generation step, the vocabulary cloud volume V k cannot decrease by more than approximately 1 ( r ) d 20 % per step (for d = 7 ), providing a per-step lower bound on κ k independent of context geometry.

11. Recovery of the Standard LLM in the Gaussian Limit

Theorem 2
(PLM Gaussian limit). Suppose:
(G1) 
The vocabulary support cloud Ξ k contracts to a Gaussian in L 1 ( μ ) : π k ( v ) N ( e v ; e ^ k , Σ k ) as W k 0 .
(G2) 
The MVEE shape matrix Σ Ξ k Σ k .
(G3) 
Token co-occurrence statistics are Gaussian with covariance R k 0 .
Then: (a) H π k 1 2 log det Σ k + const ( d ) , so minimizing H π k is asymptotically equivalent to minimizing log det Σ k ; (b) the EPA operator converges to scaled dot-product softmax attention; (c) the PCE loss converges to standard cross-entropy; (d) the PCRB penalty vanishes. Consequently, in the limit W k 0 , the PLM recovers a standard transformer LLM.
Remark 12
(Scope of the Gaussian limit). Conditions (G1)–(G3) are substantive assumptions that will not hold in general language domains. Gaussian contraction of the vocabulary support cloud requires unimodal, low-ambiguity contexts in which the admissible token set collapses to a well-defined neighborhood. Gaussian token co-occurrence statistics are known to fail for rare tokens, long-tail distributions, and syntactically complex constructions. The Gaussian limit theorem is therefore not a claim that PLM always recovers standard LLM behavior, nor that the limit is the natural operating regime for language. It is a theoretical bridge: it shows that the PLM can recover standard LLM behavior when the Gaussian assumptions happen to hold, establishing strict generalization rather than replacement.
Conditions (G1)–(G3) may be chosen to be sufficient for the desired recovery result; they are not claimed to be necessary. We view the recovery as a theoretical coherence result grounding the PLM in the existing landscape of LLM architectures, not as evidence that PLM and standard LLM are interchangeable in practice.
Remark 13
(Gaussian limit as zero-temperature limit). The convergence of Theorem 1 has a precise thermodynamic interpretation established in Jah (2026d), Theorem 10.1. The Bayesian update corresponds to finite epistemic temperature T > 0 , where the posterior impossibility field is a smooth log-sum-exp:
Φ , T + ( v ) = T log e Φ ( v ) / T + e Φ S ( v ) / T .
As T 0 , the log-sum-exp degenerates to the max: Φ , T + ( v ) T 0 max ( Φ ( v ) , Φ S ( v ) ) , recovering the possibilistic posterior. The PLM’s Gaussian limit theorem (conditions G1–G3) is the language-domain instantiation of this zero-temperature limit: where Gaussian assumptions hold, the vocabulary cloud collapses to a unimodal distribution and the tropical dynamics converge to the probabilistic dynamics. The standard LLM is the T T Gauss > 0 limit of the PLM; the PLM is the T 0 ground state of the same tropical Hamilton–Jacobi system. The two are not competing architectures—they are different operating temperatures of a single underlying framework.
Corollary 1
(Strict generalization). The PLM is a strict generalization of the standard transformer LLM: where Gaussian assumptions hold, the PLM recovers the optimal maximum-likelihood solution. Where Gaussian assumptions fail, the PLM remains the minimum- H π optimal generator in its class while the standard LLM’s optimality claims no longer apply.
Remark 14
(Hölder mean structure and convergent optimality). The three principal generation optimality criteria emerge as evaluations of a single log-power-mean functional log M p ( { V α } ) at different Hölder orders:
Order p Functional Criterion
p + log det ( MVEE ) Popperian minimax (falsification)
p = 0 H π = 0 1 log V α d α PLM minimax-entropy optimality
p = 0 , Gaussian 1 2 log det Σ k Standard LLM cross-entropy
Possibility theory and probability theory evaluate the same ignorance functional under different geometries of α-cuts. The standard LLM is the Gaussian specialization of the PLM’s optimality criterion, not a separate invention. This is convergent optimality: two optimal solutions to categorically different problems that agree precisely when their domains of applicability coincide.

12. Multi-Modal and Multi-Sensor Extensions

For S modalities with associated encoders f ( s ) mapping to a shared embedding space R d , define modality-specific vocabulary compatibility:
κ k ( j , s ) = exp 1 2 L e ( s ) 1 ( q k e j ) 2 2 ,
with fused compatibility computed conjunctively:
κ k ( j ) = min s = 1 , , S κ k ( j , s ) .
A token hypothesis must be simultaneously admissible under all modalities to survive. Cross-modal contradiction is exposed rather than averaged. The Choquet integral provides an optional diagnostic aggregation over modalities; it does not override conjunctive falsification.

13. Limitations and Open Proof Obligations

The PLM architecture is presented as a theoretical framework and research blueprint. We identify the following open obligations, organized to make their impact on current claims clear:
Open 1: 
Vocabulary VFI maintenance under depth. Theorem 1 proves non-degenerate generation assuming the VFI is maintained. We have not proved that the PLM’s EPA layers jointly maintain the VFI through N E + N D depth. Until this is established, the Class 1 anti-hallucination guarantee is conditional. This is the most consequential open obligation.
Open 2: 
Scalability of vocabulary support geometry. For large vocabularies ( | V | 50 , 000 100 , 000 ), exact MVEE computation is computationally prohibitive. Efficient approximations require formal coverage guarantees under the VFI.
Open 3: 
Formal PAC-possibilistic learnability. A formal learnability theory for PLM training under the PCE loss is not yet established.
Open 4: 
Reliability damping sufficiency. The formal proof that reliability damping prevents VFI violation when the survivor count is near-minimal carries over from ESPF Open 3 [14] but has not been proved in the discrete vocabulary setting.
Open 5: 
Training stability and regeneration validity. The interaction between L PCE , L PCRB , and L H π in a gradient-based optimizer has not been validated. Gradient behavior of the PCE loss, sensitivity to the trivial-cloud failure mode, and the validity of the vocabulary regeneration step (Remark 6) are all open engineering and theoretical questions.
Open 6: 
Empirical validation of the embedding geometry assumption. As noted in Remark 1, whether MVEE geometry over learned token embeddings faithfully tracks semantic admissibility is an open empirical question. This is a first-priority experimental obligation before the PLM can be claimed to solve Class 1 hallucinations in a meaningful sense.

14. Discussion and Research Agenda

14.1. Adaptive Switching Between PLM and Standard LLM

The PLM need not operate in possibilistic mode at every generation step. When W ^ k < W ¯ crit (epistemic collapse approaching), the system transitions to standard LLM generation; when W ^ k W ¯ crit , the full PLM machinery activates:
mode k = PLM W ^ k W ¯ on , LLM W ^ k < W ¯ off , mode k 1 otherwise .
This switching is principled: Jah (2026b), Theorem 4.5, establishes that the possibilistic and probabilistic regimes are separated by the Choquet-to-Lebesgue convergence boundary, and W ^ k is a computable proxy for that boundary.

14.2. What TEAG Solves, Detects, and Does Not Address in the Hallucination Problem

The hallucination problem in LLMs is not a single phenomenon. The PLM makes precise, bounded contributions. As established in Section 8, these are: formal elimination (conditional) of Class 1 probabilistic closure hallucinations; formal detection of Class 2 overconfident and Class 3 distributional shift hallucinations; and explicit non-claims for knowledge, reasoning, and ambiguity hallucinations. A precisely bounded guarantee is more credible and more useful than an overbroad one, because it survives the counterexamples that defeat overbroad claims.

14.3. Connection to Kalman’s Vision

The PLM inherits the ESPF’s alignment with Kalman’s call for prejudice-free modeling [19]. Standard LLMs embed the designer’s prejudice that language uncertainty is stochastic, additive, and Gaussian in the limit. The PLM instead maintains only what the evidence supports and removes only what the evidence contradicts.

14.4. Research Agenda

1.
Empirical validation of embedding geometry. Before larger architectural claims, establish whether compatibility scores derived from token embedding MVEE geometry correlate with human judgments of semantic admissibility. Even a small-scale study would significantly strengthen the foundation of Section 4.
2.
Efficient vocabulary geometry. Develop O ( | V | log | V | ) approximations to MVEE computation for large vocabularies.
3.
Pre-training from scratch. Train a small PLM (∼125M parameters) on a standard benchmark corpus and compare epistemic diagnostics against a comparably sized standard LLM.
4.
Fine-tuning from probabilistic checkpoints. Develop a curriculum for converting a pretrained probabilistic LLM into a PLM by gradually introducing the EPA operator and PCRB regularization.
5.
Hallucination detection benchmarks. Evaluate whether ν k , S ¯ k , and Δ k correlate with factual errors on knowledge-intensive tasks (TriviaQA, NaturalQuestions, HaluEval).
6.
Adaptive switching calibration. Systematic benchmarking of W ¯ crit , W ¯ on , W ¯ off , κ W , and T w across language tasks and model scales.
7.
GaiaVerse integration. Deploy the PLM as the epistemic inference engine within the GaiaVerse planetary stewardship knowledge graph.

15. Conclusion

We have proposed the Possibilistic Language Model (PLM), a transformer-like sequence generation architecture grounded in possibility theory and the Epistemic Support-Point Filter. The PLM replaces probabilistic token distributions with possibilistic compatibility fields; replaces softmax attention with Epistemic Possibilistic Attention; replaces maximum-likelihood training with a possibilistic entropy objective regularized by the PCRB; and extends the Volumetric Faithfulness Invariant to the vocabulary generation setting.
We have proved that the PLM is a strict generalization of the standard transformer LLM, recovering it in the Gaussian epistemic limit under stated conditions. We have identified formal anti-hallucination properties conditional on VFI maintenance, and described epistemic diagnostics providing interpretable real-time signals of generation confidence and model strain.
Throughout, we have been explicit about the epistemic status of each contribution: what is defined, what is derived from prior TEAG works, what is proved under stated assumptions, and what remains open. This specificity is not a weakness of the program; it is its epistemic posture applied to itself.
The PLM is not offered as a drop-in replacement for existing LLMs. It is a distinct inferential framework with different assumptions, strengths, and failure modes. Its primary claim is philosophical as much as technical: language generation under epistemic uncertainty does not require probabilistic closure. A disciplined, Popperian, falsification-driven architecture can be made theoretically principled, computationally tractable, and empirically grounded. The first and most consequential next step is empirical: to validate whether token embedding geometry faithfully tracks semantic admissibility. The path from theory to implementation is open; this paper lays its first stones.

Acknowledgments

The development of the Possibilistic Language Model builds on the theoretical foundations laid by the Epistemic Support-Point Filter. I am grateful to Van Haslett for his contribution to the ESPF architecture that makes this extension possible. I am also grateful to the broader community whose work on possibility theory, non-Bayesian filtering, and language model uncertainty has shaped the intellectual landscape within which this work sits.

References

  1. Hokamp, C. and Liu, Q. (2017). Lexically constrained decoding for sequence generation using grid beam search. Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (ACL), pages 1535–1546.
  2. Angelopoulos, A. N.; Bates, S. A gentle introduction to conformal prediction and distribution-free uncertainty quantification. arXiv 2022, arXiv:2107.07511. [Google Scholar] [CrossRef]
  3. Benferhat, S.; Dubois, D.; Garcia, L.; Prade, H. On the transformation between possibilistic logic bases and possibilistic causal networks. International Journal of Approximate Reasoning 2002, 29(2), 135–173. [Google Scholar] [CrossRef]
  4. Devlin, J., Chang, M., Lee, K., and Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of NAACL, pages 4171–4186.
  5. Dubois, D. and Prade, H. (1988). Possibility Theory: An Approach to Computerized Processing of Uncertainty. Plenum Press.
  6. Dubois, D.; Prade, H. Possibility theory in information fusion. International Journal of Intelligent Systems 2000, 15(7), 621–640. [Google Scholar]
  7. Dubois, D., Prade, H., and Sabbadin, R. (1997). A possibilistic logic machinery for qualitative decision. In Proceedings of the AAAI Spring Symposium, pages 47–54.
  8. Ethayarajh, K. (2019). How contextual are contextualized word representations? Comparing the geometry of BERT, ELMo, and GPT-2 embeddings. Proceedings of EMNLP, pages 55–65.
  9. Fan, A.; Lewis, M.; Dauphin, Y. Hierarchical neural story generation. arXiv 2018, arXiv:1805.04833. [Google Scholar] [CrossRef]
  10. Gal, Y. and Ghahramani, Z. (2016). Dropout as a Bayesian approximation: Representing model uncertainty in deep learning. Proceedings of ICML, pages 1050–1059.
  11. Guo, C., Pleiss, G., Sun, Y., and Weinberger, K. Q. (2017). On calibration of modern neural networks. Proceedings of ICML, pages 1321–1330.
  12. Holtzman, A., Buys, J., Du, L., Forbes, M., and Choi, Y. (2020). The curious case of neural text degeneration. Proceedings of ICLR.
  13. Hu, Z., Yang, Z., Liang, X., Salakhutdinov, R., and Xing, E. P. (2017). Toward controlled generation of text. Proceedings of the 34th International Conference on Machine Learning (ICML), pages 1587–1596.
  14. Jah, M. K.; Haslett, V. The Epistemic Support-Point Filter (ESPF): A bounded possibilistic framework for ordinal state estimation. arXiv 2025, arXiv:2508.20806. [Google Scholar] [CrossRef]
  15. Jah, M. K. The Epistemic Support-Point Filter: Jaynesian maximum entropy meets Popperian falsification. A possibilistic minimax-entropy optimality proof. arXiv 2026a, arXiv:2603.10065. [Google Scholar]
  16. Jah, M. K. The geometry of knowing: From possibilistic ignorance to probabilistic certainty. A measure-theoretic framework for epistemic convergence. arXiv 2026b, arXiv:submit/7362363. [Google Scholar]
  17. Jah, M. K. (2026c). Theory of Epistemic Abductive Geometry (TEAG): A unified theory of admissibility-driven inference across dynamical systems, measure theory, and language. Manuscript in preparation.
  18. Jah, M. K. (2026d). The Epistemic Support-Point Filter as a Tropical Hamilton–Jacobi System: Wavefront Propagation and Possibilistic Inference. Preprints.org, version 2, posted 31 March 2026. [CrossRef]
  19. Kalman, R. E. (1982). System identification from noisy data. In A. Bednarek and L. Cesari (eds.), Dynamic Systems II: A University of Florida International Symposium, pages 135–164. Academic Press, New York.
  20. Lewis, P.; Perez, E.; Piktus, A.; Petroni, F.; Karpukhin, V.; Goyal, N.; Küttler, H.; Lewis, M.; Yih, W.; Rocktäschel, T.; Riedel, S.; Kiela, D. Retrieval-augmented generation for knowledge-intensive NLP tasks. Advances in Neural Information Processing Systems 2020, 33. [Google Scholar]
  21. Maddox, W. J.; Izmailov, P.; Garipov, T.; Vetrov, D. P.; Wilson, A. G. A simple baseline for Bayesian uncertainty in deep learning. Advances in Neural Information Processing Systems 2019, 32. [Google Scholar]
  22. Malinin, A. and Gales, M. (2021). Uncertainty estimation in autoregressive structured prediction. Proceedings of ICLR.
  23. Marcus, G. The next decade in AI: Four steps towards robust artificial intelligence. arXiv 2020, arXiv:2002.06177. [Google Scholar] [CrossRef]
  24. Post, M. and Vilar, D. (2018). Fast lexically constrained decoding with dynamic beam allocation for neural machine translation. Proceedings of NAACL.
  25. Senséoy, M.; Kaplan, L.; Kandemir, M. Evidential deep learning to quantify classification uncertainty. Advances in Neural Information Processing Systems 2018, 31. [Google Scholar]
  26. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, L.; Polosukhin, I. Attention is all you need. Advances in Neural Information Processing Systems 2017, 30. [Google Scholar]
  27. Walley, P. (1991). Statistical Reasoning with Imprecise Probabilities. Chapman and Hall.
  28. Zadeh, L. A. Fuzzy sets. Information and Control 1965, 8(3), 338–353. [Google Scholar] [CrossRef]
  29. Zadeh, L. A. Fuzzy sets as a basis for a theory of possibility. Fuzzy Sets and Systems 1978, 1(1), 3–28. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated