Preprint
Article

This version is not peer-reviewed.

Soft Model Selection and Structured Assembly of Observables for Micro–Macro Transitions, Operational Relativity, and Emergent Thermodynamics

Submitted:

23 February 2026

Posted:

28 February 2026

You are already at the latest version

Abstract
An operational meta-model is presented for transitions among quantum, classical, relativis- tic, and thermodynamic descriptions without forcing a single master state space. Unification is performed at the level of observable predictions: each formalism produces an output in a common space Y defined by a feature map Φ (moments, spectra, correlations, or other functionals). Convex weights are assigned via a standard soft selection rule (softmax / Gibbs form) from losses, with entropic regularization and a complexity penalty (AIC/BIC/MDL) to reduce bias toward overly expressive models. Physics-facing priors are encoded through calibrated dimensionless knobs for decoherence/classicality, relativistic severity, and ther- modynamicity, yielding a regime-aware gating layer. A simple out-of-catalog diagnostic (surprise/residual monitoring) is included to flag persistent mismatch that may indicate missing observables or missing models. A minimal case study template (harmonic oscillator with a bath) and an acid test in the NISQ regime (critical decoherence) are outlined as reproducible validation pathways.
Keywords: 
;  ;  ;  ;  ;  ;  ;  ;  ;  

1. Motivation, positioning, and scope

“Essentially, all models are wrong, but some are useful.”
— George E. P. Box [1]
This manuscript does not propose a new physical theory, nor does it claim to unify quantum mechanics, classical dynamics, relativity, and thermodynamics at the level of a single state space. Instead, it proposes an operational meta-model that assembles predictions in a shared observable space. The contribution is therefore methodological: a disciplined way to combine existing regime-specific models, quantify confidence through soft weights, and detect when the available model catalog is insufficient.

Contributions (Incremental, Operational)

The contributions are modest and practical:
  • A common observable space Y is formalized via a feature map Φ , and regime transitions are treated as decisions in Y (rather than forcing incompatible state spaces into a single master model).
  • A standard entropically-regularized soft selection rule (softmax/Gibbs form) is used to obtain convex weights from losses, augmented with a complexity penalty to mitigate overfitting.
  • A small set of physics-inspired knobs (decoherence/classicality, relativistic severity, thermodynamicity) is used as operational priors that bias weights toward plausible regimes.
  • A simple out-of-catalog diagnostic (surprise/residual monitoring) is proposed to flag when the chosen Φ and the candidate model set fail to explain data.

Relation to Prior Work

The weighting mechanism is closely related to well-known ideas such as Bayesian model averaging, Gibbs posteriors/PAC-Bayes methods, and mixture-of-experts gating. The goal is not to replace these frameworks, but to provide an explicit physics-facing operational interface: (i) a shared observable space Y tailored to measurement constraints, (ii) regime priors encoded by calibrated knobs, and (iii) a simple diagnostic for model-set inadequacy. In this sense, the novelty is primarily in the packaging and operationalization for micro–macro and regime-transition problems.

Scope and Limitations

This framework is only as good as (a) the chosen observable map Φ and (b) the candidate model catalog. The knobs ( ξ , r , θ ) are not fundamental constants; they are operational summaries that must be calibrated for each experimental context. Accordingly, the method should be read as a principled decision layer on top of existing models, not as a substitute for domain-specific mechanistic modeling.

2. Level of Unification: the Output Space Y and the Map Φ

Let X be the vector of observables to be predicted. Each formalism M i produces a natural object Z i (trajectory, distribution, spectrum, quantum channel, etc.). A feature map is defined as
Φ : Z Y , X ^ i : = Φ ( Z i ) Y .
Typical examples:
  • Φ ( P ) = ( μ , σ 2 , κ 3 , , κ m ) for a distribution P.
  • Φ ( S ) = ( S ( ω 1 ) , , S ( ω m ) ) for a sampled spectrum.
  • Φ ( C ) = ( C ( τ 1 ) , , C ( τ m ) ) for sampled correlations.
In practice, Y R m .

2.1. The Choice of Y is not Neutral (an Explicit Operational Hypothesis)

Selecting Φ and Y encodes which degrees of freedom are considered relevant in the micro–macro transition. In that sense, the coarse-graining implicit in Φ is an operational hypothesis: it assumes that certain functionals (moments, spectra, correlations) capture what is essential for the regime of interest. This choice should be guided by invariances, scales, and measurement constraints.

3. Assembly: Convex Combination and Structure-Preserving Operators

3.1. Base Meta-Predictor (Convex)

A minimal assembler is defined by
X ^ = i k i X ^ i , i k i = 1 , k i 0 .

3.2. Structured assembly (when constraints exist)

When the observable has structure (e.g., normalized distributions, PSD matrices, density matrices), a naive linear combination may violate constraints (positivity, normalization, etc.). In such cases, a structure-preserving assembly operator is used:
X ^ = A ( { X ^ i } , k ) ,
where A is chosen to preserve structure. Examples:
  • Distributions: convex mixture in the probability simplex (preserves normalization and non-negativity).
  • Density matrices: convex mixture of states ρ = i k i ρ i (preserves positivity and trace 1).
  • Constrained observables: projection Π onto a valid convex set: X ^ = Π ( i k i X ^ i ) .
Equation (1) is used as the minimal reference, and (2) as its natural extension.

4. Weights from Loss: Soft Selection with Regularization and Complexity

Let E i 0 be a total loss associated with model i. Define:
k i = exp ( λ E i ) j exp ( λ E j ) , λ > 0 .

4.1. Convex Interpretation (Minimum Loss + Entropy)

Rule (3) solves
min k Δ i k i E i + 1 λ i k i ln k i ,
where Δ = { k i 0 , i k i = 1 } and the second term is entropic regularization.

4.2. Loss Definition in Y

If a reference X exists (measurement or “fine” simulation),
E i data = d ( X ^ i , X ) .
Common examples:
d RMSE ( a , b ) = 1 m = 1 m ( a b ) 2 , d KL ( P Q ) = x P ( x ) ln P ( x ) Q ( x ) .

4.3. Complexity Penalty (AIC/BIC/MDL)

To reduce bias toward overly complex models,
E i = E i data + β Ω i ,
where Ω i can be AIC/BIC/MDL [2,3,4].

4.4. Knobs as Calibrated Parametrizations

The functional forms used for ( ξ , r , θ ) are chosen for monotonicity, boundedness, and numerical stability. Other monotone parametrizations would be equally valid. In applications, these mappings should be calibrated (or learned) from validity data, measurement noise levels, and known scale separations.

5. Physics-Facing Priors from Knobs

Define dimensionless knobs (examples; operational use only):
ξ = Γ t obs ( decoherence / classicality ) ,
β S = v 2 c 2 , β G = 2 G M r c 2 , r = 1 exp [ ( β S + β G ) ] ( relativistic severity ) ,
θ = t obs τ mix N eff ( thermodynamicity ) .
Let q = e ξ . A minimal factorized prior is:
w Q R = q r , w Q = q ( 1 r ) , w R = ( 1 q ) r , w C = ( 1 q ) ( 1 r ) ,
with w · = 1 .

5.1. Combining Prior + Loss

k i = w i exp ( λ E i ) j w j exp ( λ E j ) .

6. Hardening: Knob Coupling and Out-of-Catalog Monitoring

6.1. Coupling Among Knobs (Validity Covariance)

Knobs ( ξ , r , θ ) are often correlated in real systems. Let p : = ( ξ , r , θ ) and introduce a covariance matrix Σ p 0 . A Mahalanobis-type penalty is defined by
E cpl ( p ) : = 1 2 ( p μ ) Σ p 1 ( p μ ) ,
where μ is an operational center (possibly context-dependent).

6.2. Hardened Total Loss

E i : = E i data + α E cpl ( p ) + β Ω i ,
and weights follow (10).

6.3. Surprise as a Diagnostic (Not a Physical Law)

Define residual loss E res : = d ( X ^ , X ) and
S : = E res i k i E i data .
The surprise indicator is intended as an engineering diagnostic summarizing persistent mismatch between assembled predictions and reference observations. No claim is made that it is, by itself, a thermodynamic entropy production law; any thermodynamic reading should be treated as a hypothesis to be tested in open-system settings.

7. Minimal Case Study Template and a NISQ Acid Test

7.1. Minimal Reproducibility Checklist

Any case study should report: (i) the definition of Φ and the dimension of Y , (ii) the metric d used for loss, (iii) the candidate models compared, (iv) the calibration procedure for ( λ , α , β ) and for knobs, and (v) weight trajectories k i with residual monitoring.

7.2. NISQ Acid Test: Critical Decoherence

For a qubit with dephasing characterized by T 2 ,
ξ t obs T 2 , r 0 ,
and θ depends on effective mixing/noise. The operational critical time is defined by
t crit : k Q ( t crit ) = k th , k th ( 0 , 1 ) ( e . g . 0.5 ) .

8. Qualitative Regime Figure

Preprints 200049 i001

9. Conclusions

An operational meta-model for regime-aware prediction assembly has been presented. The framework combines (i) a shared observable space, (ii) soft loss-based weighting with regularization and complexity control, (iii) physics-facing regime priors via calibrated knobs, and (iv) a diagnostic for persistent out-of-catalog mismatch. The proposal is deliberately incremental: it is intended as a practical decision layer on top of existing mechanistic models, with clear calibration and validation hooks.

References

  1. Box, G. E. P.; Draper, N. R. Empirical Model-Building and Response Surfaces; John Wiley & Sons, 1987. [Google Scholar]
  2. Akaike, H. A new look at the statistical model identification . IEEE Transactions on Automatic Control 1974, 19(6), 716–723. [Google Scholar] [CrossRef]
  3. Schwarz, G. Estimating the dimension of a model . Annals of Statistics 1978, 6(2), 461–464. [Google Scholar] [CrossRef]
  4. Rissanen, J. Modeling by shortest data description . Automatica 1978, 14(5), 465–471. [Google Scholar] [CrossRef]
  5. Hoeting, J. A.; Madigan, D.; Raftery, A. E.; Volinsky, C. T. Bayesian model averaging: A tutorial . Statistical Science 1999, 14(4), 382–417. [Google Scholar] [CrossRef]
  6. Catoni, O. PAC-Bayesian Supervised Classification: The Thermodynamics of Statistical Learning, IMS Lecture Notes–Monograph Series, 2007.
  7. Amari, S.-I. Information Geometry and Its Applications; Springer, 2016. [Google Scholar]
  8. Jacobs, R. A.; Jordan, M. I.; Nowlan, S. J.; Hinton, G. E. Adaptive mixtures of local experts . Neural Computation 1991, 3(1), 79–87. [Google Scholar] [CrossRef] [PubMed]
  9. Zurek, W. H. Decoherence, einselection, and the quantum origins of the classical. Rev. Mod. Phys. 2003, 75, 715–775. [Google Scholar] [CrossRef]
  10. Breuer, H.-P.; Petruccione, F. The Theory of Open Quantum Systems; Oxford University Press, 2002. [Google Scholar]
  11. Jaynes, E. T. Information Theory and Statistical Mechanics. Phys. Rev. 1957, 106, 620–630. [Google Scholar] [CrossRef]
  12. Zwanzig, R. Nonequilibrium Statistical Mechanics; Oxford University Press, 2001. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated