Preprint
Article

This version is not peer-reviewed.

Gaussian Boson Sampling as a Calibrated Variance-Reduction Engine: A DSFL Explanation

Submitted:

25 October 2025

Posted:

27 October 2025

You are already at the latest version

Abstract
Gaussian Boson Sampling (GBS) generates photon-count patterns whose probabilities are governed by hafnian functionals of an encoded covariance. Recent theory shows that GBS samples can approximate Gaussian expectation problems with provable sample-complexity advantages over Monte Carlo (MC) on an open, sizable subset of instances. We give a unifying explanation via the Deterministic Statistical Feedback Law (DSFL). First, GBS acts as a calibrated, admissible map that intertwines target statistics and physical sampling. Second, tuning the mean photon number is a genuine calibration step that minimizes a DSFL residual. Third, principal-angle (Friedrichs) geometry quantifies conditioning and leakage between statistical and sampling features. Fourth, a Lyapunov envelope provides non-asymptotic error decay and design rules. From these ingredients we derive bounds that recover and sharpen known GBS scalings, predict regimes where GBS outperforms MC, and specify a practical hybrid policy that switches by degree. Experiments with simulated distributions and hardware-aligned kernels support the theory and validate the tuning and selection rules.
Keywords: 
;  ;  ;  ;  ;  ;  ;  ;  ;  ;  ;  ;  

1. Introduction

Gaussian Boson Sampling (GBS) produces photon-count patterns whose probabilities are governed by hafnian functionals of a covariance encoded in a Gaussian photonic state. Beyond its foundational role in quantum advantage, GBS has recently been positioned as a practical engine for approximating Gaussian expectation problems (GEPs), where one seeks μ = E [ f ( X ) ] for X N ( 0 , B ) and analytic or polynomial f. Wick’s theorem expresses μ as a hafnian–weighted sum over multi-indices drawn from B, aligning the target with the native GBS statistics. Building on this structure, Andersen and Shan introduced degree-resolved GBS estimators and proved open, nonempty regions in which GBS achieves ( ε , δ ) sample–complexity advantages over Monte Carlo (MC), and further showed that tuning the mean photon number to the task increases the percentage of instances with advantage [1,2].

1.1. Calibrated Residuals, Principal–Angle Conditioning, and Predict–Then–Prove Guarantees

This work advances three aspects that are often treated separately. First, the residual R K ( t ) is not merely a distance: it is a calibrated, minimal, and dynamical quantity. We prove that every admissible tuning step is L 2 –nonexpansive, so the residual cannot inflate under retuning or composition. Second, conditioning is captured quantitatively by the principal angle θ F (via its cosine γ ), and this geometric control enters directly into the variance and MSE formulae, providing sharp two–sided bounds; this bridges statistical estimation, photonic physics, and numerical analysis in a single comparison geometry. Third, the framework is predict–then–prove: before sampling, it specifies where to operate (photon–number matching), how conditioning will affect dispersion (principal angles), and what decay rate to expect (Lyapunov envelope), and then establishes these prescriptions as theorems.

1.2. Problem

Practitioners need to foresee, before allocating shots, when GBS will outperform MC and how to tune the device to realize that advantage. Concretely: given a covariance B and an integrand f, can we predict in advance whether calibrated GBS will require asymptotically fewer samples than MC to reach a prescribed ( ε , δ ) accuracy, and can we prescribe a robust one-parameter tuning rule that maximizes the gain? We seek design-time criteria (not post hoc curve fits) that tell us when to use GBS, when to defer to MC, and how to switch between them by degree.

1.3. DSFL Intuition: Forecasting When GBS Beats MC

The DSFL framework predicts, a priori, when and how GBS will outperform Monte Carlo (MC) by placing the target statistics and the sampler in a single Hilbert geometry and tracking one quantity—the residual. This yields the following design-time principles:
  • Single yardstick. Target and sampler are embedded in the same space; a single residual provides a quantitative knob to optimize and monitor.
  • No inflation under tuning. Retuning, counting, and block composition act as nonexpansive maps; the residual cannot increase under admissible operations.
  • Unique, certifiable optimum. The dominant variance/MSE factor is strictly convex in the scale and attains its unique minimum at the mean–photon–number match m t B = 2 K ; this yields a robust one-line tuning rule.
  • Predictable conditioning. Alignment between sampling and blueprint features is quantified by the principal angle θ F (cosine γ ); small γ certifies low dispersion. A computable surrogate from warm-up features allows pre-run assessment.
  • Rates, not just limits. If each update removes a fixed fraction of the residual, the residual obeys a non-asymptotic exponential envelope, providing a predicted semi-log MSE slope before data collection.
  • Regime map and hybrid policy. The geometry identifies regimes where GBS is inherently favorable (high degrees) and where MC suffices (low degrees), and prescribes a degree-wise hybrid switch using the same residual/angle diagnostics.
  • Robustness. Composition and small hardware drifts degrade angles and rates smoothly; re-matching m t B restores near-optimality.
In summary, DSFL turns GBS from a black-box sampler into a calibrated, predictable engine: it specifies where to operate, how conditioning impacts dispersion, and what decay rate to expect, enabling an informed, pre-run decision on when—and by how much—GBS will outperform MC.

1.4. Thesis (DSFL View)

We show that GBS is a calibrated variance-reduction engine inside a single Hilbert comparison geometry governed by a Deterministic Statistical Feedback Law (DSFL). This provides one residual to minimize and four provable pillars: (i) Admissibility & DPI: counting and retuning are L 2 -nonexpansive and intertwine with the estimator, so the calibrated residual cannot inflate; (ii) Calibration optimality: the unique minimizer of the leading variance/MSE factor is the photon-number match  m t B = 2 K , which implements the tuning advocated and empirically optimized in [1,2]; (iii) Angle conditioning: principal (Friedrichs) angles between the sampling and blueprint subspaces yield sharp two-sided dispersion bounds ( 1 ± cos θ F ) , explaining conditioning and leakage; (iv) Lyapunov envelope: a one-step contraction turns into a non-asymptotic exponential error decay with rate α ( t ) maximized at the photon match.

1.5. Contributions

  • A unified geometry that proves a data–processing inequality (no residual inflation) for tuning and composition, and recovers the ( ε , δ ) envelopes of [1].
  • A one-line strict-convexity proof that predicts the optimal tuning rule m t B = 2 K , matching the optimized estimators in [3].
  • Principal-angle bounds that forecast dispersion and a Lyapunov inequality that delivers exponential, non-asymptotic decay—together forming practical, pre-run design rules to decide GBS vs. MC and to set t.
  • A regime map that explains, a priori, when calibrated GBS is polynomial while MC is exponential (large degree), and a hybrid, degree-wise policy for the small-degree regime.
In short, DSFL turns “folk rules’’ into theorems: match  m t B , measure angles, and expect exponential envelopes—thereby foreseeing and tuning the regimes where GBS outperforms MC, in line with and extending the results of [1,2].

2. What GBS Does (Succinct but Precise)

GBS distribution and hafnians 

Let B R N × N be symmetric with 0 < spec ( B ) < 1 . The GBS probability of a pattern I N N is
P B ( I ) = d B I ! Haf ( B I ) 2 , d B : = j = 1 N 1 λ j 2 ,
where { λ j } are eigenvalues of B. The mean photon number is
m B = j = 1 N λ j 2 1 λ j 2 .

Gaussian expectations via Wick’s theorem 

For f ( x ) = k 0 | I | = k a I x I and X N ( 0 , B ) ,
μ = E [ f ( X ) ] = k 0 | I | = 2 k a I Haf ( B I ) .
A degree partition gives μ = k = 0 K μ k with μ k = | I | = 2 k a I Haf ( B I ) .

Two practical GBS estimators 

GBS-I (importance). For μ k ( 2 ) = | I | = 2 k a I Haf ( B I ) 2 , the estimator using weights I ! a I / d B is unbiased.
GBS-P (probability). For μ k = | I | = 2 k a I Haf ( B I ) , estimate p J = P B ( J ) by counts and average p J with weights. The bias vanishes at rate 1 / n ; the leading term is explicit.

Tuning knob: photon-number matching 

Scale B t B with t ( 0 , λ 1 1 ) and choose t so that
m t B = j = 1 N t 2 λ j 2 1 t 2 λ j 2 = 2 K ,
i.e. match the sampler’s mean photon number to the target degree K.

3. Related Work

This section situates our DSFL-based treatment of Gaussian Boson Sampling (GBS) within three intertwined literatures: (i) the foundations of GBS, including sampling distributions, computational hardness, and experimental progress; (ii) GBS as a vehicle for numerical estimation, in particular its use for Gaussian expectation problems together with formal ( ε , δ ) -sample–complexity claims and percentage-of-advantage quantification; and (iii) control-theoretic and geometric principles that justify our residual-contraction, angle-conditioned bounds, and data–processing inequalities in a single Hilbertian comparison geometry. We use these threads to explain how the “folk rules’’ of GBS operation—photon-number tuning, conditioning via feature alignment, and non–asymptotic error envelopes—can be elevated to theorems, and to place our results alongside prior theory and practice.

3.1. Foundations of GBS and Hardness

The modern formulation of GBS as a sampling task driven by hafnian-valued amplitudes is due to Hamilton et al. [4] with detailed modeling and refinements in [5]. Complexity-theoretic hardness follows the linear-optical lineage of Aaronson–Arkhipov [6] and extends to experimentally informed, high-dimensional variants [7,8]. Algorithmic primitives for hafnian evaluation and classical simulation—including supercomputer-grade implementations and the thewalrus library—are documented in [9,10]. Photonic demonstrations approaching or attaining quantum advantage are reported in [11,12]. Related Gaussian models include loop-hafnians for displaced states [13] and threshold-detection (Torontoian) statistics [14], which frame hardware-relevant variants. On the application side, GBS has been explored for vibronic spectra [15,16], molecular docking [17], dense subgraph discovery and graph kernels [18,19,20,21], random point processes [22,23], and scalable time-bin/loop architectures [24,25]. These works supply the physical and computational substrate on which our estimation-theoretic analysis is built.

3.2. GBS Estimators for Gaussian Expectations

The use of GBS samples to approximate Gaussian expectation problems (GEPs)—via Wick’s theorem [26] to express Gaussian moments as hafnian-weighted functionals—was developed in [1], which introduced two concrete estimators (GBS-I and GBS-P) and proved non-asymptotic ( ε , δ ) guarantees on an open, nonempty subset of instances. An optimized variant matching the device’s mean photon number to the target degree and quantifying the percentage of the problem space with GBS advantage appears in [2]. Our DSFL framework recovers these bounds within a single comparison geometry, explains the tuning law ( m t B = 2 K ), and adds angle-conditioned rate controls. Complementary analytic identities (e.g., the hafnian master theorem [27]) and approximation tools [28,29] delineate tractable regimes and provide context for the sampling-based estimators considered here.

3.3. Hilbertian Residual Control and Data Processing

Our analysis leverages standard Hilbert-space contractivity of orthogonal projections (counting) and parameter changes (retuning) to derive a data–processing inequality (DPI) for the DSFL residual, together with principal-angle (Friedrichs) bounds for conditioning and Lyapunov-type decay envelopes. While these tools are classical (see, e.g., two-subspace/Friedrichs-angle inequalities and discrete Grönwall arguments), their joint instantiation for GBS estimation and calibration appears to be new in this context. For Monte Carlo baselines and variance-reduction perspectives that we compare against, see [30,31,32,33].
Gaussian Boson Sampling (GBS) was introduced as a photonic sampling model whose output probabilities are governed by hafnians of submatrices associated with a Gaussian covariance [4,5]. In analogy with (standard) Boson Sampling [6], a growing body of work establishes classical intractability (under plausible conjectures) not only for exact sampling but also for sampling from suitable approximations to the ideal distribution [7,8]. Algorithmic progress on hafnian evaluation—both general-purpose [9] and software implementations (e.g. The Walrus) [10]—has enabled faithful classical simulations at nontrivial scales, supporting experimental validation [11,12] and methodological development.

3.4. Applications of GBS

Proposed applications of GBS span vibronic spectroscopy [15,16], molecular docking [17], dense subgraph discovery [18], perfect matchings [19], graph similarity and kernels [20,21], point processes [22], and photonic simulations of molecular dynamics [23]. These lines have been complemented by platform-level studies of time-bin/loop architectures that aim at scalable implementations [24,25].

3.5. GBS Estimators for Gaussian Expectations

The specific connection between GBS and Gaussian expectation problems (GEPs) is made precise by Wick’s theorem [26], which expresses Gaussian moments as hafnians. Building on this structure, Andersen and Shan (i) introduced degree-resolved GBS estimators (GBS-P and GBS-I) for GEPs and proved open, nonempty regions of exponential advantage over plain Monte Carlo (MC), measured in guaranteed ( ε , δ ) sample sizes [1], and (ii) optimized the mean photon number to the task and quantified the fraction of the problem space displaying advantage, with large observed percentages and sizeable efficiency ratios [2]. Related algorithmic components include hafnian bounds (via Stirling-type inequalities) and polylogarithmic envelopes that deliver computable, nonasymptotic sample-complexity expressions.

3.6. Variance-Reduction and MC Baselines

From a classical perspective, GBS-I can be seen as an importance sampler whose proposal is the physical GBS law; GBS-P corresponds to a (biased-but-consistent) probability estimator with explicit leading error terms. These sit alongside standard MC variance-reduction techniques (importance, stratification, antithetic sampling, quasi-MC), whose strengths and limitations in high dimensions are well documented [30,31,32,33]. The present analysis adopts plain MC as the baseline, following [1,2].

3.7. Hafnian Identities and Tractable Limits

Structural identities such as the hafnian master theorem [27] provide tractable regimes for particular coefficient ensembles (e.g. z I / I ! weights), while general approximation results for hafnians and permanents (e.g. Barvinok-type bounds) furnish algorithmic surrogates outside the photonic setting [28,29]. These are complementary to the sampling-based estimators considered here.

3.8. Angles, Data Processing, and Residual Control

The present DSFL view organizes the GBS estimation task in one Hilbert geometry and leverages two ingredients that are ubiquitous in information processing and operator theory: (i) data–processing inequalities (DPI) for nonexpansive, intertwining updates (counting maps, retuning), ensuring no residual inflation under admissible operations; and (ii) principal-angle (Friedrichs) bounds that control conditioning of two intertwined subspaces (blueprint vs sampling features) and yield two-sided variance inequalities with factors ( 1 ± cos θ F ) . The Lyapunov envelope for the residual—a discrete coercivity statement converting one-step contraction into an exponential decay rate—is classical in stability theory and is instantiated here by the t-matched (photon-number–matched) tuning law. While the DSFL terminology is specific to this work, the underlying tools are aligned with standard Hilbertian projection theory and CS-decompositions.

3.9. Positioning

Relative to [3,3], the present manuscript provides a unifying explanation of when/why/how GBS outperforms MC by: (a) proving that mean–photon–number matching is the unique minimizer of the leading residual/variance factor; (b) quantifying conditioning via principal angles with explicit two-sided bounds; and (c) deriving nonasymptotic Lyapunov envelopes for error decay that recover the ( ε , δ ) sample-size laws. This framework is compatible with current hardware models [34,35,36] and classical tooling for hafnians [9,10], and it complements domain applications [15,17,18,19,20,22,23].

4. Limitations and Scope

The guarantees developed here are tailored to Gaussian expectation problems in which Wick reduction expresses the target as hafnian–weighted functionals of a real, symmetric covariance with spectrum strictly inside  ( 1 , 1 ) . This setting aligns the photonic sampler’s distribution with the blueprint geometry in a way that makes DSFL calibration, principal-angle conditioning, and Lyapunov envelopes directly applicable. Extending the framework beyond this regime entails additional structure. For non-Gaussian observables or general boson sampling models (e.g., loop-hafnians with coherent displacements, threshold detection/Torontoian statistics, or complex-valued amplitudes with sign patterns), one must re-specify the feature map, the admissible updates, and the calibration functional so that (i) an appropriate inner-product geometry captures the relevant moments, (ii) the data–processing inequality remains valid under the physical update semigroup, and (iii) bias terms induced by nonlinearity or signs are explicitly controlled. A second limitation concerns conditioning proxies: our use of principal-angle surrogates (e.g., cross-Gram norms between sampling and blueprint features) is computationally light but approximate; computing exact projectors or Friedrichs angles may be prohibitive for very large  ( N , K ) . Finally, while our analysis covers noiseless sampling and block composition via nonexpansive maps, hardware noise and calibration drift require model-specific contractions; the DSFL conclusions persist under such perturbations only insofar as a positive coercivity margin can be certified after re-matching the mean photon number.

4.1. DSFL Foresight for GBS: Verbal Logic → Mathematical Predictors

We give a from-first-principles account of how the Deterministic Statistical Feedback Law (DSFL) predicts (before any run) the behavior of Gaussian Boson Sampling (GBS): how tuning affects error, when/why it beats Monte Carlo (MC), how variance/bias scales, why mean-photon-number matching is optimal, and how composition/noise impact performance. Each verbal claim is paired with an explicit, checkable mathematical statement.

5. Stage 1: Put GBS in One Hilbert Geometry (the DSFL Stage)

DSFL places the statistical blueprint (the quantity to estimate) and the physical sampler (the device) in the same comparison geometry and measures a single misfit (residual). In our setting, GBS turns a scale parameter t (the photon-number knob) into a sample law P t B . The estimator (GBS-I or GBS-P restricted to degree K) is a calibration map  I t that reads off μ K from P t B . Alignment ⇒ small residual; misalignment ⇒ large residual.

Mathematics:

Fix a degree K. Let the blueprint vector be
s K ( a I ) | I | = 2 K R D K , D K = # { I : | I | = 2 K } .
Let the physical law at scale t be the degree- 2 K marginal
p ( t ) = p I ( t ) | I | = 2 K , p I ( t ) = d t B I ! Haf ( t B ) I 2 , d t B = j = 1 N 1 t 2 λ j 2 .
Work in H = R D K with the standard inner product. Define the readout (calibration) I t : H H by the chosen estimator:
GBS-P ( probability ) : ( I t p ) K | I | = 2 K a I p I ( t ) d t B , GBS-I ( importance on Haf 2 ) : ( I t p ) K | I | = 2 K a I I ! d t B p I ( t ) .
Define the Residual of Sameness
R K ( t ) p ( t ) I t s K H 2 .
This single scalar is the DSFL misfit we will minimize, bound, and track.

Stage 2: Admissibility ⇒ No-Inflation (DPI)

Changing t and counting more samples is nonexpansive in the comparison geometry; residuals cannot inflate just by tuning/concatenation. Any sensible schedule is safe: at worst neutral, typically improving.

Mathematics (DPI)

Let ( Φ ˜ , Φ t ) be the blueprint/physical update. Here Φ ˜ = Id while Φ t transports p p ( t ) and applies the counting map. The intertwining Φ t I t = I t Φ ˜ holds by construction. If Φ t H H 1 (counting is L 2 -nonexpansive), then
R K Φ ˜ s K , Φ t p R K ( s K , p ) .
Prediction #1 (robustness). Any well-behaved change of t, any device concatenation (gluing), and any extra sampling cannot increase the DSFL residual.

Stage 3: Optimal Tuning = Mean-Photon Matching m t B = 2 K

Both the GBS distribution and the estimator operate on multi-indices with | I | = 2 K . If the device emits photon numbers far from 2 K , mass is spent in the wrong degree, creating misalignment/variance. Matching the mean photon number to 2 K centers mass where the estimator reads, minimizing misfit and MSE.

Mathematics (Calibration Optimality)

Let
m t B = j = 1 N t 2 λ j 2 1 t 2 λ j 2 .
For GBS-P the leading MSE is
MSE K GBS P ( t ) 1 4 n t 2 K d t B 1 | I | = 2 K a I 2 I ! μ K 2 ,
and for GBS-I the variance is
Var μ ^ K GBS I ( t ) = 1 n t 4 K d t B 1 | I | = 2 K a I 2 I ! Haf ( B I ) 2 μ K ( 2 ) 2 .
Both have a factor C ( t ) { t 2 K d t B 1 , t 4 K d t B 1 } . Then
d d t log C ( t ) = 2 K t + m t B t = 0 m t B = 2 K ,
and
d 2 d t 2 log C ( t ) = 2 K t 2 + j = 1 N λ j 2 1 t 2 λ j 2 + j = 1 N 2 t 2 λ j 4 ( 1 t 2 λ j 2 ) 2 > 0 .
Therefore t with m t B = 2 K is the unique strict minimizer.
arg min t MSE / Var = { t : m t B = 2 K } .
Prediction #2 (tuning law). Set t so that m t B = 2 K ; this minimizes estimator MSE/variance and the residual R K ( t ) .

Stage 4: Conditioning and Leakage via Principal Angles

If the sampling and blueprint subspaces are aligned, readout is clean; if skew, mass leaks. DSFL quantifies this by the Friedrichs angle θ F ; γ = cos θ F acts like a conditioning factor: γ (worse), γ (better).

Mathematics (Two-Subspace Inequality)

Let P GBS ( t ) project the sampling features at scale t and P deg K the degree-K blueprint. With γ ( t ) = P GBS ( t ) P deg K = cos θ F ( t ) ,
( 1 γ ) { in , out } e 2 R K ( t ) ( 1 + γ ) e 2 .
Since empirical MSE R K (up to constants),
( 1 γ ) · ( split energy ) MSE / Var ( 1 + γ ) · ( split energy ) .
Prediction #3 (angle effect). As t t with m t B = 2 K , θ F ( t ) and bounds tighten; mis-tuning increases γ and hurts conditioning.

Stage 5: Rates via a Lyapunov Envelope

Near the matched regime, the error behaves like a dissipative system: each sample reduces a fixed fraction of residual. DSFL encodes this with a Lyapunov inequality, giving a non-asymptotic exponential envelope.

Mathematics:

Let R n = E R K ( n ) . Under admissibility (DPI) and a coercivity margin α ( t ) > 0 (monotone in alignment/angle),
R n + 1 R n α ( t ) R n R n e α ( t ) n R 0 .
Thus MSE n C e α ( t ) n .
MSE n C e α ( t ) n , α ( t ) maximal at m t B = 2 K .
Prediction #4 (rate). The decay slope α ( t ) is largest at the match; mis-tuning reduces α .

Stage 6: When GBS Beats MC (Regime Maps)

MC at degree K inherits all cross-moment mixings, exploding with effective dimension; calibrated GBS concentrates on the right degree. DSFL converts this into sample-complexity comparisons.

Mathematics (Complexity Ratio)

Chebyshev/Markov give guaranteed sizes
n GBS P ( t ) 1 ε 2 δ · t 2 K d t B 1 a I 2 / I ! μ K 2 , n GBS I ( t ) 1 ε 2 δ · t 4 K d t B 1 a I 2 Haf ( B I ) 2 / I ! ( μ K ( 2 ) ) 2 ,
versus
n MC 1 ε 2 δ · I , I a I a I Haf ( B I + I ) ( or Haf 2 ) μ 2 .
For large K or correlated growth K ζ N 2 , the MC numerator inflates exponentially (many cross indices), while GBS stays calibrated:
Large K ( or K ζ N 2 ) : n GBS = poly ( N ) , n MC = exp ( Θ ( N ) ) .
Prediction #5 (regime map). Large K⇒ GBS wins; small K⇒ use DSFL angle proxy to switch (hybrid GBS/MC).

Stage 7: What Happens When You Sweep t 

MSE versus t is U-shaped with the minimum exactly at m t B = 2 K ; sharp spectra create narrower minima (more sensitive tuning).

Mathematics:

Let g ( t ) = log t 2 K d t B 1 . Then
g ( t ) = 2 K t + m t B t , g ( t ) = 0 m t B = 2 K , g ( t ) = 2 K t 2 + j λ j 2 1 t 2 λ j 2 + j 2 t 2 λ j 4 ( 1 t 2 λ j 2 ) 2 > 0 .
Prediction #6 (t-sweep). A unique convex minimum at m t B = 2 K ; sharper eigenvalue spectra yield narrower minima.

Stage 8: Composition and Noise

Composition: concatenating blocks is admissible; residuals cannot inflate. Noise: small Gaussian perturbations contract in H ; same predictions hold with slightly reduced rate/angle; re-matching m t B restores optimality. [37,38]

Mathematics:

If each block satisfies Φ 1 and Φ I = I Φ ˜ , then for Φ = Φ ,
R ( Φ ˜ s , Φ p ) R ( s , p ) ( iterated DPI ) .
For B B + Δ with Δ small, Davis–Kahan yields
sin θ F ( B + Δ ) Δ gap , d t ( B + Δ ) = d t B · ( 1 + O ( Δ ) ) .
Hence α ( t ) decreases by O ( Δ ) ; m t ( B + Δ ) = 2 K still centers the optimum.
Small noise graceful degradation ; re - match m t B to recover near - optimality .

Stage 9: Actionable Tests

The DSFL analysis yields a concrete testing protocol that can be implemented prior to data collection and updated on-line with negligible overhead. First, the sampler must be calibrated by matching its mean photon number to the target degree. Concretely, given a covariance B with spectrum { λ j } j = 1 N ( 0 , 1 ) and a degree index K N , the scale t ( 0 , λ 1 1 ) is defined as the unique solution of the monotone equation
m t B : = j = 1 N t 2 λ j 2 1 t 2 λ j 2 = 2 K .
Setting t = t centers the sampling law on multi-indices of weight  2 K and, by the variance formulas derived above, minimizes the mean–square error for both GBS-I and GBS-P.
Second, conditioning is certified by a computable principal-angle proxy. Let X t R M × D K denote a (finite) feature matrix obtained from M synthetic or warm-up GBS feature evaluations at scale t, and let Y K R M × D K encode a basis for the degree-K blueprint subspace (e.g. normalized monomials indexed by { | I | = 2 K } ). The quantity
γ ^ ( t ) : = X t Y K 2
upper bounds the Friedrichs cosine γ = P GBS ( t ) P deg K up to discretization error. Small values of γ ^ ( t ) certify good alignment (large principal angle  θ F ) and hence favorable two-sided variance bounds, whereas large values indicate ill-conditioning and trigger either a re-tuning of t toward (26) or a temporary deferral to the Monte Carlo channel for the corresponding K.
Third, channel selection can be automated by an empirical residual criterion. Denote by R ^ K the empirical DSFL residual estimated from a fixed budget of warm-up samples (identical across candidates). Choosing the channel with smaller  R ^ K minimizes the predicted MSE under the nonexpansive dynamics that define DSFL and is therefore statistically safe. Since both GBS and MC updates are L 2 –nonexpansive in the comparison geometry, this selection step does not increase the residual in expectation.
Finally, under architectural changes (e.g. insertion of an additional optical block or concatenation in a time-bin loop), admissibility guarantees a data–processing inequality for the composite map, hence no inflation of the residual; only the calibration Equation (26) must be re-checked for the updated effective covariance and scale. In particular, the composition of nonexpansive, intertwining maps remains nonexpansive and intertwining, so the DSFL guarantees persist verbatim after gluing.

Executive Summary of Foresight

Within a single Hilbert geometry, DSFL certifies that tuning and concatenation steps are nonexpansive; hence, the residual cannot deteriorate under admissible manipulation of the sampler or of the pipeline. The unique operating point that minimizes the mean–square error and the estimator variance is characterized by the mean-photon-number matching condition m t B = 2 K , which is strictly convex in t and therefore algorithmically well-posed. Conditioning is governed by the principal angle θ F between the sampling subspace at scale t and the degree-K blueprint; the associated cosine γ = cos θ F controls sharp two-sided variance bounds and decreases as the calibration improves. In the matched regime, the expected residual satisfies a discrete Lyapunov inequality and thus admits a non-asymptotic exponential envelope MSE n C e α ( t ) n with decay rate α ( t ) maximized at the calibration point. Consequently, for high degrees (in particular when K grows as ζ N 2 ), the calibrated GBS estimators exhibit polynomial sample complexity in N, while the plain Monte Carlo baseline inherits an exponential dependence through cross-moment couplings; for small K, the angle proxy (27) recommends a hybrid policy that defers to MC until alignment becomes favorable. The dependence of the MSE on the scale t is convex with a unique minimum at m t B = 2 K ; perturbations of B (hardware noise) or blockwise composition merely reduce the alignment and rate by controllable amounts, and re-matching via (26) restores near-optimal performance. Throughout, no inflation occurs under gluing by the data–processing inequality.

Bottom Line

The preceding statements are a priori consequences of the DSFL axioms (intertwining and L 2 –nonexpansive updates) in conjunction with the Wick/GBS representation of Gaussian expectations. They condense into operational equations
m t B = 2 K , γ = cos θ F , α ( t ) increasing in alignment ,
which together predict the behavior of GBS before sampling and specify the tuning required to achieve optimal performance in practice.

6. DSFL Primer (Minimal Machinery, with Proofs)

6.1. Comparison Geometry

Let ( H , · , · ) be a (real or complex) Hilbert space equipped with the norm x : = x , x . We fix two closed subspaces S , P H (“statistical” and “physical” channels), and denote by P S , P P the corresponding orthogonal projectors.

6.2. Calibration (Interchangeability)

Definition 6.1 
(Calibration pair). A calibration is a pair of bounded operators
I : S P , J : P S ,
satisfying the interchangeability identities
I J = id P , J I = P S .
Remark 6.1. 
The relations (30) imply that I is a left inverse of J on P , and J is a right inverse of I on S . In particular, I is an isomorphism of S / ker I onto P , and J restricts to the identity on I ( S ) .

6.3. Residual of Sameness

Definition 6.2 
(Residual). Given s S (blueprint) and p P (response), the Residual of Sameness is the squared H –distance
R ( s , p ) : = p I s H 2 .
Remark 6.2. 
By construction, R ( s , p ) = 0 iff p = I s . Moreover, R is invariant under any isometry U on H that preserves ran ( I ) : if U ( ran I ) ran I , then U p U I s = U ( p I s ) = p I s .

6.4. Admissibility and the Data–Processing Inequality

Definition 6.3 
(Admissible update). A pair of bounded operators ( Φ ˜ , Φ ) with
Φ ˜ : S S , Φ : P P ,
is admissible if:
1.
Intertwining (calibration consistency):
Φ I = I Φ ˜ ;
2.
Nonexpansiveness: Φ H H 1 .
Theorem 6.4 
(Data–processing inequality (DPI)). If ( Φ ˜ , Φ ) is admissible, then for all s S and p P ,
R ( Φ ˜ s , Φ p ) R ( s , p ) .
Proof. 
Using (33),
R ( Φ ˜ s , Φ p ) = Φ p I Φ ˜ s 2 = Φ p Φ I s 2 = Φ ( p I s ) 2 Φ 2 p I s 2 p I s 2 ,
which is (34).    □
Corollary 6.5 
(Iterated DPI). If { ( Φ ˜ k , Φ k ) } k = 1 m are admissible, then so is ( Φ ˜ , Φ ) = ( k = 1 m Φ ˜ k , k = 1 m Φ k ) , and
R Φ ˜ s , Φ p R ( s , p ) .
Proof. 
Intertwining and · 1 are preserved under composition. Apply Theorem 6.4.    □

6.5. Angles and Conditioning

Let M , N H be closed subspaces with orthogonal projectors P M , P N .
Definition 6.6 
(Friedrichs angle). The Friedrichs cosine is γ ( M , N ) : = P M P N , and the Friedrichs angle  θ F ( M , N ) [ 0 , π 2 ] is defined by
cos θ F ( M , N ) : = γ ( M , N ) .
We recall a standard two–subspace inequality; we include a proof for completeness.
Proposition 6.7 
(Two–subspace bounds). Let M , N H be closed subspaces and set γ = γ ( M , N ) . For any a M , b N ,
( 1 γ ) a 2 + b 2 a + b 2 ( 1 + γ ) a 2 + b 2 .
Proof. 
By polarization,
a + b 2 = a 2 + b 2 + 2 Re a , b .
Since a M , b N , we have | a , b | = | a , P M b | a P M b γ a b . Thus
2 γ a b 2 Re a , b 2 γ a b .
Using 2 a b a 2 + b 2 and 2 a b ( a 2 + b 2 ) yields (38).    □
Proposition 6.8 
(Through–subspace contraction). Let M , N H be closed subspaces with γ = P M P N . Then for any bounded T : H H ,
P M T P N T γ .
In particular, if T 1 , then P M T P N γ = cos θ F ( M , N ) .
Proof. 
By the dual characterization of the operator norm,
P M T P N = sup x = y = 1 | P M T P N x , y | = sup x = y = 1 | T P N x , P M y | T sup x = y = 1 | P N x , P M y | .
But sup x = y = 1 | P N x , P M y | = P M P N (take the singular value decomposition of P M P N ). Hence (41).    □
Remark 6.3 
(Interpretation). In applications, M and N are the “pointer” subspaces representing, respectively, the sampling polarization and the blueprint (degree–K) polarization. The factor γ = cos θ F controls leakage and conditioning across the seam: estimates that pass through an L 2 –contractive kernel are suppressed by at least cos θ F .

6.6. Lyapunov Envelopes (Continuous and Discrete)

We formalize a standard energy–dissipation mechanism that will underlie non–asymptotic error envelopes.
Lemma 6.9 
(Differential energy inequality). Let e ( t ) H be (weakly) differentiable and set R ( t ) : = e ( t ) 2 . Assume there exists a bounded self–adjoint K = K * 0 and a remainder g ( t ) H such that
d d t R ( t ) = 2 K e ( t ) , e ( t ) + 2 e ( t ) , g ( t ) .
If, for some κ > 0 and ε 0 ,
K e , e κ e 2 , | e , g | ε e 2 for all e H ,
then
d d t R ( t ) 2 ( κ ε ) R ( t ) .
Proof. 
Combine (43) and (44):
d d t R ( t ) 2 κ e ( t ) 2 + 2 ε e ( t ) 2 = 2 ( κ ε ) R ( t ) .
   □
Theorem 6.10 
(Continuous Lyapunov envelope). Under the hypotheses of Lemma 6.9 with α : = 2 ( κ ε ) > 0 ,
R ( t ) e α ( t t 0 ) R ( t 0 ) for all t t 0 .
Proof. 
Apply Grönwall’s inequality to (45).    □
A discrete analogue covers sampling iterations.
Proposition 6.11 
(Discrete Lyapunov envelope). Let ( R n ) n N be nonnegative numbers such that for some α ( 0 , 1 ] ,
R n + 1 ( 1 α ) R n for all n N .
Then
R n ( 1 α ) n R 0 e α n R 0 .
Proof. 
Iterate (48) and use ( 1 α ) n e α n .    □
Remark 6.4 
(Connection to admissibility). In many applications (e.g. residuals computed after one sampling/update step), admissibility of the update and a sectorial coercivity provide (48) in expectation, yielding an expected exponential envelope of the form E ( R n ) e α n E ( R 0 ) .

7. Mapping GBS ⟺ DSFL

Here we place the GBS estimation problem and the DSFL control law into a single, explicit Hilbert–space model and show that the main “folk rules” for operating a Gaussian Boson Sampler—how to tune the mean photon number, when advantage over Monte Carlo (MC) appears, how conditioning depends on feature alignment, and why errors decay at a fixed rate—follow as theorems from one comparison geometry. The construction has three ingredients. First, for a fixed target degree K, we embed both the statistical blueprint (the degree–K coefficient vector of the integrand) and the physical sampling law induced by the device at scale t into the same finite–dimensional Hilbert space H K = R D K indexed by multi–indices I K = { I : | I | = 2 K } . This makes the GBS law p ( t ) and the blueprint s K comparable as vectors and allows us to measure a single misfit—the DSFL residual
R res ( K ) ( t ) = p ( t ) Π s K H K 2 .
Second, we model the estimators (GBS–P and GBS–I restricted to degree K) as bounded linear functionals I : H K R , so that “reading off” the degree–K contribution to the Gaussian expectation is nothing but a linear evaluation in H K . In this form, the basic DSFL admissibility hypotheses hold for the two primitive operations available on hardware or in simulation: (a) changing the scale t (retuning the source) and (b) replacing the law by its empirical counterpart (counting). Both operations act by contractions on H K and intertwine with the calibration I . Consequently, the DSFL data–processing inequality applies directly: no admissible update can inflate the degree–K residual.
Third, we extract quantitative laws from the geometry. A short calculus identity shows that the leading t–dependence of the residual and of the estimator variance is governed by the convex factor
C 2 K ( t ) = t 2 K d t B , with d d t log C 2 K ( t ) = m t B 2 K t ,
so the unique minimizer occurs at the photon–number match m t B = 2 K . This is the precise DSFL formulation of “mean–photon–number matching is optimal.” Conditioning and leakage across degrees are controlled by the Friedrichs principal angle θ F ( t ) between the blueprint subspace and the sampling feature subspace in H K ; standard two–subspace estimates yield two–sided variance (or squared–bias) bounds with multiplicative factors ( 1 ± cos θ F ( t ) ) . Finally, a one–step conditional contraction of the residual—guaranteed by admissibility plus a positive angle margin—produces a Lyapunov inequality and hence a non–asymptotic exponential envelope
E R res ( K ) p ( n ) ( t ) e α ( t ) n R res ( K ) p ( 0 ) ( t ) ,
with α ( t ) maximal at the photon–match.
The subsections below make this scheme explicit. We first specify the blueprint s K , the physical channel p ( t ) , and the linear estimator I inside H K . We then verify admissibility (intertwining and nonexpansiveness) for counting and retuning, yielding a degree–K data–processing inequality. Next, we prove calibration optimality—photon–number matching minimizes the residual—and derive principal–angle bounds for the estimator’s dispersion. We conclude with a Lyapunov–type result that converts the one–step contraction into an exponential decay rate, thereby recovering and clarifying the ( ε , δ ) sample–complexity laws for GBS and the observed advantage regimes over MC.

7.1. Blueprint, Physical Channel, and Calibration in One Geometry

Fix an integer K 0 and a symmetric covariance B R N × N with spectrum 0 < λ N λ 1 < 1 . Let
I K : = { I = ( i 1 , , i N ) N N : | I | : = i 1 + + i N = 2 K } , D K : = # I K .

7.1.1. Hilbert Model

We work on the comparison Hilbert space
H K : = R D K , x , y : = I I K x I y I , x 2 : = I x I 2 .

7.1.2. Blueprint

Let f ( x ) = k 0 | I | = k a I x I be the target integrand. The degree-K blueprint is the coefficient vector
s K : = ( a I ) | I | = 2 K H K .
Wick’s theorem yields the degree-K contribution to the Gaussian expectation
μ K = | I | = 2 K a I Haf ( B I ) , μ K ( 2 ) = | I | = 2 K a I Haf ( B I ) 2 ,
where B I denotes the (multi-)submatrix of B indexed by I.

7.1.3. Physical Channel

For any scale t ( 0 , λ 1 1 ) , the GBS distribution on I K is
p I ( t ) = d t B I ! Haf ( t B ) I 2 , d t B = j = 1 N 1 t 2 λ j 2 .
We regard p ( t ) : = ( p I ( t ) ) | I | = 2 K H K as the “physical state” at scale t. The (global) mean photon number of t B is
m t B : = j = 1 N t 2 λ j 2 1 t 2 λ j 2 .

7.1.4. Calibration (Estimator Readout)

We model the degree-K estimators as linear functionals on H K :
GBS-P ( probability ) : I P ( x ) : = | I | = 2 K a I x I d t B = : w P , x , w I P = a I d t B , GBS-I ( importance ) : I I ( x ) : = | I | = 2 K a I I ! d t B x I = : w I , x , w I I = I ! a I d t B .
Thus I : H K R is bounded (by Cauchy–Schwarz). For empirical laws p ^ ( t ) obtained by counting, we use the same I ; admissibility will follow from linearity and nonexpansiveness of orthogonal projections.

7.1.5. Residual

With this embedding, the DSFL residual for the degree-K task at scale t is
R res ( K ) ( t ) : = p ( t ) Π s K H K 2 .
where Π : S H K is the canonical identification Π s K = s K (we suppress the statistics subspace and view s K as an element of H K ).

7.2. Admissibility of GBS Tuning and Counting

Lemma 7.1 
(Intertwining). Let Φ ˜ = id H K and Φ be either: (i) the identity (fixed t) followed by the orthogonal projection Proj onto the simplex of probability vectors (counting), or (ii) the change of scale t t followed by Proj . Then for I { I P , I I } one has
I Φ ˜ = I , Φ Π = Π , hence Φ I = I Φ ˜ .
Proof. 
Φ ˜ = id on the blueprint space; Φ acts on physical vectors only and preserves the ambient space H K (projection and change of parameters do not move s K ). Therefore Φ Π = Π and I Φ ˜ = I , yielding the intertwining Φ I = I Φ ˜ .    □
Lemma 7.2 
(Nonexpansiveness). The operators Φ of the previous lemma satisfy Φ H K H K 1 .
Proof. 
Orthogonal projection onto a closed convex set is a contraction in H K ; parameter change t t is the identity on the ambient linear space (we reparametrize the same coordinates). Hence the composition has operator norm 1 .    □
Proposition 7.3 
(DPI for GBS tuning and counting). Let ( Φ ˜ , Φ ) be as above. Then the DPI holds for the degree-K residual:
R res ( K ) Φ ˜ s K , Φ p R res ( K ) ( s K , p ) for all s K , p H K .
Proof. 
Apply the abstract DPI to ( Φ ˜ , Φ ) , with I as specified and H = H K .    □

7.3. Calibration Optimality: Photon-Number Matching

Lemma 7.4 
(Scale factor calculus). Define
C 2 K ( t ) : = t 2 K d t B = exp 2 K log t 1 2 j = 1 N log ( 1 t 2 λ j 2 ) .
Then
d d t log C 2 K ( t ) = 1 t m t B 2 K ,
where m t B is the mean photon number. In particular, d d t log C 2 K ( t ) = 0 iff m t B = 2 K , and d 2 d t 2 log C 2 K ( t ) > 0 , so t C 2 K ( t ) is strictly convex on ( 0 , λ 1 1 ) .
Proof. 
Differentiate:
d d t log C 2 K ( t ) = 2 K t 1 2 j = 1 N 2 t λ j 2 1 t 2 λ j 2 = 1 t j = 1 N t 2 λ j 2 1 t 2 λ j 2 2 K = m t B 2 K t .
The second derivative is the sum of a positive term 2 K t 2 and strictly positive terms λ j 2 ( 1 t 2 λ j 2 ) 2 > 0 , proving strict convexity.    □
Theorem 7.5 
(Calibration Optimality). Assume R res ( K ) ( t ) admits a representation
R res ( K ) ( t ) = A C 2 K ( t ) B ,
with constants A > 0 , B R independent of t. Then the unique minimizer of t R res ( K ) ( t ) on ( 0 , λ 1 1 ) is characterized by the photon-number matching law
m t B = 2 K .
Proof. 
By the assumed form, d d t R res ( K ) ( t ) = A d d t C 2 K ( t ) ; hence critical points of R res ( K ) are those of C 2 K . Combine with the previous lemma and strict convexity to obtain uniqueness and m t B = 2 K .    □
Remark 7.1. 
For both GBS-P and GBS-I, the leading t–dependence of the MSE/variance and thus of the residual is multiplicative by C 2 K ( t ) , while the coefficient depends on ( a I , B ) but not on t. This yields the displayed form up to a harmless additive constant B (see the variance formulæ below).

7.4. Angles Quantify Conditioning: Two-Sided MSE Bounds

Let P deg K denote the orthogonal projector onto the “blueprint” subspace spanned by the canonical basis { e I } | I | = 2 K weighted by the coefficients a I (precisely, the range of the linear functional I ), and P GBS ( t ) the projector onto the sampling feature subspace at scale t (the range of I Haf ( ( t B ) I ) within H K ). Set
γ ( t ) : = P GBS ( t ) P deg K = cos θ F ( t ) .
Proposition 7.6 
(Angle–variance link). Let μ ^ K be either the GBS-P or GBS-I degree-K estimator from n samples at fixed t. Then there exist t–independent constants c ( a , B ) , c + ( a , B ) > 0 such that
( 1 γ ( t ) ) c ( a , B ) n Var ( μ ^ K ) ( 1 + γ ( t ) ) c + ( a , B ) .
Equivalently, the MSE (for GBS-P, replacing Var by the leading squared bias) is sandwiched between two multiples of the split energy, with constants modulated by ( 1 ± cos θ F ( t ) ) .
Proof sketch with precise ingredients. 
Both estimators are linear statistics of the empirical law p ^ ( t ) H K :
μ ^ K = w , p ^ ( t ) , w = w P or w I .
Then n Var ( μ ^ K ) = Σ ( t ) w , w where Σ ( t ) is the covariance operator of one draw,
Σ ( t ) = E ( ξ E ξ ) ( ξ E ξ ) , ξ H K , ( ξ = e I ) = p I ( t ) .
Since ran ( Σ ( t ) ) span { e I : | I | = 2 K } and the features live in P GBS ( t ) H K , one has (up to a bounded t–independent similarity determined by a I and B)
Σ ( t ) P GBS ( t ) P deg K D P deg K P GBS ( t ) ,
for a positive definite diagonal D encoding the t–independent weights I ! and a I . Two applications of standard two-subspace estimates (Friedrichs/CS-decomposition) yield
λ min ( D ) P GBS ( t ) P deg K 2 P deg K w 2 Σ ( t ) w , w λ max ( D ) P GBS ( t ) P deg K 2 P deg K w 2 .
Decompose w = w + w relative to P deg K ; invoke the sharp two-subspace inequalities to bound P deg K w 2 above/below by ( 1 ± γ ) w 2 and absorb t–independent factors into c ± ( a , B ) to conclude.    □

7.5. Lyapunov Envelope and Non-Asymptotic Rates

Theorem 7.7 
(DSFL Lyapunov for GBS). Let ( U , Φ ) denote one update step (blueprint-side map U and physical-side map Φ; counting and/or retuning). Fix t ( 0 , λ 1 1 ) and degree K N . Assume there exists α ( t ) ( 0 , 1 ] such that, conditionally on the past,
E R res ( K ) U s ( K ) , Φ p ( n ) ( t ) | p ( n ) ( t ) ( 1 α ( t ) ) R res ( K ) s ( K ) , p ( n ) ( t ) .
Then the mean residual decays exponentially:
E R res ( K ) p ( n ) ( t ) e α ( t ) n R res ( K ) p ( 0 ) ( t ) .
Moreover, α ( t ) is monotone improving as the principal-angle factor γ ( t ) decreases, and it is maximized at the photon-number match m t B = 2 K .
Proof. 
Taking total expectation in the conditional contraction yields
E R res ( K ) p ( n + 1 ) ( t ) ( 1 α ( t ) ) E R res ( K ) p ( n ) ( t ) .
Iterating and using ( 1 α ) n e α n gives
E R res ( K ) p ( n ) ( t ) ( 1 α ( t ) ) n R res ( K ) p ( 0 ) ( t ) e α ( t ) n R res ( K ) p ( 0 ) ( t ) .
The dependence of α ( t ) on γ ( t ) follows from the angle–variance control; its maximizer coincides with the minimizer of the leading residual factor, attained at m t B = 2 K by calibration optimality.    □
Corollary 7.8. 
If n α ( t ) 1 log R res ( K ) ( p ( 0 ) ( t ) ) / ε 2 , then E R res ( K ) p ( n ) ( t ) ε 2 . In particular, the right-hand side is minimized in t at m t B = 2 K .
Remark 7.2. 
The hypothesis is a sectorial coercivity statement: in expectation, one update removes a fixed fraction of the degree-K residual. Admissibility (data–processing) and a positive principal-angle margin guarantee such a contraction. Empirically this appears as a straight line of slope α ( t ) in semi-log MSE plots.

8. Sample-Complexity Consequences

In this section we derive, from first principles in the comparison geometry, (i) exact variance formulae for the unbiased importance estimator (GBS-I), (ii) an asymptotic mean–square error (MSE) for the probability/square–root estimator (GBS-P) under the usual sign convention, and (iii) sample–size envelopes in the ( ε , δ ) sense via Chebyshev/Markov bounds. Throughout we fix a degree K N , a covariance B with spectrum 0 < λ N λ 1 < 1 , and a scale t ( 0 , λ 1 1 ) . We also use the hafnian homogeneity
Haf ( t B ) I = t K Haf ( B I ) , ( | I | = 2 K ) ,
and the GBS law on I K : = { I : | I | = 2 K }
p I ( t ) = d t B I ! Haf ( t B ) I 2 = d t B t 2 K I ! Haf ( B I ) 2 , d t B = j = 1 N 1 t 2 λ j 2 .

8.1. GBS-I (Unbiased): Exact Variance and ( ε , δ ) Guarantee

Define the unbiased importance estimator for μ K ( 2 ) = | I | = 2 K a I Haf ( B I ) 2 by
μ ^ K GBS - I = 1 n r = 1 n Y r , Y ( I ) : = I ! a I d t B t 2 K , I p ( t ) .
Then E [ Y ] = μ K ( 2 ) by direct substitution. A single-line computation using p I ( t ) and the homogeneity gives
E Y 2 = I I K p I ( t ) Y ( I ) 2 = 1 d t B t 2 K | I | = 2 K a I 2 I ! Haf ( B I ) 2 .
Proposition 8.1 
(Variance of GBS-I). For the unbiased estimator above,
Var μ ^ K GBS - I = 1 n 1 d t B t 2 K | I | = 2 K a I 2 I ! Haf ( B I ) 2 μ K ( 2 ) 2 .
Proof. 
Since μ ^ K GBS - I is the empirical mean of i.i.d. Y with E Y = μ K ( 2 ) , we have Var ( μ ^ K GBS - I ) = Var ( Y ) / n = ( E Y 2 ( E Y ) 2 ) / n , which after the computation of E Y 2 above gives (81).    □
Corollary 8.2 
( ( ε , δ ) sample size for GBS-I). By Chebyshev,
Pr μ ^ K GBS - I μ K ( 2 ) > ε μ K ( 2 ) Var ( μ ^ K GBS - I ) ε 2 ( μ K ( 2 ) ) 2 .
Hence it suffices to take
n GBS - I 1 ε 2 δ · 1 d t B t 2 K | I | = 2 K a I 2 I ! Haf ( B I ) 2 ( μ K ( 2 ) ) 2 ( μ K ( 2 ) ) 2 .
In particular, at the photon-match m t B = 2 K (Theorem Calibration optimality), t d t B 1 t 2 K is minimized, thus minimizing the right-hand side.

8.2. GBS-P (Bias Vanishes): Asymptotic MSE via Delta Method

For the (degree-K) probability/square–root estimator we assume the standard sign convention σ I = sign ( Haf ( B I ) ) is available.1 Define
μ ^ K GBS - P = | I | = 2 K a I t K I ! σ I p ^ I ( t ) d t B = g p ^ ( t ) , g ( p ) = | I | = 2 K a I t K I ! σ I p I d t B .
Since p I ( t ) = d t B t 2 K Haf ( B I ) 2 / I ! , one has g ( p ( t ) ) = μ K . The multinomial delta method with covariance Cov ( p ^ ) = 1 n diag ( p ) p p therefore yields
n E μ ^ K GBS - P μ K 2 g ( p ) diag ( p ) p p g ( p ) , ( n ) .
A short calculation gives g / p I = 1 2 a I t K I ! σ I ( d t B 1 p I ) 1 / 2 and hence
I p I g p I 2 = 1 4 d t B t 2 K | I | = 2 K a I 2 I ! , I p I g p I 2 = μ K 2 .
We conclude:
Proposition 8.3 
(Asymptotic MSE of GBS-P). As n ,
E μ ^ K GBS - P μ K 2 = 1 4 n 1 d t B t 2 K | I | = 2 K a I 2 I ! μ K 2 + o 1 n .
Corollary 8.4 
( ( ε , δ ) sample size for GBS-P). By Markov’s inequality applied to the centered square, an asymptotic sufficient sample size is
n GBS - P 1 ε 2 δ · 1 d t B t 2 K | I | = 2 K a I 2 / I ! μ K 2 4 μ K 2 .
As in Corollary 8.2, the right-hand side is minimized at the photon match m t B = 2 K .

8.3. When GBS Outperforms Monte Carlo

Let μ ^ K MC = 1 n r = 1 n f K ( X r ) be the plain Monte Carlo estimator for the degree-K contribution with X r N ( 0 , B ) and
f K ( x ) : = | I | = 2 K a I x I .
Then E f K ( X ) = μ K and a standard Wick expansion gives
Var μ ^ K MC = 1 n | I | = 2 K | J | = 2 K a I a J Haf B I + J μ K 2 .
The quadratic form in (90) mixes all degree- 2 K multi–indices pairwise via Haf ( B I + J ) and therefore grows combinatorially in K (and, under mild spectral hypotheses, exponentially in an effective dimension), whereas the GBS prefactors (81)–(87) are controlled by the single convex factor d t B 1 t 2 K after photon–number matching.
Theorem 8.5 
(Regime separation: polynomial vs exponential). Fix B with eigenvalues bounded away from 1 and coefficients ( a I ) on the unit sphere of R D K . Then there exists c = c ( B ) > 0 such that for sequences K = K ( N ) with K ζ N 2 and ζ > 0 one has the complexity separation
n GBS = poly ( N ) whereas n MC exp c N 2 ,
for achieving the same ( ε , δ ) tolerance, provided the GBS sampler is photon–matched ( m t B = 2 K ). For fixed small K the two complexities are of the same order, and hybrid policies (MC for small K, GBS beyond a data–driven crossover) are optimal.
Proof sketch. 
Under the matching law m t B = 2 K , the GBS prefactors are minimized and remain polynomial in N (because d t B 1 = j ( 1 t 2 λ j 2 ) 1 / 2 with t < t max < λ 1 1 and t bounded away from λ 1 1 uniformly in N at fixed ζ ). In contrast, (90) sums Θ ( D K 2 ) terms with D K N 2 K and (for generic B) Haf ( B I + J ) bounded away from zero on a positive fraction of pairs, whence an exponential lower bound in K ζ N 2 . Careful bookkeeping of constants gives the stated separation.    □

8.4. Theory.

DPI, calibration optimality m t B = 2 K , angle bounds, the exact GBS–I variance, the large-n GBS–P MSE, and the ( ε , δ ) envelopes are publishable as-is.
Consequences. The exact/algebraic forms (81)–(87) and the convexity identity
d d t log d t B 1 t 2 K = m t B 2 K t
imply that photon–number matching m t B = 2 K  simultaneously (i) minimizes the leading factors in the GBS variance/MSE, (ii) sharpens the principal–angle bounds (conditioning), and (iii) maximizes the Lyapunov rate in the residual contraction. These yield non–asymptotic ( ε , δ ) envelopes for n GBS and the regime map against n MC as asserted in Theorem 8.5.

9. Design Rules from DSFL (Actionable)

This section collects operational prescriptions that follow from the comparison geometry developed above and states them as verifiable statements with proofs or proof sketches. Throughout we fix a degree K N , a covariance B, and a scale t ( 0 , λ 1 1 ) . The Hilbert space, blueprint, physical law, residual and principal-angle factor are as in the previous sections.

9.1. Calibration by Photon-Number Matching

Theorem 9.1 
(Calibration rule). Let R res ( K ) ( t ) denote the degree-K DSFL residual at scale t. Assume R res ( K ) ( t ) = A C 2 K ( t ) B with A > 0 and C 2 K ( t ) = t 2 K d t B 1 . Then t uniquely minimizes R res ( K ) ( t ) on ( 0 , λ 1 1 ) if and only if
m t B = 2 K .
Moreover, at t the leading factors in the variance/MSE bounds of the degree-K GBS estimators are minimized.
Proof. 
The identity d d t log C 2 K ( t ) = 1 t ( m t B 2 K ) and the strict convexity of t C 2 K ( t ) prove existence and uniqueness of the minimizer with stationarity condition m t B = 2 K . The leading t–dependence of n Var ( μ ^ K GBS I ) and of the large-n MSE of μ ^ K GBS P is proportional to C 2 K ( t ) by the variance/MSE formulæ in the sample-complexity section; hence the same minimizer applies.    □

9.2. Angle Check and Channel Selection

Define γ ( t ) = P GBS ( t ) P deg K = cos θ F ( t ) [ 0 , 1 ) and let c ± ( a , B ) > 0 be the t–independent constants from the angle–variance proposition.
Proposition 9.2 
(Angle gate for estimator dispersion). For the degree-K estimators at fixed t and n samples,
( 1 γ ( t ) ) c ( a , B ) n Var ( μ ^ K ) ( 1 + γ ( t ) ) c + ( a , B ) .
Consequently, if γ ( t ) γ [ 0 , 1 ) then
n Var ( μ ^ K ) ( 1 + γ ) c + ( a , B ) ,
and if γ ( t ) 1 η with η ( 0 , 1 ] then
n Var ( μ ^ K ) η c ( a , B ) .
Proof. 
The two-sided bound is the angle–variance link. The corollaries follow by monotonicity in γ .    □
Theorem 9.3 
(Angle threshold ⇒ GBS vs MC decision). Let n GBS ( t ) and n MC be guaranteed sample sizes (for a fixed ( ε , δ ) ) derived from Chebyshev/Markov bounds for the degree-K task. There exists a computable threshold γ ¯ = γ ¯ ( ε , δ , a , B , K ) > 0 such that
γ ( t ) γ ¯ n GBS ( t ) n MC .
Conversely, if γ ( t ) 1 η with η below a computable level η ̲ ( ε , δ , a , B , K ) , then n GBS ( t ) n MC .
Proof sketch. 
Insert the angle bounds into the Chebyshev/Markov forms for n GBS ( t ) (resp. the corresponding quantity for MC). The MC denominator involves the full mixed moment structure, while the GBS denominator is degree-filtered; both numerators are bounded above/below by ( 1 ± γ ) times t–independent constants. Solving the resulting inequality for γ yields the thresholds γ ¯ and η ̲ (explicit in terms of c ± ( a , B ) and the MC dispersion proxies).    □

9.2.1. Practical Surrogate

Let X t be the feature matrix whose columns are the (normalized) sampling features at scale t and Y K a basis for the degree-K blueprint range. Then γ ^ = X t Y K satisfies γ ^ γ ( t ) , so γ ^ γ ¯ is a conservative sufficient condition for choosing GBS.

9.3. Degree Decomposition and Hybrid Allocation

Let f = k 0 f k be the degree decomposition and write μ = k μ k . Suppose each degree-k component is estimated either by GBS (at its matched t k ) or by MC, yielding per-degree guaranteed sample sizes n k GBS and n k MC under a common ( ε , δ ) budget with per-degree tolerances { ε k } k K satisfying k K ε k ε .
Proposition 9.4 
(Greedy degree-wise policy is optimal under monotone margins). Assume n k GBS and n k MC are positive and the advantage margins
Δ k : = n k MC n k GBS
are nondecreasing in k. Then the policy that assigns GBS to the largest k for which Δ k > 0 and MC otherwise minimizes the total sample size k K n k among all degree-wise assignments obeying the same tolerance split.
Proof. 
Under the monotonicity hypothesis, the sequence of savings { Δ k } is ordered; an exchange argument shows any assignment deviating from the “assign GBS to the largest positive Δ k ” rule can be improved by swapping one GBS and one MC choice, strictly reducing total cost. Iterating yields the stated greedy optimum.    □
Corollary 9.5 
(A posteriori hybrid switch). If an a priori monotonicity check is unavailable, an empirical DSFL criterion suffices: after a warm-up with n 0 samples per degree,
R res ^ GBS ( k ) R res ^ MC ( k ) use GBS at degree k ; else use MC .
By the Lyapunov contraction and angle bounds, this rule is asymptotically consistent.

9.4. Seam Stability Under Composition (Gluing) and Leakage Quantification

Consider a pipeline of L updates ( U , Φ ) , = 1 , , L , where each U acts on the blueprint side and each Φ on the physical side (retuning, counting, linear optical blocks etc.).
Proposition 9.6 
(No inflation under gluing). If every step is DSFL-admissible, i.e. Φ U = U Φ and Φ 1 , then the composite update ( U , Φ ) = ( U L U 1 , Φ L Φ 1 ) satisfies
R res ( K ) U s ( K ) , Φ p R res ( K ) s ( K ) , p for all s ( K ) , p .
Proof. 
The data–processing inequality holds at each step:
R res ( K ) U s , Φ p R res ( K ) ( s , p ) .
Induction on yields the composite bound.    □
Proposition 9.7 
(Angle leakage bound). Let θ F ( ) be the Friedrichs angle between the degree-K blueprint subspace and the sampling feature subspace after step ℓ. If each Φ is a contraction onto a feature subspace, then
cos θ F ( L ) min 1 , = 1 L cos θ F ( ) .
In particular, if cos θ F ( ) η with η < 1 , then θ F ( L ) stays bounded away from 0 and the angle–variance bounds remain uniformly coercive along the pipeline.
Proof sketch. 
For orthogonal projectors P , Q one has P Q = cos θ F ( P , Q ) . CS-decomposition and sub-multiplicativity give P GBS ( L ) P deg K P GBS ( L ) P GBS ( L 1 ) P GBS ( 1 ) P deg K . Each factor is bounded by the corresponding cosine; the asserted 1 bound follows from the inequality 1 ( 1 x ) x applied to x = 1 cos θ F ( ) .    □

9.4.1. Consequence

Under calibrated retuning at each seam (so that m t B = 2 K ), the principal angle cannot collapse in finite length pipelines formed by admissible blocks whose individual misalignments are summable; thus residuals do not inflate and the Lyapunov envelope persists.

10. Experiments

This section describes synthetic and hardware-aligned numerical studies designed to validate the DSFL predictions. We specify the experimental design, the estimators and metrics, and we state formal consistency properties that connect empirical readouts (MSE curves, angle surrogates, advantage percentages) to the theoretical quantities in the preceding sections. Throughout, degree-K quantities are taken at the photon-match m t B = 2 K unless otherwise indicated; empirical laws p ^ ( n ) ( t ) are formed from n samples.

10.1. Simulated GBS and Synthetic ( B , a I )

10.1.1. Design.

Fix grids
N N , K N , and choose ( N , K ) N × K .
For each ( N , K ) :
  • Draw eigenvalue spectra { λ j } j = 1 N from prescribed classes:
    ( flat ) λ j = λ ( 0 , 1 ) , ( polynomial ) λ j = c j β , ( exponential ) λ j = c e β j ,
    with amplitude c ( 0 , 1 ) and decay β > 0 chosen so that 0 < λ N λ 1 < 1 .
  • Draw a random Haar orthogonal U O ( N ) and set B = U diag ( λ 1 , , λ N ) U .
  • Sample the degree-K blueprint on the unit sphere:
    s K = ( a I ) | I | = 2 K Unif S D K 1 , D K = # { I : | I | = 2 K } .
  • Compute the matched scale t by solving m t B = 2 K ; equivalently, find the unique t ( 0 , λ 1 1 ) such that j = 1 N t 2 λ j 2 / ( 1 t 2 λ j 2 ) = 2 K .

10.1.2. Estimators and Metrics

For each configuration and for a range of sample sizes n:
  • Form μ ^ K GBS - I ( n , t ) and μ ^ K GBS - P ( n , t ) from n i.i.d. synthetic GBS draws at t (via the hafnian-defined pmf).
  • Form the plain Monte Carlo estimator μ ^ K MC ( n ) using n i.i.d. X N ( 0 , B ) and Wick reconstruction to isolate degree K if desired.
  • Estimate MSEs and empirical residuals by repetitions:
    MSE ^ ( ) : = 1 R r = 1 R μ ^ ( r ) μ K 2 , R res ^ ( K ) ( p ^ ( n ) ( t ) ) : = p ^ ( n ) ( t ) s K H K 2 .
  • Report guaranteed sample sizes n ^ GBS - I , n ^ GBS - P , n ^ MC by inverting Chebyshev/Markov envelopes calibrated with empirical variances.
  • Compute efficiency ratios and advantage percentages:
    Eff . ratio = n ^ MC n ^ GBS , Adv. % = 1 T t = 1 T 1 { n ^ GBS n ^ MC } × 100 % .

10.1.3. Theoretical Guarantees

Proposition 10.1 
(Consistency of advantage estimators). Fix ( N , K ) and a spectrum class with 0 < λ j < 1 . Let n ^ GBS and n ^ MC be defined from empirical variances with R repetitions and n samples per repetition at matched t . Then
n ^ GBS a . s . n GBS , n ^ MC a . s . n MC ,
where n are the population Chebyshev/Markov sample sizes. Consequently, the empirical advantage percentage converges almost surely to the true percentage under the same configuration ensemble.
Proof sketch. 
Strong laws for empirical second moments of linear statistics of i.i.d. draws imply almost-sure convergence of variance estimators. Continuity of the inversion map ( v c v / ( ε 2 δ ) ) yields the stated limits; the law of large numbers over configuration draws implies convergence of frequencies.    □

10.2. Ablations Testing DSFL Predictions

10.2.1. T-Sweep (U-Shape)

For each fixed ( N , K ) and s K , sweep t in a neighborhood of t and evaluate R res ^ ( K ) ( p ^ ( n ) ( t ) ) and MSE ^ ( μ ^ K ( t ) ) .
Proposition 10.2 
(Empirical convexity and minimum at the match). Under the simulators above, for any fixed ( N , K ) and sufficiently large n , R ,
t E R res ^ ( K ) ( p ^ ( n ) ( t ) ) and t E MSE ^ ( μ ^ K ( t ) )
are approximately convex with a unique minimum at t = t (the solution of m t B = 2 K ). Moreover, their second derivatives with respect to log t are strictly positive up to o n ( 1 ) errors.
Proof sketch. 
The residual and variance/MSE leading factors are proportional to C 2 K ( t ) = t 2 K d t B 1 ; Lemma (scale calculus) showed d 2 d t 2 log C 2 K ( t ) > 0 and stationarity d d t log C 2 K ( t ) = 0 iff m t B = 2 K . Empirical averages converge to their expectations by SLLN, hence the observed convex U-shape and unique minimum at t up to sampling error.    □

10.2.2. Angle Surrogate vs Variance

Construct features X t and blueprint basis Y K , compute γ ^ = X t Y K , and correlate γ ^ with empirical n Var ^ ( μ ^ K ( t ) ) across t.
Proposition 10.3 
(Monotone trend). For fixed ( N , K ) and large n , R , the map t n Var ^ ( μ ^ K ( t ) ) is bounded between two affine functions of γ ^ ( t ) :
( 1 γ ^ ( t ) ) c ^ ( a , B ) n Var ^ ( μ ^ K ( t ) ) ( 1 + γ ^ ( t ) ) c ^ + ( a , B ) ,
with c ^ ± c ± in probability as R .
Proof sketch. 
Replace the projector norm γ ( t ) by the computable surrogate γ ^ ( t ) γ ( t ) ; the angle–variance proposition yields the population inequalities. Empirical variances converge to their population values, and the Gram-based surrogate is continuous in sample averages of features, giving the displayed bounds up to o n , R ( 1 ) errors.    □

10.2.3. Hybrid Crossover

Partition the blueprint across degrees k K max and estimate the total samples needed by degree-wise GBS vs MC. Plot the crossover index k × where the hybrid policy switches.
Proposition 10.4 
(Crossover index is increasing with N and decreasing with decay). Let k × ( N ; β ) be the smallest degree where n k GBS n k MC under a polynomial (or exponential) eigen-decay parameter β. Then for sufficiently large N and fixed tolerance ( ε , δ ) ,
N k × ( N ; β ) 0 , β k × ( N ; β ) 0 ,
i.e. larger systems favor GBS at lower degrees, while stronger decay (smaller effective dimension) delays the crossover.
Proof sketch. 
The MC dispersion includes mixed-moment growth with the effective dimension and increases rapidly with N when decay is weak; GBS dispersion at the matched degree depends primarily on degree-k features. Comparing the two sample-size envelopes and differentiating w.r.t. the structural parameters yields the stated monotonicities.    □

10.3. Hardware-Aligned Kernels (Optional)

10.3.1. Models

Implement time-bin/loop architectures by stochastic kernels that compose (i) squeezing maps defining t B , (ii) interferometer unitary draws, and (iii) photon-number resolving measurements (or threshold versions). Insert calibrated loss channels modeled as contractions on H K .

10.3.2. Stress Tests

For each pipeline, concatenate L calibrated blocks and measure R res ^ ( K ) after each seam, as well as γ ^ ( t ) .
Proposition 10.5 
(No inflation under concatenation with calibrated retuning). Assume each block is DSFL-admissible and retuned to its local match m t B = 2 K . Then the empirical residual is nonincreasing along the pipeline up to sampling error:
E R res ^ ( K ) p ^ ( n ) E R res ^ ( K ) p ^ 1 ( n ) + o n ( 1 ) , = 1 , , L .
Proof. 
Apply the DPI for each admissible block (projection/contraction and intertwining), and use convergence of empirical residuals to their expectations.    □
Proposition 10.6 
(Robustness to small losses). Let a loss channel of strength η [ 0 , 1 ) act between calibrated blocks. Then the matched scale t solves the perturbed equation m t B η = 2 K and the contraction rate α ( t ) degrades by at most O ( η ) ; consequently, the empirical semi-log MSE slope remains within O ( η ) of the noiseless value for fixed n.
Proof sketch. 
A lossy Gaussian channel induces B B η with B η B = O ( η ) ; Davis–Kahan perturbation gives | γ η γ | = O ( η ) , while the scale calculus shows t shifts continuously in η . The Lyapunov rate α is monotone in the angle margin, hence perturbs by O ( η ) .    □

10.3.3. Reporting

For each setting we present: (i) MSE and residual vs n (semi-log slopes), (ii) MSE vs t (U-shapes with minimum at t ), (iii) variance vs angle surrogate curves, (iv) advantage heatmaps over ( N , K ) , and (v) pipeline residual traces across seams with/without loss.

11. Concluding Discussion

This work reframes Gaussian Boson Sampling (GBS) as a calibrated variance–reduction engine operating in a single comparison geometry. Within that geometry, the Deterministic Statistical Feedback Law (DSFL) supplies three pillars that, taken together, elevate the “folk rules’’ of GBS into theorems: (i) calibration optimality, which identifies photon–number matching m t B = 2 K as the unique minimizer of the leading variance/MSE factor t 2 K d t B 1 ; (ii) principal–angle conditioning, which yields sharp two–sided dispersion bounds with multiplicative constants ( 1 ± cos θ F ) ; and (iii) a Lyapunov envelope that converts one–step contraction into a non–asymptotic exponential error decay MSE n C e α ( t ) n with α ( t ) maximal at the photon match. These statements are not merely sufficient conditions; they jointly explain why GBS beats Monte Carlo (MC) in high–degree regimes, how to tune the device to realize the advantage, and what rates to expect in finite samples.
On the algorithmic side, the DSFL lens unifies unbiased (GBS–I) and biased–but–consistent (GBS–P) estimators, shows that both inherit the same convex t–dependence, and clarifies the role of degree partitioning: once the target is decomposed into degree components, GBS naturally concentrates effort where MC pays an exponential price through cross–moment couplings, while a hybrid policy remains optimal for small degrees. On the systems side, DSFL’s data–processing inequality guarantees no residual inflation under counting, retuning, or block concatenation; principal–angle control quantifies leakage across interfaces; and perturbation arguments (Davis–Kahan type) explain the graceful degradation under small hardware noise together with recovery by re–matching m t B .
The implications are practical. First, calibration reduces to solving a strictly convex, one–dimensional equation for t, hence is robust and fast; this makes photon–number matching a pre–run, certifiable step. Second, a computable angle surrogate (e.g., a cross–Gram norm between sampling features and blueprint basis) provides an a priori gate for channel selection (GBS vs MC), enabling hybrid deployment without post–hoc tuning. Third, the Lyapunov envelope gives finite–n guarantees that can be monitored online by empirical DSFL residuals, turning semi–log MSE plots into quantitative health checks.
There are natural extensions. For non–Gaussian observables or generalized boson–sampling models (loop–hafnians, threshold detection), one must redesign the feature map and admissible updates so that the DPI persists and bias terms remain controlled; the same Hilbertian template applies. Angle surrogates can be refined—e.g. via randomized subspace embeddings—to scale to very large ( N , K ) while preserving principal–angle fidelity. Finally, the experimental program suggested here—t–sweeps for convex U–shapes, variance–vs–angle curves, and block–wise residual traces—offers falsifiable diagnostics for hardware pipelines, and a path to controller–in–the–loop calibration where t and the interferometer are adjusted to maximize the Lyapunov rate α ( t ) subject to device constraints.
In sum, DSFL furnishes a predict–then–prove methodology for GBS: it predicts the optimal operating point ( m t B = 2 K ), the conditioning ( θ F ), and the rate ( α ( t ) ) before data are collected; it proves that admissible updates cannot worsen the calibrated misfit; and it turns those predictions into actionable rules that persist under composition and mild noise. This synthesis clarifies when and why GBS delivers sample–complexity advantages over MC, and how to tune photonic pipelines to realize them in practice.

Funding

No external funding was received for this work.

Data and code availability

This work uses only synthetic data produced by the algorithms described in the paper; no external datasets were used or analyzed. A fully reproducible package containing (i) source code, (ii) parameter files for every figure, (iii) exact random seeds, and (iv) an exportable runtime environment (both environment.yml and requirements.txt) will be deposited at Zenodo under an archival DOI (to be added at acceptance) and mirrored on GitHub (read-only tag matching the DOI). The repository includes a one-click reproducibility script: which regenerates all tables and figures in the manuscript. Tested platforms: Linux (x86_64) and macOS (arm64) with Python ≥ 3.10; core dependencies: NumPy, SciPy, Matplotlib, and JAX/Numba (optional, for accelerated runs). Code will be released under an OSI-approved license (MIT).

Acknowledgments

The author affirms sole authorship and received no external funding.

Conflicts of Interest

The author declares no competing interests.

Appendix A. Crisp Theorem Statements (Ready to Formalize)

Appendix A.1. Standing Setup

Let ( H , · , · ) be a Hilbert space with norm x = x , x , and closed subspaces S , P H . A calibration is a pair of bounded maps
I : S P , J : P S ,
satisfying the interchangeability identities
I J = id P , J I = P S ,
with P S the orthogonal projector onto S . For s S and p P , the Residual of Sameness is
R ( s , p ) : = p I s H 2 .
An admissible update is a pair ( Φ ˜ , Φ ) with Φ ˜ : S S , Φ : P P such that
Φ I = I Φ ˜ and Φ H H 1 .
For angle bounds, let M , N H be closed subspaces with orthogonal projectors P M , P N , and Friedrichs cosine
γ ( M , N ) : = P M P N [ 0 , 1 ] , cos θ F ( M , N ) = γ ( M , N ) .

DPI (GBS Admissibility)

Theorem A.1 
(Data–processing inequality). If ( Φ ˜ , Φ ) is admissible, then for all s S and p P ,
R ( Φ ˜ s , Φ p ) R ( s , p ) .
Proof. 
Using Φ I = I Φ ˜ ,
R ( Φ ˜ s , Φ p ) = Φ p I Φ ˜ s 2 = Φ ( p I s ) 2 Φ 2 p I s 2 p I s 2 .
Corollary A.2 
(Composition). If { ( Φ ˜ k , Φ k ) } k = 1 L are admissible, then ( Φ ˜ , Φ ) = ( k = 1 L Φ ˜ k , k = 1 L Φ k ) is admissible and
R ( Φ ˜ s , Φ p ) R ( s , p ) .
Proof. 
Intertwining and · 1 are preserved under composition; apply Theorem A.1. □

Calibration optimality ⇒ mtB = 2K

Appendix A.2. GBS Specialization

Let B R N × N be symmetric with 0 < λ N λ 1 < 1 , and fix a degree K N . For a scale t ( 0 , λ 1 1 ) , define
d t B : = j = 1 N 1 t 2 λ j 2 , C 2 K ( t ) : = t 2 K d t B , m t B : = j = 1 N t 2 λ j 2 1 t 2 λ j 2 .
Assume the degree-K residual admits the leading form R res ( K ) ( t ) = A C 2 K ( t ) B with A > 0 independent of t.
Lemma A.3 
(Scale calculus). d d t log C 2 K ( t ) = m t B 2 K t , and d 2 d t 2 log C 2 K ( t ) = 2 K t 2 + j = 1 N λ j 2 1 t 2 λ j 2 + j = 1 N 2 t 2 λ j 4 ( 1 t 2 λ j 2 ) 2 > 0 .
Proof. 
Differentiate log C 2 K ( t ) = 2 K log t 1 2 j log ( 1 t 2 λ j 2 ) termwise. □
Theorem A.4 
(Calibration optimality). The unique minimizer t ( 0 , λ 1 1 ) of t R res ( K ) ( t ) is characterized by the photon–number match
m t B = 2 K .
Equivalently, t minimizes the leading t–factors in the variance/MSE of the degree-K GBS estimators.
Proof. 
By Lemma A.3, critical points of C 2 K satisfy m t B = 2 K and C 2 K is strictly convex, hence the minimizer is unique. The variance/MSE leading factors for GBS-I and GBS-P are proportional to C 2 K ( t ) ; thus they share the same minimizer. □

What is provable once clearly stated assumptions are added (action items)

  • Angle ⇒ rate monotonicity. We state that the Lyapunov rate α ( t ) improves as γ ( t ) = P GBS ( t ) P deg K decreases. Action: Include Lemma B.2, which bounds the rate between linear functions of the angle margin:
    c ( 1 γ ( t ) ) α ( t ) c + ( 1 γ ( t ) ) ,
    with t–independent 0 < c c + < . The proof is a two–subspace argument using the through–subspace contraction (Proposition 6.8) and the sharp bounds in Proposition A.7.
  • Residual factorization. We use R ( K ) ( t ) = A C 2 K ( t ) B to map calibration to residual minimization. Action: Add Lemma A.5 (Degree–K residual factorization) showing that all t–dependence factors through t 2 K and d t B 1 for the degree–K laws. A concise proof sketch is included in the appendix, relying on hafnian homogeneity Haf ( ( t B ) I ) = t K Haf ( B I ) and linearity of the readouts.
Lemma A.5 
(Residual factorization for degree-K). Fix K N , a covariance B with spectrum in ( 0 , 1 ) , and coefficients ( a I ) | I | = 2 K . Let t ( 0 , λ 1 1 ) , set d t B = j = 1 N 1 t 2 λ j 2 and
C 2 K ( t ) : = t 2 K d t B .
Consider the degree-K DSFL residual R res ( K ) ( t ) : = p ( t ) I s K H K 2 built from either readout
GBS-P : I P x = | I | = 2 K a I x I d t B , GBS-I : I I x = | I | = 2 K a I I ! d t B x I .
Then there exist t–independent constants A > 0 and B R (depending only on ( a I ) and B) such that
R res ( K ) ( t ) = A C 2 K ( t ) B .
Sketch. 
For | I | = 2 K , hafnian homogeneity gives Haf ( ( t B ) I ) = t K Haf ( B I ) , whence p I ( t ) = d t B t 2 K Haf ( B I ) 2 / I ! . Thus every coordinate of p ( t ) carries the common scalar factor d t B t 2 K . The degree-K readouts introduce only the normalization d t B 1 (and I ! in GBS-I), which does not depend on t beyond d t B . Expanding p ( t ) I s K 2 shows the t–dependence factors through t 2 K d t B 1 = C 2 K ( t ) , with all inner products collected into t–independent constants A > 0 and B R . A fully expanded computation (including the multinomial covariance form for GBS-P and the exact second moment for GBS-I) is provided in Appendix A. □
Corollary A.6 
(Calibration ⇒ residual minimization). Under the hypotheses of Lemma A.5, the unique minimizer of t R res ( K ) ( t ) on ( 0 , λ 1 1 ) is characterized by m t B = 2 K .
Proof. 
By Lemma A.5, R res ( K ) ( t ) = A C 2 K ( t ) B with A > 0 . The scale calculus shows d d t log C 2 K ( t ) = m t B 2 K t and t C 2 K ( t ) is strictly convex on ( 0 , λ 1 1 ) . Hence the unique minimizer of R res ( K ) ( t ) satisfies m t B = 2 K . □
Proof. 
By hafnian homogeneity, p I ( t ) = d t B t 2 K Haf ( B I ) 2 / I ! . Hence p ( t ) 2 = I p I ( t ) 2 = d t B 2 t 4 K I Haf ( B I ) 4 / I ! 2 = d t B t 2 K · t 2 K d t B I Haf 4 / I ! 2 . The parenthesized factor is independent of t after regrouping; similarly, the cross term p ( t ) , Π s K = I p I ( t ) ( Π s K ) I is proportional to d t B t 2 K times a t–independent sum, and Π s K 2 is t–independent. Expanding p ( t ) Π s K 2 yields the claimed affine dependence on C 2 K ( t ) = t 2 K d t B 1 with A > 0 . □

Angle bounds

Proposition A.7 
(Two-subspace inequality). Let M , N H be closed subspaces with γ = P M P N [ 0 , 1 ] . For any a M and b N ,
( 1 γ ) a 2 + b 2 a + b 2 ( 1 + γ ) a 2 + b 2 .
Proof. 
Expand a + b 2 = a 2 + b 2 + 2 a , b , and use | a , b | = | a , P M b | a P M b γ a b , then 2 a b a 2 + b 2 . □

Appendix A.3. GBS specialization (degree-K).

Let P deg K be the blueprint projector and P GBS ( t ) the sampling-feature projector at scale t, and define γ ( t ) = P GBS ( t ) P deg K = cos θ F ( t ) .
Theorem A.8 
(Angle–dispersion link). For the degree-K estimators from n samples at fixed t, there exist t–independent c ( a , B ) , c + ( a , B ) > 0 such that
( 1 γ ( t ) ) c ( a , B ) n Var ( μ ^ K ) ( 1 + γ ( t ) ) c + ( a , B ) ,
with Var understood as the variance for GBS-I and as the leading asymptotic variance (or squared-bias term) for GBS-P. Equivalently, MSE is sandwiched between two multiples of the split energy, with multiplicative factors ( 1 ± cos θ F ( t ) ) .
Proof sketch. 
Write the estimator as μ ^ K = w , p ^ ( t ) with w fixed and p ^ ( t ) the empirical law. Then n Var ( μ ^ K ) = Σ ( t ) w , w , where Σ ( t ) is the single-sample covariance. The range of Σ ( t ) lies in P GBS ( t ) H , while w decomposes into P deg K w + w . Apply Proposition A.7 twice (once to control leakage into P GBS ( t ) , once for the split relative to P deg K ) and absorb t–independent weights into c ± ( a , B ) . □

Lyapunov bound

Theorem A.9 
(Discrete Lyapunov envelope). Let R res ( K ) be the degree-K residual and suppose there exists α ( t ) ( 0 , 1 ] such that, conditionally on the past,
E R res ( K ) U s ( K ) , Φ p ( n ) ( t ) | p ( n ) ( t ) ( 1 α ( t ) ) R res ( K ) s ( K ) , p ( n ) ( t ) ,
for an admissible one-step update ( U , Φ ) . Then
E R res ( K ) p ( n ) ( t ) e α ( t ) n R res ( K ) p ( 0 ) ( t ) .
Moreover, α ( t ) increases as γ ( t ) decreases, and is maximized at the calibration match m t B = 2 K .
Proof. 
Taking total expectation yields E R n + 1 ( 1 α ) E R n with R n : = R res ( K ) ( p ( n ) ( t ) ) . Iteration gives E R n ( 1 α ) n R 0 e α n R 0 . By Theorem A.8, improved alignment (smaller γ ) strengthens contraction; by Theorem A.4, the match maximizes this gain. □
Corollary A.10 
( ( ε , δ ) sample-size envelope). If n α ( t ) 1 log R res ( K ) ( p ( 0 ) ( t ) ) / ε 2 , then E R res ( K ) p ( n ) ( t ) ε 2 . At the match m t B = 2 K , the right-hand side is minimized in t.
Proof. 
Solve e α ( t ) n R res ( K ) ( p ( 0 ) ( t ) ) ε 2 for n. □

Appendix B. Explaining the Elements: sDoF/pDoF, Interchangeability, DSFL and the Residual of Sameness R

This section fixes the primitives used throughout. We first embed the statistical degrees of freedom (sDoF) and physical degrees of freedom (pDoF) in one comparison Hilbert space; then we define the single observable (Residual of Sameness); finally we state the admissibility gate that makes the residual monotone and, when dynamics are present, yields a Lyapunov envelope.

Appendix B.1. Interchangeability (Calibration)

Let ( H , · , · ) be a Hilbert space carrying two closed subspaces S (statistical channel) and P (physical channel). A calibration is a pair of bounded maps
I : S P , J : P S
satisfying
I J = id P , J I = P S ,
with P S the orthogonal projector onto S . Thus J provides calibrated statistical representatives of p P , I provides calibrated physical images of s S , and composing in either channel returns the original.

Appendix B.2. One Observable: the Residual of Sameness

For s S and p P ,
R ( s , p ) : = p I ( s ) H 2
is the squared distance from p to ran I . It vanishes iff p = I ( s ) and is invariant under isometries preserving ran I .

Appendix B.3. Admissibility and the Data–Processing Inequality

A pair ( Φ ˜ , Φ ) with Φ ˜ : S S and Φ : P P is admissible if
Φ I = I Φ ˜ , Φ H H 1 , Φ ˜ H H 1 .
Then the Hilbertian data–processing inequality holds:
R ( Φ ˜ s , Φ p ) = Φ ( p I s ) H 2 p I s H 2 = R ( s , p ) .
Hence no admissible update inflates the calibrated mismatch. In operator–algebraic models, J is an orthogonal conditional expectation, and (34) follows from Kadison–Schwarz.

Appendix B.4. Propagation with a Gap: the DSFL Lyapunov Envelope

Assume a residual identity
d d t R ( t ) = 2 K e , e H + 2 e , g H , e : = p I s , K = K * 0 ,
and bounds
K e , e H κ e H 2 , | e , g H | ε e H 2 , κ > ε 0 .
Then
d d t R ( t ) α R ( t ) , α : = 2 ( κ ε ) > 0 , R ( t ) e α ( t t 0 ) R ( t 0 ) .
Thus log R decays at a single rate α fixed by the sector’s coercivity margin.

Appendix B.5. Operational Arrow (Accumulated Improvement)

Define
S R ( t ) : = log R ( t ) R ( 0 ) = 0 t α ( τ ) d τ .
It is nondecreasing and quantifies total calibrated mismatch removed, independently of the choice of clock.

Appendix B.6. Minimal Checklist for Applications

  • Fix H , S , P ; choose ( I , J ) with (30).
  • Verify (A19) for each step; (34) then guarantees monotonicity.
  • If dynamics are available, identify ( κ , ε ) in (A22) to obtain α in (A23).

Appendix B.7. Generic Two–Channel Application Template

Let s ( t ) S and p ( t ) P with calibration I : S P . Set e ( t ) : = p ( t ) I s ( t ) , R ( t ) : = e ( t ) H 2 . If
R ˙ ( t ) = 2 K e , e H + 2 e , g H , K e , e H κ e H 2 , | e , g H | ε e H 2 ,
then R ˙ 2 ( κ ε ) R and R ( t ) e 2 ( κ ε ) ( t t 0 ) R ( t 0 ) . This dynamic envelope pairs with the static monotonicity R ( Φ ˜ s , Φ p ) R ( s , p ) whenever ( Φ ˜ , Φ ) is admissible.

Appendix B.8. Terminology and Mechanism: sDoF/pDoF, DSFL, and the Dual Feedback Loop

Appendix B.8.1. What are sDoF and pDoF?

The statistical degrees of freedom (sDoF) are the blueprint variables s S that encode the model, prior, or control law; the physical degrees of freedom (pDoF) are the measured/realized responses p P . Both live in the same comparison Hilbert space ( H , · , · ) via a calibration ( I , J ) obeying
I J = id P , J I = P S .
Equation (A26) ensures two-way coherence: J reads a calibrated blueprint from a physical state, and I writes a calibrated physical image of a statistical state. The Residual of Sameness
R ( s , p ) = p I s H 2
measures the mismatch between pDoF and the calibrated sDoF within H .

Appendix B.8.2. Why “DSFL”

Deterministic Statistical Feedback Law (DSFL) is shortened to DSFL because updates are deterministic (linear/contractive), the driver is statistical (sDoF), the mechanism is feedback (via the residual e : = p I s ), and it functions as a law (one inequality governs all steps). For any admissible update ( Φ ˜ , Φ ) ( Φ I = I Φ ˜ , Φ , Φ ˜ 1 ),
R ( Φ ˜ s , Φ p ) = Φ ( p I s ) H 2 p I s H 2 = R ( s , p ) ,
so the residual is L 2 –monotone.

Appendix A.8.3. Dual Feedback Loop (Discrete Time)

There are two tightly coupled feedback paths in the same geometry:
S-loop : s k + 1 = Φ ˜ s k , P-loop : p k + 1 = Φ p k , e k + 1 = p k + 1 I s k + 1 ,
with Φ I = I Φ ˜ and Φ , Φ ˜ 1 . By (A28), R ( s k + 1 , p k + 1 ) R ( s k , p k ) sample–wise. If, in addition, a Lyapunov identity with a gap holds, R ˙ α R , then R k C e α k (exponential envelope).

Appendix B.8.4. Flow form (Continuous Time)

Writing e ( t ) = p ( t ) I s ( t ) , assume
d d t R ( t ) = 2 K e , e H + 2 e , g H , K e , e κ e 2 , | e , g | ε e 2 ,
with κ > ε 0 . Then d d t R 2 ( κ ε ) R and R ( t ) e 2 ( κ ε ) ( t t 0 ) R ( t 0 ) . The accumulated drop S R ( t ) = log R ( t ) R ( 0 ) = 0 t 2 ( κ ε ) d τ is nondecreasing and quantifies the total calibrated mismatch removed.

Appendix B.8.5. Practical Interpretation

The P-loop evolves pDoF through an L 2 –contractive map that respects calibration, reducing e on the physical side; the S-loop refines sDoF so that I s tracks the evolving p. Because both loops live in H and intertwine ( I , J ) , they cannot fabricate residual—only dissipate or transport it. In applications with slow drift, one may update the calibration ( I , J ) ( I ( m ) , J ( m ) ) on a coarser timescale as long as I ( m ) J ( m ) = id P , J ( m ) I ( m ) = P S ; between updates the inner dual loop remains admissible and residual–monotone.

What Is Fully Proved (or Reducible to Standard Facts)

Appendix B.8.6. Data–Processing Inequality (DPI) in the DSFL Geometry

Claim. Admissible updates (intertwining Φ I = I Φ ˜ and nonexpansive Φ H H 1 ) cannot inflate the residual.
Status. Proved in the text (Theorem “DPI”) by a one–line Hilbert–space contraction; stability under composition is immediate.

Appendix B.8.7. Calibration Optimality mtB = 2K

Claim. The unique minimizer of the leading variance/MSE factor is the photon–number match.
Status. Proved by strict convexity on ( 0 , λ 1 1 ) of
log C 2 K ( t ) = 2 K log t 1 2 j log 1 t 2 λ j 2 , d d t log C 2 K ( t ) = m t B 2 K t .
Scope. Optimal for any quantity whose t–dependence factors through C 2 K ( t ) : the leading terms of the GBS-I variance and GBS-P MSE, and the DSFL residual representation R res ( K ) ( t ) = A C 2 K ( t ) B (when that representation holds).

Appendix B.8.8. Angle (Friedrichs) Bounds

Claim. Two–sided MSE/variance control with factors ( 1 ± cos θ F ) .
Status. Proved via the two–subspace inequalities and P M P N = cos θ F ; explicit derivations included.

Appendix B.8.9. Exact Variance for GBS-I; Asymptotic MSE for GBS-P

GBS-I. Exact variance formula proved from the pmf and hafnian homogeneity.
GBS-P. Large-n leading term proved by the multinomial delta method (gradient and covariance computed).
Scope/assumptions. For GBS-P we assume the usual sign-known convention and interior probabilities/smoothness so the delta method applies.

Appendix B.8.10. Chebyshev/Markov (ε,δ) Guarantees

Claim. Invert variance/MSE to obtain guaranteed sample sizes.
Status. Proved once the variance/MSE statements hold (and μ 0 ), by direct Chebyshev/Markov bounds.

Appendix B.8.11. Lyapunov Envelopes (Discrete/Continuous)

Claim. If a one–step conditional contraction holds, the expected residual decays exponentially.
Status. Proved as an implication via discrete (and continuous) Grönwall.
Scope. The contraction is the hypothesis; the envelope follows rigorously.

What Is Provable Once Clearly Stated Assumptions Are Added

Appendix B.8.12. Residual Representation

Where used. To map calibration optimality to residual minimization.
Status. Provable for degree-K laws and the two estimators, since the only nontrivial t–dependence enters through t 2 K and d t B 1 . One factors t out of (i) the multinomial covariance and (ii) the linear readout; a short lemma closes the argument.

Appendix B.8.13. “Angle ⇒ rate” Linkage

Claim. The Lyapunov rate α ( t ) improves as γ ( t ) = cos θ F ( t ) decreases.
Status. Provable once the conditional contraction constant is written as a monotone functional of the split energy controlled by ( 1 ± γ ) . With the projector bounds in hand, a brief comparison lemma formalizes this.

Appendix B.8.14. Composition/Gluing Stability and Small-Noise Robustness

Claim. No inflation under concatenation; small perturbations degrade angles/rates by O ( Δ ) .
Status. Concatenation: already follows from DPI under composition (proved). Robustness: Davis–Kahan/Wedin gives
| sin θ F ( B + Δ ) sin θ F ( B ) | = O Δ gap ,
hence γ perturbs by O ( Δ ) and rates follow from the previous item.

Appendix B.8.15. Asymptotic Normality (CLT) for GBS-P

Claim. Beyond the leading MSE constant, a full CLT holds.
Status. Provable by standard smooth-functionals-of-the-multinomial arguments under mild regularity (already used for the delta-method step).

What Is Solid but We Cite Rather than Re-Prove Here

Corollary B.1 
(CLT for GBS-P). Let I K = { I : | I | = 2 K } and write m : = | I K | . Fix a scale t ( 0 , λ 1 1 ) and assume the GBS law
p ( t ) = p I ( t ) I I K ( 0 , 1 ) m , I I K p I ( t ) = 1 ,
i.e. all degree-K outcome probabilities are strictly interior. Let the degree-K GBS-P functional be
g ( p ) : = I I K a I t K I ! σ I p I d t B , d t B = j = 1 N 1 t 2 λ j 2 ,
with σ I = sign ( Haf ( B I ) ) (or absorbed into a I when Haf ( B I ) 0 ). Let p ^ ( t ) be the empirical law based on n i.i.d. samples from p ( t ) , and μ ^ K GBS - P : = g ( p ^ ( t ) ) . Then
n μ ^ K GBS - P μ K d N 0 , g ( p ) diag ( p ) p p g ( p ) ,
where
g p I ( p ) = 1 2 a I t K I ! σ I 1 d t B p I 1 / 2 ( I I K ) .
In particular, the asymptotic variance simplifies to
g ( p ) diag ( p ) p p g ( p ) = 1 4 1 d t B t 2 K | I | = 2 K a I 2 I ! μ K 2 ,
which matches the leading constant in Proposition 8.3.
[39,40]
Proof. 
Write S = ( S I ) I I K Multinomial ( n , p ) and p ^ = S / n . The interior condition p I ( 0 , 1 ) implies the classical multinomial central limit theorem:
n ( p ^ p ) d N 0 , Σ ( p ) , Σ ( p ) : = diag ( p ) p p .
Since g is C 1 on the open simplex (each coordinate appears under a square root and p I > 0 ), the delta method yields
n g ( p ^ ) g ( p ) d N 0 , g ( p ) Σ ( p ) g ( p ) .
The displayed gradient follows from g ( p ) = I c I p I with c I = a I t K I ! σ I / d t B , so g / p I = 1 2 c I p I 1 / 2 . Finally,
I p I g p I 2 = 1 4 I c I 2 1 = 1 4 1 d t B t 2 K | I | = 2 K a I 2 I ! , I p I g p I 2 = μ K 2 ,
because g ( p ) = μ K by construction (Wick + homogeneity). Hence g Σ ( p ) g = I p I ( g / p I ) 2 I p I g / p I 2 = 1 4 d t B 1 t 2 K a I 2 / I ! μ K 2 .
Remark B.1 
(Boundary cases). If some coordinates satisfy p I = 0 , the map p g ( p ) is not differentiable at p; one may either (i) remove those indices from I K (they are p-null) and re-normalize, or (ii) work with a directional delta method under sparsity assumptions. The interior condition in the corollary avoids these technicalities.

Appendix B.8.16. Open, Nonempty Regions of Advantage; Percentage-of-Advantage

Content (Andersen–Shan 2025). Large-n leading term for GBS-P (matching ours), improved tuning via m t B = 2 K , ( ε , δ ) sample-size envelopes for both estimators, existence of nonempty open subsets with advantage over MC, and percentage-of-advantage values over ( a , B ) under Haar–Vandermonde measures.
Status. We rely on these theorems and cite them explicitly.

What Is Still Heuristic/Needs Additional Work

Appendix B.8.17. Principal-angle surrogate γ ^ = X t Y K 2 as a certified upper bound.

Claim. The computable Gram surrogate upper-bounds the true Friedrichs cosine.
Status. Heuristic but tighten-able: if X t spans the sampling feature subspace and Y K spans the blueprint subspace (properly normalized), then X t Y K P GBS ( t ) P deg K . A short approximation lemma with sampling-error bounds can make this rigorous.

Appendix B.8.18. Regime Separation (Poly vs. Exp).

Claim. For K ζ N 2 , n GBS = poly ( N ) while n MC = exp ( Θ ( N ) ) .
Status. Proof sketch given; a fully rigorous version needs fixed ensembles (eigenvalues in [ c , 1 ε ] ; coefficients on the unit sphere) and a positive-mass lower bound for the MC mixed-hafnian quadratic form. With these in place, the exponential lower bound follows by counting/Stirling estimates (in line with Andersen–Shan open-set results).

Appendix B.8.19. One-Step Conditional Contraction (Hypothesis in Lyapunov Theorem)

Claim. Each update removes a fixed fraction of the residual in expectation.
Status. Currently an assumption (empirically supported by semi-log linearity). A device-level proof would require an explicit Markov kernel and a uniform spectral gap; feasible but model-specific.

Appendix B.8.20. One-Step Conditional Contraction (Hypothesis in Lyapunov Theorem)

Claim. Each update removes a fixed fraction of the residual in expectation.
Status. Currently an assumption (empirically supported by semi-log linearity). A device-level proof would require an explicit Markov kernel and a uniform spectral gap; feasible but model-specific.
Lemma B.2 
(Rate improves as angle opens). Let R res ( K ) ( t ) = p I s 2 be the degree-K residual at scale t and let ( U , Φ ) be an admissible one-step update (intertwining and Φ 1 ). Assume the conditional one-step contraction hypothesis of Theorem A.9: there exists α ( t ) ( 0 , 1 ] such that
E R res ( K ) U s , Φ p | s , p 1 α ( t ) R res ( K ) ( t ) .
Then there exist t–independent constants 0 < c c + < (depending only on the fixed weights a I and B through bounded positive operators) such that
c ( 1 γ ( t ) ) α ( t ) c + ( 1 γ ( t ) ) , γ ( t ) = P GBS ( t ) P deg K = cos θ F ( t ) .
In particular, α ( t ) increases as γ ( t ) decreases and is maximized at the photon–number match m t B = 2 K .
Proof. 
Write the residual as R res ( K ) ( t ) = e 2 with e : = p I s H K . One step gives
e + : = Φ p I U s = Φ ( p I s ) + Φ I s I U s = 0 by intertwining = Φ e ,
hence R res ( K ) + = e + 2 = Φ e 2 . Taking conditional expectation and comparing to e 2 ,
α ( t ) = 1 E Φ e 2 e 2 .
Decompose e = e + e with respect to the blueprint/sampling pair,
e : = P deg K e , e : = ( I P deg K ) e ,
and note that Φ acts through the sampling feature subspace (by admissibility) so that the “through–subspace’’ contraction applies:
Φ e = P GBS ( t ) Φ P deg K e + P GBS ( t ) Φ e Φ P GBS ( t ) P deg K e + Φ P GBS ( t ) e .
Using Φ 1 and the two–subspace bounds (Proposition A.7),
P GBS ( t ) P deg K e γ ( t ) e , P GBS ( t ) e e .
Squaring and taking expectation (conditionally on e) yields
E Φ e 2 γ ( t ) 2 e 2 + e 2 1 ( 1 γ ( t ) 2 ) e 2 e 2 e 2 .
By the two–subspace lower bound,
( 1 γ ( t ) ) e 2 + e 2 e + e 2 = e 2 e 2 e 2 1 γ ( t ) 2 ,
where the constant 1 2 is an absolute (non-optimized) choice; any fixed c ( 0 , 1 ) would do after renormalizing constants. Hence
E Φ e 2 e 2 1 1 2 ( 1 γ ( t ) 2 ) ( 1 γ ( t ) ) ( 1 γ ( t ) ) ,
which implies an upper bound
α ( t ) c ( 1 γ ( t ) ) ,
for some c > 0 independent of t (absorbing the numerical constants and the fixed weights). A matching upper bound follows by reversing the roles of the projectors in the two–subspace inequalities:
P GBS ( t ) P deg K e 1 γ ( t ) 2 e E Φ e 2 1 ( 1 γ ( t ) 2 ) e 2 e 2 e 2 ,
and using e 2 e 2 1 1 γ ( t ) (another consequence of Proposition A.7) gives
α ( t ) c + ( 1 γ ( t ) )
for some c + < independent of t. The monotonicity of α ( t ) in γ ( t ) is immediate, and maximization at m t B = 2 K follows since γ ( t ) is minimized (best alignment) at the photon–match by the calibration optimality and angle–conditioning discussion. □

Appendix B.9. Bottom Line

Rock solid now: DPI; calibration optimality m t B = 2 K ; angle bounds; exact GBS-I variance; large-n GBS-P MSE; ( ε , δ ) inversions; Lyapunov envelope (conditional on a one-step contraction); composition stability; small-noise robustness (Davis–Kahan).
Provable with short add-ons: residual factorization A C 2 K B ; monotone “angle ⇒ rate” linkage; CLT for GBS-P; certified Gram surrogate bound.
Backed by the literature: nonempty open regions with advantage and percentage-of-advantage computations (Andersen–Shan 2025).
Still to complete: a model-independent proof of the one-step contraction and a fully quantified “poly vs. exp” lower bound with explicit constants under clear ensembles.

References

  1. Andersen, J.E.; Shan, S. Using Gaussian Boson Samplers to Approximate Gaussian Expectation Problems, 2025, [arXiv:quant-ph/2502.19336].
  2. Andersen, J.E.; Shan, S. Estimating the Percentage of GBS Advantage in Gaussian Expectation Problems, 2025, [arXiv:quant-ph/2502.19362].
  3. Andersen, J.E.; Shan, S. Using Gaussian Boson Samplers to Approximate Gaussian Expectation Problems, 2025, [arXiv:quant-ph/2502.19336].
  4. Hamilton, C.S.; Kruse, R.; Sansoni, L.; Barkhofen, S.; Silberhorn, C.; Jex, I. Gaussian Boson Sampling. Phys. Rev. Lett. 2017, 119, 170501.
  5. Kruse, R.; Hamilton, C.S.; Sansoni, L.; Barkhofen, S.; Silberhorn, C.; Jex, I. Detailed study of Gaussian Boson Sampling. Phys. Rev. A 2019, 100, 032326.
  6. Aaronson, S.; Arkhipov, A. The Computational Complexity of Linear Optics. In Proceedings of the STOC ’11, 2011, pp. 333–342.
  7. Deshpande, A.; Mehta, A.; Vincent, T.; Quesada, N.; Hinsche, M.; Ioannou, M.; Madsen, L.; Lavoie, J.; Qi, H.; Eisert, J.; et al. Quantum computational advantage via high-dimensional Gaussian boson sampling. Science Advances 2022, 8, eabi7894. [CrossRef]
  8. Bulmer, J.F.F.; Bell, B.A.; Chadwick, R.S.; Jones, A.E.; Moise, D.; Rigazzi, A.; Thorbecke, J.; Haus, U.U.; Vaerenbergh, T.V.; Patel, R.B.; et al. The boundary for quantum advantage in Gaussian Boson Sampling. Science Advances 2022, 8, eabl9236. [CrossRef]
  9. Björklund, A.; Gupt, B.; Quesada, N. A faster hafnian formula for complex matrices and its benchmarking on a supercomputer. Journal of Experimental Algorithmics 2019, 24, 1–19. [CrossRef]
  10. Gupt, B.; Izaac, J.; Quesada, N. The Walrus: a library for hafnians, Hermite polynomials and Gaussian Boson Sampling. J. Open Source Software 2019, 4, 1705.
  11. Zhong, H.S.; Wang, H.; Deng, Y.H.; Chen, M.C.; Peng, L.C.; Luo, Y.H.; Qin, J.; Wu, D.; Gong, S.Q.; Su, H.; et al. Quantum computational advantage using photons. Science 2020, 370, 1460–1463. [CrossRef]
  12. Madsen, L.S.; Laudenbach, F.; Falamarzi Askarani, M.; Rortais, F.; Vincent, T.; Bulmer, J.F.F.; Miatto, F.M.; Neuhaus, L.; Helt, L.G.; Collins, M.J.; et al. Quantum computational advantage with a programmable photonic processor. Nature 2022, 606, 75–81. [CrossRef]
  13. Quesada, N.; Helt, L.G.; Izaac, J.; Arrazola, J.M.; Shahrokhshahi, R.; Myers, C.R.; Sabapathy, K.K. Simulating realistic non-Gaussian state preparation. Physical Review A 2019, 100, 022341, [arXiv:quant-ph/1810.03706]. [CrossRef]
  14. Quesada, N.; Arrazola, J.M.; Killoran, N. Gaussian boson sampling using threshold detectors. Physical Review A 2018, 98, 062322, [arXiv:quant-ph/1807.01639]. [CrossRef]
  15. Huh, J.; Guerreschi, G.G.; Peropadre, B.; McClean, J.R.; Aspuru-Guzik, A. Boson sampling for molecular vibronic spectra. Nature Photonics 2015, 9, 615–620. [CrossRef]
  16. Huh, J.; Yung, M.H. Vibronic boson sampling: Generalized Gaussian boson sampling for molecular vibronic spectra at finite temperature. Scientific Reports 2017, 7, 7462. [CrossRef]
  17. Banchi, L.; Fingerhuth, M.; Babej, T.; Ing, C.; Arrazola, J.M. Molecular docking with Gaussian Boson Sampling. Science Advances 2020, 6, eaax1950. [CrossRef]
  18. Arrazola, J.M.; Bromley, T.R. Using Gaussian boson sampling to find dense subgraphs. Physical Review Letters 2018, 121, 030503. [CrossRef]
  19. Brádler, K.; Dallaire-Demers, P.L.; Rebentrost, P.; Su, D.; Weedbrook, C. Gaussian boson sampling for perfect matchings of arbitrary graphs. Physical Review A 2018, 98, 032310. [CrossRef]
  20. Brádler, K.; Friedland, S.; Izaac, J.; Killoran, N.; Su, D. Graph isomorphism and gaussian boson sampling. Special Matrices 2021, 9, 166–196. [CrossRef]
  21. Schuld, M.; Brádler, K.; Israel, R.; Su, D.; Gupt, B. A quantum hardware-induced graph kernel based on GBS, 2019, [arXiv:quant-ph/1905.12646].
  22. Jahangiri, S.; Arrazola, J.M.; Quesada, N.; Killoran, N. Point processes with Gaussian boson sampling. Physical Review E 2020, 101, 022134. [CrossRef]
  23. Sparrow, C.; Martín-López, E.; Maraviglia, N.; Neville, A.; Harrold, C.; Carolan, J.; Joglekar, Y.N.; Hashimoto, T.; Matsuda, N.; O’Brien, J.L.; et al. Simulating the vibrational quantum dynamics of molecules using photonics. Nature 2018, 557, 660–667. [CrossRef]
  24. He, Y.; Ding, X.; Su, Z.E.; Huang, H.L.; Qin, J.; Wang, C.; Unsleber, S.; Chen, C.; Wang, H.; He, Y.M.; et al. Time-bin-encoded boson sampling with a single-photon device. Physical Review Letters 2017, 118, 190501. [CrossRef]
  25. Motes, K.R.; Gilchrist, A.; Dowling, J.P.; Rohde, P.P. Scalable boson sampling with time-bin encoding using a loop-based architecture. Physical Review Letters 2014, 113, 120501. [CrossRef]
  26. Wick, G.C. The evaluation of the collision matrix. Physical Review 1950, 80, 268–272. [CrossRef]
  27. Kocharovsky, V.V.; Kocharovsky, V.V.; Tarasov, S.V. The hafnian master theorem. Linear Algebra and its Applications 2022, 651, 144–161. [CrossRef]
  28. Barvinok, A. Approximating Permanents and Hafnians. Discrete Analysis 2017, 2017, 1–34. [CrossRef]
  29. Barvinok, A. Polynomial time algorithms to approximate permanents and mixed discriminants within a simply exponential factor. Random Structures & Algorithms 1999, 14, 29–61. [CrossRef]
  30. Robert, C.P.; Casella, G. Monte Carlo Statistical Methods; Springer, 1999. [CrossRef]
  31. Rubinstein, R.Y.; Kroese, D.P. Simulation and the Monte Carlo Method, 3 ed.; Wiley, 2016. [CrossRef]
  32. Owen, A.B. Monte Carlo Theory, Methods and Examples. https://artowen.su.domains/mc/, 2013. Accessed 2025-02-28.
  33. Owen, A.B. Practical Quasi-Monte Carlo Integration. https://artowen.su.domains/mc/practicalqmc.pdf, 2023. Accessed 2025-02-28.
  34. Madsen, L.S.; Laudenbach, F.; Askarani, M.F.; Rortais, F.; Vincent, T.; Bulmer, J.F.F.; Miatto, F.M.; Neuhaus, L.; Helt, L.G.; Collins, M.J.; et al. Quantum computational advantage with a programmable photonic processor. Nature 2022, 606, 75–81. [CrossRef]
  35. Zhong, H.S.; Wang, H.; Deng, Y.H.; Chen, M.C.; Peng, L.C.; Luo, Y.H.; Qin, J.; Wu, D.; Ding, X.; Hu, Y.; et al. Quantum computational advantage using photons. Science 2020, 370, 1460–1463. [CrossRef]
  36. Deutsch, F. Best Approximation in Inner Product Spaces; Springer, 2001.
  37. Davis, C.; Kahan, W.M. The Rotation of Eigenvectors by a Perturbation. SIAM Journal on Numerical Analysis 1970, 7, 1–46. [CrossRef]
  38. Wedin, P.Å. Perturbation bounds in connection with singular value decomposition. BIT Numerical Mathematics 1972, 12, 99–111. [CrossRef]
  39. van der Vaart, A.W. Asymptotic Statistics; Cambridge Univ. Press, 1998.
  40. Billingsley, P. Probability and Measure, 3rd ed.; Wiley, 1995.
1
When all Haf ( B I ) 0 the convention is trivial; otherwise this corresponds to the usual loop-hafnian or sign–augmented readout.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated