1. Introduction
Economic models are typically defined by systems of equilibrium restrictions, laws of motion, and expectation operators. Solving such models—whether as decentralised equilibria or as planner problems—amounts to finding zeros of a nonlinear map (or fixed points of an induced operator) in a high-dimensional function space [
1,
2]. A large numerical toolbox has therefore developed for DSGE computation, including projection and collocation methods, time iteration, policy iteration, and Newton or quasi-Newton solvers [
1,
2,
3,
4]. Despite this maturity, solver performance claims remain difficult to compare across papers and codebases, because they are often made in different residual representations and different norms.
The practical issue is that “convergence” is a geometric statement. A residual trace depends on which equilibrium conditions are measured, how they are discretised, and how the resulting components are scaled and aggregated. Changing variable normalisation, equation scaling, or aggregation weights can change solver rankings without changing the underlying economic fixed point [
1]. This motivates a disciplined measurement layer: before comparing solvers, one must freeze the residual geometry in which deviation from equilibrium is quantified.
This paper introduces such a frozen geometry through a DSFL (Deterministic Statistical Feedback Law) residual calculus. The central object is a quadratic
residual of sameness , where
e is a declared defect vector (equilibrium violations in a chosen representation) and
is a declared symmetric positive definite (SPD) ruler that aggregates, scales, and types those violations. Once
are fixed, convergence becomes a typed statement in a single normed space: a numerical solver is evaluated by the defect dynamics it induces in the frozen norm, and solver comparisons become reproducible because all residual traces are measured in the same geometry [
5,
6].
1.1. Problem Statement: Solver Comparisons Are not Typed
A practical gap motivates the paper. DSGE solvers are commonly compared by wall-clock time, ad hoc Euler-error aggregations, unreported scaling conventions, or tolerance thresholds that depend on the norm used for stopping. But if two implementations use different residual definitions or different implicit rulers, then their “errors” are not commensurate objects. A mathematically honest comparison therefore requires a frozen record: one must declare (i) the defect representation e (which equilibrium conditions are measured and how) and (ii) the ruler W (which defines the norm in which contraction and convergence are claimed). This paper adopts that discipline as the core typing requirement.
1.2. Scope and Standing Assumptions
Scope.
We do not propose new DSGE algorithms. Instead, we provide a solver-agnostic certificate layer that attaches to any solver once the defect representation and ruler are declared. The focus is on local (or tail) convergence diagnostics in a frozen norm, including forced variants that tolerate numerical noise and inexact solves. We illustrate the framework on a standard small New Keynesian (NK) model solved by several solver families.
Typing data.
The frozen record consists of: a defect map with defect , a neutral convention (optional projection to a neutral-free subspace), and an SPD ruler W on inducing and . All claims in the paper are explicitly typed by this data.
Regularity.
For the local linearisation arguments and Newton-type certificates we assume Fréchet differentiability of
F in a neighbourhood of the solution and a local chart/identifiability condition (neutral-free invertibility), as recorded in Assumption A1. These assumptions are standard in nonlinear operator equations and root-finding theory [
3,
7,
8,
9].
1.3. What Is Known and the Gap
The mathematical language for contraction and decay in fixed normed geometries is classical in semigroup theory and numerical analysis [
5,
6]. The DSGE numerical toolbox (projection/collocation, time iteration, policy iteration, Newton and quasi-Newton methods) is also well developed [
1,
2,
3,
4]. What is less standard is a
single measurement geometry that makes solver performance claims comparable across implementations and exposes transient pathologies caused by non-normal update operators. Eigenvalue-based diagnostics can miss substantial finite-iteration amplification when updates are non-normal, even when they are asymptotically stable [
10]. This paper supplies a frozen-geometry certificate layer that targets precisely these comparability and diagnosis failures.
1.4. Approach and Method Overview
Our approach has three steps.
- (i)
Frozen residual geometry.
We fix the equilibrium defect map and declare an SPD ruler W, producing the quadratic residual . Neutral conventions (redundant equations, normalisations) are handled explicitly by projection or augmentation.
- (ii)
Solvers as induced defect dynamics.
Each solver step induces defect dynamics on the frozen defect space. Contraction in the frozen norm yields explicit residual envelopes; forcing terms produce robust tail envelopes and principled stopping rules.
- (iii)
Spectral/Gram certificates and non-normal diagnostics.
For linear or linearised updates, the contraction factor
is encoded by a Gram operator
. This yields a solver-agnostic rate score
and companion diagnostics that detect transient amplification when updates are non-normal [
10,
11].
1.5. Main Contributions and Results
We provide four contributions.
- C1
Typed residual geometry for equilibrium computation. We formalise a frozen record for DSGE equilibrium problems: defect map, neutral convention, and SPD ruler W that defines a single quadratic residual used consistently across solvers.
- C2
Contraction and forced-contraction certificates. We prove that frozen-norm contraction of the induced defect update implies explicit geometric residual envelopes, and we extend this to forced envelopes that tolerate numerical noise and inexact solves, yielding principled stopping rules [
5,
6,
9].
- C3
Gram-spectrum solver certificates. For linear or linearised updates we introduce the Gram operator
and show that its top eigenvalue equals
, providing a computable, solver-agnostic contraction certificate and a comparable rate score
[
11].
- C4
Non-normal transient diagnostics. We show how the same frozen-geometry machinery separates tail contraction from transient amplification, explaining why eigenvalue-only diagnostics can fail to predict practical behaviour in non-normal regimes [
10].
1.6. Evidence and Instantiations
We instantiate the framework on a standard small New Keynesian model with a fixed discretisation and compare three solver families (time iteration, policy iteration, Newton/quasi-Newton) by reporting residual traces in the same frozen ruler, certificate rates, and transient amplification indicators. The purpose is a computational audit: to demonstrate that typed certificate numbers predict observed convergence rankings and remain interpretable under solver-dependent transients.
1.7. Roadmap (How to Read the Paper)
The paper is organised as a certificate pipeline: we first freeze the measurement geometry, then identify the induced defect dynamics of solvers, then prove contraction/forcing certificates and spectral diagnostics, and finally instantiate the framework on a standard DSGE testbed.
Part I: Frozen record (typing data).
Section 2 fixes the equilibrium defect map
, the defect
, and the SPD ruler
that defines the quadratic residual
. Optional neutral conventions (redundant equations, normalisations) are handled by an explicit projection or augmentation. This section supplies the core methodological requirement: solver claims are meaningful only after the residual representation and its norm have been declared and frozen.
Part II: Solvers as induced defect dynamics.
Section 3 rewrites standard DSGE solvers as update maps
and expresses their performance in the frozen geometry via the induced defect evolution
(exactly in linear settings and locally after linearisation). The output is the object to be certified: the induced defect update operator (or its local Jacobian) in the frozen
W-norm.
Part III: Certificate spine (proved theorems).
Section 4 proves the analytical results used throughout: (i) frozen-norm contraction implies explicit residual envelopes; (ii) forcing/inexactness yields tail-robust envelopes and principled stopping rules; (iii) the Gram operator
encodes
and provides a solver-agnostic spectral certificate; and (iv) non-normality is diagnosed through transient amplification indicators that eigenvalues alone can miss. These results are theorem-level and are reusable: once a solver is expressed as a defect update in the frozen geometry, the envelopes follow.
Part IV: Comparable solver metrics.
Section 5 defines the solver-agnostic comparison score
and clarifies what it measures (typed contraction per iteration in the frozen geometry) and what it does not measure (welfare, wall-clock time, or global convergence). Companion diagnostics are recorded for transient amplification.
Part V: Numerical instantiation.
Section 6 instantiates the framework on a small New Keynesian model with a fixed discretisation. The defect evaluator, ruler
W, and initialisation are held fixed while solver families are varied (time iteration, policy iteration, Newton/quasi-Newton). The section reports residual traces in the frozen norm together with certificate numbers (rate and transient indicators) to demonstrate solver-comparable, reproducible diagnostics.
Part VI: Discussion and outlook.
Section 8 discusses what the certificate layer adds to DSGE computation (reproducibility by typing, diagnosis of non-normal transients, and stopping rules under forcing), and records explicit limitations (locality, ruler design, and high-dimensional approximation issues). The conclusion summarises the certificate pipeline and lists theorem-shaped upgrades (welfare-aligned rulers, uniform-in-refinement stability, and certified forcing budgets).
2. Frozen Residual Geometry for DSGE Equilibrium Problems
This section fixes, once and for all, the representation of equilibrium, the residual (defect) map, and the norm in which deviation from equilibrium is measured. The methodological point is classical but often left implicit: solver statements such as “converges”, “is accurate”, or “is faster” are not invariant facts — they are
typed claims that depend on (i) what is called the unknown and (ii) how violation of equilibrium is aggregated and scaled. Accordingly, we treat the residual map and the ruler as theorem data. This mirrors the role of a chosen norm in semigroup stability and numerical analysis: convergence rates, error bounds, and conditioning are only meaningful relative to a declared geometry [
5,
6,
12,
13].
2.1. Equilibrium as a Defect Equation
Let
be a Banach space of
candidate representations for a DSGE equilibrium. The choice of
is solver-dependent: for projection and perturbation methods,
z may be a vector of coefficients for policy functions; for time-iteration and endogenous-grid methods,
z may encode a grid representation of decision rules; for stacked-residual approaches,
z may already be a vector of discretised equilibrium conditions [
1,
14,
15,
16,
17].
Let
be a real Hilbert space (finite-dimensional in most computational implementations) equipped with inner product
. We represent the equilibrium conditions as a (generally nonlinear)
defect map
The vector
stacks all equilibrium restrictions in the chosen representation: Euler equations, pricing relations, implementability constraints, policy rules, resource constraints, and (if present) auxiliary conditions used to pin down expectations or enforce laws of motion. For any candidate
we define the
defect
which measures deviation from equilibrium in the frozen representation.
Differential structure.
Many convergence theorems and certificates require local smoothness. We therefore record a standing regularity class.
Assumption A1
(Local differentiability of the defect map). There is an open set such that F is Fréchet differentiable on , with derivative , and is locally Lipschitz on .
Assumption A1 is standard in Newton–Kantorovich analyses and in local conditioning statements for nonlinear systems [
3,
7,
8,
9]. In many DSGE constructions
F is composed of smooth primitives (utility, production, transition densities) and smooth expectation operators under dominated convergence; nondifferentiabilities can be handled by smoothing or by switching to semismooth frameworks, but we do not pursue that here [
18].
2.2. Residual and Ruler
To compare solvers, we must choose
how to aggregate the equilibrium violations in
. We therefore fix a symmetric positive definite (SPD) linear operator
and define the associated weighted inner product and norm
The
quadratic residual is
This
R is the single scalar object to which we will attach accuracy statements and solver certificates. The role of
W is analogous to preconditioning: it fixes which components of the equilibrium system matter more and how units are reconciled [
3,
4].
Interpretation of W (typical choices).
- (i)
Steady-state scaling. Let
D be a diagonal scaling matrix built from steady-state magnitudes of each equation. Then
yields a dimensionless residual, mitigating dominance by large-unit conditions [
1].
- (ii)
Variance weighting. If
stacks moments or forecast errors,
W can be chosen as an inverse covariance proxy (or block-diagonal approximation), turning
R into a generalized least squares objective [
19].
- (iii)
Welfare-based weighting. If some equations correspond to implementability or planner optimality,
W can encode marginal welfare weights or Lagrange-multiplier magnitudes, aligning residual decrease with welfare loss to first order [
1].
Frozen-Geometry Principle
Throughout, W is treated as theorem data. All convergence claims are to be read as statements about the decay of in this fixed norm. Changing W changes what is being measured, and therefore changes what “convergence” means.
2.2.1. Norm Equivalence Bookkeeping
If a solver is analysed in one norm but reported in another, explicit equivalence constants are required. In finite dimensions, any two norms are equivalent, but constants matter for rates and stopping rules [
11,
12]. For SPD
W, we have the sharp comparison with the ambient norm
induced by
:
Lemma 1
(Weighted norm equivalence).
Let with inner product and let with eigenvalues . Then for all ,
and hence and
Proof. Diagonalise
with orthogonal
Q and
. Then
. Bounding
by
and
gives (
6). □
Lemma 1 is the precise mechanism by which we will allow norm changes: they are admissible only through declared constants, never silently.
2.2.2. Residual Geometry and First-Order Stationarity
Although the equilibrium problem is , many solvers can be cast as residual-reduction schemes. We therefore record the derivative of R and the corresponding normal equations.
Lemma 2
(Gradient of the quadratic residual).
Assume with Euclidean inner product and F is differentiable at z. Let . Then
Moreover, any stationary point of R satisfies the weighted normal equation
Proof. Differentiate and use the chain rule: . □
Equation (
8) clarifies the difference between
solving and
minimising: local minima of
R can exist with nonzero defect if
lies in a near-null direction of
. This is precisely why the identifiability and neutral-direction discussion in subsec:neutrals matters, and why we keep
W explicit (it changes both conditioning and stationarity).
2.3. Neutral Directions and Identifiability (Optional)
Some DSGE representations contain neutral directions: changes in z that leave the economic allocation unchanged or change it only by a normalisation (e.g. the numéraire), or redundancy in stacked equilibrium conditions (e.g. one market-clearing equation is implied by the others under Walras’ law). If neutrals are not removed, a solver may appear to “stall” because it is moving along an equivalence class rather than reducing economically meaningful defect.
2.3.1. Quotient Formulation
Let
be a closed subspace of neutral defect directions (redundant equations), and let
be the orthogonal projector onto
. Define the
effective defect
and the corresponding neutral-free residual
All contraction and certification statements can then be stated for
without ambiguity. This is the Hilbert-space way to “drop” redundant equations while keeping a consistent ruler [
12].
2.3.2. Parameter Non-Identification in z
Neutrality can also arise in the
unknown representation: distinct
z encode the same allocation due to gauge-like choices (e.g. scaling/numéraire, normalisations of value functions, or redundant coefficients in an approximation basis). A clean mathematical treatment is to declare an equivalence relation
(same economic object) and work on the quotient
, but in computation it is more common to fix a gauge (normalisation) or to add a pinning condition. Formally, one may add a
gauge-fixing map
and solve the augmented system
with an augmented ruler
. This converts non-identification into an explicit design choice (the gauge map and its weight), making solver comparisons well-typed.
Why the main results assume neutral-free geometry.
To keep the exposition focused, the core results of the paper are stated under the simplifying assumption that the declared representation has already removed neutrals (either by projection or by augmentation). When neutrals are present, every statement below is read with F replaced by (or by ), and with the corresponding ruler (or ). Nothing essential changes; what changes is that one must explicitly state the projection/gauge as part of the frozen record.
Forward reference (use in solver theorems).
Later, when we state solver comparison results, we will treat each algorithm as producing a sequence
and we will compare them through the same frozen scalar
(or
). This permits theorems of the following general form: under local regularity and identifiability, a solver decreases the frozen residual at a provable rate, and stopping rules become meaningful because the ruler has been declared in advance [
4,
8,
9].
3. DSGE Solvers as Update Maps
This section rewrites standard DSGE solution methods in a single algebraic language:
update maps acting on a declared
defect (residual) representation. The objective is not to propose new algorithms, but to identify the operator (or local Jacobian) that each method induces, because that operator is what determines contraction, non-normal transients, and solver comparability in a frozen ruler. The presentation is deliberately typed: we distinguish the
economic model data (equilibrium conditions) from the
numerical method (update rule) and from the
measurement geometry (the SPD ruler
W). This is the discipline that allows solver comparisons to be meaningful across codes and implementations [
1,
2,
5,
6].
3.1. Equilibrium Problems as Operator Equations
A DSGE model solution can be posed as an operator equation on a Banach/Hilbert space of candidate objects (e.g. policy functions, decision rules, or stacked time-
t objects such as
). Let
be the model’s
solution-object space and let
be the
equilibrium operator, mapping a candidate solution object
to a defect vector
in a defect space
. Typical choices include Euler-equation defects, intratemporal first-order conditions, implementability constraints, and resource constraints; for log-linear models
F is linear, while for global nonlinear methods
F is nonlinear and evaluated pointwise on a grid or basis expansion [
1,
2].
Defect space and frozen ruler.
We assume throughout that the defect space
is a real Hilbert space with inner product
and that the analyst declares a symmetric positive definite ruler
W on
. The induced norm is
This is the
frozen measurement geometry: all contraction and rate statements in this section are in
. The choice of
W is part of the theorem data; changing
W changes what “fast” or “stable” means [
5,
6].
Equilibrium.
An equilibrium solution is a zero of the equilibrium operator:
We emphasise that solvers act on
z, but our certificates (later) will be expressed in terms of the induced dynamics of the defect
.
3.2. Solver Steps as Maps on Defect Space
Let
denote a single solver step:
This induces a
defect update map via the composition
In general
T depends on the representation of
and on how
F is evaluated, but once the frozen record (defect representation and ruler) is fixed, the performance question becomes a question about the dynamics
either exactly (in linear problems where
F and
S reduce to affine maps) or locally (after linearisation around
).
Why defect maps (rather than z maps).
Two implementations can use different parameterisations of z (different bases, different normalisations, different state scalings) while representing the same economics. Defects provide a canonical comparison target: if the defect representation is fixed, then all solvers are judged in the same space with the same ruler W. This is the typed notion of comparability underlying the certificate layer.
3.3. Local linearisation and Jacobians
To expose solver dynamics near an equilibrium, we linearise the defect update. Assume F and S are Fréchet differentiable in a neighbourhood of .
Definition 1
(Defect Jacobian of a solver).
Let satisfy and define as a fixed point of the solver step (i.e. ). Thedefect Jacobian
induced by S at is the bounded linear operator satisfying
whenever is invertible as a map (local chart condition).
The slightly abstract form (
18) separates three components: (i) the model map
F, (ii) the solver step
S, and (iii) the coordinate choice linking
and
. In many DSGE applications one works directly with
e-based updates, in which case
in the usual sense.
Remark 1
(When is not invertible). In models with neutral directions (e.g. indeterminacy, redundant equations, normalisation conventions), may fail to be invertible. The correct typed object is then the Jacobian on the quotient or on the projected subspace (“neutral-free” defect space), obtained by choosing neutrals and a projector . We postpone this neutral discipline to the frozen-record section; in this section we work in the simplest setting where the local chart is available.
3.4. Contraction in a Frozen SPD Ruler
All solver comparisons here are statements in the frozen
W-geometry. For a bounded linear map
, define the induced operator norm
where
is the Euclidean/operator norm in coordinates. The equivalence (
19) is standard for SPD-weighted norms.
Theorem 1
(Local
W-contraction implies a certified residual envelope).
Assume that near the defect update satisfies
for some and some . Then for all iterates that remain in the ball we have the geometric decay
If moreover T admits an additive perturbation model (tolerances, numerical noise)
then
Proof. Under (
20),
, hence by induction
and (
21) follows by squaring. For (
22), iterate the inequality:
. □
Corollary 1
(Linear case: Jacobian bound is sufficient).
If T is linear, , and , then (21) holds globally with .
Proof. For linear
, the Lipschitz property (
20) holds with
for all
e by definition of the operator norm. □
Interpretation.
Theorem 1 formalises the point that solver performance is, at root, a statement about contractivity of the induced update in the declared ruler. Different solver families correspond to different induced
J (and different higher-order terms away from the equilibrium), hence different
q in (
20). Later sections will turn this into computable certificates.
3.5. Canonical Solver Families and Their Induced Defect Maps
We now record the induced defect maps for standard solver families. The purpose is not exhaustive algorithmics, but a clean identification of the T (or J) each method implies.
3.5.1. Fixed-Point (Time) Iteration
Many global methods express equilibrium as a fixed point
(e.g. policy function operators). Time iteration applies the forward map:
A common defect is
, so
. Then the induced defect map is
and locally (if
is differentiable) the Jacobian at a fixed point
is
whenever
is invertible. In the common case where one tracks the
state error rather than the defect
, the local dynamics reduce to
. Either way, the spectral and non-normal properties of
determine local speed [
1,
2].
What changes across implementations.
Two codes can implement the same but scale variables differently, or aggregate equation errors differently. That changes the effective ruler W on the defect space and can change even when the economics are the same. This is exactly the motivation for freezing W before comparing solvers.
3.5.2. Policy Iteration / Howard Improvement (Partial Inversion)
Policy iteration (Howard improvement) alternates between
policy evaluation (solve linear(ised) conditions under a fixed policy guess) and
policy improvement (update the policy by re-optimisation under the implied value/marginals). In deterministic dynamic programming, policy iteration is known to accelerate value iteration by effectively applying a partial inverse of
on the Bellman residual; analogous ideas appear in DSGE policy-function solvers when intratemporal blocks can be solved (exactly or approximately) conditional on intertemporal objects [
1,
20].
Abstractly, write the defect as
and split variables as
where
u is the block to be “evaluated” and
v the block to be “improved”. A single policy-iteration step can be written as
where
represents the evaluation solve (often a linear system solve or a conditional expectation solve) and
the improvement map. The induced defect Jacobian takes a block form that includes an explicit
solve (inverse) in the
u-block, which is precisely why policy iteration can contract faster than a pure forward iteration. Because the exact block formulas depend on the model representation (Euler equation form, state reduction, basis choice), we record the general principle:
Policy iteration replaces part of the forward map by a (possibly approximate) inverse of the evaluation operator, hence the induced defect Jacobian is a preconditioned map with potentially much smaller than the forward iteration.
This preconditioning viewpoint will align policy iteration with Newton-type methods below.
3.5.3. Newton and Quasi-Newton Steps as Preconditioned Defect Updates
Suppose the equilibrium is posed directly as a root-finding problem
with
F differentiable. Newton’s method updates
To see the induced defect update, linearise around
where
and
is invertible. Let
. Then
and
so Newton is locally quadratic in state error.
In defect coordinates, the update is even simpler: to first order,
so the defect Jacobian is
at the solution. Thus Newton’s method is, in the frozen geometry,
infinitesimally contractive at the fixed point.
Quasi-Newton replaces
by an approximate inverse
:
Linearising at
gives the local state-error iteration
and the induced
defect Jacobian becomes
The message is clear: quasi-Newton is a
preconditioned defect iteration with defect map approximately
. Hence
is small precisely when the preconditioner
is a good right-inverse in the frozen geometry. This viewpoint is standard in numerical analysis [
3,
4] and will be used later to build a posteriori contraction checkers.
3.5.4. Projection and Galerkin Solvers (Basis-Induced Maps)
Global DSGE solvers often restrict the solution object to a finite-dimensional approximation space
(polynomials, splines, Smolyak grids, neural bases) and impose equilibrium by collocation or Galerkin conditions [
1]. In that case the solver step is a composition of (i) projection to
, (ii) a finite-dimensional update (Newton, time iteration, etc.), and (iii) lifting back to
. The induced defect map is then a
compressed update on the image of
. This is where refinement ladders enter naturally: increasing
K changes the representation, and any meaningful convergence claim must be made in a geometry that does not move with
K. We record this principle now and formalise it later.
3.6. A Theorem-Shaped Summary: Solver Families Differ by Induced J
The preceding subsections can be condensed into a single theorem-shaped statement emphasising what we will certify.
Theorem 2
(All standard solver families induce a local defect update operator).
Let encode the DSGE equilibrium conditions and let be a solver step (time iteration, policy iteration, Newton/quasi-Newton, or a projection-based variant). Assume F and S are Fréchet differentiable at a solution with and , and assume a local chart condition (invertibility on the neutral-free subspace). Then there exists a bounded linear operator such that, for sufficiently small,
Moreover:
- (i)
For forward (time) iteration, J is (conjugate to) the Jacobian of the forward policy operator.
- (ii)
For policy iteration, J contains a partial inverse (evaluation solve) and is therefore a preconditioned forward map.
- (iii)
For Newton, at the solution (quadratic local defect decay).
- (iv)
For quasi-Newton, for a preconditioner .
Proof. The existence of a local linearisation (
33) follows from Fréchet differentiability of
in defect coordinates (or from the chain rule in
z-coordinates under the chart condition), giving
with
and
. The method-specific characterisations are immediate from the standard first-order expansions of the corresponding
S (
Section 3.5.1,
Section 3.5.2 and
Section 3.5.3). □
Why this matters for the next section.
Theorem 2 is the bridge from “which solver do you use?” to “what contraction do you induce in the frozen ruler?”. Once the induced J is identified (or bounded), the certificate layer reduces performance claims to inequalities involving and its non-normal structure. That is where Gram-operator certificates and a posteriori checkers enter.
3.7. Pointers to Implementation Practice (Typed Diagnostics)
Two practical consequences follow immediately from the frozen-geometry viewpoint.
(1) Residual logging must be W-typed.
If a code logs “Euler errors” but rescales equations or aggregates them differently across runs, then the logged sequence is not comparable. A meaningful solver audit must log (or sufficient statistics to reconstruct it) and compute in the same ruler across solvers.
(2) Jacobian estimation can be certificate-driven.
Even when J is not formed explicitly, one can estimate bounds on from Jacobian blocks, from finite-difference probes, or from iterate data. These become the a posteriori “checkers” later; the present section identifies what object the checker must approximate: the induced defect update operator in the fro
4. Contraction Certificates in a Frozen Norm
This section provides the analytical spine needed for solver-agnostic certification. The guiding idea is simple: once the residual geometry
is frozen, a solver update induces an operator on defects, and
contractivity in the frozen norm yields explicit, checkable convergence envelopes. We develop (i) deterministic contraction certificates, (ii) forcing-robust envelopes that tolerate numerical noise and finite-precision stopping, and (iii) spectral/Gram diagnostics that turn solver behaviour into comparable numbers. The results are standard in numerical analysis and operator theory, but we state them in the frozen-geometry language so that they can be used as reusable certificate components [
4,
8,
9,
10,
12].
4.1. Setup: Defect Dynamics Induced by a Solver
Let
be the defect map and
the defect as in sec:frozen-geometry. Consider an iterative solver producing iterates
via
where
is the algorithm map (Newton, quasi-Newton, time-iteration, projection, policy iteration, etc.). The associated defect sequence is
and residual sequence is
To connect solver behaviour to contraction, we work directly with the
induced defect update. Whenever
F is locally invertible on the relevant manifold (or after quotienting out neutrals), we may represent defect evolution as
where
T is a (possibly state-dependent) defect update map and
collects forcing terms: finite-precision effects, inexact linear solves, stochastic quadrature noise, truncation from basis updates, or deliberate damping/regularisation that is not exactly representable as a pure defect map. When
T is state-dependent we interpret (
36) locally (near equilibrium) with uniform bounds on the relevant neighbourhood. This “defect-dynamics” viewpoint is the discrete-time analogue of stability analysis for dynamical systems and is the natural home for a posteriori certificates [
3,
8,
9].
4.2. Contractivity and Residual Envelopes
We first state the clean, deterministic contraction certificate in the frozen W-norm.
Theorem 3
(Frozen-norm contraction certificate).
Let be the frozen defect space. Assume that in a neighbourhood of the equilibrium defect the induced defect update satisfies
for some constant , and that the defect iterates remain in . Then the defect and residual sequences satisfy, for all ,
Proof. By (
37) and induction,
. Squaring yields the residual bound. □
What the theorem is (and is not).
Theorem 3 is deliberately “typed”: it certifies convergence
in the frozen norm, not in an arbitrary reporting scale. It does not claim global convergence; it claims that
if the solver stays in the contraction neighbourhood, then the residual decays at a certified geometric rate. This is the standard local-analytic structure behind Newton–Kantorovich and fixed-point convergence theorems, rewritten in residual language [
7,
8,
9].
4.2.1. Forcing-Robust Envelopes (Numerical Noise and Inexact Solves)
Practical solvers do not implement the exact map T. Linear systems are solved inexactly, quadrature is truncated, basis updates introduce discretisation error, and finite precision induces small perturbations. We therefore state a forcing-robust certificate that provides a stable envelope and a principled stopping rule.
Theorem 4
(Forced contraction envelope in the frozen norm).
Let and suppose the defect dynamics satisfy
Then for all ,
In particular, if for all k, then
Proof. Unroll the recursion (
39):
. If
, the geometric series gives (
41). □
Stopping rule implied by the certificate.
If an implementation can bound (e.g. via linear-solve residuals or quadrature error indicators), then Theorem 4 yields a principled stopping tolerance: once the observed defect reaches the plateau scale , further iterations cannot improve the residual beyond the forcing floor without reducing the forcing. This is the precise sense in which the certificate is tail-robust.
4.3. Gram Operator and Spectral Certificate
For linear or linearised updates, the contraction factor in the frozen norm is encoded by a weighted operator norm. This admits a spectral representation through a Gram operator.
4.3.1. Weighted Operator Norm
Let
be linear. Define the frozen (induced) operator norm
A one-step contraction certificate is precisely
. In matrix terms, if
,
, so
reducing the weighted norm to a Euclidean norm after a similarity transform [
11].
4.3.2. Gram Operator and Contraction Spectrum
Definition 2
(Gram operator in the frozen geometry).
Let be linear and . Define
Lemma 3
(Basic properties of
).
The operator is self-adjoint and positive semidefinite. Moreover,
and the singular values of are the square-roots of the eigenvalues of .
Proof. Write
so that
. Hence
is self-adjoint PSD and
. Using (
43) gives (
45). □
Lemma 3 shows that the Gram spectrum is a solver-agnostic contraction fingerprint: its top eigenvalue is the squared contraction factor, while the full spectrum diagnoses anisotropy (there may be directions that contract quickly and others that are nearly neutral).
4.3.3. Linearisation: Local Certificates from JACOBIANS
Near equilibrium, many solvers behave like a linear map on the defect. Suppose
near 0; then
implies local contraction. This is the standard linearisation principle for fixed points and Newton-like methods, expressed in the frozen geometry [
8,
9].
4.4. Mirror Implication and a Posteriori Diagnosis
We now connect time-domain evidence (observed residual decay) to operator structure. The “mirror” direction says: if a linear defect update exhibits uniform exponential decay, then its Gram spectrum must lie below that decay rate; conversely, Gram bounds limit all possible exponential decay rates.
Theorem 5
(Mirror implication: decay ⇒ Gram bound (linear case)).
Let be linear and let . Assume there exist constants and such that for all ,
Then
In particular, if (46) holds with then .
Proof. Set
in (
46) to get
for all
e, hence
by definition (
42). Then apply (
45). □
Theorem 6
(Spectral certificate bounds all admissible decay rates).
Let T be linear and . If for some one has for all , then necessarily
Proof. With , , hence . Use Lemma 3 to rewrite in Gram form. □
A posteriori diagnosis from data.
If an algorithm is run and yields observed ratios
then any upper bound
over a tail window
defines an empirical contraction candidate. Theorems 4–6 explain how to interpret this: a stable tail with
supports a
candidate certificate rate
, while deviations in early iterations are consistent with transient amplification (next subsection). Turning empirical
into a
provable requires bounding linearisation error and forcing terms, which is solver- and model-specific, but the logical interface is fixed by the frozen geometry.
4.5. Non-Normality and Transient Amplification
Eigenvalues alone do not control finite-time behaviour when
T is non-normal. This matters in DSGE practice because Newton, quasi-Newton, and damped time-iteration can generate highly non-normal updates, especially when the Jacobian is ill-conditioned or when damping/line search mixes directions [
3,
4].
4.5.1. Why Eigenvalues Are Insufficient
Let
denote the spectral radius. Even if
, the norms
can be large for moderate
k when
T is non-normal, producing transient growth before asymptotic decay. This is governed by pseudospectra and the numerical range, not only the eigenvalues [
10]. In particular, a solver may exhibit an early increase in
even though it is asymptotically contractive.
4.5.2. Gram Certificate Separates Transient and Tail
The Gram top eigenvalue equals and therefore controls the worst-case one-step amplification in the frozen norm, regardless of eigenstructure. Consequently:
If , then every step is contractive in the frozen norm and no transient growth is possible.
If but , transient growth is possible (and in general expected) even though the iterates may converge eventually.
This dichotomy is exactly what a certificate must record: transient episodes do not invalidate a tail certificate if the forced envelope in Theorem 4 can still be established on a tail window, but they do warn that the solver is operating in a non-normal regime where early stopping rules can be misleading [
10].
4.5.3. A Computable TRANSIENT Flag
For a linear(ised) update estimate (e.g. from secant information or Jacobian linearisation), compute . If , then there exists a direction in defect space that is amplified in one step in the frozen norm; this is a structural explanation for early residual growth. If while the tail ratios are , the interpretation is that the solver is close to neutral in some directions but contracts on average due to nonlinear effects and/or adaptive damping.
5. Solver-Agnostic Comparison Metrics
This section defines a
single solver-comparison metric that is meaningful only after the frozen record has been fixed: (i) the equilibrium defect representation
e, (ii) the neutral conventions (if any), and (iii) the SPD ruler
W on defect space. Once these typing data are declared, solvers differ only through their induced defect update maps
T (exactly in linear settings, locally after linearisation). Our objective is to extract from
T a scalar score that (a) has a certified interpretation as a contraction rate in the frozen geometry, (b) is comparable across solver families, and (c) exposes non-normal transient behaviour rather than hiding it behind eigenvalue-only diagnostics. The guiding principle is semigroup/numerical-stability orthodoxy: decay and stability are statements in a fixed normed geometry, and the correct scalar score is therefore a norm-induced contraction margin [
5,
6,
10].
5.1. The Frozen Geometry and the Induced Operator Norm
Let
be the defect space and
the declared ruler. The induced norm and residual are
For a bounded linear operator
, define the induced operator norm
The equality to a Euclidean operator norm in coordinates is standard for SPD-weighted norms and is what makes the metric computable.
Typing requirement (no moving goalposts).
If
W is changed across solver runs (by rescaling equations, renormalising variables, or changing aggregation rules), then
changes even when the underlying dynamics are identical. Thus (
51) is meaningful only under the frozen-ruler discipline.
5.2. Definition of the Solver-Agnostic Rate Score
Consider a defect update map T (either the exact defect map in a linear setting, or a local Jacobian J in a nonlinear setting). In either case, the contraction factor in the frozen geometry is the operator norm (or ).
Definition 3
(Algorithmic contraction rate in a frozen ruler).
Let be a bounded linear defect update operator in the frozen ruler W. Define thesolver-agnostic comparison metric
When T is a local linearisation (Jacobian) valid in a neighbourhood of the equilibrium, we interpret as alocal
certified rate.
Basic interpretation.
If then and one expects geometric decay of the residual in the frozen norm; larger means faster certified contraction per iteration. If then and the metric correctly signals that no contraction guarantee is available in that geometry (even if the method converges in some other norm or from some other initialisation).
Why the logarithm.
Taking converts multiplicative per-iteration contraction into an additive “rate” scale. If a solver has contraction factor , then after k iterations the envelope is . Thus is the natural discrete-time analogue of an exponential decay rate.
5.3. Certified Meaning: Residual Envelopes and Iteration Complexity
We now state the theorem that gives (
55) its operational meaning.
Theorem 7
(Solver-agnostic rate implies a certified residual envelope).
Let be linear and suppose . Then for the defect iteration ,
Moreover, to achieve it suffices that
Proof. By definition of the induced operator norm,
, hence
by induction. Squaring yields the residual bound. Since
, the residual envelope is
. Solving
gives (
54). □
Local nonlinear interpretation.
If the solver induces a nonlinear map
with a local Lipschitz constant
in
on a ball around the solution, then the same envelope holds as long as the iterates remain in that ball. This is the standard contractive-map logic in a fixed normed geometry [
5,
6].
5.4. Invariance and What “Solver-Agnostic” Actually Means
The metric (
55) is solver-agnostic in the following precise sense.
Proposition 1
(Invariance under
W-orthogonal coordinate changes).
Let be invertible and W-orthogonal in the sense that . Let be the same linear map expressed in the transformed coordinates. Then
Proof. Compute using (
51):
Since
, we have
orthogonal in the Euclidean norm, hence it preserves
and the norm is unchanged. □
What is not invariant.
If an implementation rescales equations or changes aggregation weights, it changes W itself. Then the metric changes, and it should change: the meaning of “small residual” changed. Thus solver-agnostic here does not mean “invariant under arbitrary reweighting”; it means comparable once the geometry is frozen.
5.5. Non-Normality: Why Eigenvalues Are not Enough
A common solver-comparison practice is to look at eigenvalues of a Jacobian. This is insufficient in non-normal settings: even if
, one can have
for intermediate
k, causing transient residual growth and practical instability. This phenomenon is classical in numerical linear algebra and is quantified by pseudospectra and the Kreiss constant [
10].
Definition 4
(Transient amplification index).
For a linear update T define the (finite-horizon) amplification factor in the frozen ruler
For observed iterates we use the empirical proxy .
Metric vs. diagnostics.
is a tail contraction score; is a transient diagnostic. Both are needed for solver auditing: a method can have a large (fast tail contraction) yet be practically poor because transient growth pushes it out of the local basin of contraction. This is why we report both quantities in the numerical study.
5.6. Forced and Inexact Variants (Tolerances as Budgets)
Real solvers use finite tolerances, inexact linear solves, and floating-point arithmetic. In the certificate language these appear as forcing terms. A simple model is
Let
.
Proposition 2
(Forced envelope in the frozen geometry).
If , then the forced iteration (58) satisfies
In particular, if , then .
Proof. Iterate and sum the resulting geometric series. □
Interpretation.
Inexactness does not invalidate the metric; it adds a budget term. This is one reason to prefer over wall-clock comparisons: it separates algorithmic contraction from numerical budgets that depend on hardware, linear algebra choices, and stopping policies.
5.7. What the Metric Does and Does not Measure
What it measures.
measures
certified contraction per iteration of the solver-induced defect update in the declared norm. It is therefore: (i) a reproducible scalar for comparing solver dynamics, (ii) interpretable as an iteration-complexity rate via (
54), and (iii) meaningful across solver families because it is defined on the common defect space.
What it does not measure.
It is not a welfare criterion and it is not a statement about economic plausibility. It also does not include implementation-dependent cost per iteration. A solver with smaller may still be faster in wall-clock time if its per-iteration cost is much lower. Conversely, a solver with large may be impractical if it requires expensive Jacobians or dense linear solves. The correct use is therefore diagnostic: tells you how good the update is in the frozen geometry, while separate engineering metrics tell you how expensive it is to apply.
Why this is the right abstraction for the paper.
The paper’s central claim is not “Newton is best” or “time iteration is best”; it is that solver comparisons become honest only when they are typed by a frozen residual geometry.
is the minimal scalar that carries that typing information and that connects directly to contraction inequalities in semigroup stability theory [
5,
6] and to non-normality diagnostics in numerical linear algebra [
10].
6. Numerical Study: A Small New Keynesian Model
This section instantiates the frozen-geometry, solver-agnostic certificate framework on a canonical
small New Keynesian (NK) model. The economic specification is held fixed throughout; only the solver family is varied. The purpose is a computational audit: we demonstrate that (i) residual trajectories are comparable only once the defect representation and ruler are frozen, (ii) contraction certificates computed in the same ruler predict the observed convergence ranking, and (iii) transient amplification due to non-normal update operators is detectable by explicit diagnostics. The NK baseline and its log-linear equilibrium conditions are standard in modern monetary theory [
21,
22,
23].
6.1. Model and Defect Representation
We work with a Markov solution representation on a low-dimensional state and evaluate equilibrium conditions by collocation on a fixed grid. The same defect evaluator and the same SPD ruler W are used for every solver. This enforces the paper’s typing discipline: performance comparisons refer to one declared measurement geometry rather than to solver-specific scaling conventions.
6.1.1. Equilibrium Conditions (Small NK block)
Let
denote the output gap,
inflation, and
the nominal interest rate. Let
be the natural-rate (demand) disturbance and
a cost-push shock. We adopt the standard three-equation NK block:
with parameters
. Equations (
60)–(62) can be derived as the log-linear approximation of a monopolistic-competition economy with nominal rigidities and a simple monetary-policy rule; we use them as a canonical testbed for solver comparison [
21,
22,
23].
Shock dynamics.
We model the exogenous state as
, with AR(1) dynamics
This choice yields a compact, widely used state representation and makes expectation evaluation explicit.
Markov solution.
We seek a stationary Markov policy
which solves (
60)–(62) when
and expectations are taken over the shock transition (
63).
6.1.2. Discretisation and Approximation Class
We discretise the state space and approximate expectations by quadrature, so that the equilibrium conditions become a finite system of nonlinear equations in a finite-dimensional approximation class.
State grid.
Fix a grid for on a bounded rectangle . All solvers use the same grid nodes and the same ordering.
Quadrature.
Fix a quadrature rule
approximating the Gaussian innovations
. Define the successor state map
and the quadrature expectation operator
Approximation class.
We parameterise the policy by coefficients
in a fixed basis
:
where the
are
-valued basis functions or, equivalently, three scalar bases with stacked coefficients. Examples include tensor Chebyshev polynomials, splines, or other standard projection spaces used in DSGE numerics [
1,
2].
Typing discipline.
The grid , quadrature rule, and basis are frozen throughout the numerical study. This prevents representation changes from contaminating solver comparisons.
6.1.3. Defect Vector and Frozen SPD Ruler
We now define the defect evaluator and the frozen ruler W.
Pointwise residuals.
For any state
define, using (
66),
These are the Euler/Phillips/policy-rule defects evaluated consistently for every solver.
Stacked defect vector.
Define the stacked defect vector in
:
An approximate equilibrium in the chosen approximation class is a root
.
Frozen ruler.
Fix an SPD ruler
. A transparent and reproducible choice is block-diagonal across states:
where
are state weights (e.g. quadrature/cell weights or stationary-density weights) and
are equation weights. The weights are part of the frozen record; changing them changes the meaning of “small defect”.
Scalar residual.
Define the scalar residual
Every solver will be compared by the evolution of
and by contraction certificates computed in this same
W.
6.2. Solvers Compared
We compare three solver families that cover the dominant paradigms in DSGE practice: fixed-point (time) iteration, policy iteration (Howard-type improvement / partial inversion), and Newton/quasi-Newton root finding [
1,
2,
3,
4]. Each solver generates a coefficient sequence
and hence a defect trajectory
.
6.2.1. Time Iteration (TI)
Time iteration treats expectations under the
previous policy and updates the current policy accordingly. Operationally, define the
frozen-continuation residuals by evaluating successor-state terms with
:
Then a TI update is defined by the least-squares solve
This produces a robust outer fixed-point method whose contraction properties depend on the composition of expectations, projection, and policy evaluation in the chosen basis/grid [
1].
6.2.2. Policy Iteration (PI) / Howard Improvement
Policy iteration accelerates TI by applying a partial inverse to an evaluation block. In dynamic programming, Howard improvement replaces repeated forward application of a contraction by an exact (or multi-sweep) policy evaluation solve, yielding faster convergence [
1,
20]. In the NK implementation we adopt a Howard-style variant: for a given outer iterate
, perform
evaluation sweeps of the frozen-continuation system (or its least-squares variant) before updating the continuation policy.
Formally, define an evaluation operator
that maps a current coefficient vector to one evaluation sweep under continuation
. Then PI corresponds to
with
recovering TI. The structural point for this paper is that PI induces a defect update with a smaller local contraction factor in the frozen geometry because
acts like a polynomial (and in the limit, a partial inverse) of the underlying evaluation operator.
6.2.3. Newton and Quasi-Newton (N/QN)
Newton’s method treats
as a root-finding problem:
When
and
is square and invertible, this reduces to the classical Newton step
. Quasi-Newton replaces
(or its normal-equation inverse) by an approximate inverse updated from iterates, such as Broyden-type updates [
3,
4].
Near a solution , Newton has a vanishing local defect Jacobian (quadratic local defect decay), while quasi-Newton behaves like a preconditioned defect iteration whose certified rate depends on the quality of the approximate inverse in the frozen ruler.
6.2.4. Common Initialisation and Stopping
To ensure fairness, all solvers share: (i) the same grid , quadrature, and basis, (ii) the same initial guess , (iii) the same residual tolerance in the frozen norm , and (iv) the same maximum iteration cap. Runtime is reported only secondarily: the primary comparison is in the typed residual geometry.
6.3. Certificates and Observed Behaviour
For each solver run we compute three classes of quantities: (i) the residual trajectory , (ii) contraction certificates (Jacobian-based when available, otherwise iterate-based), and (iii) transient amplification indicators. We now formalise these objects and record the theorem-level implications that justify their interpretation.
6.3.1. Residual Decay in the Frozen Geometry
The primary observable is the residual sequence
Because
W is fixed,
is comparable across solvers. We plot
against
k to visualise exponential envelopes and transient growth.
6.3.2. Jacobian-Based Contraction Certificates
Let
T denote the induced defect update map
and let
J be a local linearisation
near the equilibrium. We define the certified contraction factor
When
, it yields a local contraction guarantee and thus a predicted exponential envelope.
Theorem 8
(Certified local envelope from
).
Assume there exists such that for all the defect update satisfies with . Then any iterate sequence that remains in this ball satisfies
Proof. This is the discrete-time contraction envelope in a fixed norm; it follows by iterating
and squaring, as in standard semigroup/contraction arguments [
5,
6]. □
How J is obtained.
In implementations, J can be obtained either by differentiating the solver map (available for Newton-type methods through ) or by finite-difference probing of T on the defect space. We report the method used for each solver in the reproducibility block.
6.3.3. Iterate-Based a Posteriori Rate Estimates
When Jacobians are unavailable or expensive, we compute iterate-based certificates from residual ratios:
Define a tail contraction estimate after burn-in
:
If the method enters a locally linear regime,
approximates
from above. This is an empirical but typed object: it depends only on
and is comparable across solver families.
Proposition 3
(Observed tail decay implies a structural restriction). If for some and all one observes , then any local linearisation J consistent with the tail regime must satisfy .
Proof. If
in the tail regime and
on a set of tail iterates spanning a neighbourhood, then by definition of the induced operator norm
. This is the discrete-time analogue of “Lyapunov decay bounds the spectrum” in typed stability theory [
5,
6]. □
6.3.4. Transient Amplification and Non-Normality
A key motivation for certificate-based auditing is that eigenvalues can miss non-normal amplification. Even when
, one can have
for intermediate
k, producing transient residual growth that is relevant at finite iteration counts [
10].
Definition 5
(Empirical amplification factor).
Fix a horizon . The empirical amplification factor is
Large together with a small tail estimate is the operational signature of non-normal transient behaviour: the solver is asymptotically contractive in the frozen norm but may be practically sensitive to initialisation.
Optionally, when a Jacobian estimate
J is available we also compute a pseudospectral proxy via resolvent sampling in the
W-geometry (a finite-grid approximation to Kreiss-type quantities), following standard non-normal analysis [
10].
6.3.5. Expected Qualitative Ranking and What Is Being Tested
The test is not “which solver wins in general” but whether the typed certificate numbers predict the observed behaviour in a controlled setting. The qualitative expectation is:
- (i)
Newton (when well-initialised) exhibits the fastest local decay (quadratic in state error; vanishing linear defect term), hence the smallest certified contraction factor [
3,
4].
- (ii)
Policy iteration accelerates time iteration by partial inversion, typically reducing the induced
relative to TI [
1,
20].
- (iii)
Time iteration is often slowest and most prone to transient amplification because expectation and projection compose into a non-selfadjoint update in the chosen approximation space [
1,
10].
The numerical study reports to test these predictions while holding the economic problem and geometry fixed.
6.4. Reproducibility
This subsection lists the minimal artefacts needed to reproduce the numerical study as a typed audit. The aim is that an independent reader can reconstruct the defect sequence and compute the same certificate numbers in the same ruler.
6.4.1. Minimal Computational Artefacts
A reproducible package consists of:
- (i)
Model file. Equations (
60)–(62), parameters
, and shock law (
63).
- (ii)
Grid and quadrature declaration. Grid bounds and nodes
; quadrature nodes/weights for (
66); seeds if Monte Carlo quadrature is used.
- (iii)
Approximation specification. Basis functions , coefficient layout, and evaluation routines for .
- (iv)
Defect evaluator. Implementation of (
68)–(
71) returning
and
.
- (v)
Frozen ruler. An explicit representation of
W as in (
72) (or its Cholesky factor), including equation weights and state weights.
- (vi)
Solver implementations. TI update (
77); PI variant (
78) with declared
m; Newton/quasi-Newton update (
79) with declared linear-solve tolerances and stopping conditions.
- (vii)
Certificate checker. Computation of
,
,
, and
from (
80), (
83)–(
85), and (optionally) computation of
via (
81) when a Jacobian estimate is available.
6.4.2. Reporting Protocol
To make results comparable across machines and codebases, we report:
- (i)
the frozen typing data and the defect definition;
- (ii)
the shared initialisation and stopping rule in ;
- (iii)
for each solver: iterations to tolerance, the trajectory , and the certificate numbers and (and if computed);
- (iv)
runtime as a secondary quantity, clearly separated from the typed contraction scores.
This reporting protocol enforces the separation between (a) economics, (b) geometry, and (c) solver dynamics, which is the central methodological claim of the paper [
1,
2].
7. Proposision
This paper advocates a disciplined answer to a recurring problem in computational macroeconomics: solver performance claims are often not comparable because they are made in different residual representations and different norms. By freezing the defect representation and the ruler W, and by expressing solver behaviour as contractive (or forced) dynamics on a common defect space, we obtain certificates that are (i) mathematically typed, (ii) reproducible across codes, and (iii) informative about both asymptotic rates and transient pathologies. In this section we clarify what this buys, what it does not buy, and which theoretical and empirical upgrades are most valuable next.
7.1. What the Framework Changes in Practice: Reproducibility by Typing
Typed claims versus untyped claims.
A statement such as “solver A converges faster than solver B” is only meaningful after fixing: (i) what is called the defect (Euler errors, policy-function fixed-point error, stacked collocation residuals, etc.), (ii) how those components are aggregated and scaled (the ruler
W), and (iii) which neutral conventions are modded out (redundant equations, numéraire normalisations, identifiability fixes). This paper makes that typing explicit and treats it as theorem data, in line with stability theory in fixed normed geometries [
5,
6,
12].
Reproducibility as a mathematical property.
Once
are frozen, the reported objects
are invariant to many implementation details: different parameterisations of
z that lead to the same defect vector, different internal scalings in the solver that do not change the frozen defect evaluator, and different stopping heuristics. This is the same reason preconditioner quality is meaningfully reported only relative to a declared norm: without the norm, conditioning statements are not comparable [
4,
11].
Concrete reporting protocol.
A minimal reproducible report is:
This is deliberately orthogonal to runtime, which depends on hardware, linear algebra backends, and engineering decisions. The goal is not to replace runtime benchmarking; it is to prevent runtime benchmarking from being interpreted as a proxy for convergence quality when geometries differ [
1,
2].
7.2. Economic Meaning of the Ruler W
W is part of the model-to-evidence contract.
The ruler W encodes which equilibrium violations are “large” and which are “small”. In applied DSGE work, different communities implicitly choose different rulers: Euler equation errors, maximum norm across states, consumption-equivalent welfare loss, or likelihood/GMM-type weighting. Freezing W does not force a single choice; it forces the choice to be declared so that solver comparisons remain meaningful.
7.2.1. Three Principled Design Patterns for W
- (i)
Dimensionless accuracy ruler. Let
D be a scaling built from steady-state magnitudes or local variances of each equation component. Then
yields a dimensionless residual. This is the canonical “engineering comparability” ruler and is close in spirit to preconditioning [
1].
- (ii)
Statistical ruler (GMM/likelihood weighting). If
is a moment or forecast-error vector, choosing
W proportional to an inverse covariance yields a generalised least squares residual. This connects certificates to statistical efficiency criteria and makes the residual interpretable as a quadratic form in measurement space [
19].
- (iii)
Welfare-aligned ruler. In models with an explicit welfare objective, W can be chosen so that upper-bounds (or approximates) welfare loss to second order around the optimum. Then contraction in R becomes a meaningful proxy for welfare improvement in the local regime. Such alignments are model-dependent and require an explicit derivation, but they are conceptually clean.
7.2.2. Ruler Choice as an Optimisation Problem
Treating
W as design data exposes a legitimate optimisation problem: choose
W to best predict out-of-sample error, to best align with welfare loss, or to best separate stable from unstable solver regimes. This problem is related to metric learning and to optimal preconditioning, but with an explicit constraint that
W must be frozen for comparability once chosen [
4]. The key methodological point is that this optimisation cannot be hidden inside the solver comparison: changing
W changes the question.
7.3. What the Certificates Diagnose Beyond “Speed”
Separation of tail contraction and transient risk.
The two scalars
serve different purposes.
measures certified tail contraction per iteration when the linear(ised) defect update is valid;
diagnoses finite-horizon amplification that can destroy practical convergence by pushing iterates out of the contraction neighbourhood. This separation is essential in non-normal regimes, where eigenvalues alone can be misleading [
10].
Newton-type solvers and the “good-but-dangerous” regime.
Newton and quasi-Newton methods can be exceptionally fast near the solution (small
), yet fragile far away because linearisation is poor and transient amplification is common, especially under damping and line-search mixing [
3,
4]. A certificate report that includes both
and
makes this visible: fast asymptotic contraction can coexist with large early risk.
Diagnosing ill-conditioning and near-neutral directions.
If the Gram spectrum of is highly anisotropic, then some defect directions are nearly neutral. This is a structural diagnosis: it identifies that the algorithm is not strongly contracting in all economically relevant directions under the chosen ruler. In DSGE practice, this often corresponds to weak identification, near-indeterminacy, or redundancy in the stacked conditions, all of which can be handled by the neutral discipline (projection or gauge-fixing), but should not be ignored.
7.4. Relation to Existing DSGE Numerical Practice
This paper does not replace the classical DSGE numerical toolbox. Rather, it sits above it as a verification layer.
Relation to standard texts and methods.
Projection, collocation, time iteration, policy iteration, and Newton/quasi-Newton methods are standard [
1,
2,
3,
4]. What is added here is an explicit, reusable geometry-and-certificate discipline: the defect evaluator and ruler are declared as first-class objects, and solver behaviour is reduced to contractivity statements in that frozen norm.
Relation to identification and determinacy.
Neutral directions can arise from redundancy (Walras’ law), normalisations (numéraire), or determinacy structure. The certificate logic does not decide determinacy; it decides how residuals should be measured and compared once the analyst chooses a representation. If determinacy is in question, the neutral-free projection formulation provides a clean place to state what is being modded out and what is being certified.
7.5. Limitations
The certificate calculus is intentionally structural. Its limitations are therefore explicit.
7.5.1. Locality: Far-from-Equilibrium Nonlinearity
The strongest certificates in this paper are local: they certify contraction in a neighbourhood where either a Lipschitz bound holds or a linearisation is accurate. Far from equilibrium,
T can be strongly state-dependent, the basin of attraction can be complicated, and contraction in a single fixed norm may fail even when the solver converges from practical initialisations. This is not a defect of the framework; it is a faithful reflection of nonlinear root-finding and fixed-point theory [
8,
9].
7.5.2. Curse of Dimensionality and Representation Dependence
Global DSGE solvers face the curse of dimensionality. Our framework does not remove it. Instead, it clarifies what must remain invariant under approximation changes: if one compares solvers across different discretisations or bases, one must add a refinement-ladder analysis that keeps the meaning of residual fixed (or pays explicit equivalence constants). This is the analogue, in numerical PDE analysis, of stating stability and consistency in compatible norms before discussing convergence [
1].
7.5.3. Ruler Design Is not Automated
Choosing W is a modelling decision. The framework does not claim that there is a single correct W; it claims that comparisons are meaningless without declaring one. Automating ruler choice is an open research direction (see subsec:discussion-open).
7.5.4. Empirical a Posteriori Certificates Require Careful Forcing Budgets
When certificates are computed from run logs rather than from structural Jacobian bounds, they are empirical. To upgrade them to proofs one must bound the forcing terms (inexact solves, quadrature error, truncation error). This is feasible in principle—numerical analysis provides a language for a posteriori error estimation—but implementing it in full generality for large DSGE pipelines is nontrivial [
9].
7.6. Open Problems and Research Directions
We list the most valuable upgrades as theorem-shaped questions.
7.6.1. Optimal Rulers and Welfare Alignment
Problem 1
(Designing
W for welfare-aligned certificates).
Given a DSGE model with a welfare objective and a defect representation , construct an SPD ruler W and constants such that, in a neighbourhood of equilibrium,
where denotes (a quadratic approximation of) welfare loss.
A solution would make certificate decay directly interpretable in welfare units rather than as a purely computational diagnostic.
7.6.2. Uniform-in-Refinement Certificates for Discretisation Ladders
Problem 2
(Ladder-stable solver comparability).
Let denote a sequence of defect maps induced by refinement (grid/basis level K) with a fixed ruler W. Give sufficient conditions under which a family of solver updates satisfies
and under which the residuals converge to the continuous-model residual as .
This is the precise statement needed to compare solvers across changing discretisations without moving goalposts.
7.6.3. Non-Normality Control and Basin Enlargement
Problem 3
(Quantitative transient control in W-geometry). Provide computable upper bounds on the amplification factor in terms of structural properties of the solver linearisation (e.g. block structure, damping parameters, approximate inverses), and relate those bounds to basin-of-attraction estimates.
This would connect certificate diagnostics to robust globalisation strategies for Newton-type DSGE solvers.
7.6.4. Empirical Forcing Budgets
Problem 4
(Certified forcing budgets for DSGE pipelines). Construct practical a posteriori estimators such that the defect evolution satisfies with provable probability/confidence guarantees (when stochastic quadrature or Monte Carlo expectations are used).
This would bridge the current gap between empirically observed decay and provable forced envelopes.
7.7. Broader Implications: Economics as a Residual Geometry
The central conceptual message is that “equilibrium computation” is not only a numerical task; it is a measurement task. A DSGE model is a map from candidate objects to a defect space. Choosing W is choosing what counts as deviation. Once that choice is made explicit, solver design, numerical diagnostics, and reproducible comparison become clearer.
There is also a methodological implication for empirical work: if the same residual geometry is used to compare solvers, to choose discretisations, and to report approximation error, then the entire computational pipeline becomes auditable in a single metric language. This is particularly important as DSGE computation increasingly integrates heterogeneous components (simulation, filtering, estimation, occasionally machine-learned surrogates), where metric drift can otherwise occur silently.
8. Concluding Discussion and Outlook
This paper proposes a simple discipline with large practical consequences for computational macroeconomics: freeze the residual geometry before making any solver claim. Once the equilibrium defect representation and the SPD ruler are declared, solver behaviour becomes a typed statement about the induced defect dynamics in the single frozen norm . The resulting certificates separate three ingredients that are often conflated in practice: (i) the economic equilibrium problem (the operator F), (ii) the measurement geometry (the ruler W and neutrals), and (iii) the algorithmic dynamics (the update map T). This separation is the core methodological contribution: it makes solver comparisons reproducible and makes failures diagnosable rather than anecdotal.
8.1. What the Certificate Layer Adds to DSGE Computation
Reproducibility by typing.
In standard DSGE practice, two codes can implement the same economics but report different error numbers because they scale equations differently, aggregate residuals differently, or implicitly change which deviations count as important. The frozen-ruler discipline fixes this: a solver is evaluated by
in one declared geometry. The price is that
W must be reported; the payoff is that comparisons become meaningful across labs and implementations [
1,
2].
A single certified rate score.
The comparison metric is a structural number: it depends only on the induced defect update in the frozen geometry. When , it yields an explicit residual envelope and an iteration-complexity bound; when it fails, it correctly signals that no contraction guarantee is available in that geometry. This is precisely the kind of “proof-carrying” diagnostic one wants when solver rankings are used for research claims or policy-relevant computations.
Diagnosis beyond eigenvalues.
Non-normality is common in Newton-type and damped fixed-point DSGE solvers. In such cases eigenvalues do not control transient behaviour; early residual growth can occur even if the method is asymptotically convergent. The Gram-operator viewpoint and the amplification diagnostics separate transient risk from tail contraction, explaining why some solvers are “fast when they work” but fragile to initialisation or damping choices [
10].
Stopping rules that respect numerical reality.
Real implementations have forcing terms: inexact linear solves, quadrature error, truncation, and floating-point effects. Forced-envelope certificates yield principled stopping criteria: once the residual reaches the forcing plateau scale, further iterations cannot reduce
without tightening budgets. This is a clean way to relate solver settings (tolerances) to certified residual outcomes [
4,
9].
8.2. Economic Meaning of Choosing the Ruler W
The ruler W is not a cosmetic scaling; it is part of the contract between the model and what is counted as deviation. Different choices correspond to different scientific questions.
Three canonical interpretations.
- (i)
Engineering comparability: steady-state scaling or dimensionless aggregation, designed to prevent one equation from dominating purely due to units [
1].
- (ii)
Statistical comparability: inverse-covariance weighting when defects are moment/forecast errors, aligning
R with generalised least squares in measurement space [
19].
- (iii)
Welfare comparability: local welfare-aligned weighting, where R is designed to approximate welfare loss near the optimum (model-dependent and requiring an explicit derivation).
Consequent Warning
A change in W can legitimately change the ranking of solvers. This is not a paradox; it means the question changed. The framework forces that change to be explicit, which is a prerequisite for honest comparison.
8.3. Limitations and Scope Wall
The certificate layer is deliberately structural, and its limitations are correspondingly explicit.
Locality.
The strongest statements are local: they certify contraction in a neighbourhood where Lipschitz bounds hold or linearisation is accurate. Far from equilibrium, solver dynamics can be strongly nonlinear and the basin of attraction can be complicated; no single frozen norm can guarantee global convergence without additional globalisation arguments [
8,
9].
High-dimensional state spaces.
The framework does not remove the curse of dimensionality in global methods. What it does provide is a disciplined way to keep the meaning of residual fixed across discretisation changes; a full uniform-in-refinement (ladder) theory requires additional approximation analysis beyond the scope of this paper [
1].
Ruler design is not automated.
Choosing W is modelling data. The paper provides the calculus to propagate a choice into reproducible certificates, but it does not claim that there is a unique correct ruler for all applications.
8.4. Outlook: Upgrades that Turn the Framework into a Standard Auditing Layer
The most valuable next steps are theorem-shaped and operational.
(O1) Welfare-aligned certificate geometry.
Construct model-specific rulers W for which bounds (or approximates) welfare loss near equilibrium. This would convert the present computational diagnostic into a welfare-relevant guarantee.
(O2) Uniform-in-refinement certificates.
Extend the analysis to a refinement ladder in grid/basis level K and prove conditions under which contraction rates and forcing budgets remain stable as the discretisation is tightened. This is the precise step needed to compare solvers across “better models” without moving goalposts.
(O3) Quantitative non-normality control.
Develop computable upper bounds on transient amplification in the frozen geometry (e.g. via resolvent/pseudospectral proxies) and relate them to robust basin-of-attraction enlargements, yielding principled damping and safeguarding strategies [
10].
(O4) Practical forcing budgets.
Provide systematic a posteriori estimators for forcing terms coming from quadrature, truncation, and inexact linear algebra, so that observed decay can be upgraded to a provable forced-envelope certificate [
9].
9. Ethics Approval
This article does not contain any studies with human participants or animals performed by the author.
10. Declaration of Generative AI and AI-Assisted Technologies in the Writing Process
During preparation of this manuscript, the author used ChatGPT (OpenAI) in an assistive role for tasks such as drafting and editing text, formatting formulas and statements in , and exploring alternative formulations of arguments and proofs. All AI-assisted content and suggestions were reviewed, edited, and critically assessed by the author, who takes full responsibility for the final form of all scientific claims, mathematical statements, proofs, and conclusions. No generative system was used to fabricate, alter, or selectively filter empirical or numerical data, and no proprietary, confidential, or unpublished information was provided to any AI system.
Notation
| Symbol |
Type / Domain |
Definition / Meaning (standing assumptions) |
| Spaces, maps, and frozen geometry (core) |
|
Banach space |
Space of candidate equilibrium objects z (e.g. policy coefficients, collocation values). |
|
Hilbert space |
Defect space (stacked equilibrium violations), with inner product . |
| F |
map |
Defect map ; defect is . |
|
element of
|
Defect (equilibrium violation vector) at z. |
| W |
SPD operator |
Frozen ruler on : bounded, self-adjoint, strictly positive (). |
|
scalar |
Weighted inner product . |
|
scalar |
Frozen norm . |
|
scalar |
Residual of sameness: . |
|
projection |
Orthogonal projector onto the neutral-free subspace of (only when neutrals are present). |
|
map |
Neutral-free defect map (used when is invoked). |
|
scalar |
Neutral-free residual , with the restricted ruler. |
| Solver-induced dynamics (core) |
| S |
map |
Solver step on : . |
| T |
map |
Induced defect update (local or exact): (and forced variant ). |
| J |
bounded operator |
Local defect Jacobian: near an equilibrium. |
|
operator norm |
Induced norm . |
| q |
scalar |
Contraction factor in frozen norm, typically or . |
|
element of
|
Forcing term (inexact solves, truncation, numerical noise) in . |
|
scalar |
Forcing magnitude bound used in forced envelopes. |
| Certificates and spectral objects (core) |
|
operator |
Gram operator (PSD, self-adjoint). |
|
scalar |
Top eigenvalue of a self-adjoint operator/matrix. |
|
scalar |
Solver-agnostic rate: . |
|
scalar |
Empirical amplification . |
|
scalar |
Observed one-step ratio (when ). |
|
scalar |
Tail contraction estimate . |
| New Keynesian numerical study (used in Section 6) |
|
scalar |
Output gap. |
|
scalar |
Inflation. |
|
scalar |
Nominal interest rate (policy rate). |
|
scalar |
Natural-rate (demand) disturbance. |
|
scalar |
Cost-push disturbance. |
|
vector |
State . |
|
scalars |
AR(1) persistence parameters for . |
|
scalars |
Innovation scales for . |
|
scalars |
NK parameters: discount factor, intertemporal elasticity parameter, Phillips slope. |
|
scalars |
Taylor-rule coefficients. |
|
finite set |
State grid . |
| M |
integer |
Number of grid points in . |
| L |
integer |
Number of quadrature nodes for expectations. |
|
nodes/weights |
Quadrature nodes and weights for the Gaussian innovations. |
|
operator |
Quadrature expectation operator . |
|
vector |
Policy-parameter vector in the chosen basis representation . |
|
vector |
Stacked NK defect vector on the grid (IS/PC/TR residuals concatenated). |
Master Definitions (used in this manuscript)
| Entry |
Definition / Formula |
Role in the paper |
| Frozen residual geometry (core) |
| Frozen record |
(and optionally ). |
Typing data that makes all solver claims comparable. |
| Defect (equilibrium violation) |
. |
Canonical deviation variable (what solvers are contracting). |
| Residual of sameness |
. |
Single scalar ledger used for envelopes and certificates. |
| Neutral-free residual (optional) |
, with . |
Removes redundant/neutral directions when present. |
| Solver-induced defect dynamics (core) |
| Solver step |
on . |
Abstract form of all solver families compared. |
| Induced defect update |
(exact in linear settings, local after linearisation). |
Object whose contraction is certified in the frozen ruler. |
| Local defect Jacobian |
near . |
Local linear proxy determining tail convergence and transients. |
| Forced defect update |
, with . |
Captures inexact solves, truncation, and numerical noise. |
| Certificates and spectral objects (core) |
| Contraction certificate |
(or locally ). |
Sufficient condition for geometric decay of . |
| Forced envelope |
for . |
Tail-robust bound and principled stopping rule via plateau. |
| Gram operator |
. |
Encodes . |
| Solver-agnostic rate |
. |
Comparable per-iteration contraction score. |
| Transient amplification (empirical) |
. |
Flags non-normal finite-iteration amplification. |
Funding
This research received no specific grant from any funding agency in the public, commercial, or not-for-profit sectors.
Data Availability Statement
No datasets were generated or analysed in this work. All results are of a theoretical and analytical nature and are fully contained in the manuscript.
Acknowledgments
The author affirms sole authorship of this work. First-person plural (“we”) is used purely for expository clarity. No co-authors or collaborators contributed to the conception, development, analysis, writing, or revision of the manuscript. The author received no external funding and declares no institutional, ethical, or competing interests.
Conflicts of Interest
The author declares that there are no competing interests.
References
- Judd, Kenneth L. Numerical Methods in Economics; MIT Press: Cambridge, MA, 1998. [Google Scholar]
- Heer, Burkhard; Maussner, Alfred. Dynamic General Equilibrium Modelling: Computational Methods and Applications; Springer: Berlin, 2009. [Google Scholar]
- Dennis, John E.; Schnabel, Robert B. Numerical Methods for Unconstrained Optimization and Nonlinear Equations; SIAM: Philadelphia, 1996. [Google Scholar]
- Nocedal, Jorge; Wright, Stephen J. Numerical Optimization, 2 edition; Springer: New York, 2006. [Google Scholar]
- Pazy, Amnon. Semigroups of Linear Operators and Applications to Partial Differential Equations of Applied Mathematical Sciences; Springer: New York, 1983; volume 44. [Google Scholar]
- Klaus-Jochen Engel and Rainer Nagel. One-Parameter Semigroups for Linear Evolution Equations, volume 194 of Graduate Texts in Mathematics. Springer, New York, 2000.
- Kantorovich, Leonid V. On newton’s method for functional equations English translations appear in various collections. Doklady Akademii Nauk SSSR 1948, 59, 1237–1240. [Google Scholar]
- Ortega, James M.; Rheinboldt, Werner C.
Iterative Solution of Nonlinear Equations in Several Variables
; Computer Science and Applied Mathematics; Academic Press: New York, 1970. [Google Scholar]
- Deuflhard, Peter. Newton Methods for Nonlinear Problems: Affine Invariance and Adaptive Algorithms; Springer: Berlin, 2011; volume 35. [Google Scholar]
- Trefethen, Lloyd N.; Embree, Mark. Spectra and Pseudospectra: The Behavior of Nonnormal Matrices and Operators; Princeton University Press: Princeton, NJ, 2005. [Google Scholar]
- Horn, Roger A.; Johnson, Charles R. Matrix Analysis, 2 edition; Cambridge University Press: Cambridge, 2013. [Google Scholar]
- Kreyszig, Erwin. Introductory Functional Analysis with Applications; John Wiley & Sons: New York, 1978. [Google Scholar]
- Luenberger, David G. Optimization by Vector Space Methods; John Wiley & Sons: New York, 1969. [Google Scholar]
- Judd, Kenneth L. Projection methods for solving aggregate growth models. Journal of Economic Theory 1992, 58(2), 410–452. [Google Scholar] [CrossRef]
- den Haan, Wouter J.; Marcet, Albert. Accuracy in simulations. Review of Economic Studies 1994, 61(1), 3–17. [Google Scholar] [CrossRef]
- Christiano, Lawrence J.; Fisher, Jonas D. M. Algorithms for solving dynamic models with occasionally binding constraints. Journal of Economic Dynamics and Control 2000, 24(8), 1179–1232. [Google Scholar] [CrossRef]
- Boragan Aruoba, S.; Fernández-Villaverde, Jesús; Rubio-Ramírez, Juan F. Comparing solution methods for dynamic equilibrium economies. Journal of Economic Dynamics and Control 2006, 30(12), 2477–2508. [Google Scholar] [CrossRef]
- Facchinei, Francisco; Pang, Jong-Shi. Finite-Dimensional Variational Inequalities and Complementarity Problems, volume I–II; Springer: New York, 2003. [Google Scholar]
- Hansen, Lars Peter. Large sample properties of generalized method of moments estimators. Econometrica 1982, 50(4), 1029–1054. [Google Scholar] [CrossRef]
- Rust, John. Optimal replacement of gmc bus engines: An empirical model of harold zurcher. Econometrica 1987, 55(5), 999–1033. [Google Scholar] [CrossRef]
- Clarida, Richard; Galí, Jordi; Gertler, Mark. The science of monetary policy: A new keynesian perspective. Journal of Economic Literature 1999, 37(4), 1661–1707. [Google Scholar] [CrossRef]
- Woodford, Michael. Interest and Prices: Foundations of a Theory of Monetary Policy; Princeton University Press: Princeton, 2003. [Google Scholar]
- Galí, Jordi. Monetary Policy, Inflation, and the Business Cycle; Princeton University Press: Princeton, 2008. [Google Scholar]
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).