Submitted:
09 September 2025
Posted:
10 September 2025
You are already at the latest version
Abstract
Keywords:
MSC: 46E35; 41A25; 35Q68; 42B35; 68T07; 58J20; 58B34; 65D15; 81T75
1. Introduction
- We introduce a hypermodular-symmetric operator framework (ONHSH) that coherently integrates hyperbolic activations, arithmetic-informed spectral damping, and curvature-sensitive kernels, enabling PDE operator learning on anisotropic, curved, and modularly structured domains.
- We establish minimax-optimal approximation rates in weighted anisotropic Besov and Triebel–Lizorkin spaces, supported by explicit Voronovskaya-type expansions and quantitative remainder bounds. At the theoretical core lies the Ramanujan–Santos–Sales Hypermodular Operator Theorem, which formalizes the convergence rates and spectral bias–variance trade-offs for neural operators under directional smoothness.
- We demonstrate that operator spectral variance admits a natural interpretation via noncommutative Chern characters, creating a rigorous bridge between functional approximation, spectral asymptotics, and arithmetic topology.
1.1. Research Scope and Methodological Positioning
- Geometric Adaptivity: Moving beyond models confined to flat or mildly deformed Euclidean settings [4,5], ONHSH employs curvature-sensitive kernels that adapt to hyperbolic and anisotropic manifolds. This design is motivated by functional spaces on spheres and balls [24] and enriched by tools from spectral geometry [25].
- Spectral Modularity: By embedding modular arithmetic into the spectral filtering process, ONHSH captures oscillatory dynamics and aliasing effects that classical FNO variants [13,15] cannot fully represent. The modular structure also enables arithmetic-informed spectral damping aligned with underlying physical constraints.
- Function-Space Theoretic Rigor: ONHSH is firmly grounded in the approximation theory of anisotropic and mixed-smoothness function spaces, notably Besov and Triebel–Lizorkin classes [16,19]. At the core of this framework lies the Ramanujan–Santos–Sales Hypermodular Operator Theorem, which establishes minimax-optimal convergence rates and formalizes the spectral bias–variance trade-off for neural operators under directional smoothness. This provides a principled bridge between neural operator design and harmonic analysis [17,22].
1.2. Conceptual Diagram of the ONHSH Architecture
- Curved kernels control spatial localization and capture anisotropic geometry.
- Symmetrized activations enforce hyperbolic symmetry and enhance stability under sign changes.
- Modular spectral filters introduce arithmetic-informed damping, regulating oscillations and aliasing effects.
- Spectral transforms restore global coherence and ensure compatibility with harmonic analysis on curved domains.
2. Mathematical Foundations
2.1. Anisotropic Besov Spaces
2.1.1. Interpretation
- Deficit quantification: measures local -directional irregularity,
- Scale sensitivity: Integration over captures decay of smoothness deficits at fine scales,
- Directional synthesis: Summation over j aggregates mixed smoothness.
2.1.2. Functional Analytic Properties.
- The factor quantifies the smoothness deficit in direction ;
- The integration in assesses the rate of regularity decay at small scales;
- The summation across aggregates the total mixed smoothness.
2.2. Norm Equivalence via K-Functionals
2.3. Characterization by Smoothness Moduli
- ,
- ,
- For each j, as .
- Quasi-Banach Structure: For , is a quasi-norm satisfyingwith constant depending on . Completeness holds for all .
-
Anisotropic Scaling Invariance: For , define the dilation operator . Then:This symmetry is intrinsic to architectures preserving directional scaling laws, such as ONHSH.
2.4. Characterization via Directional Smoothness Moduli
- (i)
- ;
- (ii)
- ;
- (iii)
- where and as ;
- (iv)
- for each j and .
3. Anisotropic Embedding Theorems
- Sharpness of (30): If for some j, then the univariate Sobolev embedding fails in that coordinate. Consider the example , where , , and . Then , but due to the local singularity at 0.
- Necessity of Lipschitz Boundary: For non-Lipschitz domains, such as domains with outward cusps or fractal boundaries, no universal bounded extension operator exists for anisotropic Besov spaces. In such settings, the geometry of may obstruct the preservation of local moduli of smoothness under extension.
3.1. Compactness of the Anisotropic Embedding
Part 1:
- ,
- for ,
- for ,
Case 1: ().
Case 2: ().
Summary:
Part 2: Continuous embedding:
4. Anisotropic Besov Embedding on Compact Riemannian Manifolds
- The atlas is finite;
- The transition maps have uniformly bounded derivatives;
- The global Besov norm is equivalent to the collection of local norms.
5. Embedding Theorems in Function Spaces
5.1. Embedding on Bounded Lipschitz Domains
5.2. Embedding on Compact Riemannian Manifolds
6. Approximation Theory
6.1. Directional Moduli of Smoothness
- (i)
- Seminorm properties: The functional defines a seminorm in for each fixed , and satisfies the following:where denotes the space of all polynomials of degree at most in the variable .
- (ii)
- Derivative bound: If , the Sobolev space of functions with weak derivatives up to order r in , then the directional modulus satisfies the following upper estimate:where .
- (iii)
- Jackson-type estimate: There exists a constant , independent of f and n, such thatwhere,denotes the best -approximation error of f by univariate polynomials of degree less than n in the variable , keeping all other coordinates fixed.
- (i)
- The mapping defines a seminorm on the function space, and satisfies the scaling relation:
- (ii)
- If , then:where denotes the r-th weak derivative in the direction j, and is a constant depending only on r.
- (iii)
- Conversely, for any , there exists a polynomial-type approximation operator (constructed via mollification in the j-th variable) such that:where depends only on the kernel used and the order r.
6.2. Modular Spectral Multipliers: Kernel Estimates, Compactness, and Hyperbolic Invariance
- Kernel representation and estimates: The integral kernelsatisfies, for all multi-indices , and for some constants independent of n:for every integer . In particular, with rapid decay in spatial variables enhanced by the damping .
- Compactness on : For any , the operator is compact. Indeed, since , is an integral operator with kernel in for every , ensuring Hilbert–Schmidt (or nuclear) type properties in , and boundedness plus compactness in by Schur’s test and smoothing arguments.
-
Approximation and convergence: As , we have:Moreover, the rate of convergence satisfiesfor some constants depending on the anisotropic Besov regularity vector .
-
Hyperbolic invariance and neural operators: The modular multiplier respects anisotropic scaling symmetries aligned with the hyperbolic geometry induced by the normConsequently, the operators commute (or intertwine) with a hyperbolic group action on , i.e.,where,with anisotropy weights . This invariance property makes natural building blocks for hyperbolically invariant neural operators incorporating anisotropic spectral filtering consistent with the geometry of the data domain.
6.3. Spectral Damping and Phase-Space Localization
Implications and Phase-Space Compactness
- In PDE approximation, it guarantees that the learned neural operator retains control over the resolution scale while avoiding amplification of high-frequency noise;
- In inverse problems, the compactness provides natural regularization, mitigating instability associated with ill-posedness;
- In neural architectures, it supports sparse parameterization and efficient training, especially in anisotropic or non-Euclidean domains.
7. Symmetrized Hyperbolic Activation Kernels
7.1. Definition and Core Properties
7.2. Fourier Analysis and Spectral Localization
7.3. Even-Order Moments and Asymptotic Scaling
8. Asymptotic Expansion of the Approximation Operator
- all odd-order moments vanish: ;
- all even-order moments up to are finite: , for .
8.1. Moment Structure and Symmetry Summary
- (i) Odd symmetry. The activation kernel is odd with respect to the origin:
- (ii) Vanishing odd moments. All odd-order moments of the kernel vanish due to its odd symmetry:
- (iii) Even moments. The even-order moments of the kernel are given explicitly by:
- (iv) Asymptotic expansion of the integral operator. The operator admits the following asymptotic expansion in terms of even derivatives of f:
Explanation of terms
- The odd symmetry in (175) ensures that the kernel changes sign under spatial inversion, which in turn enforces the cancellation of all odd-order contributions in Taylor expansions.
- The vanishing of odd moments (176) is a direct consequence of the odd symmetry and implies that only even-order derivatives of f contribute to the leading terms in the operator expansion.
- The even moments are explicitly computed in (177) based on the analytical form of the kernel. These constants depend on the parameters (scaling factor), (hyperbolic modulation), and a structural constant arising from the base function (e.g., a mollified or scaled tanh).
- The asymptotic expansion (178) reflects the accuracy of the approximation as , with leading-order contributions given by even derivatives of f, weighted by the corresponding moments . The residual error is of order , under the assumption .
9. Spectral Variance and Voronovskaya-Type Expansions
9.1. Geometric Interpretation
9.2. Bias–Variance Trade-Off
9.3. Hyperbolic Symmetry Invariance
Lorentz Group and Minkowski Geometry
Kernel Invariance under Lorentz Transformations
Modular–Hyperbolic Coupling and Periodicity
Spectral and Representation-Theoretic Consequences
10. Hyperbolic Symmetry Invariance
Setup and notation
Kernel hypothesis
Remarks on measure-preservation and determinant
Modular–hyperbolic kernel: invariance subtleties
-
Lattice-stabilizing subgroup: If belongs to the subgroup , then the map permutes . In that case we may rename the summation index and use the same change-of-variables argument as above to obtainThus invariance is retained on the arithmetic subgroup .
- General Lorentz maps: If , the lattice is not preserved, and the sum in (218) is mapped to a sum indexed by , which is typically not the same set as . Therefore the pointwise invariance fails in general; however, the modular Gaussian factor provides rapid decay so that the operator still regularizes high-frequency lattice modes and can be analyzed spectrally using Poisson summation and arithmetic harmonic analysis.
Spectral and representation-theoretic consequences
Remarks
11. Anisotropic Sobolev Embedding
11.1. (A) Embedding Under the Balanced Anisotropic Condition
11.2. (B) Coordinatewise Sufficient Condition with Explicit Constants
Remarks on (A) vs (B).
- The coordinatewise condition (238) used in (B) is a simple, easily checked sufficient hypothesis and gives an explicit constant via the geometric series . This suffices in many applications.
- The balanced condition (221) in (A) is more flexible: it allows some coordinates to have small smoothness provided others compensate. The proof in (A) uses shell/scale counting and geometric decay; to obtain a fully sharp anisotropic criterion one refines the counting estimate (228) and the scale bound (227) and often works in mixed-norm ℓ-spaces. If you want, I can convert the argument in (A) into a fully quantitative statement with explicit constants (this requires a more careful combinatorial estimate of and the constants in (227)).
12. Spectral Refinement via ONHSH Operators
12.1. Fourier Multiplier Representation
12.2. Significance of the Spectral Decay
12.3. ONHSH-Enhanced Sobolev Embedding Theorem
13. Nonlinear Approximation Rates
13.1. Duality in Anisotropic Besov Spaces
14. Hyperbolic Symmetry Invariance
14.1. Lorentz Group Action on Tempered Distributions
14.2. Equivalence of Anisotropic Symbols Under Lorentz Transformations
14.3. Lorentz Invariance of the Anisotropic Besov Norm
15. Symmetrized Hyperbolic Activation Kernels
15.1. Base Activation Function
- (i)
- Strict monotonicity: for every ;
- (ii)
- Asymptotic limits:
- (iii)
- Modular duality: For all ,
- (iv)
- Zero at shifted origin:
- (i)
-
Strict monotonicity. Differentiating with respect to x, we use the chain rule on the hyperbolic tangent function:Since the hyperbolic secant satisfies for all , and given , it follows thatHence, is strictly increasing on .
- (ii)
-
Asymptotic limits. For , we rewrite asby dividing numerator and denominator by . Since as , we haveSimilarly, for , dividing numerator and denominator by yieldsSince as , it follows that
- (iii)
-
Modular duality. By direct substitution,Multiplying numerator and denominator by , we obtain
- (iv)
15.2. Central Difference Kernel
- (i)
- Modular antisymmetry: For all ,
- (ii)
- Exponential decay: There exists a constant such that for all ,
- (i)
- Modular antisymmetry. By definition of and applying the modular duality property of , Prop. (iii), we have
- (ii)
-
Exponential decay. Note that the central difference kernel can be expressed via the fundamental theorem of calculus as the average derivative over the interval :From the derivative formula (312) and recalling the explicit form,Using the exponential decay of , there exist constants depending on and q such thatTherefore, for ,By the triangle inequality and monotonicity of the exponential,This establishes the exponential decay of for large .
15.3. Symmetrized Hypermodular Kernel
- (i)
- Even symmetry: for all ;
- (ii)
- Strict positivity: for all ;
- (iii)
- Vanishing of all odd moments:
- (iv)
- Normalization:
- (i)
-
Even symmetry: By definition (331) and the modular antisymmetry property of from Theorem ??(i), we haveThis shows is an even function.
- (ii)
- Strict positivity: Since is strictly increasing, its difference quotient is strictly positive for all x. The same holds for , so their average is strictly positive:
- (iii)
- Vanishing odd moments: Because is even by (334), the product is an odd function. Integrating any odd function over the entire real line yields zero:
- (iv)
-
Normalization: Using the integral representation of given byand Fubini’s theorem to interchange integrals, we computeConsequently,
15.4. Regularity and Spectral Decay
- (i)
- Smoothness:
- (ii)
- Derivative decay: For every , there exist constants and such that
- (iii)
- Fourier decay: For every , there exists such that
15.5. Regularity and Spectral Decay in the Multivariate Anisotropic Setting
- (i)
- Smoothness:
- (ii)
- Anisotropic derivative decay: For every multi-index , there exist constants and such that
- (iii)
- Anisotropic Fourier decay: For every , there exists such that
15.6. Fractional Smoothness Gain via Real Interpolation
15.7. Consequences for Approximation Rates
15.8. Moment Structure and Modular Correspondence
15.9. Multivariate Anisotropic Moment Structure and Modular Correspondence
- (a)
- (Absolute convergence) the series in (390) converges absolutely for every fixed ; in fact, for any there exists with
- (b)
- (Modular / Eisenstein representation) writing the Eisenstein-type generating seriesthe moment can be expressed as a q-series convolutionin the sense used in the text (cf. Theorem 28). This equality is equivalent to (390).
- (c)
- (Consistency with moment bounds) the factorial growth bounds for moments obtained from spatial exponential decay of are consistent with representation (390) via standard bounds .
15.10. Multidimensional Kernel
15.11. Geometric Interpretation
- The base parametrizes the modular deformation parameter q.
- The fiber over a point is the function space generated by and its derivatives in x.
15.12. Geometric Interpretation
- The base encodes the modular parameter q;
- The fiber is the function space generated by and its x-derivatives.
15.13. Geometric Interpretation
- Base:, encoding the modular parameter q;
- Fiber:, the function space generated by and its derivatives in x.
15.14. Geometric Interpretation: Chern–Eisenstein Integral
15.15. Geometric Interpretation at Level N: Chern Character, Area, and Dirichlet L-Values
Chern–Weil at level N.
Hyperbolic area via index.
Eisenstein viewpoint and Dirichlet L-values.
A compact closed form for .
16. Minimax Convergence in Anisotropic Besov Spaces
16.1. Anisotropic Besov Norm and Directional Smoothness
16.2. Statement of the Minimax Theorem
17. Main Convergence Theorem for ONHSH
18. Geometric Interpretation of Chern Characters
- is a finite-dimensional smooth manifold (the parameter/moduli space);
- for each the operator is a smoothing operator on and depends smoothly on s in the topology of trace-class (or, more generally, in a nuclear operator topology guaranteeing the manipulations below);
- when we refer to Tr we mean an admissible trace (ordinary trace when operators are trace-class; a Dixmier-type singular trace when operators lie in the weak ideal and are measurable in the sense of Connes).
18.1. Operator Bundle, Connection and Curvature
18.2. Chern Character in the Operator Setting
18.3. Index Integrals on Arithmetic Quotients
18.4. Non-Commutative Index Pairing and Dixmier Traces
18.5. Consequences and Interpretation
- The operator-valued curvature measures the failure of the operator family to be flat in parameter space; concretely it records noncommutativity of parameter derivatives (see (495)).
- Provided the family is smoothing (or satisfies nuclearity/Schatten estimates), the forms are well-defined closed differential forms and define cohomology classes; the formal exponential is the ensuing characteristic class (Chern character) of the operator bundle.
- When the parameter manifold descends to an arithmetic quotient , integration of over produces index-type invariants with arithmetic significance; under ellipticity these coincide with classical analytical indices.
- In the noncommutative (spectral) picture, Dixmier traces extract the residue part of spectral asymptotics and implement the index pairing between K-theory and cyclic cohomology, thereby translating approximation-theoretic spectral data into topological/arithmetic invariants.
18.6. Detailed One-Dimensional Example
18.7. Rigorous Membership in Operator Ideals, Schatten Estimates, and Regularization
18.8. When the Base Is Noncompact and Convolutional Symmetry Holds: Regularization and Dixmier Traces
- Localization / compactification. Insert cutoffs with pointwise (for instance supported in a ball of radius R). Study the family , which has kernel compactly supported in and therefore lies in . Analyze asymptotics as and extract invariant coefficients (differences, densities). This is the standard approach for defining “trace per unit volume” or renormalized traces.
- Spectral regularization (heat / zeta). Introduce an auxiliary elliptic operator H (for instance ) with discrete-like spectral asymptotics upon confinement or via functional calculus, and definefor . For many operators A (including convolutional families after suitable weighting), the small-t expansion of has an asymptotic expansion whose coefficients carry geometric content. Zeta-regularization proceeds by defininganalytically continuing and extracting residues or finite parts at particular points; the Dixmier trace corresponds to the coefficient of the log-term in the small-t expansion and can be recovered from the residue of at the critical dimension.
18.9. Concluding Proposition and Practical Checklist
- (a)
- uniformly in s (or has sufficient polynomial decay in both x and y so that weighted bounds hold); or
- (b)
- uniformly in s (Schwartz-class kernels); or
- (c)
- after localization by compact cutoff , the localized operators satisfy (a) or (b) uniformly in R and s, and the renormalized limits exist as ,
19. Schatten Estimates and Heat-Kernel/Zeta Regularization
- is a finite-dimensional smooth manifold (parameter space).
- For each the operator is given by an integral kernel on , and the map is smooth into a function space specified below.
- When we write Tr we mean either the ordinary trace (for trace-class operators) or an admissible singular trace (Dixmier trace) when the weaker ideal is the relevant setting.
19.1. Rewritten and Numbered Preliminaries
19.2. Explicit Schatten-norm Estimates: Strategy and Results
19.3. Explicit Schatten-Norm Estimates for the 1D Hypermodular Kernel
19.4. Heat-Kernel and Zeta Regularization for the 1D Example
- Localize the operator (cutoff) or otherwise ensure is well-defined for .
- Compute or estimate the small-t asymptotic expansion of .
- Identify the coefficient (if present) or the constant term corresponding to the critical dimension.
- Obtain the zeta function by Mellin transform and read off the residue at ; this residue equals and, up to normalization, yields the Dixmier trace.
19.5. Concrete Remark on Constants and Normalizations (Practical Guidance)
19.6. Practical Checklist for Implementation
- Verify Schwartz-type decay (or weighted bounds) of and its parameter derivatives. If true, direct trace-class statements apply (see (515)).
- If the kernel is convolutional and translation invariant, introduce cutoffs , compute localized traces, and study the asymptotics to obtain density per unit volume (see (521)).
- For noncompact settings where only weak decay holds, compute , expand for small t and extract the coefficient to determine the Dixmier residue (recipe above).
- When numerics are intended, approximate diagonal integrals such as (539) using quadrature over a sufficiently large computational domain and monitor convergence as the cutoff grows.
20. Hypermodular Kernel Construction
- Hyperbolic deformation: governed by a spatial scaling parameter , which controls concentration in the physical domain via Gaussian localization.
- Modular deformation: governed by a spectral parameterwhich enforces spectral suppression in a way compatible with modular symmetries.
20.1. Spectral Damping Properties
- (1)
-
Superexponential decay: For all ,In particular, for any ,
- (2)
- Besov space stability: If with and , thenwhere is independent of n.
21. Geometric Interpretation of Chern Characters
21.1. Geometric and Topological Meaning
21.2. Explicit Schatten-Norm Estimates
21.3. Heat-Kernel and Zeta-Regularization in 1D
21.4. Multidimensional Heat-Kernel Asymptotics and Index Invariants
- (i)
- , even, strictly positive, and normalized:
- (ii)
- Spatial decay: For every there exists such that
- (iii)
- Fourier decay: For every there exists such that
- Leibniz rule applied to and
- Derivative bounds:
- Optimization:
- Exponent : Originates from the interplay between spectral decay and anisotropic tile growth .
- Constant sharpness: The formula for reflects the balance between kernel decay () and modular spectral damping ().
- Minimax sharpness: The rate matches the intrinsic approximation limit for mixed smoothness.
- Geometric invariance: When and the tiling respects hyperbolic symmetry, commutes with .
22. Application: Thermal Diffusion Benchmark
- ONHSH: integrates symmetric hyperbolic activations, modular spectral damping, and curvature-sensitive convolution kernels, reflecting both geometric adaptivity and arithmetic-informed regularization.
- Fourier Neural Operator (FNO) [1]: employs global Fourier filters with exponential decay in the spectral domain.
- Geo-FNO [4]: introduces coordinate deformations that account for geometric variability before spectral filtering.
- NOGaP [6]: incorporates a probabilistic spectral filter with Gaussian perturbations to encode uncertainty.
- Convolutional Baseline: local averaging with fixed kernels, representing classical low-pass filtering.
- Gaussian Smoothing: isotropic smoothing implemented via convolution with Gaussian kernels.
22.1. Numerical Analysis of Error Metrics
23. Analysis of Neural Operators
23.1. ONHSH: A Promising Framework for Hypermodular and Anisotropic Domains
- Relativistic partial differential equations (PDEs) on Lorentzian manifolds,
- Thermal diffusion in modular and arithmetic-enriched domains,
- High-frequency dynamics in anisotropic media.
- Optimizing the hyperbolic symmetry parameters for improved empirical performance,
- Exploring adaptive modular damping strategies to mitigate over-smoothing,
- Leveraging the operator’s inherent Lorentz invariance for relativistic applications.
23.1.1. Strengths of ONHSH
- Mathematical Rigor: ONHSH is built upon a robust theoretical framework, ensuring minimax-optimal approximation rates in anisotropic Besov spaces.
- Geometric Adaptivity: Its hyperbolic symmetry and curvature-sensitive kernels make it inherently suitable for non-Euclidean geometries, including relativistic PDEs and modular domains.
- Spectral Flexibility: The modular spectral damping mechanism allows for fine-grained control over oscillatory behavior, making it adaptable to high-frequency dynamics.
23.1.2. Challenges and Future Directions
- Parameter Sensitivity: ONHSH’s performance is highly dependent on the selection of hyperbolic symmetry parameters and modular damping factors. Future work should focus on automated parameter optimization to enhance its practical applicability.
- Computational Overhead: The complexity of ONHSH’s architecture may introduce computational challenges. However, advancements in parallel computing and GPU acceleration could mitigate these issues.
23.2. Geo-FNO: The Benchmark for Geometric Adaptivity
- MAE
- MSE
- RMSE
23.3. FNO, NOGaP, Convolution, and Gaussian: Reliable but Limited
- MAE
- MSE –
- RMSE –
24. Comparative Summary
| Operator | MAE | MSE | RMSE | Key Strengths |
|---|---|---|---|---|
| Geo-FNO | Geometric adaptivity, high accuracy | |||
| ONHSH | Theoretical rigor, hyperbolic symmetry | |||
| FNO | Stability, global spectral basis | |||
| NOGaP | Uncertainty quantification | |||
| Convolution | Simplicity, computational efficiency | |||
| Gaussian | Smoothness, noise reduction |
- Relativistic partial differential equations (PDEs) on Lorentzian manifolds,
- Thermal diffusion in modular and arithmetic-enriched domains,
- High-frequency dynamics in anisotropic media.
25. Algorithmic Pipeline
- Data Generation. A synthetic three-dimensional thermal diffusion field was generated using sinusoidal initial conditions and exact analytical solutions of the heat equation. This setup ensures controlled smoothness through a tunable frequency parameter, providing a precise ground-truth reference for subsequent evaluations. The generated data captures both isotropic and anisotropic diffusion regimes, enabling a comprehensive assessment of operator performance under varying geometric and spectral conditions.
-
Operator Layers. Multiple operator-based models were implemented to propagate the initial thermal conditions and approximate the solution field. The evaluated architectures include:
- ONHSH: The proposed Hypermodular Neural Operator with Hyperbolic Symmetry, integrating curved convolutional kernels, hyperbolic activations, and modular spectral filters. This architecture is designed to adapt to anisotropic and curved domains, leveraging the Ramanujan-Damasclin Hypermodular Operator Theorem for minimax-optimal approximation rates.
- FNO: The Fourier Neural Operator, which employs global spectral filtering to capture long-range dependencies in structured domains.
- Geo-FNO: A geometric variant of FNO that incorporates domain deformations prior to spectral filtering, enhancing adaptability to non-Euclidean geometries.
- NOGaP: The Neural Operator-induced Gaussian Process, which combines operator learning with probabilistic perturbations for uncertainty quantification.
- Baselines: Classical methods such as convolutional averaging and Gaussian smoothing were included to provide a reference for traditional approaches.
-
Error Metrics. The predicted thermal fields were quantitatively assessed against the exact solution using standard error norms, see Eqs. (586–588). These metrics provide complementary insights into performance:
- MSE captures the global variance and sensitivity to outliers.
- MAE reflects absolute deviations and robustness to noise.
- RMSE offers a balanced measure of root-mean-square stability.
-
Visualization. High-quality comparative visualizations were generated using the viridis colormap, optimized for thermal emphasis and perceptual uniformity. Two complementary visualization strategies were employed:
- Three-dimensional scatter plots to illustrate volumetric diffusion structures and spatial gradients.
- Two-dimensional mid-plane slices enriched with isothermal contour lines to highlight anisotropic gradients and local variations.

26. Introduction to the ONHSH Algorithm
26.1. Theoretical Foundations
- Minimax-optimal approximation rates in anisotropic Besov spaces, ensuring best-possible convergence under directional smoothness.
- Spectral bias–variance trade-offs, providing precise characterizations of approximation errors across frequency regimes.
- Geometric adaptivity through curvature-sensitive kernels that intrinsically follow domain geometry.
- Noncommutative connections, linking spectral variance phenomena to principles of noncommutative geometry.
26.2. Algorithmic Components
- Symmetrized Hyperbolic Activation:which ensures Lorentz invariance and stability under non-Euclidean transformations.
- Modular Spectral Filtering:designed to incorporate arithmetic-informed damping for precise control of oscillatory modes.
- Curvature-Sensitive Kernels:which adaptively capture intrinsic geometric variations within the domain.
26.3. Comparative Advantages
26.4. Implementation Pipeline and Applications
- Generation of three-dimensional thermal diffusion datasets with controlled smoothness profiles.
- Application of the ONHSH operator, integrating hyperbolic activations and modular filtering mechanisms.
- Evaluation of performance using rigorous error metrics (MSE, MAE, RMSE), supported by theoretical validation.
- Production of high-quality visualizations, employing perceptually uniform color maps such as viridis.
26.5. Key Benefits
- Guaranteed minimax-optimal approximation rates in anisotropic settings.
- Natural adaptability to highly complex and curved geometries.
- Stable control of high-frequency dynamics via modular spectral filtering.
- Inherent Lorentz invariance, enabling compatibility with relativistic frameworks.
- Strong empirical robustness across challenging PDE benchmarks.
26.6. ONHSH Algorithm with Ramanujan–Santos–Sales Hypermodular Operator Theorem Integration
| Algorithm 1 ONHSH Implementation Incorporating Ramanujan–Santos–Sales Theorem |
|
26.7. Theorem Integration Notes
- Minimax-Optimal Rates: The modular spectral filter enforces the convergence rate from the Ramanujan–Santos–Sales Hypermodular Operator Theorem.
-
Anisotropic Besov Spaces: The implementation implicitly works in where:
- with ,
- Embedding into is guaranteed (Theorem 4).
- Spectral Bias-Variance Trade-off: The parameter q controls the trade-off as formalized in:where .
- Geometric Adaptivity: The curved kernel implementation respects the Lorentz invariance and Riemannian manifold.
- Modular Correspondence: The spectral filter’s construction follows:linking to the arithmetic topology.
27. Quantitative and Qualitative Analysis of Numerical Results
27.1. Quantitative Analysis
27.1.1. MSE vs. Grid Size
- The ONHSH operator exhibits systematically higher errors compared to Geo-FNO, which sets the accuracy benchmark for problems in complex geometric domains. However, the error for ONHSH remains stable and comparable to FNO and NOGaP, particularly for larger grid sizes.
- The error for ONHSH increases from approximately to as the grid size grows from 18 to 30, indicating moderate sensitivity to spatial discretization.
- The Convolution and Gaussian operators show significantly lower and stable errors but are limited to simple domains and fail to capture the geometric and spectral complexity addressed by ONHSH.
- The ONHSH operator starts with an error of approximately at , which increases to about at . This growth is more pronounced at early times, stabilizing at later times.
- The Geo-FNO operator maintains a consistently low error, reinforcing its effectiveness in smooth geometric domains.
- The FNO and NOGaP operators exhibit intermediate behavior, with errors growing similarly to ONHSH but with lower absolute values.
27.2. Qualitative Analysis
27.2.1. Advantages of ONHSH
- Geometric Adaptability: The integration of curved kernels and hyperbolic symmetry enables ONHSH to effectively capture the geometry of anisotropic and curved domains, overcoming limitations of traditional operators such as FNO and Convolution.
- Theoretical Rigor: Grounded in the Ramanujan–Santos–Sales Hypermodular Operator Theorem, ONHSH guarantees minimax-optimal approximation rates in anisotropic Besov spaces, providing a solid mathematical foundation for its application.
- Modular Spectral Filtering: The incorporation of modular spectral filters allows for refined control over oscillatory behaviors, which is essential for problems involving high-frequency and arithmetic structures.
27.2.2. Comparison with Other Operators
- Geo-FNO: While Geo-FNO exhibits lower errors, its applicability is limited to domains with smooth deformations. ONHSH, on the other hand, is designed for domains with intrinsic curvature and extreme anisotropy.
- FNO and NOGaP: These operators offer a balance between accuracy and generality but lack the geometric adaptability and theoretical rigor of ONHSH.
- Convolution and Gaussian: Limited to simple domains, these methods serve as classical baselines but are unsuitable for complex domain problems where ONHSH excels.
28. Results
28.1. Problem Setup and Evaluation Protocol
28.2. Quantitative Accuracy on Thermal Diffusion
28.3. Resolution and Time Studies
28.4. Qualitative Comparisons
28.5. Takeaways for ONHSH
29. Conclusions
Author Contributions
Funding
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
| Acronyms | |
| ONHSH | Hypermodular Neural Operators with Hyperbolic Symmetry |
| PDE | Partial Differential Equation |
| FNO | Fourier Neural Operator |
| FSO | Fourier-Sobolev Operator |
| NOGaP | Neural Operator-induced Gaussian Process |
| Mathematical Symbols | |
| f, | Input/output functions in operator learning |
| , | Neural operators at discretization level n |
| Anisotropic kernel with curvature and modularity q | |
| Symmetrized hyperbolic activation kernel | |
| Base hyperbolic activation function | |
| Central difference kernel | |
| Anisotropic Besov space with regularity vector | |
| , | Shimura variety and upper half-plane |
| Chern character of operator family | |
| Curvature form | |
| Spectral variance term | |
| Macaev ideal for Dixmier traces | |
| r-order directional difference operator | |
| Directional modulus of smoothness | |
| Key Parameters | |
| Curvature scaling factor (controls spatial localization) | |
| q | Modular deformation parameter () |
| Anisotropic smoothness index in direction j | |
| (bottleneck smoothness) | |
| (embedding gain coefficient) | |
| c, C | Exponential decay constants () |
| Operators and Spaces | |
| , | Fourier transform and inverse |
| Norm in anisotropic Besov space | |
| -norm | |
| Inner product/duality pairing | |
| , | Trace and Dixmier trace |
| Lorentz group of hyperbolic symmetries | |
| ↪ | Continuous embedding |
| ≍ | Norm equivalence |
| ⊗ | Tensor product (kernel construction) |
| ∧ | Wedge product (differential forms) |
| -norm | |
| Norm in anisotropic Besov space | |
| Inner product (or duality pairing) | |
| , | Partial derivatives with respect to coordinates , |
| Trace operator | |
| ∼ | Asymptotic equivalence |
| ∧ | Wedge product in differential geometry |
| Special Functions | |
| Eisenstein series | |
| Divisor sum | |
| Riemann zeta function | |
| Damping factor | |
| Symbols and Nomenclature | |
| f | Target function or solution of the PDE |
| Neural operator indexed by discretization level n | |
| Symmetrized activation kernel with parameters and q | |
| Base hyperbolic function with modular and curvature control | |
| Central difference kernel | |
| , | Fourier transform and its inverse |
| Anisotropic Besov space with regularity vector | |
| Shimura variety or geometric parameter space | |
| Vector bundle over | |
| Chern character of bundle E | |
| Modular-invariant volume form | |
| Euclidean domain of dimension d | |
| Greek Letters | |
| Curvature parameter controlling spatial decay | |
| q | Modular deformation parameter |
| Local spectral covariance associated with | |
| , | Spatial and spectral spread (uncertainty) |
| Gamma function in moment formulas | |
| Indices and Notation | |
| Coordinate indices in | |
| n | Resolution or discretization index |
| d | Spatial dimension |
| Smoothness index in anisotropic direction j | |
| Norm and summability parameters in Besov spaces | |
| Harmonic mean of anisotropic smoothness indices | |
Appendix A. Standing Hypotheses and Auxiliary Lemmas
Appendix A.1. Kernel and Multiplier Hypotheses
- (H1)
-
Schwartz regularity. For each , . Equivalently, for every multiindex and integer there exists withThis guarantees absolute convergence of Fourier transforms, moment integrals, and allows the exchange of limits in asymptotic expansions.
- (H2)
- Finite moments. There exists (or larger, if higher-order Voronovskaya expansions are required) such that for all ,is finite and depends smoothly on . These moments appear explicitly in bias terms of asymptotic expansions.
- (H3)
- Parameter regularity. The Schwartz seminorms of vary smoothly in . Differentiation in and q can be interchanged with integration whenever an integrable majorant exists. This ensures well-defined parametric differentiation of operators in proofs of stability and minimax bounds.
- (H4)
-
Spectral multiplier decay. The Fourier multiplier satisfies, for some , and all multiindices ,This guarantees smoothing, compactness, and Schatten-class membership of the resulting operators.
Appendix A.2. Geometric and Operator Hypotheses (Chern/Index Arguments)
- (G1)
- The operator families considered (Laplace-type or elliptic pseudodifferential operators on M) are essentially self-adjoint, classical elliptic of positive order, and have discrete spectrum with .
- (G2)
- Heat-kernel expansion and zeta continuation. As ,with local invariants (curvature, symbol coefficients). The spectral zeta function admits meromorphic continuation to with only simple poles at prescribed locations. These hypotheses are standard (see Gilkey, Seeley, Connes–Moscovici) and ensure the analytic validity of index-theoretic and Chern-character identities.
Appendix A.3. Function-Space Hypotheses
- (F1)
- The anisotropic smoothness vector satisfies for all j whenever embedding into continuous functions is required (matching Theorem 3 of the main text). In the presence of critical indices , one either excludes that index from embedding claims or strengthens hypotheses (via VMO/logarithmic refinements).
Appendix A.4. Auxiliary Lemmas
Appendix A.5. Citation Guide
- Use Lemma A.1 when interchanging summation and integration in asymptotic expansions.
- For Voronovskaya-type expansions, state explicitly the dependence on moments and invoke (H1)–(H3) to bound remainders.
- For spectral/zeta manipulations, cite (G1)–(G2) and refer to Appendix B for detailed spectral-analytic background.
References
- Li, Z., Kovachki, N., Azizzadenesheli, K., Liu, B., Bhattacharya, K., Stuart, A., & Anandkumar, A. (2020). Fourier neural operator for parametric partial differential equations. arXiv preprint arXiv:2010.08895. [CrossRef]
- Lu, L., Jin, P., Pang, G., Zhang, Z., & Karniadakis, G. E. (2021). Learning nonlinear operators via DeepONet based on the universal approximation theorem of operators. Nature machine intelligence, 3(3), 218-229. Nature Machine Intelligence, 3(3), 218–229. [CrossRef]
- Serrano, L., Le Boudec, L., Kassaï Koupaï, A., Wang, T. X., Yin, Y., Vittaut, J. N., & Gallinari, P. (2023). Operator learning with neural fields: Tackling pdes on general geometries. Advances in Neural Information Processing Systems, 36, 70581-70611.
- Li, Z., Huang, D. Z., Liu, B., & Anandkumar, A. (2023). Fourier neural operator with learned deformations for pdes on general geometries. Journal of Machine Learning Research, 24(388), 1-26. https://www.jmlr.org/papers/v24/23-0064.html.
- Wu, H., Weng, K., Zhou, S., Huang, X., & Xiong, W. (2024, August). Neural manifold operators for learning the evolution of physical dynamics. In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (pp. 3356-3366). [CrossRef]
- Kumar, S., Nayek, R., & Chakraborty, S. (2024). Neural Operator induced Gaussian Process framework for probabilistic solution of parametric partial differential equations. Computer Methods in Applied Mechanics and Engineering, 431, 117265. [CrossRef]
- Luo, D., O’Leary-Roseberry, T., Chen, P., & Ghattas, O. (2023). Efficient PDE-constrained optimization under high-dimensional uncertainty using derivative-informed neural operators. arXiv preprint arXiv:2305.20053. [CrossRef]
- Molinaro, R., Yang, Y., Engquist, B., & Mishra, S. (2023). Neural inverse operators for solving PDE inverse problems. arXiv preprint arXiv:2301.11167. [CrossRef]
- Middleton, M., Murphy, D. T., & Savioja, L. (2025). Modelling of superposition in 2D linear acoustic wave problems using Fourier neural operator networks. Acta Acustica, 9, 20. [CrossRef]
- Bouziani, N., & Boullé, N. (2024). Structure-preserving operator learning. arXiv preprint arXiv:2410.01065. [CrossRef]
- Sharma, R., & Shankar, V. (2024). Ensemble and Mixture-of-Experts DeepONets For Operator Learning. arXiv preprint arXiv:2405.11907. [CrossRef]
- Lanthaler, S., Mishra, S., & Karniadakis, G. E. (2022). Error estimates for deeponets: A deep learning framework in infinite dimensions. Transactions of Mathematics and Its Applications, 6(1), tnac001. [CrossRef]
- Alesiani, F., Takamoto, M., & Niepert, M. (2022). Hyperfno: Improving the generalization behavior of fourier neural operators. In NeurIPS 2022 Workshop on Machine Learning and Physical Sciences.
- Tran, A., Mathews, A., Xie, L., & Ong, C. S. (2021). Factorized fourier neural operators. arXiv preprint arXiv:2111.13802. [CrossRef]
- Long, D., Xu, Z., Yuan, Q., Yang, Y., & Zhe, S. (2024). Invertible fourier neural operators for tackling both forward and inverse problems. arXiv preprint arXiv:2402.11722. [CrossRef]
- Triebel, H. (1983). Theory of function spaces, Birkhauser, Basel.
- Bourgain, J., & Demeter, C. (2015). The proof of the l 2 decoupling conjecture. Annals of mathematics, 351-389. https://www.jstor.org/stable/24523006.
- Hansen, M. (2010). Nonlinear approximation and function space of dominating mixed smoothness (Doctoral dissertation). https://nbn-resolving.org/urn:nbn:de:gbv:27-20110121-105128-4.
- Runst, T., & Sickel, W. (2011). Sobolev spaces of fractional order, Nemytskij operators, and nonlinear partial differential equations (Vol. 3). Walter de Gruyter.
- DeVore, R. A., & Lorentz, G. G. (1993). Constructive approximation (Vol. 303). Springer Science & Business Media.
- Butzer, P. L., & Nessel, R. J. (1971). Fourier analysis and approximation, Vol. 1. Reviews in Group Representation Theory, Part A (Pure and Applied Mathematics Series, Vol. 7). [CrossRef]
- Schmeisser, H. J., & Triebel, H. (1987). Topics in Fourier analysis and function spaces. (No Title).
- Rômulo Damasclin Chaves Dos Santos, Jorge Henrique de Oliveira Sales. Neural Operators with Hyperbolic-Modular Symmetry: Chern Character Regularization and Minimax Optimality in Anisotropic Spaces. 2025. https://hal.science/hal-05199221.
- Dai, F. (2013). Approximation theory and harmonic analysis on spheres and balls.
- Baez, J. C. (2019). Foundations of Mathematics and Physics One Century After Hilbert: New Perspectives.
- Moscovici, H. (2010). Local index formula and twisted spectral triples. Quanta of maths, 11, 465-500.
- Tsybakov, A. B. (2008). Nonparametric estimators. In Introduction to Nonparametric Estimation (pp. 1-76). New York, NY: Springer New York.
- Meyer, Y. (1992). Wavelets and operators (No. 37). Cambridge university press.






| Feature | ONHSH | FNO | Geo-FNO | Classical |
|---|---|---|---|---|
| Anisotropic Adaptivity | yes | no | no | no |
| Curved Domain Support | yes | no | yes | no |
| Modular Spectral Control | yes | no | no | no |
| Theoretical Guarantees | yes | no | no | no |
| Hyperbolic Symmetry | yes | no | no | no |
| Minimax-Optimal Rates | yes | no | no | no |
| Operator | MAE | MSE | RMSE |
|---|---|---|---|
| Geo-FNO | |||
| ONHSH | |||
| FNO | – | – | |
| NOGaP | |||
| Conv. | |||
| Gaussian |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).