1. Introduction
The intersection of
degenerate partial differential equations (PDEs),
neural network symmetrization, and
turbulence modeling offers a rich ground for mathematical innovation. Recent advances in
inverse problems for degenerate PDEs [
1] and
physics-informed neural networks (PINNs) [
2] have enabled the integration of
adaptive diffusion mechanisms into machine learning models. Yet, a
unified theoretical framework connecting these areas remains largely unexplored.
1.1. Degenerate PDEs and Inverse Problems
Degenerate PDEs naturally arise in contexts such as
anisotropic diffusion [
3],
geometric singularities [
4], and
phase transitions [
5]. Cannarsa et al. [
1] established
Lipschitz stability for reconstructing degeneracy points in parabolic equations of the form
from boundary measurements of
. Previous works on
inverse source problems [
6] and
coefficient identification [
7] mainly addressed
scalar, one-dimensional domains, leaving multi-dimensional, vector-valued problems largely uncharted.
1.2. Neural Symmetrization
Equivariant neural networks [
8] and the broader field of
geometric deep learning [
9] embed symmetry principles into machine learning models. Conventional group-convolution approaches [
10] struggle with
continuous symmetries such as
and
anisotropic phenomena, including turbulent shear layers. Neural operators [
11] and PINNs [
2] offer promise for turbulence modeling, but often lack
structural guarantees like rotation equivariance, energy conservation, or adaptivity to localized singularities.
1.3. Turbulence Modeling
Classical turbulence models—
LES [
12] and
RANS [
13]—rely heavily on
empirical closures, which poorly capture
intermittency and
anisotropic dissipation. Data-driven neural closures [
14,
15] improve predictive performance but can violate fundamental physical constraints. Our approach enforces incompressibility via
where
is a
degeneracy-aware neural operator designed to respect the underlying PDE structure.
1.4. Contributions
We introduce spectral degeneracy operators (SDOs)—differential operators that encode both physical symmetries and adaptive singularities—and demonstrate their application to:
Neural symmetrization, through SDO-based activation and layer design,
Turbulence closure modeling, via data-driven calibration and spectral filtering,
Inverse problem formulation, for reconstructing degeneracy points from sparse or boundary observations.
The key theoretical contributions of this work are:
Generalized spectral decomposition for vector-valued SDOs (
Section 2),
Lipschitz stability results for inverse calibration in turbulence models (
Section 3),
A
neural-turbulence correspondence theorem, connecting learned SDO parameters to underlying turbulent structures (
Section 3).
2. Spectral Degeneracy Operators (SDOs)
2.1. Definition and Spectral Properties
Let
be a bounded Lipschitz domain. Denote by
the
degeneracy points and
the
degeneracy exponents. Define the
spectral degeneracy operator (SDO) as
where
The natural energy space associated with
is
with inner product
Theorem 2.1 (Spectral Decomposition of SDOs).
Let and . Then the operator
with domain , isself-adjoint,positive semi-definite, and has acompact resolventon .
Moreover, there exists a countable set of eigenpairs forming an orthonormal basis of such that
where each satisfies the 1D Bessel-type Sturm–Liouville problem
The eigenvalues admit the asymptotic behavior
where denotes the -th positive zero of the Bessel function of the first kind.
Proof. Assume
. Then the operator acts as
Thus, the
d-dimensional spectral problem reduces to
d decoupled 1D weighted Sturm–Liouville problems (
8).
Each 1D operator
is self-adjoint with respect to the weighted inner product
The operator is positive semi-definite because
Compactness of the resolvent follows from standard embeddings
for singular-weighted Sobolev spaces [
1].
By tensorizing the 1D eigenfunctions
, we obtain
with eigenvalues
Orthonormality in
follows from the product structure and orthonormality of each 1D set.
For each 1D Sturm–Liouville problem with weight
, standard singular Bessel function theory [
16] implies
Summing over
i yields the asymptotic formula (
9).
Compact resolvent and orthonormality guarantee that forms a complete orthonormal basis of , establishing the spectral decomposition. □
2.2. Neural Symmetrization via SDOs
Define
SDO layers by
with trainable
and
. Let
be the Green’s function:
Definition 2.2 (SDO-Net).
AnSDO-Net
contains layers of the form
with Lipschitz activation σ, weight , bias , and trainable .
Lemma 2.3 (Weighted SDO Solve).
Let and . For any , the boundary-value problem
admits a unique solution , where the weighted Sobolev space is defined as
Moreover, there exists a constant such that
Proof. Consider the bilinear form
The weak form of (
13) is: find
such that
By definition,
Also,
shows coercivity.
The Lax–Milgram theorem then guarantees a unique
solving (
15).
From (
15) with
and Cauchy–Schwarz inequality:
Using a weighted Poincaré inequality [
1]:
we obtain
which proves (
14).
Alternatively, using the spectral decomposition
from Theorem 2.1, one can write
which also satisfies the same norm estimate due to orthonormality and positivity of
. □
Theorem 2.4 (Well-Posedness of SDO Layers).
Let , and let the SDO layer be defined by
where σ is Lipschitz continuous with constant , is a linear operator, and . Then:
Proof. From Lemma 2.3, for any
, the problem
admits a unique solution
with
Setting
, we obtain existence and uniqueness of
before activation.
Applying the Lipschitz continuous activation
with constant
gives
Using Lemma 2.3:
which yields (
17).
Let
and
. Then
Taking the
norm and applying the stability estimate from Lemma 2.3:
Since
depends continuously on
and
, we have
which establishes (
18). □
Lemma 2.5 (Spectral Decomposition of SDO Layer).
Let be the SDO at layer l as in Theorem 2.1, and let . Then admits the spectral decomposition
and the action of the operator satisfies
where are the eigenpairs of as in Theorem 2.1.
Proof. From Theorem 2.1, is self-adjoint, positive semi-definite, and has compact resolvent in . Therefore, the spectral theorem for compact self-adjoint operators applies, ensuring the existence of a countable orthonormal basis of eigenfunctions with corresponding eigenvalues .
Since
forms an orthonormal basis of
, any
can be expanded as
with convergence in
. This proves (
19).
Applying
to the expansion termwise gives
where we used the eigenvalue equation
. Convergence is in
.
The decomposition allows efficient representation of SDO layers and is crucial for spectral analysis of deep networks. The dependence of and on and is continuous due to the parameter-dependence properties of SDOs (see Theorem 2.1 and Theorem 2.4). This completes the proof. □
Corollary 2.6 (Spectral Stability under Training).
Let be a bounded Lipschitz domain. For each layer l, denote by
acting on , and let be the eigenvalues enumerated in nondecreasing order with associated orthonormal eigenfunctions . Assume that between two training steps t and the parameters change by
Then, for each fixed ,
In particular, the spectrum of each mode remains stable under sufficiently small parameter updates during training.
Remark 2.7 (Physics-Informed Kernels and Modewise Filtering).
Using the Green’s function representation, each layer admits the expansion
Thus each SDO layer acts as a modewise filter whose anisotropy adapts to the learned parameters. When the eigenvalues vary smoothly with , the gains of the dominant modes change only by for small updates, ensuring stable physics-informed filtering across layers.
3. Main Results
3.1. Universality for Turbulence Closure
Theorem 3.1 (Universality of SDO-Nets).
Let be a bounded Lipschitz domain, and let be a turbulence closure operator mapping velocity fields to stress tensors. Then, for any compact set and any , there exists an SDO-Net such that
with weakly divergence-free:
Proof. By Theorem 2.1, the set of eigenfunctions
of the spectral degeneracy operator
forms a complete orthonormal basis of
. Thus, for each component
of the closure tensor, and for any
, we can approximate
with coefficients
and error arbitrarily small in
norm by choosing
N sufficiently large.
Each coefficient map
is a continuous functional on the compact set
. By the **universal approximation theorem** for feedforward networks [
17], for any
, there exists a neural network
such that
Let
be any tensor field. By the Hodge decomposition (or Helmholtz projection) in bounded Lipschitz domains, there exists a unique decomposition
with
and potential
vanishing on
. Applying this projection to the neural approximation of the spectral expansion (
26), we obtain a divergence-free SDO-Net:
where
denotes the
-orthogonal projection onto divergence-free tensor fields.
By triangle inequality,
Each term can be made smaller than
by choosing sufficiently large
N and accurate neural networks
. Hence
satisfies (
24) and is weakly divergence-free by construction (
29).
□
Remark 3.2 (Interpretation). This theorem shows that **SDO-Nets provide a universal approximator** for turbulence closure operators while exactly enforcing incompressibility. The spectral degeneracy basis allows adaptive representation of anisotropic and localized structures, critical for turbulent flows.
3.2. Inverse Calibration of Degeneracy Points
Theorem 3.3 (Lipschitz Stability of Degeneracy Points).
Let be a bounded Lipschitz domain, and let be solutions to the degenerate Navier–Stokes system
with degeneracy points and identical initial/boundary conditions.
Assume that boundary measurements satisfy
Then there exist constants and , depending on Ω, , and the spectral gap of the SDO, such that
Proof. We outline a rigorous four-step argument:
Set
. Subtracting the two PDEs (
31) gives
Multiply (
34) by
and integrate over
:
where
is the nonlinear advection term.
Apply Cauchy–Schwarz and Young inequalities:
By Theorem 2.1, the SDO has a positive first eigenvalue
:
Combining (
35)–(
37) and bounding the nonlinear term via
, we obtain a Grönwall inequality:
Using the boundary measurement condition (
32) and standard elliptic estimates (or Carleman inequalities for SDOs [
1]), we can control the interior norm
in terms of
. Thus, solving (
38) gives
for some
, completing the proof. □
Remark 3.4 (Interpretation). This result guarantees Lipschitz-type stability for inverse calibration of degeneracy points: small changes in boundary fluxes lead to controlled shifts in the inferred centers . It forms the theoretical foundation for learning adaptive attention centers in SDO-Nets from partial or boundary measurements.
3.3. Neural-Turbulence Correspondence
Theorem 3.5 (Neural-Turbulence Correspondence – Refined).
Let be an SDO-Net trained to minimize the residual energy functional
Assume:
The dataset is dense in the function space of resolved velocities as .
The SDO-Net satisfies the Lipschitz stability property from Theorem 3.3.
The loss functional is equi-coercive and lower semicontinuous with respect to the degeneracy points .
Then, as , the learned degeneracy points converge to the true turbulence structures in :
Proof. The proof is based on a three-step argument combining consistency, stability, and variational convergence:
By the Germano identity [
18], the true subgrid stress satisfies
ensuring that the SDO-Net can approximate the exact residual as
.
Applying Theorem 3.3 to the mapping
guarantees that small residual errors in
induce controlled deviations in the inferred
:
Let
. By equi-coercivity and lower semicontinuity,
Combining (
43) and (
44) yields
Hence, the SDO-Net learns degeneracy points that asymptotically align with true turbulent structures. □
Remark 3.6 (Practical Implication). The theorem rigorously justifies the use of trainable degeneracy points in SDO-Nets: during large-data training, these points act as learned attention centers, automatically localizing key turbulence structures in a physically consistent way.
4. Results
We summarize here the main theoretical and computational contributions developed in the preceding sections:
Spectral theory of SDOs. We proved self-adjointness, compact resolvent, and a complete tensor-product basis of eigenfunctions with Bessel-type asymptotics, enabling efficient spectral representations of SDO layers.
Stability and inverse calibration. Using min–max principles and Carleman-type estimates we derived Lipschitz-type stability bounds for recovering degeneracy points from boundary measurements in degenerate Navier–Stokes systems.
Universal approximation for turbulence closure. We established that SDO-Nets approximate arbitrary divergence-free closure operators on compact subsets of , combining spectral expansions with neural coefficient maps.
Neural–turbulence correspondence. We proved convergence of trainable degeneracy points to true turbulent structures as the amount of training data increases, rigorously justifying their interpretation as learned attention centers.
These results jointly demonstrate that SDO-based architectures retain the expressive power of modern neural operators while inheriting structural guarantees from the underlying PDEs.
5. Conclusions
We have introduced and analyzed Spectral Degeneracy Operators as a bridge between degenerate PDE theory and physics-informed neural networks. Our analysis shows that SDO layers act as stable, adaptive, and interpretable spectral filters capable of representing anisotropic and intermittent features of turbulence without violating conservation laws. The inverse stability results provide a theoretical foundation for data-driven calibration of degeneracy points, and the universality theorem establishes the approximation power of SDO-Nets for turbulence closures. Future work will extend these ideas to fully three-dimensional, time-dependent turbulent flows, incorporate stochastic parameterizations, and explore SDO-based architectures in other domains such as geophysical fluid dynamics and nuclear-reactor thermo-hydraulics.
Symbols and Nomenclature
|
Bounded Lipschitz domain. |
|
Spatial coordinates. |
|
Degeneracy (attention) centers — trainable layer parameters. |
|
Degeneracy exponents controlling anisotropy. |
|
Diagonal degeneracy matrix . |
|
Spectral degeneracy operator on . |
|
Weighted Sobolev space associated with . |
|
k-th eigenvalue of (nondecreasing order). |
|
Corresponding eigenfunction, orthonormal in . |
|
Green’s function solving with homogeneous Dirichlet boundary. |
|
Input / output at layer l of an SDO-Net. |
|
Linear weight operator and bias at layer l. |
|
Lipschitz activation function. |
|
True turbulence closure operator (velocity ↦ stress tensor). |
|
SDO-Net approximation of . |
|
Helmholtz–Hodge projection onto divergence-free tensor fields. |
|
k-th positive zero of the Bessel function . |
|
Update norm between training steps:
|
Notation. Bold lowercase denotes vectors; bold uppercase denotes operators or matrices. Inner product . Norms default to unless otherwise stated. Superscript denotes the training iteration index.
Acknowledgments
Santos gratefully acknowledges the support of the PPGMC Program for the Postdoctoral Scholarship PROBOL/UESC nr. 218/2025. Sales would like to express his gratitude to CNPq for the financial support under grant 30881/2025-0.
References
- Cannarsa, P., Doubova, A., & Yamamoto, M. (2024). Reconstruction of degenerate conductivity region for parabolic equations. Inverse Problems, 40(4), 045033. [CrossRef]
- Raissi, M., Perdikaris, P., Karniadakis, G. E., & Shrimali, B. (2021). CS598: Physics-Informed Neural Networks: A deep learning framework for solving forward and inverse problems involving nonlinear PDEs.
- DiBenedetto, E. (2012). Degenerate parabolic equations. Springer Science & Business Media.
- Díaz, J. I. (1985). Nonlinear partial differential equations and free boundaries. Elliptic Equations. Research Notes in Math., 1, 106.
- Oleinik, O. (2012). Second-order equations with nonnegative characteristic form. Springer Science & Business Media.
- Hussein, M. S., Lesnic, D., Kamynin, V. L., & Kostin, A. B. (2020). Direct and inverse source problems for degenerate parabolic equations. Journal of Inverse and Ill-Posed Problems, 28(3), 425-448. [CrossRef]
- Kamynin, V. L. (2018). On inverse problems for strongly degenerate parabolic equations under the integral observation condition. Computational Mathematics and Mathematical Physics, 58(12), 2002-2017. [CrossRef]
- Finzi, M., Stanton, S., Izmailov, P., & Wilson, A. G. (2020, November). Generalizing convolutional neural networks for equivariance to lie groups on arbitrary continuous data. In International conference on machine learning (pp. 3165-3176). PMLR.
- Bronstein, M. M., Bruna, J., LeCun, Y., Szlam, A., & Vandergheynst, P. (2017). Geometric deep learning: going beyond euclidean data. IEEE Signal Processing Magazine, 34(4), 18-42. [CrossRef]
- Cohen, T., & Welling, M. (2016, June). Group equivariant convolutional networks. In International conference on machine learning (pp. 2990-2999). PMLR.
- Li, Z., Kovachki, N., Azizzadenesheli, K., Liu, B., Bhattacharya, K., Stuart, A., & Anandkumar, A. (2020). Neural operator: Graph kernel network for partial differential equations. arXiv preprint arXiv:2003.03485. arXiv:2003.03485. [CrossRef]
- Sagaut, P. (2006). Large eddy simulation for incompressible flows: an introduction. Berlin, Heidelberg: Springer Berlin Heidelberg.
- Cant, S., & Pope, S. B. (2000). Turbulent Flows.
- Beck, A. D., Flad, D. G., & Munz, C. D. (2018). Deep neural networks for data-driven turbulence models. arXiv preprint arXiv:1806.04482. arXiv:1806.04482. [CrossRef]
- Xiao, M. J., Yu, T. C., Zhang, Y. S., & Yong, H. (2023). Physics-informed neural networks for the Reynolds-Averaged Navier–Stokes modeling of Rayleigh–Taylor turbulent mixing. Computers & Fluids, 266, 106025.
- Moore, C. N. (1945). GN Watson, A treatise on the theory of Bessel functions.
- Cybenko, G. (1989). Approximation by superpositions of a sigmoidal function. Mathematics of control, signals and systems, 2(4), 303-314. [CrossRef]
- Germano, M., Piomelli, U., Moin, P., & Cabot, W. H. (1991). A dynamic subgrid-scale eddy viscosity model. Physics of fluids a: Fluid dynamics, 3(7), 1760-1765. [CrossRef]
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).