1. Introduction
Curvature is one of the central notions of Riemannian geometry and plays a distinguished rôle in both mathematics and physics. In its classical incarnation, sectional curvature measures the second-order deviation of geodesics in a two-dimensional direction, while scalar curvature is obtained by averaging over all such directions. From a more global perspective, curvature can also be understood in terms of holonomy: transporting a vector around a small closed loop and comparing the result with the original vector. In a flat space the transported vector returns to itself, whereas in a curved space a non-trivial discrepancy appears, and this discrepancy encodes the curvature of the underlying connection.
In parallel, information geometry has developed a rich differential-geometric framework on spaces of probability distributions and quantum states, where Riemannian metrics arise from statistical divergences such as the Kullback–Leibler or Jensen–Shannon divergences. The associated Fisher metrics and their variants have been used to define information-geometric analogues of geodesics, connections and curvature, with applications ranging from statistics and machine learning to quantum information theory; see, e.g., [
1].
These two viewpoints—Riemannian curvature and information geometry—suggest a natural question:
Can curvature be reconstructed, or even defined, purely in terms of informational data and their transformations, without a priori access to a smooth Riemannian structure?
In the present paper we do not attempt to solve this reconstruction problem in full generality. Instead, we assume a smooth Riemannian manifold and a state bundle with a compatible connection as part of the input, and show how their curvature can be encoded and approximated in purely informational terms.
This question has received substantial attention in the context of discrete and non-smooth geometries. In the setting of graphs and Markov chains, several notions of “discrete curvature” have been introduced, most notably the coarse Ricci curvature à la Ollivier [
7], Forman’s combinatorial curvature [
4], and Regge-type curvature concentrated on simplices in piecewise flat manifolds [
8]. These constructions either derive curvature from a discrete metric structure or from combinatorial and measure-theoretic data. In all cases, an important theme is the
consistency problem: under suitable sampling or refinement hypotheses, do these discrete curvatures converge to the classical Riemannian curvatures of a smooth limit space?
In previous work, the author considered the following situation: given a family of local informational states attached to points of a Riemannian manifold , one can endow a sampling graph with the Jensen–Shannon metric induced by the pairwise divergence of these states. A Regge-type scalar curvature estimator built from this metric was shown, under explicit metric and sampling assumptions, to converge in the weak-* sense to the scalar curvature of g. This provides a scalar curvature estimator derived purely from a non-geodesic, informational metric.
The present paper aims at a refinement of a different nature. Instead of constructing curvature from distances alone, we introduce a new notion of curvature that is based on holonomy of informational channels. Concretely, we consider a discrete sampling of a manifold in which each vertex carries a space of informational states (classical or quantum), and each oriented edge is equipped with a channel transporting states between neighbouring vertices. By composing these channels along small loops we obtain a discrete notion of holonomy acting on the state space at a base point. We then measure how far this holonomy deviates from the identity, using an informational divergence such as the Jensen–Shannon divergence and the associated informational distance , and normalize by the geometric area of the loop. The resulting quantity is what we call the informational holonomy curvature.
Intuitively, one can think of this construction as follows. At each point we choose a reference state of the system. We then let this state travel along a small triangular loop by applying, in sequence, the channels assigned to each edge. If the underlying “informational geometry” were flat, we would expect the state to come back unchanged. Any systematic deviation between the initial state and the state obtained after completing the loop signals the presence of curvature. By comparing these two states with a suitable divergence (or equivalently, with the associated informational distance) and normalizing by the area of the loop, we obtain a discrete curvature associated with that loop. By averaging over families of loops that approximate a given two-dimensional direction, we obtain a quantity that plays the rôle of a sectional curvature in an informational setting.
From the Riemannian viewpoint, this is reminiscent of the classical description of sectional curvature via holonomy of the Levi–Civita connection in the tangent bundle. The novelty here is that the objects being transported are not tangent vectors but informational states, and the parallel transport is implemented by channels rather than by a linear connection on a vector bundle. Nevertheless, we shall show that, under appropriate assumptions, the resulting informational holonomy curvature can be expressed in terms of the curvature of a connection on a state bundle over , and in geometric models induced by the Levi–Civita connection it reduces, in spaces of constant curvature, to a constant multiple of .
Main Contributions
We now summarize the main contributions of this work.
-
We introduce a general framework in which to study informational holonomy on a discrete sampling of a Riemannian manifold. The basic data consist of:
a sampling graph embedded in a manifold ,
a state space attached to each vertex (for instance, a space of probability distributions or density matrices),
and a family of channels associated with each oriented edge .
Composition of these channels along loops in gives rise to discrete holonomy operators on the fibres .
-
Given a divergence
D on each state space
(in particular, the Jensen–Shannon divergence), we define the
informational holonomy defect of a loop
based at
x as
where
is the holonomy operator obtained by composing channels along
, and
is a reference state at the base point
x. We also consider the associated distance defect
For sufficiently small loops bounding an area
, we define the
informational holonomy curvature of the loop by
By averaging over discrete loops that approximate a two-plane , we obtain an informational sectional curvature at scale .
On the continuum side, we consider a smooth Riemannian manifold
equipped with a smooth bundle of state spaces
and a connection-like structure which assigns, to nearby points
, channels
compatible with
g. Under natural regularity and compatibility conditions, we define a
continuous informational holonomy curvature by transporting a reference state around infinitesimal loops tangent to a given two-plane
and measuring, via a divergence, the infinitesimal deviation from the identity. Our first main result shows that this quantity can be written as
where
depends linearly on the curvature of the connection on
along
. In particular, when the connection is induced from the Levi–Civita connection via a linear isometric representation,
is a scalar invariant of the restriction of
to
, and in spaces of constant sectional curvature it is proportional to
.
Second, building on previous work on scalar curvature estimators derived from informational metrics, we show that the discrete informational sectional curvature
converges, as
, to the continuous quantity
above. More precisely, we obtain an estimate of the form
where
is the error scale appearing in Theorem 2 (see (
10)). Here
is the sampling radius,
and
quantify directional anisotropy and area-approximation errors of the triangle families, and
encodes the channel-consistency error from Assumption 9(2). In particular, in spaces of constant curvature this yields convergence to a constant multiple of
.
Finally, we discuss several model constructions and examples. In particular, we consider classical Fisher-type models in which the state bundle is induced by a statistical model on M, and we construct channels by transporting distributions along short geodesic segments. On spaces of constant curvature, we show how the resulting informational holonomy curvature reproduces, up to a constant factor, the expected constant sectional curvature.
Together, these results provide what we view as a natural notion of curvature in a setting where the primary objects are informational states and channels rather than tangent vectors and linear connections. In contrast with purely metric-based discrete curvatures, the informational holonomy curvature fundamentally exploits the dynamical aspect of information transport.
Relation to Previous Work
This work lies at the interface between several active research directions.
First, in Riemannian geometry and general relativity, Regge calculus and its refinements provide discrete models of curvature in piecewise flat manifolds, where curvature is concentrated on codimension-2 simplices and can be described in terms of deficit angles and holonomy [
8]. Second, in the study of graphs and Markov chains, discrete Ricci and scalar curvatures have been introduced using optimal transport, entropy convexity, and combinatorial Laplacians [
4,
7]. Third, information geometry provides natural Riemannian metrics and connections on spaces of probability measures and quantum states, together with associated notions of curvature [
1].
Our construction is directly inspired by the Regge and holonomy viewpoints, but it is formulated in an intrinsically informational setting. Rather than starting from a discrete metric or Laplacian, we start from a network of informational channels and a divergence on each fibre. At the technical level, our convergence results build on discrete-to-continuum analysis of curvature estimators on sampling graphs, including previous work on scalar curvature estimators derived from informational metrics.
A simple model to keep in mind throughout the paper is provided by the classical Jensen–Shannon divergence on probability simplices (
Section 7.1), which satisfies Assumption 1 and induces fibre metrics proportional to the Fisher information metric. This example shows that the abstract hypotheses on the fibre divergences are realised in a familiar information-geometric setting.
Organization of the Paper
The paper is organized as follows. In
Section 2 we recall the necessary background on Riemannian geometry, information geometry, and holonomy, and we set up the continuous model of state bundles and channels. In
Section 3 we introduce the discrete sampling framework, define discrete holonomy operators on state spaces, and state the main assumptions on sampling, metric approximation and channel regularity.
Section 4 contains the precise definitions of the informational holonomy defect and informational holonomy curvature, both at the level of individual loops and in the averaged sectional form.
In
Section 5 we formulate and prove our main continuous theorem relating informational holonomy curvature to the curvature of a connection on the state bundle.
Section 6 is devoted to the discrete-to-continuum convergence theorem. Finally,
Section 7 presents model constructions and examples, and
Section 8 discusses possible extensions and applications.
2. Geometric and Informational Background
In this section we recall the basic geometric notions and fix the continuous framework in which the informational holonomy curvature will be defined. We start with standard Riemannian preliminaries, then introduce the state bundle and divergences, and finally describe a continuous notion of informational transport and holonomy.
2.1. Riemannian Preliminaries
Let
be a smooth, connected, oriented Riemannian manifold of dimension
. We denote by
the associated geodesic distance, by
∇ the Levi–Civita connection of
g, and by
the Riemann curvature tensor, taken with the convention
For
and a 2-dimensional subspace
, the
sectional curvature of
g at
is defined by
where
is any basis of
with
. This definition is independent of the choice of basis.
We shall frequently use normal coordinate charts. For
we denote by
the exponential map at
x, defined on a maximal star-shaped neighbourhood
of 0 where
is a diffeomorphism onto its image. There exists
such that for every
the open ball
is contained in the normal neighbourhood
and is geodesically convex: any two points
are joined by a unique minimizing geodesic contained in
.
Given three points
sufficiently close to each other and contained in such a convex normal neighbourhood, there is a unique geodesic triangle
formed by the minimizing geodesic segments
. We denote by
the Riemannian area of
and by
,
and
the interior angles at
, respectively. The
angle defect of the triangle is
The following classical fact relates angle defect and sectional curvature; see, for example, [
2] [Chapter 6].
Lemma 1
(Angle defect and sectional curvature)
. Let be a smooth Riemannian manifold. For each compact set there exist constants and such that the following holds. For any and any geodesic triangle contained in , let be the plane spanned by the initial velocities of the geodesics from x to y and from x to z. Then
where and the implicit constant in the term is bounded by .
In particular, for sequences of triangles shrinking to x with diameters of order and areas of order , the ratio converges to .
Holonomy gives an alternative description of sectional curvature. Let
denote the holonomy group of the Levi–Civita connection at
x. For any piecewise smooth closed loop
based at
x there is an associated parallel transport operator
obtained by solving the parallel transport equation along
. If
bounds a small geodesic triangle tangent to a plane
, then one has the expansion
where
is the area of the triangle, and
is a linear map depending linearly on the curvature tensor
restricted to
; see, e.g., [
5] [Chapter II]. We do not need the precise form of
; it will be enough to assume the existence of expansions of the form (
2) in the informational setting below.
2.2. State Spaces and Informational Divergences
We now recall basic notions from information geometry. For the purposes of this work it is convenient to formulate the discussion at the level of a general smooth state space, although throughout we keep classical probability distributions as a canonical example.
Definition 1
(State space). A state space is a smooth manifold S whose points represent informational states of a system. Typical examples include:
-
(i)
An open subset of the simplex of strictly positive probability vectors on a finite set :
-
(ii)
An open submanifold of the space of faithful density matrices on a finite-dimensional Hilbert space (quantum states).
We endow S with two pieces of structure:
The divergence D is not assumed to be a distance (it may fail to be symmetric and may not satisfy the triangle inequality), but we require that it induces the Riemannian metric in the usual way.
Assumption 1
(Divergence and information metric). We assume that D is of class in a neighbourhood of the diagonal and satisfies:
-
1.
for all ;
-
2.
for all (vanishing first derivative in the second argument along the diagonal);
-
3.
the Hessian of D in the second argument at the diagonal recovers the Riemannian metric , i.e.
The last condition means that for
t close to
s one has the second-order expansion
where
is any tangent vector such that
in a normal coordinate chart on
S.
Remark 1
(Jensen–Shannon divergence)
. A central example in this work is the Jensen–Shannon divergence
on the simplex of probability vectors on a finite set; see [6]. Let and let
denote the open probability simplex. For , the Jensen–Shannon divergence is defined by
where is the Shannon entropy (with a fixed choice of logarithm). It is well known that is symmetric and non-negative and that
defines a genuine metric on [3]. In particular, with the normalization used throughout this paper,
is also a genuine metric on .
Moreover, the second-order expansion of around the diagonal induces a multiple of the Fisher information metric on . In other words, there exists a constant such that, for p fixed and q close to p,
where is as in (3). Thus Assumption 1 holds with proportional to the Fisher metric.
Assumption 1 ensures that the divergence D and the metric are compatible in the sense of information geometry: D is a “potential” whose Hessian yields the local quadratic structure. The precise constant relating D to will play no essential rôle; it will simply be absorbed into the constant c appearing in the main theorems.
2.3. State Bundles over a Riemannian Manifold
We now couple the Riemannian manifold with the state space S.
Definition 2
(State bundle)
. A state bundle
over is a smooth fibre bundle
with typical fibre S, together with:
for each , a smooth identification of the fibre with S;
a smooth family of fibrewise Riemannian metrics , where each is a Riemannian metric on obtained from a copy of under the identification ;
-
a smooth family of divergences , where each
satisfies Assumption 1 with respect to .
For notational simplicity, we often suppress the explicit identifications and denote the metric on by and the divergence by , keeping in mind that they vary smoothly with x.
Remark 2.
In the simplest classical setting, one may take to be the trivial bundle with fibre , endowed with the product smooth structure. Then each fibre is naturally identified with S, and one can set and independently of x. The non-trivial bundle case is more appropriate, for instance, in quantum settings or in statistical models where the parameterization of states varies with x.
The state bundle provides the configuration space for informational states over M. To speak about transport and holonomy of states, we need a notion of connection on .
2.4. Connections and Continuous Informational Transport
The key geometric input in the continuous theory is a connection producing parallel transport on the state bundle. In order to place the holonomy expansion used later on a fully standard footing, we henceforth restrict the continuous framework to the associated-bundle setting of principal connections.
Assumption 2
(Associated-bundle model and regularity). There exist:
-
(i)
A Lie group G acting smoothly on the typical fibre S by Riemannian isometries of . We denote the action by .
-
(ii)
A principal G-bundle endowed with a principal connection , with curvature .
-
(iii)
-
The state bundle is the associated bundle
and the connection on is the Ehresmann connection induced by ω.
Moreover, all geometric structures are understood on compact subsets of M, so that (after choosing local trivializations) the coefficients of ω, and their first derivatives are uniformly bounded on compact sets.
Under Assumption 2, the principal connection induces an Ehresmann connection on by declaring a vector in to be horizontal if it is represented by a pair with horizontal in (i.e. ) and in S. This yields a smooth horizontal distribution and, therefore, parallel transport along curves in M.
Definition 3
(Parallel transport in the associated state bundle)
. Let be a piecewise curve with and . Let and denote by the ω-horizontal lift of γ with . There exists a unique element such that . Then the induced parallel transport on is the map
Parallel transport satisfies the functorial properties
whenever the concatenations are defined. In particular, any piecewise smooth closed loop
based at
x yields a holonomy map
.
Lemma 2
(Parallel transport is a fibrewise Riemannian isometry)
. Assume the associated-bundle setting of Assumption 2, and that the G-action on is by Riemannian isometries. Then for every piecewise curve with , , the induced parallel transport map is a Riemannian isometry between the fibres:
In particular, is locally distance-preserving and globally distance-preserving on each fibre.
Proof. Fix and let be the -horizontal lift of with . By Definition 3, there is such that . Since acts by a Riemannian isometry of , it preserves the Riemannian distance on S, and hence preserves the Riemannian distance on the fibres and . □
Assumption 3
(Compatibility of divergences with transport)
. For each compact set there exist constants and such that the following holds. For every piecewise curve with and , and every satisfying , one has
where denotes the Riemannian distance in the fibre .
Remark 3.
If the fibre divergence is induced by a single divergence D on S which is G-invariant (i.e. for all ), then (5) holds with equality (hence ). In general we treat (5) as a modelling axiom controlling how the chosen divergences vary under parallel transport, with constants uniform on compact sets.
2.5. Informational holonomy in the continuous setting
In the continuous setting, holonomy of the state bundle is defined exactly as in the Riemannian case: for a piecewise smooth closed loop
based at
x we have a holonomy map
We interpret
as an
informational channel acting on the state space at
x: starting from a state
, one transports
s along
using the connection and compares the resulting state
with the original one.
To quantify the deviation from trivial holonomy, we introduce reference states and informational defects.
Definition 4
(Reference states)
. A reference state field
is a smooth section
of the state bundle, i.e. . For each , the point will serve as the base state with respect to which informational changes are measured.
Definition 5
(Continuous informational holonomy defect)
. Let μ be a reference state field and let γ be a piecewise smooth closed loop in M based at x. The informational holonomy defect
of γ at x is
where is the divergence on the fibre .
We also consider the associated (local) informational distance defect:
In this continuous framework we will be interested in loops which are the boundaries of small geodesic triangles. Let
and let
be a two-dimensional subspace. For
sufficiently small, we consider the geodesic triangle with vertices
and denote by
the corresponding closed loop obtained by traversing the geodesic segments
in order. Its Riemannian area is
.
Definition 6
(Continuous informational holonomy curvature)
. Let be as above. For and a two-dimensional subspace , consider geodesic triangles based at x and tangent to Π with area . We say that thecontinuous informational holonomy curvature
at exists if the limit
exists and is independent of the particular way in which the triangle shrinks to x within Π.
In later sections we will show, under appropriate assumptions, that this limit exists and is proportional to , with a constant factor depending on the chosen connection, divergence, and reference state field.
Remark 4.
The definition above is closely analogous to the classical definition of sectional curvature via holonomy of the Levi–Civita connection, with the crucial difference that we compare states in a non-linear state space using an informational divergence/distance, rather than tangent vectors using a linear norm. Assumptions 1 and 3 ensure that, to second order, the informational defect behaves quadratically in the infinitesimal holonomy, while the associated distance defect scales linearly in the holonomy displacement, hence is the natural object to normalize by area.
The continuous framework developed in this section provides the conceptual target for the discrete constructions that follow. In the next section we introduce sampling graphs, discrete channels and discrete holonomy operators, which will serve as the basis for our definition of discrete informational holonomy curvature.
3. Discrete Sampling, Channels and Holonomy
In this section we introduce the discrete framework that will serve as the basis for our definition of informational holonomy curvature. We consider sampling graphs embedded in a Riemannian manifold, discrete state spaces attached to vertices, channels attached to edges, and discrete holonomy operators obtained by composing channels along loops. The assumptions formulated here are discrete counterparts of the continuous structures described in
Section 2.
3.1. Sampling Graphs on a Riemannian Manifold
Let
be a smooth Riemannian manifold of dimension
. For each small parameter
we are given a finite subset
of sampling points and a simple undirected graph
where
is the set of edges. We denote by
the fact that
.
The vertex sets are assumed to be asymptotically dense and quasi-uniform in M, in the following sense.
Assumption 4
(Quasi-uniform sampling). There exist positive constants and a sequence as such that, for all sufficiently small ε:
-
1.
(Separation) For all distinct one has
-
2.
(Covering) For every there exists such that
Thus the sampling points form a Delone set at scale . We shall refer to as the sampling radius.
The edge set is assumed to connect points that are at distance of order .
Assumption 5
(Local connectivity). There exists a constant such that, for all sufficiently small ε, one has:
-
1.
If , then .
-
2.
The degrees of the graph are uniformly bounded: there exists such that every vertex satisfies
Under Assumptions 4 and 5, each connected component of has uniformly bounded local complexity and provides a reasonable discrete approximation of at scale . In particular, for sufficiently small, any point and any direction admit neighbours of x whose geodesic directions approximate up to an error of order .
In order to average quantities defined at vertices, we will occasionally associate a volume weight to each vertex .
Assumption 6
(Volume approximation). For each ε there exist positive weights , , such that:
-
1.
There exists independent of ε with
-
2.
For every one has the quadrature convergence
Assumption 6 is satisfied, for instance, if are defined as the Riemannian volumes of a Voronoi tessellation associated with .
3.2. Discrete State Spaces and Divergences
We now discretize the state bundle
introduced in
Section 2.3. For each sampling point
we consider a fibre
representing the possible states at
. In the simplest setting one may take
to be the fibre of the continuous state bundle at
, but we keep the notation
to emphasize the discrete nature of the sampling.
Assumption 7
(Discrete fibres and divergences). For each and each we are given:
-
1.
-
a smooth manifold and a smooth identification
with the continuous fibre at ;
-
2.
a Riemannian metric on obtained as the pullback of the metric under ;
-
3.
-
an informational divergence
such that, after transporting it to via , Assumption 1 holds uniformly in and ε.
For notational simplicity, we will usually drop the superscript and write , and , keeping in mind the dependence on through the underlying sampling set .
We also sample the continuous reference state field .
Assumption 8
(Discrete reference states)
. For each we are given a reference state such that, under the identification , one has
Thus the discrete fibres, metrics, divergences and reference states are obtained by restriction of the continuous state bundle to the sampling set .
3.3. Discrete Channels and Local Consistency
We now assign discrete informational channels to the edges of the sampling graph
. For each oriented edge
with
we are given a channel
Again, we often drop the superscript
when no confusion arises.
The channels are required to be local and to approximate the parallel transport maps in the continuous state bundle. Let denote the unique minimizing geodesic segment in joining to , which lies in a convex normal neighbourhood when is sufficiently small. By Assumption 5 we may (and do) assume that each edge length is bounded by a constant multiple of .
Assumption 9
(Locality and consistency of channels). There exist constants and such that, for all sufficiently small ε and all oriented edges :
-
1.
-
(Locality) The channel depends only on the geometry of in a neighbourhood of the geodesic segment and satisfies
for all , i.e. is (locally) Lipschitz with respect to the divergences.
-
2.
(Consistency with continuous parallel transport) Let be the continuous parallel transport map along the geodesic segment joining to . There exist constants and such that, for all ,
The exponent ensures that the error in approximating continuous parallel transport along an edge is of order strictly higher than the edge length, which will imply that the error in approximating holonomy around small loops is of order strictly higher than the area of the loop.
Remark 5.
In many concrete models one can take or , depending on how the channels are constructed from the underlying connection on . For the purposes of the convergence theorems, any suffices.
3.4. Discrete Holonomy Operators
Given the channels on edges, we can define discrete holonomy operators by composition along loops in the graph .
Definition 7
(Discrete paths and loops)
. A discrete path
in of length is a sequence of vertices
such that for all . The path is closed
or a loop
if .
We denote by the set of all discrete paths and by the set of all loops based at a given vertex .
To each oriented edge along a path we associate the channel . The channel associated with a path is the composition of the edge channels in order.
Definition 8
(Discrete transport and holonomy)
. Let be a discrete path in . The discrete transport
along γ is the map
If γ is a loop based at , i.e. , we call the discrete holonomy operator
of γ and denote it by
The discrete transport operators satisfy the obvious composition rules: if
and
are two paths with matching endpoint and starting point, then
If
denotes the reversed path, then
with the approximation becoming exact if the channels satisfy an exact involutive property. In our setting we will only need approximate inversion properties at the infinitesimal level, which follow from Assumption 9.
In the sequel we will focus on loops associated with small discrete triangles.
Definition 9
(Discrete triangles)
. A discrete triangle
in is an ordered triple of distinct vertices such that all three edges , and belong to . The associated oriented loop is
We denote by the set of all discrete triangles in .
For each discrete triangle
we define the corresponding holonomy operator
3.5. Triangle Geometry and Area Approximation
In order to compare discrete holonomy around triangles with continuous holonomy around geodesic triangles, we need to relate the combinatorial triangles in to small geodesic triangles in and to assign an appropriate area to each discrete triangle.
Given a discrete triangle
we consider the unique geodesic triangle
formed by the minimizing geodesics between
. For
sufficiently small, Assumption 5 ensures that all edge lengths
,
and
are bounded by a constant multiple of
, so that
is contained in a convex normal neighbourhood. We denote by
the Riemannian area of this geodesic triangle.
In purely informational settings one may not have direct access to , but for the analytical convergence results we only require that the area used in the discrete curvature is a good approximation of .
Assumption 10
(Area approximation)
. For each triangle we are given a non-negative number , called its discrete area
, such that there exists a sequence with
for all when ε is sufficiently small.
The factor reflects the typical area of a small triangle with edge lengths of order .
3.6. Families of Triangles Approximating Two-Planes
To define discrete sectional curvatures, we will need to average over families of discrete triangles that approximate a given point and two-plane in the manifold. This requires an additional isotropy assumption on the sampling.
Let be a fixed compact subset. For each and each two-dimensional subspace , we wish to consider families of discrete triangles based at vertices close to x whose geodesic realisations are small and whose edge directions approximate in an approximately isotropic fashion.
Assumption 11
(Directional sampling and triangle families). There exists a sequence such that the following holds. For each compact there exists such that, for all , for every and every two-dimensional subspace , one can choose:
Assumption 11 is a discrete isotropy condition ensuring that the graph carries enough small, uniformly non-degenerate triangles to sample each two-plane uniformly in the limit . It is analogous to assumptions used in the analysis of discrete Laplace–Beltrami operators and curvature estimators on random or quasi-uniform point clouds.
The structures and assumptions introduced in this section—sampling graphs, discrete fibres and divergences, channels, holonomy operators, triangle areas and isotropic triangle families—provide the discrete environment in which we will define the informational holonomy curvature. In the next section we use these ingredients to formulate precise discrete and continuous curvature quantities and to state the main convergence results.
4. Informational Holonomy Curvature: Definitions
In this section we define the informational holonomy defect and the associated curvature, both in the discrete setting of sampling graphs and in the continuous state-bundle setting. A key point, already implicit in Assumption 1, is that the divergence on each fibre induces a Riemannian metric and therefore a natural local distance. The curvature will be defined using this distance, which is linear in the holonomy displacement to first order, rather than the divergence itself, which only captures a quadratic effect.
4.1. Informational Distances and Defects for Discrete Loops
We begin by extracting from each fibre divergence a distance-like function.
Recall that, for each and each vertex , we have a fibre endowed with a Riemannian metric and a divergence satisfying Assumption 1 (after transport to the continuous fibre). We define:
Definition 10
(Discrete informational distance)
. For each vertex and we define the informational distance
For the purposes of the estimates below, one may replace by the Riemannian distance on , which is locally equivalent to by Lemma 4. We implicitly make this replacement whenever the triangle inequality is invoked.
In general is only guaranteed to be a local distance function near the diagonal (we do not assume the triangle inequality). In the Jensen–Shannon setting, coincides (up to the global constant factor ) with the usual Jensen–Shannon distance on the probability simplex.
We now assign informational defects to discrete loops. Recall that, for each loop
based at
, we have a holonomy operator
defined by composition of edge channels along
, and a reference state
.
Remark 6
(Riemannian vs. informational distance)
. For later estimates it will be convenient to work with the genuine Riemannian distance on each fibre , which we denote by . By Lemma 4 and compactness, and are locally equivalent: there exist constants such that, whenever is sufficiently small,
In particular, all notions of “defect” and “curvature” defined using are unchanged, up to uniform multiplicative constants, if one replaces by . From this point on, whenever the triangle inequality is invoked we implicitly work with ; the symbol may be read as either distance, since they are locally equivalent.
Lemma 3
(From divergence contraction to distance contraction)
. Let D be a nonnegative divergence and define . If a map Φ satisfies
then
Proof. Immediate from and monotonicity of the square root. □
Definition 11
(Discrete informational holonomy defects). Let be a discrete loop based at . We define:
For curvature purposes, and contain equivalent information, but the distance defect scales linearly with the holonomy displacement and is therefore the natural quantity to normalize by the area of small loops.
As in
Section 3.4, we now specialise to loops associated with discrete triangles. For a triangle
, with associated loop
and holonomy operator
we use the shorthand
Definition 12
(Triangle defects)
. For a discrete triangle , the divergence defect
and distance defect
of the triangle based at are given by
4.2. Discrete Informational Holonomy Curvature of Triangles
To obtain a curvature quantity from the defects, we normalize by the area associated with each triangle. Assumption 10 provides a discrete area which approximates the Riemannian area of the geodesic triangle with vertices .
Definition 13
(Triangle-wise informational holonomy curvature)
. Let be a discrete triangle. The informational holonomy curvature
of the triangle is
with the convention that if .
Thus measures the informational distance travelled by the reference state per unit area when it is transported around the discrete triangle .
In order to define a discrete sectional curvature at a point and a two-dimensional direction, we now average the triangle-wise curvature over suitable families of triangles.
Let
be a fixed compact set. For each
and each 2-plane
, Assumption 11 provides, for
small enough, a vertex
close to
x and a finite family of discrete triangles
which are small, non-degenerate, have edge directions close to
, and are approximately isotropically distributed in
.
Definition 14
(Discrete informational sectional curvature)
. Let be compact, , and a two-dimensional subspace. For small enough, let and be as in Assumption 11. The discrete informational sectional curvature
at scale ε associated with is
If the denominator vanishes, we set by convention.
In words, is the area-weighted average of the triangle-wise informational holonomy curvature over all triangles in .
Remark 7
(Choice of base vertex and triangle family)
. The definition of in (7) involves several auxiliary choices: for each and two-plane we pick a nearby vertex with and a finite family of triangles as in Assumption 11. A priori, different admissible choices could lead to different values of .
Under Assumptions 4–11, however, Lemmas 10 and 11 imply that any two such choices produce values that differ by at most , with as in (10). In particular, the limit in Theorem 2 is independent of these auxiliary choices.
4.3. Continuous Informational Holonomy Curvature Revisited
We now recall the continuous framework and align the definitions with the discrete case by introducing the corresponding informational distance on each fibre.
Let be a Riemannian manifold, a state bundle with fibre metrics and divergences as in Definition 2 and Assumption 1, equipped with a connection satisfying Assumption 3. Let be a smooth reference state field.
For each
and
, we define the continuous informational distance
By Assumption 1,
is locally equivalent to the Riemannian distance induced by
on the fibre (see Lemma 4 in
Section 5).
For
and a two-dimensional subspace
, consider pairs of tangent vectors
of sufficiently small norm and the associated geodesic triangle with vertices
Let
denote the closed loop
obtained by traversing the geodesic segments in order, and let
be its Riemannian area. The continuous holonomy map
is defined by parallel transport along
.
Definition 15
(Continuous informational holonomy defects)
. The continuous divergence defect
and continuous distance defect
of the loop at x are
We are interested in the behaviour of as the triangle shrinks to x within the plane .
Definition 16
(Continuous informational sectional curvature). Let be as above, and fix and a two-dimensional subspace . For sufficiently small we denote by the associated piecewise-geodesic loop and by its informational distance defect at x (cf. Definition 15).
We say that the continuous informational sectional curvature at exists if there are constants and and, for each , vectors such that:
-
1.
-
the corresponding geodesic triangle
is contained in a normal neighbourhood of x and its vertices converge to x as ;
-
2.
-
the triangle has uniformly non-degenerate shape in
, in the sense that
so that the side lengths and the area are uniformly comparable to r and , respectively, independently of r.
Whenever this holds and, for every such admissible family , the limit
exists and has the same value, we call this common value the continuous informational sectional curvature
at .
In
Section 5 we show, under Assumptions 1 and 3, that this limit exists for all
and all two-planes
, and that it can be expressed explicitly in terms of the curvature of the connection on
along
.
4.4. Main Curvature Theorems
We can now formulate the main results of this work, which will be proved in
Section 5 and
Section 6. The first theorem relates the continuous informational sectional curvature to the curvature of the connection on the state bundle. The second theorem shows that the discrete informational sectional curvature converges to this continuous quantity as the sampling becomes dense.
Throughout this subsection we fix a compact subset and tacitly restrict attention to points .
Theorem 1
(Continuous informational holonomy curvature). Let be a smooth Riemannian manifold, a state bundle with fibre metrics and divergences satisfying Assumption 1, endowed with a connection satisfying Assumption 3, and let be a smooth reference state field.
Then, for every and every two-dimensional subspace , the continuous informational sectional curvature exists in the sense of Definition 16 and can be written as
where is a vector depending linearly on the curvature of the connection on along Π. In particular, when the connection on is induced by the Levi–Civita connection via a linear isometric representation, is a scalar invariant built from the Riemann curvature tensor restricted to Π and is proportional to in spaces of constant sectional curvature.
Remark 8
(Dependence on the informational structure). The quantity should not be thought of as a curvature of the Riemannian manifold alone. It depends on the choice of state bundle , on the fibre divergences (equivalently, on the fibre metrics ), on the Ehresmann connection used to define parallel transport, and on the reference state field μ. Different choices over the same may lead to different informational holonomy curvatures. In particular, two bundles with the same base but different fibre geometries or connections need not produce the same values of .
We now turn to the discrete setting. In addition to the continuous structures above, we assume that the manifold is sampled by graphs and that discrete fibres, divergences, reference states, channels, areas and triangle families are given as in Assumptions 4–11.
For convenience, we introduce a sequence of positive numbers
that captures the various discretization errors. More precisely, we set
Here
is the sampling radius from Assumption 4,
and
are the anisotropy and area errors from Assumptions 11 and 10, and
controls the channel consistency scale induced by Assumption 9(2) with exponent
.
Theorem 2
(Discrete-to-continuous convergence of informational holonomy curvature). Let be as in Theorem 1, and let , together with discrete fibres, divergences, reference states, channels, areas and triangle families, satisfy Assumptions 4–11. Let be compact.
Then there exist constants and such that, for all , for every and every two-dimensional subspace , the discrete informational sectional curvature is well defined and satisfies
where is given by (10). In particular,
Consequently, as , the discrete informational sectional curvatures converge uniformly on compact subsets of M and on the Grassmannian of two-planes to the continuous informational sectional curvature .
The proof of Theorem 2, given in
Section 6, proceeds in two steps. First, we compare the discrete holonomy operators
around small triangles with the continuous holonomy operators associated with the corresponding geodesic triangles, using Assumption 9 and the local consistency of the sampling. Second, we exploit the isotropy of the triangle families from Assumption 11 to show that the area-weighted average (
7) converges to the continuous limit (
9), with an error controlled by
.
Overview of assumptions.
For the reader’s convenience we briefly summarise the rôle of the hypotheses. Assumption 1 requires each fibre divergence to admit a second-order expansion whose quadratic part induces the fibre Riemannian metric ; together with Assumption 3 this ensures that informational distances behave, locally and under parallel transport, like the Riemannian distances of the fibre metrics. Assumption 2 is a structural regularity assumption guaranteeing that the Ehresmann connection on arises from a smooth principal connection.
On the discrete side, Assumptions 4 and 5 encode that the sampling graphs
form quasi-uniform discretisations of
with bounded degree and uniformly controlled edge lengths, whereas Assumption 6 provides vertex weights approximating the Riemannian volume. Assumption 7 specifies discrete fibres and divergences approximating the continuous ones, and Assumption 8 does the same for the reference states. Assumption 9 postulates channels that are local and Lipschitz and whose first-order behaviour approximates continuous parallel transport, with an error measured by
. Finally, Assumption 10 ensures that the discrete triangle areas
approximate the Riemannian areas, and Assumption 11 provides, for each
, families of triangles that sample directions in
in an almost isotropic way. The quantity
in (
10) summarises the various errors contributed by these assumptions.
Definitions 13, 14 and 16, together with Theorems 1 and 2, provide the conceptual and analytical core of the informational holonomy curvature framework. In the next sections we make these statements precise by deriving the continuous holonomy-curvature relation and then establishing the discrete-to-continuous convergence.
5. Continuous Holonomy Curvature and Connection Curvature
In this section we prove Theorem 1. We work in the continuous framework of
Section 2:
is a smooth Riemannian manifold,
is a state bundle with fibre metrics and divergences satisfying Assumption 1, endowed with a connection satisfying Assumption 3, and
is a smooth reference state field.
The core of the argument consists of two ingredients:
the second-order expansion of the divergence on each fibre, which implies that the informational distance is locally equivalent to the Riemannian distance induced by ;
the first-order (in area) expansion of the holonomy map of the connection on around small geodesic triangles, controlled by the curvature of the connection.
Combining these, we obtain a linear-in-area behaviour for the distance defect and hence the existence and explicit form of the continuous informational sectional curvature.
Throughout this section we fix a compact set , and all constants will be uniform over and over two-planes .
5.1. Local Expansion of the Informational Distance
We recall that for each
and
the informational distance is defined by
where
is the divergence on the fibre
(see (
8)). The following lemma makes precise the local behaviour of
in terms of the Riemannian metric
on the fibre.
Lemma 4
(Local expansion of the informational distance)
. Let be compact. There exist constants and such that for every , for every , and for every with , the following holds. Let
where is the Riemannian exponential map on . Then
In particular, there exist constants such that, for all such v,
Proof. Fix
and
. By Assumption 1, in a normal coordinate chart for
centred at
s, the divergence
satisfies
where
and
is a remainder term with
for
, with
independent of
x and
s in compact sets (smoothness and compactness of
K).
By definition,
Let
. Then
and
For
we have
Choose
such that
. Then for
we have
For
the Taylor expansion of
yields
where
for some universal constant
. Taking
, we obtain
Using
, we get
Therefore,
for some constant
depending only on
and
. This proves (
12).
The two-sided inequality follows from (
12) by taking
L sufficiently small and absorbing the quadratic term into the linear one. □
5.2. Curvature of the Connection and Small-Loop Holonomy
We now recall the curvature of an Ehresmann connection on and its relation with holonomy around small loops. We adopt a local viewpoint sufficient for our application to small geodesic triangles.
Let be the horizontal subspace at and the vertical subspace. For each vector field X on M there exists a unique horizontal lift on such that for all .
The curvature of the connection is the vertical-valued 2-form
on
defined by
where
are vector fields on
M and the superscript
denotes projection onto
in the decomposition
. This definition is independent of the choice of extensions of
.
For each
and
, the restriction of
gives a bilinear alternating map
by setting
, where
are any vector fields extending
in a neighbourhood of
x. This is well defined and smooth in
.
We now state the holonomy expansion around small geodesic triangles. The result is a specialization of standard holonomy-curvature relations (see, for example, [
5] Chapter II, Sections 3–4]), adapted to our two-dimensional situation and with explicit attention to the dependence on the area of the triangle.
Let
and
be a two-dimensional subspace. For
sufficiently small, we set
and consider the geodesic triangle with vertices
. Let
denote the closed loop obtained by traversing the geodesic segments
in order, and let
be the Riemannian area of the geodesic triangle.
Fix
. Parallel transport along
yields a holonomy map
sending
to a point
.
Lemma 5
(Small-loop holonomy expansion). Assume the associated-bundle framework of Assumption 2. Let be compact. Then there exist constants and such that for every , every 2-plane , every pair with , and every , the following holds.
Let , , let be the Riemannian area of the geodesic triangle , and let be the piecewise-geodesic loop . Denote and . Then
where depends smoothly on and linearly on the curvature of the inducing principal connection, and the remainder satisfies
More explicitly: fix an orientation of Π and let be any oriented g-orthonormal basis of Π. Let be the curvature element determined by the principal curvature (in any local gauge) evaluated on at x. Let be the infinitesimal action (fundamental vector fields) of G on S. Then, identifying via any choice of ,
and this definition is independent of the chosen gauge because is -equivariant and the action of G on S is by isometries.
Proof. We work on a convex normal neighbourhood of x contained in a compact set K and choose a local section . In this gauge, the principal connection is represented by a -valued 1-form on U, with curvature , whose coefficients are and uniformly bounded on K (by Assumption 2).
Let
be the geodesic triangle surface with boundary
. The holonomy of the principal connection around
is an element
obtained by the path-ordered exponential of
A along
. In this standard principal-connection setting, the non-abelian Stokes/holonomy–curvature expansion (see, e.g., [
5], Chapter II, §3–§4) yields
for
ℓ small, with constants uniform for
and
varying in the Grassmannian. Moreover, since
F is
, Taylor expansion on
gives
where
denotes the curvature evaluated at
x on an oriented orthonormal basis
of
(and the
term is uniform on
K). Combining with (
15) yields
Now pass to the associated bundle. Choose
and identify
by
. Under this identification, the holonomy acts by the
G-action:
Consider the smooth map
defined for
small. Since the
G-action is smooth,
is
and satisfies
(the fundamental vector field). Therefore, Taylor expansion at 0 gives
with a uniform constant for
and
ranging in compact subsets of the fibres.
Apply (
17) to
from (
16). Since
and
, we obtain
because the quadratic term
is
and hence dominated by
. This yields (
13) and (
14) with
. □
Remark 9.
The vector depends linearly on the curvature restricted to Π. In particular, when is an associated bundle to a principal bundle with a connection induced from the Levi–Civita connection via an isometric representation, is obtained by applying the differential of the representation to the Riemann curvature tensor restricted to Π.
5.3. Proof of Theorem 1
Fix
and a two-dimensional subspace
. Let
be a family as in Definition 16, with
and with uniformly non-degenerate shape in
. Set
and let
be the boundary loop of the geodesic triangle
. Finally set
By Lemma 5 (applied with
) we have, for
r small,
where
and
. Since the triangles have uniformly non-degenerate shape in
, there exists
such that
for
r small. Consequently,
where the implicit constant is uniform for
x in compact sets and for
.
Now the continuous distance defect satisfies
Applying Lemma 4 with
and
yields
Dividing by
and letting
we obtain
This limit is independent of the chosen shrinking family
(subject to the non-degeneracy condition), and therefore
exists and equals
.
5.4. Geometric Models Induced from the Levi–Civita Connection
We briefly justify the last assertion of Theorem 1. Assume that is an associated bundle to the orthonormal frame bundle , with structure group acting on the model fibre S by isometries of . Let denote this representation and let ∇ be the Levi–Civita connection on ; it induces an Ehresmann connection on .
Let
and
be a two-plane. Denote by
the curvature endomorphism of ∇ restricted to
(equivalently, the image of
under the identification
). The curvature of the induced connection on
is obtained by applying the differential
fibrewise, and therefore there exists a smooth linear map
such that
In particular,
is a scalar invariant determined by the restriction of
to
.
If
has constant sectional curvature
, then for every
x and
one has
, where
is the infinitesimal generator of the
-rotation in
(normalized so that
). Combining this with (
21) yields
Under the additional natural hypothesis that the model is
-equivariant and the reference field
is chosen compatibly with the symmetry, the factor
is independent of
x and
, so that
reduces to a constant multiple of
, as claimed.
6. Discrete-to-Continuous Convergence of Informational Holonomy Curvature
In this section we prove Theorem 2. We work under Assumptions 4–11 and the continuous hypotheses of Theorem 1. Throughout, is a fixed compact set and all constants are uniform over and over two-planes .
The proof proceeds in three steps:
we compare discrete and continuous holonomy on individual small triangles (Sub
Section 6.1);
we convert this comparison into a bound between discrete and continuous
triangle-wise informational holonomy curvature (Sub
Section 6.2);
we pass to the averaged,
sectional quantity by exploiting the isotropic triangle families from Assumption 11 and the area approximation from Assumption 10 (Sub
Section 6.3).
6.1. Comparison of Discrete and Continuous Holonomy on Small Triangles
Fix
small, a vertex
and a triangle
. Let
be the associated discrete loop and
the discrete holonomy operator based at
.
On the continuous side, consider the geodesic triangle
in
and the loop
obtained by traversing the minimizing geodesic segments
in order. Let
be the corresponding holonomy map induced by the connection on
. Via the identification
from Assumption 7, we may view both
and
as acting on the same fibre.
For notational simplicity, we replace by the identity and treat as , understanding that all maps and distances are transported accordingly. Thus, in what follows, denotes and denotes the informational distance .
We first estimate how well the discrete holonomy operator approximates the continuous one near the reference state.
Lemma 6
(Staying in the local regime along short compositions). Let be compact. Assume:
-
(i)
The fibre divergences satisfy Assumption 1 uniformly on K.
-
(ii)
The continuous model is in the associated-bundle setting Assumption 2, so that parallel transport preserves fibre Riemannian distances (Lemma 2).
-
(iii)
The discrete channels satisfy Assumption 9(2) in the fibre Riemannian distance with exponent .
-
(iv)
The channel locality Assumption 9(1) holds.
Then there exist constants , and such that for all the following holds.
Let be any oriented edge with , and let be the minimizing geodesic from to . For any with one has the local Lipschitz bound
Moreover, for any (nondegenerate) discrete triangle in K, define the continuous and discrete intermediate states starting from by
Then each pair remains in the local regime:
and, quantitatively,
where is the sampling radius.
Proof.
Step 1: uniform local equivalence and choice of . Since
K is compact and
is smooth,
is compact in
. By Assumption 1 and smooth dependence of the fibre metrics/divergences, there exists
and constants
such that for any
and any
with
,
Step 2: local Lipschitz of in . Assumption 9(1) gives
. By Lemma 3, this implies
. If
then (
24) yields
which is (
22) with
.
Step 3: staying in the local regime and the bound (23). For
, Assumption 9(2) (applied at
) gives
Choose
so that
for all
. Then
lies in the local regime.
Assume inductively that
and
for
. Using the triangle inequality in
,
The first term is bounded by Assumption 9(2):
. The second term equals
by Lemma 2. Therefore,
For this is , closing the induction. □
Lemma 7
(Discrete vs continuous holonomy). Let be compact. There exist constants and such that, for all , for every vertex and every discrete triangle with all vertices in K and satisfying the non-degeneracy scale condition of Assumption 11(1), the following holds. Let and let . Then:
Proof. The first assertion is immediate from Assumption 5: each edge has length , so any triangle with vertices connected by edges has all side lengths bounded by a constant multiple of . Adjusting constants yields .
For the second assertion, write the continuous holonomy map as a composition of continuous parallel transport along the three geodesic edges:
where
denotes the minimizing geodesic segment from
to
.
We use a telescoping argument. Set
Then
and
.
By Assumption 9(2), the edge-wise channel error is controlled in the
fibre Riemannian distance : after adjusting constants, for every oriented edge
and every
,
We therefore perform the telescoping estimate entirely in the distances
(so that the triangle inequality holds globally), and only at the end pass back to the informational distance
using the local equivalence of Remark 6 (which applies since the final displacement is
).
Moreover, by Assumption 3(1), parallel transport is a fibrewise Riemannian isometry; in particular, it preserves fibre distances:
We estimate recursively:
Next,
and similarly,
Combining the three bounds yields
Finally, since
, Remark 6 applies for
small, and after adjusting constants we obtain the same bound in the informational distance:
we show that the discrete informational sectional
Now write
By the uniform non-degeneracy in Assumption 11(1), the triangle area satisfies
, and since
we also have
(after adjusting constants). Hence
which is exactly (
25). □
6.2. A Triangle-Wise Discrete-to-Continuum Bound
Fix
small,
, and a discrete triangle
with vertices in
K satisfying the non-degeneracy condition of Assumption 11(1). Recall the triangle-wise curvature
and the corresponding continuous quantity
where
denotes the plane spanned by the geodesic initial directions from
to
and
to
(and
is any shrinking family of geodesic triangles tangent to
). In practice we will compare the discrete loop
with the continuous holonomy around the
geodesic triangle
.
Let
where
is the loop traversing the geodesic edges
.
Lemma 8
(Triangle-wise defect comparison)
. There exist constants and such that, for all and all triangles with vertices in K satisfying Assumption 11(1), one has
where .
Proof. By the triangle inequality for the fibre Riemannian distance (Remark 6) and the local equivalence with
, we have
The right-hand side is bounded by Lemma 7(
25), which gives (
26). □
We now incorporate the area approximation. Recall from Assumption 10 that
satisfies
Lemma 9
(Triangle-wise curvature comparison)
. There exist constants and such that, for all and all triangles with vertices in K satisfying Assumption 11(1), one has
Proof. Write
Add and subtract
:
For the first term, Lemma 8 gives
By Assumption 11(1) and Assumption 10,
is uniformly comparable to
from below and above (for
small), hence the ratio is bounded and the first term is
.
For the second term, note that by the holonomy expansion (Lemma 5) and the distance expansion (Lemma 4), there is a uniform constant
C such that
for triangles in
K sufficiently small. Therefore,
Using Assumption 10 and the uniform lower bound
(Assumption 11(1)), together with the comparability
, we obtain that this term is
. Combining the bounds yields (
27). □
6.3. From Triangle-Wise to Sectional Curvature by Averaging
We now pass from the triangle-wise comparison to the averaged sectional quantity
defined in (
7).
Fix and a two-plane . By Assumption 11, for sufficiently small we can choose a vertex with and a finite non-empty family of triangles based at satisfying:
uniform scale and non-degeneracy at scale ;
planarity up to with respect to ;
approximate directional isotropy in .
We write the discrete sectional curvature as an area-weighted average:
Similarly, define the corresponding continuous average over the same
geodesic triangles:
Lemma 10
(Averaging stability)
. There exist constants and such that, for all , for every and every two-plane ,
Proof. By Lemma 9, for each triangle in the family we have
Multiplying by
and summing over the family gives
Using uniform comparability
on the family (Assumption 11(1) and Assumption 10), we may replace
by 1 at the cost of an additional
relative error. Dividing by the denominators and using again that
yields (
28). □
We now compare with the intrinsic continuous quantity .
Lemma 11
(Planarity and base-point stability)
. There exist constants and such that, for all , for every and every two-plane ,
Proof. First, by smoothness of in (Theorem 1 and smooth dependence of ), moving the base point from x to induces an error bounded by .
Second, by Assumption 11(2), the planes spanned by the geodesic directions of each triangle in the family are within angle of . Again by smoothness in (uniform on K), replacing by induces an additional error .
Finally, the quantity is an average over finitely many triangles of uniformly comparable shape and size , hence the above pointwise stability bounds propagate to the average with the same order. □
6.4. Conclusion of the Proof of Theorem 2
Combining Lemmas 10 and 11, we obtain
uniformly for
and
. Recalling the definition
from (
10), this is exactly (
11). This completes the proof of Theorem 2.
7. Examples and Model Constructions
In this section we discuss several model constructions that illustrate the notion of informational holonomy curvature and the hypotheses of our convergence theorem. We first describe a basic classical choice of state space and divergence (the Jensen–Shannon model), then introduce a natural geometric state bundle built from tangent distributions and parallel transport, and finally discuss spaces of constant curvature and discrete sampling schemes.
7.1. The Classical Jensen–Shannon Model
We begin with a simple and concrete choice of state space and divergence, which fits into the general framework of
Section 2 and
Section 4.
Fix a finite set
and let
denote the open probability simplex. We endow
S with:
It is well known that
is symmetric and non-negative and that
defines a genuine metric on
S [
3]. With our normalization
, we set
which is also a genuine metric on
S. Moreover, the second-order expansion of
at the diagonal yields a constant multiple of the Fisher metric:
where
is the tangent vector such that
and
is a constant.
Thus Assumption 1 is satisfied with proportional to the Fisher metric and . In particular, the informational distance is locally equivalent to the Riemannian distance induced by on S.
Given a Riemannian manifold
, the simplest associated state bundle is the trivial bundle
with fibre
independent of
. To obtain a non-trivial informational holonomy curvature, however, one needs a connection on
whose curvature reflects the geometry of
. The trivial product connection on
has zero curvature and yields vanishing holonomy and hence vanishing informational holonomy curvature. Thus, in interesting examples, the state bundle and its connection must be constructed from the Levi–Civita connection in a non-trivial manner. In this trivial product situation the connection on
is taken to be the product of the Levi–Civita connection on
with the trivial connection on
S, so that parallel transport acts as the identity on the fibre. Consequently, the Jensen–Shannon divergence is exactly preserved under parallel transport, i.e.
for every curve
in
M and every
. Thus Assumption 3 holds with
, and Assumption 9(2) may be realised with vanishing channel-consistency error: in this example the term
in (
10) does not contribute to the bound of Theorem 2 if the discrete channels
are chosen to coincide with the exact parallel transport on
.
7.2. A Geometric State Bundle from Tangent Distributions
We now describe a natural geometric construction in which the state bundle is built from probability distributions on tangent spaces and the connection is induced by parallel transport in .
Let
be a smooth Riemannian manifold of dimension
n. For each
, consider the tangent space
with its Euclidean inner product
. Let
be a chosen smooth manifold of probability measures on
, for instance:
the manifold of non-degenerate Gaussian measures on ,
or a finite-dimensional exponential family of probability measures with smooth densities with respect to Lebesgue measure on .
To fix ideas, one may take to be the set of Gaussian measures on , with mean and covariance matrix in some fixed compact subset of the positive definite cone.
We define the state bundle
by gluing the fibres
smoothly via the tangent bundle structure. The Riemannian metric
on each fibre is taken to be the Fisher information metric associated with the chosen statistical model on
, and the divergence
is taken to be the Jensen–Shannon divergence between distributions in
.
To define the connection on
, we use parallel transport in
. Let
be a smooth curve with
and
, and let
denote the parallel transport map associated with the Levi–Civita connection of
g. We define the parallel transport on
along
by pushing forward measures under
:
In particular, if
is Gaussian, then
so that the family
is preserved by parallel transport. The resulting parallel transport maps satisfy the functoriality conditions (
4), and thus define an Ehresmann connection on
.
The curvature of this connection is induced by the curvature of : the curvature of the Levi–Civita connection acts on via the Riemann curvature tensor , and this, in turn, induces a curvature 2-form on by differentiation of the pushforward action on . In particular, if the base manifold has zero curvature, then the induced connection on is flat and the informational holonomy curvature vanishes.
The reference state field can be chosen, for instance, as the isotropic Gaussian with mean and covariance at each , for some fixed . This is invariant under orthogonal transformations of , which simplifies the structure of .
Under this construction, Assumptions 1 and 3 are satisfied: the Jensen–Shannon divergence on each induces the Fisher metric, the pushforward by isometries preserves the Fisher metric and the second-order expansion of , and the connection on is metric along the fibres. Thus the continuous informational holonomy curvature is well defined and determined by the curvature tensor .
7.3. Spaces of Constant Curvature
We now consider the case where has constant sectional curvature and specialise the construction of the previous subsection. The aim is to illustrate how the informational holonomy curvature reflects the constant curvature of the base manifold.
Let
be complete, simply connected, and of constant sectional curvature
. Thus
is isometric to the Euclidean space
(if
), the round sphere
(if
), or the hyperbolic space
(if
). The Riemann curvature tensor satisfies
for all
and
, where
is the Riemannian inner product.
We equip
M with the geometric state bundle
of tangent distributions described in
Section 7.2, with fibres consisting of Gaussian measures on
and divergence given by the Jensen–Shannon divergence. We choose the reference state field
to be the isotropic Gaussian
on
, with fixed
independent of
x.
The isotropy of
implies that, for any
and any two 2-planes
, there exists an isometry of
mapping
to
. The induced action on the state bundle
preserves the connection, the fibre metric and the divergence, and sends
to itself. Thus, for each fixed
x, the map
must have constant norm on the Grassmannian of 2-planes at
x, and this norm can depend only on
and on the parameters of the state bundle (e.g.
and the choice of divergence). In particular, there exists a constant
such that
for all
and all 2-planes
.
When
, the Levi–Civita connection is flat and the parallel transport maps
along closed loops are the identity. Hence the induced connection on
is flat, the holonomy maps on
are trivial, and
, so
Alternatively, one can considerFor
, the curvature of the Levi–Civita connection is non-zero and so is the curvature of the induced connection on
; therefore,
is non-zero and
.
By Theorem 1, the continuous informational sectional curvature is
which is constant in
. In particular, the informational holonomy curvature detects the constant curvature of
modulo the scale factor
coming from the choice of state bundle and divergence. In spaces of constant curvature,
is thus a constant multiple of
.
7.4. Discrete Sampling Schemes
Finally, we discuss concrete choices of sampling graphs, areas and triangle families that satisfy the assumptions of
Section 3 and enable the application of Theorem 2.
Neighbour graphs and edges
Given
, one natural choice of graph is the
-neighbourhood graph: for a suitable radius
satisfying
one sets
Alternatively, one can consider
k-nearest neighbour graphs with
k fixed or slowly increasing as
. Under standard conditions, these constructions satisfy Assumption 5: edges connect points at distance of order
, and the vertex degrees are uniformly bounded.
Discrete areas and triangle families
Given a sampling graph embedded in M, one can define the set of discrete triangles as in Definition 9. For each triangle , the discrete area can be chosen, for example, as:
the Euclidean area of the triangle formed by the images of in a normal coordinate chart centred at ;
or the area of a piecewise flat triangle obtained by approximating the metric g locally by its value at .
In both cases, standard Taylor expansions in normal coordinates show that
so Assumption 10 is satisfied with
.
Families of triangles satisfying Assumption 11 can be constructed by selecting, for each vertex and each approximate direction in , a finite number of neighbouring vertices whose geodesic directions approximate a given 2-plane in an approximately isotropic fashion. In random sampling models, the law of large numbers ensures that the empirical distribution of edge directions becomes asymptotically isotropic, with deviations captured by a parameter .
Discrete channels from continuous transport
Finally, the discrete channels
on edges
can be defined by approximating the continuous paralleland local equivalence of transport maps
along the minimizing geodesics
. In the geometric state bundle of
Section 7.2, this amounts to approximating the pushforward of tangent distributions by the parallel transport
.
For instance, one can set
whenever
is uniquely defined and computable, in which case Assumption 9(2) holds with
. In numerical settings where
and
are approximated by finite-difference schemes or local polynomial approximations of
g, consistency estimates of the form
can be obtained for suitable
, leading to a non-zero but convergent
.
Under these constructions, all the assumptions of
Section 3 and
Section 4 are satisfied, and Theorem 2 applies. Thus the discrete informational sectional curvature
computed from sampling graphs, approximate geodesic triangles, and discrete channels converges, as
, to the continuous informational sectional curvature
determined by the geometric data
.
8. Discussion and Outlook
The constructions developed in this work provide a framework for defining and estimating curvature from informational holonomy. Starting from a Riemannian manifold and a state bundle endowed with fibrewise divergences and a compatible connection, we defined a continuous informational holonomy curvature associated with a point and a two-plane by measuring, via the informational distance induced by the divergence, the leading (area–linear) effect of transporting a reference state around small geodesic triangles. We then showed that, under explicit sampling, area-approximation, and channel-consistency assumptions, a purely discrete estimator constructed on graphs embedded in M converges to as the sampling scale , with a quantitative error bound controlled by the discretization scale.
8.1. Summary of the Framework
The continuous construction hinges on three ingredients:
a state bundle whose fibres represent informational states (classical or quantum), equipped with a fibre Riemannian metric and a divergence whose second-order expansion induces this metric;
an Ehresmann connection on compatible with the fibre metrics and divergences, so that parallel transport acts as a fibrewise isometry to first order and preserves the informational structure infinitesimally;
a reference state field serving as a basepoint for measuring informational defects.
Holonomy of the connection on along small geodesic triangles based at x and tangent to produces a displacement of in which, by holonomy–curvature expansions, is proportional to the triangle area to first order. The informational distance associated with the fibre divergence then yields a scalar quantity per unit area, identified as the continuous informational sectional curvature (Theorem 1).
On the discrete side, we considered quasi-uniform sampling graphs on M, endowed with:
discrete fibres and divergences at vertices approximating the continuous fibres and divergences;
edge channels approximating continuous parallel transport along short geodesic segments;
discrete areas for triangles and triangle families that are asymptotically isotropic in prescribed directions.
The resulting discrete holonomy operators yield distance defects , and the triangle-wise curvatures are defined by normalising by . Averaging over produces a discrete sectional curvature which converges, at a rate governed by , to the continuous quantity .
8.2. Relation to Classical and Discrete Curvature Notions
The informational holonomy curvature sits at the intersection of several strands of work on curvature:
Riemannian sectional curvature. In classical Riemannian geometry, sectional curvature can be characterised in terms of angle defects, Jacobi fields, or holonomy of the Levi–Civita connection. Our framework replaces the linear tangent bundle by a (generally non-linear) state bundle and linear norms by informational distances induced by divergences. When the connection on is induced by the Levi–Civita connection via a linear isometric representation, the vector in Theorem 1 is obtained as a linear image of the restriction of to , and becomes an invariant of this restriction. In spaces of constant sectional curvature, this reduces to a constant multiple of (equivalently, of when ).
Discrete and combinatorial curvature. Various notions of curvature for graphs and discrete spaces have been proposed, including Ollivier–Ricci curvature, Forman curvature, and Regge-type discretisations. The discrete informational holonomy curvature differs from these in two key aspects: it is based on holonomy of a bundle connection (rather than on pairwise comparisons of neighbourhood measures or purely combinatorial angle/defect data), and it uses divergences on state spaces attached to vertices (rather than solely distances in the ambient manifold or graph). In particular, it blends geometric information about with an informational structure in the fibres.
Curvature in information geometry. In information geometry, Fisher metrics and
-connections yield Riemannian and affine structures on statistical manifolds, and their curvature encodes statistical properties of models. The present construction can be viewed as a “mixed” curvature: it is controlled by the curvature of a connection on a state bundle over a geometric base, while the informational structure enters via the choice of fibre divergence and reference state. The Jensen–Shannon model of
Section 7.1 provides a particularly transparent example where the divergence has a direct information-theoretic meaning.
8.3. Limitations and Choices of State Bundle
The informational holonomy curvature is not a curvature of alone: it depends on the choice of state bundle, divergence, connection, and reference state field. Different choices can therefore produce different curvature functionals over the same base manifold. This flexibility is both a strength and a limitation.
On the one hand, it allows the notion of curvature to be adapted to an application: classical probability distributions on finite sets with Jensen–Shannon divergence, Gaussian distributions on tangent spaces, or quantum density matrices with quantum Jensen–Shannon or other quantum divergences all fit naturally into the framework. On the other hand, it raises the question of which choices are canonical, or geometrically natural, for a given problem.
A natural option in a purely geometric setting is the geometric state bundle built from tangent distributions (
Section 7.2), whose connection is induced canonically by parallel transport in
and whose fibres behave naturally under base isometries. In data-driven or statistical settings, other choices may be more appropriate, for instance, state spaces encoding empirical distributions of local observations, feature vectors, or structured data attached to points in
M.
From a foundational perspective, the present work treats the state bundle and its connection as given. Understanding how to construct such bundles in a canonical or data-driven way, and how the resulting varies across different constructions, remain interesting open questions.
8.4. Potential Applications and Further Directions
We conclude by mentioning several directions in which the informational holonomy curvature framework may be developed further.
Data analysis and manifold learning.
In applications where only a point cloud in M is observed, possibly together with empirical distributions or feature states at each point, the discrete curvature provides a way to estimate curvature-like quantities that combine geometric and informational structure. Compared to purely metric estimators based on distances or angles, informational holonomy curvature incorporates how local states are transported along the graph via channels, which may reflect dynamics, diffusion, or parallel transport in latent spaces. Analysing statistical properties and robustness of such estimators in the presence of noise and finite-sample effects is a natural next step.
Other divergences and connections.
While we focused on divergences whose second-order expansion induces a Riemannian metric (e.g. Jensen–Shannon), one could consider more general f-divergences or Bregman divergences, potentially leading to non-Riemannian local geometry on fibres. Extending the holonomy curvature construction to such settings would require an appropriate notion of distance defect and a careful analysis of higher-order terms. Similarly, one may study families of connections on (for example, -connections in information geometry) and compare the corresponding informational holonomy curvatures.
Quantum and non-commutative models.
The state-bundle viewpoint naturally accommodates quantum state spaces, where fibres consist of density matrices on finite-dimensional Hilbert spaces and divergences are given by quantum generalisations of Jensen–Shannon or relative entropy. In such settings, the connection on may encode both geometric parallel transport and quantum channels acting along paths in M. Extending the convergence analysis to non-commutative state bundles, and understanding how informational holonomy curvature reflects underlying quantum geometric structure, are promising directions.
Algorithmic and numerical aspects.
From a practical perspective, computing requires:
constructing a sampling graph and identifying triangle families ;
specifying discrete channels and evaluating their composition along loops;
computing divergences and distances in the fibre state spaces.
Each of these steps has algorithmic consequences, and different applications may favour different trade-offs between accuracy and complexity. Designing efficient algorithms for informational holonomy curvature in high-dimensional state spaces, and testing them on simulated and real data, would help assess the practical relevance of the notion.
In summary, the informational holonomy curvature introduced here provides a bridge between classical Riemannian curvature, graph-based approximations, and information-theoretic structures on state spaces. It offers a geometrically grounded way of measuring how “information” twists under transport around small loops. The results of this paper establish its mathematical foundation and discrete-to-continuous consistency; its full potential will likely emerge in concrete applications and in further theoretical developments linking geometry, probability and information.
Author Contributions
The author contributed solely to all aspects of this work: conceptualization, methodology, formal analysis, writing, and revision.
Funding
This research received no external funding.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Conflicts of Interest
The author declares no conflict of interest.
References
- Amari, S.; Nagaoka, H. Methods of Information Geometry; Translations of Mathematical Monographs, Vol. 191; American Mathematical Society: Providence, RI, USA, 2000.
- do Carmo, M.P. Riemannian Geometry; Birkhäuser: Boston, MA, USA, 1992. [Google Scholar]
- Endres, D.M.; Schindelin, J.E. A new metric for probability distributions. IEEE Trans. Inf. Theory 2003, 49, 1858–1860. [Google Scholar] [CrossRef]
- Forman, R. Bochner’s method for cell complexes and combinatorial Ricci curvature. Discrete Comput. Geom. 2003, 29, 323–374. [Google Scholar] [CrossRef]
- Kobayashi, S.; Nomizu, K. Foundations of Differential Geometry; Interscience Publishers: New York, NY, USA, 1963; Vol. I. [Google Scholar]
- Lin, J. Divergence measures based on the Shannon entropy. IEEE Trans. Inf. Theory 1991, 37, 145–151. [Google Scholar] [CrossRef]
- Ollivier, Y. Ricci curvature of Markov chains on metric spaces. J. Funct. Anal. 2009, 256, 810–864. [Google Scholar] [CrossRef]
- Regge, T. General relativity without coordinates. Nuovo Cimento 1961, 19, 558–571. [Google Scholar] [CrossRef]
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).