1. Introduction
The standard
CDM model treats the universe as a Markovian dynamical system. Once the matter content and cosmological parameters are specified, the Friedmann equations close on a small set of variables at each cosmic time, and the expansion history
is compressed into a single number, an effective age
obtained by integrating
back to the initial singularity [
1,
2,
3]. In this setting it is natural to ask “How old is the universe?” and to quote
Gyr as a summary of the background history.
This view already assumes that the dynamics is memoryless. The present Hubble rate depends only on instantaneous densities and pressures, not on a record of past states. The age is then a Markovian quantity: it is the length of the backward extrapolation in a model that does not remember its own history.
There is, however, no deep reason why large scale cosmology must be Markovian. Non-local and non-Markovian effects have been discussed in many contexts, including back reaction, effective field theory and modified gravity (e.g., [
4,
5,
6]). If the expansion rate at time
t depends on a weighted integral over past states, then the relevant concept is not just “age” but also a
memory horizon, a timescale over which past dynamics remains dynamically relevant for the present.
The Infinite Transformation Principle (ITP) is a recently proposed phenomenological framework in which the large scale dynamics is described by a small set of effective parameters that encode internal energy, structural memory and their coupling, and that naturally lead to non-Markovian behaviour at late times [
7]. When the ITP parameters are constrained by late time expansion and growth data, the fits favour a long delay scale in the memory kernel. This suggests that the present expansion behaves as if it remembers its own past over timescales much longer than the usual Hubble age.
This paper develops that statement in operational form. The main steps are:
Section 2 introduces a simple non-Markovian extension of the Friedmann equations, defines an effective memory horizon
and explains how it differs from the usual age
.
Section 3 summarises the late time datasets used to constrain the ITP parameters, including
and
measurements [
8,
9,
10,
11].
Section 4 outlines the inference pipeline, the mapping from ITP parameters to kernel parameters, and the estimation of
from posterior samples.
Section 5 presents representative constraints on the delay parameter and the memory horizon, and tests their sensitivity to kernel shape and priors.
Section 7 discusses the interpretation of
, its relation to the usual notion of age, and its place within broader cosmological models, including cyclic scenarios and inflationary histories [
12,
13,
14].
The goal is not to replace Gyr with a new “true age” of the universe. Instead, is treated as a derived quantity in a Markovian model, and as a derived quantity in a non-Markovian model. Both are projections of a deeper dynamics. The main claim is that, in a non-Markovian fit of the ITP type, current late time data prefer a universe whose present expansion rate carries effective memory of its past over several Hubble times.
4. Methods: From ITP Parameters to
4.1. ITP Parameterisation and Inference
The Infinite Transformation Principle describes the effective large scale dynamics of the universe in terms of a small set of parameters that control internal energy, structural memory and their coupling [
7]. In the early universe tests these were written as
where
is the present matter density fraction,
the Hubble constant,
describe a two bin non-Markovian correction, and
encodes a dimensionless delay in units of the Hubble time.
Given a parameter set the model predicts:
a background expansion history ;
an effective dark energy equation of state and density , reconstructed numerically;
a growth history , obtained by solving the standard growth equation in the effective background.
The inference proceeds as follows. An ensemble MCMC sampler is used to explore with broad priors. For each proposed point the background is evolved, and are computed, the linear growth factor and are obtained, and the joint likelihood of the , distance and growth data is evaluated. Posterior samples and summary statistics for are then retained for further analysis.
4.2. Reconstructing the Effective Kernel
In the ITP framework the non-Markovian correction can be written either as an explicit kernel in time or as an effective fluid with a time dependent equation of state. In practice the inference is done in the second picture: the code outputs and on a redshift grid for each posterior sample of .
To define
these effective quantities are mapped back onto a simple kernel based description. One convenient route is to treat the non-Markovian part as an effective fluid whose pressure depends on a convolution of some internal variable
with a kernel
K,
Under mild assumptions about , the reconstructed can be matched by a parametric family of kernels with a small number of degrees of freedom, including a characteristic timescale controlled by . For the exponential family, the mapping is straightforward and leads to .
Once the kernel parameters for a sample are known, the cumulative memory function
in Equation (
5) can be evaluated and the associated
can be found by solving
numerically, using
as fiducial.
4.3. Estimating the Memory Horizon
For each posterior sample:
- 1.
the kernel parameters are extracted from ;
- 2.
the cumulative memory function at is computed;
- 3.
the smallest T with is recorded as for that sample.
This gives a posterior distribution for . For comparison, the Hubble time and the CDM age for the same can be computed.
The mapping from to is close to linear in the exponential family, so it is useful to track the posterior for directly and to study its sensitivity to the prior. This is done using a piecewise constant (two bin) parameterisation of together with different top hat priors on .
5. Results
5.1. Two Bin ITP Fits and the Delay Parameter
The main constraints on the delay parameter come from a two bin ITP fit to a combined dataset, with allowed to take independent values in a low redshift and a high redshift bin and with a flat prior on between a lower and an upper bound.
With a broad prior
, the posterior favours a long delay solution with
where quoted uncertainties are
credible intervals. Even at the lower
edge the delay remains well above unity,
, which already suggests a memory scale longer than a single Hubble time.
In response to a natural robustness question, the same two bin fit was repeated with a tightened prior
, corresponding to an approximately Markovian limit. Under this short delay prior the posterior shifts to
The posterior for now saturates the upper bound of the prior, and the background parameters are driven into a region with high , low and a sign flip between the two bins.
The comparison shows that the preference for in the broad prior run is not a prior artefact. When the model is forced into a short delay regime, the fit simply piles against the ceiling and compensates by contorting . The short delay solution is therefore strained, while the broad prior solution remains compatible with standard late time constraints.
5.2. Sensitivity to the Delay Prior
To explore how far the delay can extend without conflicting with the data, the analysis was repeated with successively wider priors,
,
,
and
. The resulting posterior summaries are given in
Table 1. In each case the background parameters remain stable, while the distribution of
shifts towards longer delays and broadens:
Once the upper bound on is relaxed beyond unity, the background parameters quickly stabilise at and . Only the upper tail of responds to the prior volume. As the prior is widened, the lower limit on increases: for the run at , and for the and runs the ranges move deeper into the long delay regime. The upper bound on is prior limited for the widest priors, but the existence of a lower bound well above unity is robust.
For the main analysis it is therefore natural to adopt as a fiducial prior. It is the narrowest uninformative prior that fully contains the high probability region and avoids boundary effects. Wider priors leave the inferred expansion and growth histories unchanged and mainly broaden the high tail.
5.3. From Delay to Memory Horizon
For an exponential kernel and
the effective memory horizon is
. For the two–bin ITP fits with
of order 6–12 and
this corresponds to
while the corresponding Hubble age is
–14 Gyr. Thus, in the ITP fit, the present expansion behaves as if it retains dynamical memory over many Hubble times.
5.4. Sensitivity to Kernel Assumptions
The inferred memory horizon is robust to reasonable choices of kernel family, early time priors and the precise definition of . Three representative kernel shapes were tested: a simple exponential, a stretched exponential tail, and a power law with an exponential cut off. In each case the data were refit and the resulting was computed.
All three families yield memory horizons well in excess of a Hubble time. For the exponential kernel, representative best–fit solutions have , i.e., tens of Hubble times. Even the most conservative models remain in a long–memory regime, . Tighter priors on the high redshift behaviour, for example restricting to lie close to at , slightly compress the posterior on but do not drive it back to a Markovian regime. Varying between , and rescales the quoted at the tens of per cent level but leaves the qualitative conclusion unchanged.
These tests, together with the prior study summarised in
Table 1, show that the existence of a long memory horizon,
, is a robust feature of the ITP fit to current late time data. What is constrained is the
minimum memory horizon, which is conservatively several Hubble times, while the exact upper extent of the delay remains prior limited within the kernel family explored here.
5.5. Relation to Inflation and Pre Big Bang Models
Inflationary models were originally introduced to address geometric puzzles of the hot Big Bang picture, such as the horizon and flatness problems [
12,
13,
14]. By positing a brief epoch of accelerated expansion, inflation dilutes curvature and relics and stretches quantum fluctuations to super horizon scales. In most inflationary scenarios the large scale dynamics remains effectively Markovian in the sense used here: the background evolution is described by a local equation of state or a scalar field potential, and the cosmic history is still an initial value problem on a spacelike slice.
The present work does not attempt to replace inflation or to decide what preceded the hot Big Bang. The question being addressed is different. Given a non-Markovian phenomenological framework at late times, what does the data say about the range of past times that remain dynamically relevant for the present expansion? The ITP kernel is reconstructed entirely from late time and growth data, without assuming a specific inflationary potential or a particular pre Big Bang scenario.
The effective memory horizon inferred here, –80 Gyr in representative fits, should therefore be read as an operational property of a non-Markovian response kernel, not as a claim that “the universe is 70 Gyr old” or that inflation did not occur. Inflation may still have taken place as part of the early history. The point is that, even after any such early phase, the late time universe can behave as a system with structural memory over several Hubble times, and current low redshift data are compatible with, and in fact prefer, such behaviour when it is allowed.
5.6. Five–Parameter Early–Universe Robustness Test
To check whether the preference for long memory survives in a more flexible early–universe fit, a five–parameter ITP model was run on the combined
compilation using a script. In this setup the parameter vector is
where
controls the strength of the non–Markovian correction,
sets the characteristic delay of the memory kernel in Hubble units, and
rescales the growth amplitude.
The fiducial run 5-parametre run yields the following median parameters with asymmetric
credible intervals:
The total chi–square at the median point is modest, for the combined dataset, indicating a good fit.
Two features are worth stressing. First, the background parameters remain in a conventional late–time range and are consistent, within the quoted uncertainties, with the two–bin ITP runs discussed above. The non–Markovian correction is positive and of order , showing that a non–zero memory term is preferred when the model is allowed to explore it. Second, the delay parameter again lies firmly in the long–memory regime: despite the broad posterior, even the lower edge of the interval sits well above unity, , while the median value is . In other words, the five–parameter early–universe fit independently reproduces the main conclusion of the two–bin analysis: short–delay, effectively Markovian behaviour () is disfavoured by the combined data when a non–local response is permitted.
For a simple exponential kernel with characteristic scale
, this range of
corresponds to an effective memory horizon
for
, i.e., many Hubble times. The precise mapping from
to
is kernel–dependent, but the qualitative result is robust: within this class of ITP models the late–time expansion behaves as if it retains dynamical memory over timescales far longer than the canonical
Gyr inferred in a Markovian
CDM framework. To make the contrast between the long–memory solutions and the forced short–delay regime more transparent,
Table 2 summarises the posterior medians and
credible intervals for the key ITP parameters in three representative fits: the fiducial two–bin model with a broad delay prior, the same two–bin model forced into an approximate Markovian limit
, and the five–parameter early–universe run discussed in
Section 5.6. Across these experiments the background parameters
remain stable in the long–delay solutions and become distorted only when
is artificially restricted, while the delay parameter itself consistently prefers
whenever the model is allowed to explore that region of parameter space.
5.7. Residual Comparison with a Best–Fit CDM Model
To benchmark the non-Markovian ITP fit against a standard CDM background, a direct residual comparison was carried out using script. For each and data point the code computes the individual contribution to under the best–fit CDM model and under the fiducial ITP model, and then identifies the measurements with the largest change in local goodness–of–fit.
On global statistics the two models perform comparably on both the
and
blocks. Using the joint fits summarised in Eqs. (30)–(39) and
Table 3, the baseline
CDM solution yields
, while the five–parameter ITP fit gives
for the same
compilation. In other words, the history–dependent model is statistically competitive with standard
CDM on this dataset; its main role here is to expose and constrain an explicit delay scale rather than to improve the global goodness–of–fit.
The code also lists the individual measurements with the largest change in local when going from CDM to ITP. For the block, the most affected points are low– to intermediate–redshift measurements at , , , and a higher–redshift point at , with single–point contributions in the range – under both models. For the block, the largest changes cluster around –, where several data points already contribute strongly to the global in CDM (e.g., –14 per point), and the ITP fit typically increases these contributions by order unity.
Figure 1.
Real–data fit comparing the best–fitting CDM model (solid orange) with the ITP background (dashed green) using the compilation. The data points show the cosmic–chronometer measurements with uncertainties. The two curves are almost indistinguishable at the current level of observational precision.
Figure 1.
Real–data fit comparing the best–fitting CDM model (solid orange) with the ITP background (dashed green) using the compilation. The data points show the cosmic–chronometer measurements with uncertainties. The two curves are almost indistinguishable at the current level of observational precision.
Taken together, this comparison shows that the preference for long memory in the ITP framework does not arise from a handful of outlying points or from a dramatically better global fit to either or . Instead, the non–Markovian model achieves a comparable background fit while introducing an explicit delay scale; that delay is then pushed into the long–memory regime by the joint constraints from expansion and growth, even though the overall remains slightly lower for a standard CDM interpretation of the same dataset.
Figure 2.
Comparison of the best–fitting CDM (solid orange) and ITP (dashed green) models from the joint + analysis. Left: expansion history with cosmic–chronometer measurements. Right: growth rate with redshift–space distortion data. The ITP curve uses the reconstructed effective equation of state and the best–fitting growth amplitude .
Figure 2.
Comparison of the best–fitting CDM (solid orange) and ITP (dashed green) models from the joint + analysis. Left: expansion history with cosmic–chronometer measurements. Right: growth rate with redshift–space distortion data. The ITP curve uses the reconstructed effective equation of state and the best–fitting growth amplitude .
5.8. CDM age from the same dataset
For reference it is useful to compute the effective
CDM age of the universe implied by the same
compilation used for the ITP runs. This was done with the script
age_lcdm_from_Hz_growth_bestfit_v1.py, which evaluates the age integral
with
corresponding to
for the present calculation.
The best–fit
CDM parameters for this dataset are
for which the Hubble time is
. The dimensionless age integral evaluates to
for
, yielding an effective age
This number should not be over–interpreted as a precise estimate, since it depends on the exact compilation, the choice of and the neglect of radiation at very high redshift. Its main role here is to provide a consistent baseline: when the same data are read through a Markovian lens, they imply an effective age of order 12–14 Gyr, whereas in the non–Markovian ITP framework the corresponding memory horizon inferred from the delay parameter is several times larger. In that sense the memory horizon constrained in this paper is not a replacement for the CDM age, but an additional timescale that quantifies how far back the present expansion remains dynamically entangled with its own past.
5.9. Parameter Correlations in the Five–Parameter ITP Run
To assess whether the inferred long delay
is simply a proxy for degeneracies with the background parameters or growth amplitude, a correlation analysis was performed on the five–parameter ITP chain (ITP–EARLY–HZ–5PARAM–20260118–110133). The sampled parameter vector is
where
controls the non–Markovian correction,
the delay scale in Hubble units, and
rescales the growth amplitude. From
posterior samples the standard deviations are
and the corresponding correlation matrix is
Two aspects are noteworthy. First, the usual background degeneracy between and is clearly visible (correlation coefficient ), and both parameters are anti–correlated with the growth amplitude as expected (, ). The non–Markovian amplitude is moderately entangled with the background (, ), reflecting the fact that a change in the kernel strength can partially mimic a change in the late–time expansion rate.
Second, and most important for the present work, the delay parameter is only weakly correlated with all other parameters: , , and . This shows that the preference for a long delay in the five–parameter fit is not simply a reparametrisation of the usual – degeneracy, nor is it driven by a tight coupling between and the kernel amplitude or growth normalisation. Within this parametrisation the delay scale is effectively an independent direction in parameter space: the data select even when , , and are allowed to vary freely. As a result, the inference of a long memory horizon is best understood as a genuine constraint on the temporal extent of the response kernel, rather than as a side–effect of background or growth parameter degeneracies.
6. Joint and fits
First fit a baseline flat
CDM model to the combined
and
compilation using a three–parameter state
. The MCMC run with 64 walkers and 12 000 steps (post burn–in
) yields the median constraints
with goodness of fit
For the five–parameter ITP model with a history–dependent effective equation of state, fit
to the same data. The corresponding run (64 walkers, 12 000 steps,
after burn–in) gives
with
The total is essentially unchanged relative to CDM, so the ITP model is statistically competitive on this dataset while shifting the preferred expansion rate from to and slightly favouring a non–zero memory depth and transition scale
Figure 3.
Comparison of the best–fitting CDM (solid orange) and ITP (dashed green) models from the joint + analysis. Left: expansion history with cosmic–chronometer measurements. Right: growth rate with redshift–space distortion data. The ITP curve uses the reconstructed effective equation of state and the best–fitting growth amplitude .
Figure 3.
Comparison of the best–fitting CDM (solid orange) and ITP (dashed green) models from the joint + analysis. Left: expansion history with cosmic–chronometer measurements. Right: growth rate with redshift–space distortion data. The ITP curve uses the reconstructed effective equation of state and the best–fitting growth amplitude .
6.1. Growth Amplitude and Effective Age
Given the background solution from the 5–parameter ITP fit, we recomputed the linear growth history by solving the growth ODE with the tabulated
and fitting a single amplitude
to the
data. Fixing
to their joint–fit medians and allowing only
to vary yields
consistent with the joint–fit value
within
. In other words, once the background history is fixed by the ITP kernel, a single growth–amplitude parameter remains sufficient to describe the current
compilation.
Using the same
table the study also computes the effective cosmic age,
with
. For the best-fit ITP the study obtains
This is close to the age implied by the best–fit CDM solution from the same data ( Gyr), confirming that the history–dependent ITP reconstruction does not require an exotic cosmic age to match the low–redshift expansion and growth constraints.
6.2. Planck Constraints with and Without ITP Memory
For comparison with CMB–only constraints the study analysed Planck 2018 TTTEEE+lowE+lensing with an external SZ prior, first in baseline CDM and then in an ITP–inspired wCDM extension.
Resuming the
CDM + psi
Cobaya chain gives, after 30% burn–in (
),
with minimum
for the Planck likelihood.
Allowing a constant effective equation of state
w in the ITP mapping and reusing the same likelihood configuration yields
with
, only
worse than the
CDM baseline despite the extra parameter. The preferred
w remains consistent with
at the
level, so Planck alone neither demands nor rules out the mild deviation from
CDM encoded in the ITP kernel. What it does show is that an ITP–like history dependence can be introduced without spoiling the excellent Planck fit.
6.3. Localized Kernel Fit Without Explicit w(z) Reconstruction
To test whether the inferred kernel behaviour depends on the intermediate representation, the low–redshift analysis was repeated without constructing any explicit table. Instead, the localized kernel parameters were treated as the primitive degrees of freedom, and a single joint fit was performed directly to the combined and data in the five–parameter space . This removes the growth pipeline layer and isolates the question of whether the data constrain the kernel parameters themselves.
The posterior from this direct fit lies in the same region as the baseline implementation and returns
At the posterior median, the goodness of fit is , , and . The corresponding CDM fit to the same late–time data gives , i.e., . Thus, on late–time expansion and growth data alone, the localized–kernel ITP fit and the CDM baseline are statistically indistinguishable at current precision, with a slight preference for CDM in raw .
Because the localized–kernel fit introduces two additional parameters relative to the three–parameter late–time CDM baseline, it is also useful to report standard information criteria on the same dataset. With data points ( and ), the Gaussian approximations and yield and in favour of CDM. On this dataset the localized kernel therefore does not improve model economy, even though it remains compatible with the data.
The parameter–dependence structure is consistent with the earlier analysis. The memory amplitude is moderately anti–correlated with (), reflecting a trade–off between instantaneous expansion and the size of the late–time correction, while is nearly uncorrelated with the horizon parameter (). The horizon parameter itself exhibits only weak correlations with other parameters (), indicating that the inference separates “strength” and “horizon” as largely independent kernel features.
Finally, integrating the localized–kernel background to high redshift () yields an effective age , essentially identical to the CDM value () obtained from the same late–time dataset. In this operational sense, the localized kernel acts as a late–time non–Markovian correction without materially changing the global age inferred from integration.
7. Discussion
7.1. Memory Horizon Is Not Age
It is tempting to treat a number such as Gyr as a revised age of the universe. The analysis here does not justify that interpretation. The usual age remains a Markovian quantity, defined by integrating backwards in a model where the expansion rate carries no explicit memory of earlier states.
The memory horizon is a different kind of object. It measures how much of the present expansion, in a non-Markovian framework, can be traced to earlier internal energy and structural history. A long means that the universe behaves as if it is still dynamically entangled with states far beyond the standard Hubble time, but it does not assign a unique starting time to the whole system.
In that sense and are complementary summaries of the same deeper dynamics. A finding that is a signal that the Markovian framing was incomplete, not that the age was simply miscalculated.
7.2. Connection to Cyclical Models
Long memory horizons arise naturally in many cyclic or bouncing cosmologies, in which successive phases of expansion and contraction leave imprints on later observables (e.g., [
15,
16,
22]). The ITP framework was motivated by a broader picture in which the universe evolves through repeated phases of transformation and reorganisation [
7].
The results here are structurally compatible with such scenarios. If late time observables carry structural memory from earlier phases, then a non-Markovian kernel with support over several Hubble times is a natural outcome. The present fits do not prove that the universe is cyclic, but they show that current data allow, and in fact favour, a long memory regime of the kind that cyclic models often assume. It is worth noting that the long–delay solutions preferred by the ITP fits echo, at cosmological scale, the “processor–to–reservoir” transition seen in galaxy–scale homeostasis tests. In the homeostatic potential framework, galaxies cross a structural gate beyond which internal state and stored history dominate over instantaneous environment. The present results suggest that late–time cosmology enters an analogous reservoir regime: the expansion rate is more strongly constrained by an accumulated memory kernel than by strictly local, Markovian closure. In this sense, the ITP memory horizon and the galaxy–scale flip are two manifestations of the same underlying transition from processor to reservoir dynamics.
7.3. Limits of Age-Based Reasoning
From the point of view of an internal observer, age is always inferred from local gradients and correlations. It is a projection of limited information onto a single number. If the underlying dynamics is non-Markovian, then this projection can miss important structure. An internal observer can meaningfully ask how far back local patterns remain correlated, but cannot easily assign an absolute “birthday” to the system without additional assumptions about what lies beyond the observed memory horizon.
The memory horizon defined in this work provides an explicit way to separate these questions. Late time data can constrain the range of past states that remain dynamically relevant in a non-Markovian cosmology, without making strong claims about what happened before that range. This is a modest but useful shift in emphasis. It suggests that, in a universe with structural memory, age is not the only clock that matters.