Preprint
Article

This version is not peer-reviewed.

New Issues About the Central Limit Theorem Based on a More Comprehensive Approach to Probability: Developments and Future Perspectives

Submitted:

04 January 2026

Posted:

05 January 2026

You are already at the latest version

Abstract
A more comprehensive approach to probability studies an infinite number of probability laws that are formally admissible. Thus, an infinite number of weighted averages can be handled. Areinterpretation of the central limit theorem is accordingly shown. The deviations or errors from a fixed value are calculated. It is proved that they are normally distributed. Furthermore, such deviations are invariant with respect to geometric translations identifying repeated samples. In this paper, the way of understanding the statistical model to which a specific and pragmatic distribution is compared is not a functional scheme in the continuum, but it is itself a specific and pragmatic distribution. It is possible to enlarge the reasoning, so developments and future perspectives that underlie the reinterpretation of the central limit theorem are discussed.
Keywords: 
;  ;  ;  ;  ;  

1. Introduction

A random variable is a real-valued function on S, where S is a non-empty set. It is a function from S into the set R of real numbers such that the pre-image of any interval of R is an event in S. A random variable can also be called a random quantity whenever the atomistic meaning of the notion of event is accepted instead of the collective one ([1]). Therefore, an event is not a repeatable fact, but it is a single case that can happen or not when uncertainty ceases ([2]). There are different types of repeatable facts. Hence, a repeatable fact may be a finite class of trials, or an infinite sequence of trials, or an indefinite category of trials. Another meaning of probability is associated with repeatable facts ([3]). If uncertainty does not cease, then the true value of a given quantity is not known for a certain individual, even if it is intrinsically well-determined, that is, individuated without any possibility of error. The term random is therefore used to denote a given quantity, whose true value is unknown for a given individual. Uncertainty does not cease whether there is an incomplete state of information and knowledge associated with a given individual. If each element of S is a numerical value expressing an alternative that can be true or false after observing the outcome of a given experiment, then S is reasonably a finite set. In this paper, we focus on it.
It is necessary to distinguish two things that should be distinct: the logic of certainty and the logic of uncertainty. The logic of certainty is studied by mathematics only. The logic of uncertainty obeys the rules of probability theory. However, such objective rules are moved by psychological and subjective sensations and expectations. Thus, extralogical sensations and expectations enter into the logic of uncertainty. Unfortunately, these two logics are not distinct according to the most widely used approaches to probability at the present time. They are merged. The failure to crystallize this essential distinction could depend on the fact that it is likely not to understand the difference between saying that something is pragmatically certain and saying that it is absolutely certain. Saying that something is pragmatically certain means that it is possible to give a very high probability to its occurrence, even if its occurrence is still uncertain. Conversely, saying that something is absolutely certain means that the fact that it has occurred is sure. The lack of a clear distinction between the two above logics is also seen as a trick that is played in such a way that probabilistic statements can be considered as true statements in an objective sense. The need to distinguish the two above logics can be summarized in the following slogan: prevision is not prediction. Previsions to which the logic of uncertainty leads a given individual consist of a distribution of his psychological and subjective expectations among all possible alternatives. Such a distribution happens in accordance with his degree of belief in the occurrence of each alternative. Only the distinction between possible and impossible alternatives falls inside the logic of certainty. Conversely, a prediction or prophecy consists of the statement that something will not happen, even if it is logically possible, or that something will happen, even if it is not logically certain. The domain of the logic of certainty is not therefore left to make a prediction. On the contrary, this domain is left to make a prevision. One has therefore to enter into a new field to make a prevision that is based on subjective opinions. This field obeys the rules of the logic of uncertainty. This field obeys the rules of the calculus of probability. After knowing the outcome of a given experiment, a prediction can be said to be correct or incorrect. A prevision, it does not matter what happens, cannot be said to be correct or incorrect. This is because a prevision is always based on the judgment of a given individual at a given moment. Such a judgment depends on his variable state of information and knowledge at that moment. If this state changes, then all previsions based on it also change. In this paper, we focus on fair evaluations or estimations of quantities such that undesirable decisions or choices are avoided. For this reason, a mathematical model considering the deviations or errors of each possible alternative from a specific and coherent prevision is set. A more comprehensive approach to probability says that coherence reduces to finite additivity and non-negativity. If a finite partition of events appears, where these events are the possible values for a random quantity, then the sum of their probabilities has coherently to be equal to 1. If a partition is infinite, in the sense that a countable infinity of possible values x h , h = 1 , 2 , , for a random quantity X appears, then specific probabilities denoted by p h , which are either positive or zero, can be attributed to x h , h = 1 , 2 , , by the individual who evaluates x h ( h = 1 , 2 , ) . Such probabilities p h might even all be zero. One writes
h p h = 1 p * 1 , ( 0 p * 1 ) .
Let I be a set or interval. One could say, if only the x h and p h are known, that it is P ( X I ) = h p h ( x h I ) whenever the set denoted by I contains a finite number of elements. Conversely, if the set denoted by I contains an infinite number of elements, then one could only say
h p h ( x h I ) P ( X I ) h p h ( x h I ) + p * .
This is because the probability p * can always be thought as deriving from elements whose number is infinite. Other approaches to probability generally postulate that countable additivity or σ -additivity holds, as for Lebesgue measure, and that the field over which the probability is studied is the whole of a Boolean algebra ([4,5]). On the other hand, every Boolean algebra gives rise to a Boolean ring and vice versa. There exists a cryptomorphism between them. Hence, other approaches to probability could also postulate that the field over which the probability is studied is a ring of events or a σ -ring of events. Any event whatsoever is always equal to 1 or 0 according to the logic of certainty. Here, every event is a particular random quantity, so the field over which the probability is studied is a finite dimensional linear space over R . This latter is the space where random quantities, meant as finite partitions of events, can be handled. In particular, linear spaces over R having different dimensions are studied. Hence, what is completely arbitrary is the dimension of a given linear space over R . One can pass from a ring to a linear space over R via the notion of an Abelian group. This notion underlies many fundamental algebraic structures, such as rings, algebras, and linear spaces. A given finite dimensional linear space over R is a linear system.
Section 2 contains preliminary notions on random sampling. Section 3 studies means and variances associated with linear combinations of two or more normal random variables. A reinterpretation of the central limit theorem is shown in Section 4, where a specific proposition is proved. Section 5 contains an analysis of the effects resulting from a specific reinterpretation of the central limit theorem. Finally, conclusions and future perspectives, where this latter are related to the study of particular relationships between variables, are contained in Section 6.

2. Preliminaries

This research paper contains a more complete and satisfactory answer to the following question: what can we expect from a random sample drawn from a known population? It is especially interesting to assess the consequences of this answer. This assessment has not yet been made in the literature. This assessment is therefore unknown at the present time. A sample where each individual of the parent population has the same probability of being sampled is said to be a random sample. There are different ways to achieve a random sampling ([6]). It is known that it is possible to sample with or without replacement. If we sample with replacement, then the n observations in a random sample are independent. The population remains fixed because we replace each individual before drawing the next. A random sampling can mathematically be specified as follows. First, it is possible to think of the population as a set of numbered balls contained in an urn that are mixed and sampled. If the population is extremely large, then the corresponding urn will only be very big. There are no conceptual difficulties associated with this simplification. Second, if the first ball is drawn at random, its number can be considered as a random variable X 1 that assumes a value within all the values ​​of the population. All the values of the population can therefore be studied together with their absolute and relative frequencies. The second ball that is drawn at random is denoted by X 2 , the third one is denoted by X 3 , and so on. A sample with n independent observations X 1 , X 2 , , X n is expressed by ( X 1 , X 2 , , X n ) and it has the following property. The probability distribution of each X i , i = 1 , , n , is the population probability distribution denoted by p ( x ) . If p 1 denotes the probability function of X 1 , p 2 denotes the probability function of X 2 , …, p n denotes the probability function of X n , then it is possible to write
p 1 ( x ) p 2 ( x ) p 3 ( x ) p n ( x ) p ( x ) ,
where ≡ means identically equal for all the x values of the population. Each observation has the mean μ and standard deviation σ of the population ([7]). It is known that (3) holds if we sample with replacement. Nevertheless, (3) remains valid even if we sample without replacement. This happens when the parent population is finitely large or even infinite ([8,9,10]).

3. Linear Combinations of Two or More Normal Random Variables: Means and Variances

Let X 1 and X 2 be two normal random variables. Any linear combination of them is given by
Z = a X 1 + b X 2 ,
with a and b that are real numbers. It is known that even Z is a normal random variable. If the average or mean of Z is expressed via the expectation operator E , then one writes
E ( Z ) = a E ( X 1 ) + b E ( X 2 ) .
Conversely, the variance of Z is given by
var ( Z ) = a 2 var ( X 1 ) + b 2 var ( X 2 ) + 2 a b cov ( X 1 , X 2 ) .
If X 1 and X 2 are independent, then the variance of Z is the following
var ( Z ) = a 2 var ( X 1 ) + b 2 var ( X 2 ) .
More generally, if the n observations X 1 , X 2 , , X n in a random sample are normal, then any linear combination of them is normal. We focus on linear combinations such that the sum of the coefficients of each combination is equal to 1, so one writes
Z = p 1 X 1 + p 2 X 2 + + p n X n ,
with p 1 + p 2 + + p n = 1 . Furthermore, one has 0 p i 1 , i = 1 , , n . In particular, the sample mean Z ¯ is given by
Z ¯ = 1 n X 1 + 1 n X 2 + + 1 n X n ,
where one has 1 n + 1 n + + 1 n = 1 n · n = 1 . The first moment of the sample mean is the expected value of Z ¯ . It is known that one has
E ( Z ¯ ) = μ .
The second moment of the sample mean is the variance of Z ¯ , so one has
var ( Z ¯ ) = σ 2 n .
It follows that the standard deviation of Z ¯ is
σ n .
This typical deviation of Z ¯ from its target μ coincides with the estimation error. It is called the standard error of Z ¯ , so one has
Standard   Error   of   Z ¯ = σ n .
The larger the value of n, the smaller the standard error of Z ¯ becomes. Thus, the standard error of Z ¯ shrinks as the sample size n increases. Finally, if the parent population, whose variance is not equal to + , is normal, or the sample size is large, then in either case the sampling distribution of Z ¯ has an approximately normal shape. If it is n = 10 or n = 20 , then sample sizes will often be large enough. One writes
Z ¯ N ( μ , σ 2 n ) .
It is known that the central limit theorem has always attracted the attention of many researchers ([11]). Many researchers have often attempted to enlarge what the central limit theorem intrinsically contains ([12,13,14,15,16]). One can be very precise in making deductions about the sample mean whenever a random sample is drawn from a known population. Because of averaging, the sample mean is not as extreme as the individuals in the population ([17]). Its great convenience is its fairness, in the sense that an unbiased estimation of the mean μ of the underlying population is carried out ([18,19]). Furthermore, sampling uncertainty can suitably be handled.

4. A Reinterpretation of the Central Limit Theorem

A specific weighted average has been denoted by Z ¯ . Nevertheless, there exists an infinite number of weighted averages that can be calculated. This happens when a more comprehensive approach to probability is used ([20]). In this approach, every weighted average is a coherent prevision of a random variable ([21]). This variable is normal because it is obtained via a linear combination of n normal random variables. Deviations or errors from a specific weighted average can be calculated as follows. First, a specific mean is calculated. Second, the errors are calculated with respect to this specific mean. Deviations or errors are normally distributed, with mean that is equal to zero and variance that is equal to k σ 2 = σ D 2 . One writes
E ( D ) = 0 ,
so positive and negative errors are distributed around the value zero, and
var ( D ) = σ D 2 = k σ 2 ,
where k is the sum of quadratic weights. Here, D denotes a particular variable identifying deviations as its possible values. It is possible to prove the following:
 Proposition 1.
If a more comprehensive approach to probability is used, then an infinite number of weighted averages of the n independent observations X 1 , X 2 , , X n can be calculated. A specific random variable denoted by D is studied and its moments are such that one has E ( D ) = 0 and var ( D ) = σ D 2 = k σ 2 .
 Proof. 
Given
Z = p 1 X 1 + + p n X n ,
with p 1 + + p n = 1 , and 0 p i 1 , i = 1 , , n , it is possible to consider the possible values of another variable denoted by D. Such values are the possible deviations from E ( Z ) . The possible values of D are therefore the elements of the set denoted by
I ( D ) = { X 1 E ( Z ) , , X n E ( Z ) } .
Here, E ( Z ) = E ( Z ¯ ) = μ . This is because one has
E ( Z ) = p 1 E ( X 1 ) + + p n E ( X n ) = p 1 μ + + p n μ = ( p 1 + + p n ) μ = μ .
It is possible to write
p 1 ( X 1 E ( Z ) ) + + p n ( X n E ( Z ) ) ,
where the weights of (19) are the same of the ones characterizing Z . One has
p 1 E ( X 1 E ( Z ) ) + + p n E ( X n E ( Z ) ) = p 1 ( μ μ ) + + p n ( μ μ ) = 0 ,
so it turns out to be E ( D ) = 0 . Conversely, from
var p 1 ( X 1 E ( Z ) ) + + p n ( X n E ( Z ) ) ,
it follows that one writes
p 1 2 var ( X 1 ) + + p n 2 var ( X n ) = ( p 1 2 + + p n 2 ) σ 2 ,
with ( p 1 2 + + p n 2 ) = k , so it is var ( D ) = σ D 2 = k σ 2 . □
The sum of quadratic weights p i 2 , i = 1 , , n , is between 0 and 1. It is never equal to zero. It is equal to 1 only in one case, that is, when all weights except one are equal to zero and there is accordingly only one weight that is equal to 1. Note the following:
 Remark 1.
Since Z and Z ¯ are two linear combinations of n independent random variables such that the sum of the corresponding non-negative coefficients or weights is always equal to 1, one has E ( Z ) = E ( Z ¯ ) = μ . Hence, this result does not depend on the condition according to which one has to observe p 1 = 1 n , , p n = 1 n , even if one writes 1 n + + 1 n = 1 n · n = 1 .

5. An Analysis of the Effects Resulting from a Specific Reinterpretation of the Central Limit Theorem

There exists an infinite number of probability laws that can be used to handle the n observations X 1 , X 2 , , X n in a random sample. These laws are all formally admissible. They characterize a more comprehensive approach to probability based on subjective probabilities ([22,23]). The n observations X 1 , X 2 , , X n in a random sample are n random events, where each event is a possible value for a random quantity. This latter is linearly dependent from n random entities. The possible values for a random quantity are logically independent. Thus, there exists an infinite number of probability laws that can be associated with the elements of the set E of events, denoted by E 1 , E 2 , , E n in general. Such events identify a finite partition. One focuses on the set
{ x 1 , x 2 , , x n } .
This is because X 1 takes on a specific value denoted by x 1 , …, X n takes on a specific value denoted by x n . In the absence of uncertainty only the indicators of each event E i , i = 1 , , n , can be considered. The indicators of each event E i , i = 1 , , n , take on 1 or 0 as their values. The values of the indicators of each event E i , i = 1 , , n , can therefore be two idempotent numbers only. However, there is uncertainty here, so one focuses on x 1 , x 2 , , x n together with their probabilities p 1 , p 2 , , p n . If one writes
p 1 x 1 + + p n x n ,
with p 1 + + p n = 1 , and 0 p i 1 , i = 1 , , n , then the prevision, denoted by P , or mathematical expectation of a random quantity is studied. In this approach, it is possible to use P in place of E although P and E have the same properties. There exists a unique notion, denoted by a unique symbol, which in general we call prevision P and, in the case of events, also probability P . If the possible values for a random quantity are put on the real line which is a geometric line isomorphic to the set R of real numbers, then all probability laws applied to x 1 , x 2 , , x n that are formally admissible identify a closed line segment. This latter is a one-dimensional convex set.

5.1. Probability Laws That Are Formally Admissible

Let
P ( X ) = p 1 x 1 + + p n x n
be the mathematical expression of the prevision or mathematical expectation of X, where X is a random quantity identifying a linear combination of n events expressed by
X = x 1 | E 1 | + + x n | E n | .
We do not yet know which of the n events that are taken into account in (26) will be true in the absence of uncertainty. For this reason, n probabilities are introduced to calculate P ( X ) . All probability laws that are formally admissible are infinite in number. They must be coherent. Since the n observations X 1 , X 2 , , X n in a random sample are seen as n mutually disjoint events, every coherent probability law that can be chosen has to satisfy the fundamental property of coherence according to which P ( X ) must not be less than the lower bound of the set given by
{ x 1 , x 2 , , x n } ,
nor greater than the upper bound of it. Check the following:
 Example 1.
Let x 1 = 5 , x 2 = 7 , x 3 = 10 , and x 4 = 11 be the 4 observations in a random sample. They are the possible values for a random quantity denoted by X. Such values characterizing a bounded quantity from above and below are taken into account together with 4 non-negative weights. They are p 1 = 0.2 , p 2 = 0.3 , p 3 = 0.4 , and p 4 = 0.1 , where it turns out to be p 1 + + p 4 = 1 , and 0 p i 1 , i = 1 , , 4 . Thus, these weights identify a probability law that is formally admissible. Since it is P ( X ) = 8.2 , it is possible to calculate the errors or deviations from P ( X ) . They are d 1 = 3.2 , d 2 = 1.2 , d 3 = 1.8 , and d 4 = 2.8 . One has
p 1 d 1 + + p 4 d 4 = 0
because of a fundamental property qualifying the deviations or errors from the supposed center of a distribution. If the deviations from P ( X ) are not all negative, then a probability law is formally admissible. Similarly, if the deviations from P ( X ) are not all positive, then a probability law is formally admissible. In particular, given x 1 = 5 , x 2 = 7 , x 3 = 10 , and x 4 = 11 , if a probability law is expressed by p 1 = 0 , p 2 = 0 , p 3 = 0 , and p 4 = 1 , then it is formally admissible because the corresponding deviations are not all negative. One of them coincides with 0. Similarly, if a probability law that is referred to the same possible values for X is expressed by p 1 = 1 , p 2 = 0 , p 3 = 0 , and p 4 = 0 , then it is formally admissible because the corresponding deviations are not all positive. One of them coincides with 0. Conversely, given x 1 = 5 , x 2 = 7 , x 3 = 10 , and x 4 = 11 , if it is P ( X ) = 4 , then the corresponding deviations are all positive. Similarly, if it is P ( X ) = 13 , then the corresponding deviations are all negative. Here, in either case, the fundamental property of coherence according to which P ( X ) must not be less than the lower bound, given by x 1 = 5 , of the set expressed by
{ x 1 , x 2 , , x 4 } ,
nor greater than the upper bound of it, given by x 4 = 11 , does not hold. If we put x 1 = 5 , x 2 = 7 , x 3 = 10 , and x 4 = 11 on the real line, then all probability laws that are formally admissible in a first stage identify a closed line segment. Its endpoints are two extreme points, expressed by x 1 = 5 and x 4 = 11 respectively, of a one-dimensional convex set.
Note the following:
 Remark 2.
If we sample with replacement, then it is possible that two or more observations in a random sample are equal. If this happens, then nothing changes, in the sense that a formally admissible probability law is such whenever the sum of the corresponding probabilities is equal to 1. For instance, let x 1 = 6 , x 2 = 8 , x 3 = 12 , x 4 = 14 , x 5 = 8 , and x 6 = 16 be the 6 observations in a random sample. A formally admissible probability law is such that one writes p 1 = 0.1 , p 2 = 0.05 , p 3 = 0.35 , p 4 = 0.2 , p 5 = 0.05 , and p 6 = 0.25 . On the other hand, if we set x 1 = 6 , x 2 = 8 , x 3 = 12 , x 4 = 14 , x 5 = 16 , and x 6 = 0 together with p 1 = 0.1 , p 2 = 0.1 , p 3 = 0.35 , p 4 = 0.2 , p 5 = 0.25 , and p 6 = 0 , then nothing changes. The two ways of writing the set of ordered pairs given by [ x i , p i ] , i = 1 , , 6 , are equivalent according to the language of calculus.
A more comprehensive approach to probability based on subjective probabilities clearly distinguishes between two aspects ([24]). The logical aspect must be clearly distinguished from the empirical one. Other approaches to probability tend to merge these two aspects. In particular, here the logical aspect is such that the sum of n probabilities must be equal to 1. Conversely, the empirical aspect is such that n 1 probability laws can formally be admitted in a first stage. This is because every probability law reflects a specific subjective opinion associated with a given individual. There exists an infinite number of possible opinions about the evaluations of probability. On the other hand, it is possible that a particular subjective opinion coincides with the one of many or all people. If this happens, then nothing changes. This is because the ambit where subjective probability is available is not subject to any limitation. Other approaches to probability, typically those for which probability is not a degree of belief of an individual in the occurrence of a specific event, but it is a conventional notion that is subject to rigid mathematical laws only, say that a unique way of identifying those probabilities through which a mathematical calculation is carried out has to be considered.

5.2. A Deceptive Dichotomy Between Statistics and Probability

An event is always a logical entity, so it is always a proposition. On the other hand, this does not exclude that an event may coincide with a set individuated by a proposition. Check the following:
 Example 2.
A finite population, whose size is N = 800 , is taken into account for studying a specific quantitative issue. Let H be a continuous variable. It is normally distributed, with mean μ = 168 centimeters and standard deviation σ = 13 centimeters. Its values potentially belong to the set R of real numbers. This latter is an interval denoted by ( , + ) = R . Every event is a set individuated by a proposition. Its probability can uniquely be determined. Since the probability that H lies in the interval [ a , b ] is equal to the area under f ( x ) between x = a and x = b , where f ( x ) is the probability density function of H, one has
P ( a H b ) = a b f ( x ) d x .
The standardized variable corresponding to H is defined by
Z = H μ σ ,
where Z is also a normal distribution, with μ = 0 and standard deviation σ = 1 . The probability density function for Z is obtained by setting z = ( h μ ) / σ , so one writes
ϕ ( z ) = 1 2 π e z 2 / 2 .
The main interval can be divided into equal-width subintervals. The area under the standard normal distribution ϕ ( z ) between 0 and z 0 is given by
P ( 0 Z z ) = 0 z ϕ ( z ) d z = Φ ( z ) ,
so it is possible to calculate the absolute frequencies of a frequency distribution. This happens whenever N = 800 is multiplied by the probability that corresponds to a specific subinterval. The number of subintervals is finite. It has to be a reasonable compromise between too much detail and too little. Each subinterval midpoint is chosen in such a way that it is a whole number. Each subinterval midpoint represents all observations that are contained in the corresponding subinterval. It is also possible to calculate the relative frequencies, so we can pass from a frequency distribution to a random quantity. Each subinterval midpoint is a possible value for a discrete random quantity. We can use the relative frequencies as probabilities to calculate the prevision of this random quantity.
Note the following:
 Remark 3.
A certain subjective opinion may have a specific reason for being held. This opinion can be based on the observation of frequencies. For this reason, it can be said to be objectively true. However, nothing changes. It is still a subjective opinion which is rigorously studied like all other subjective opinions.
 Remark 4.
A frequency distribution and a random quantity are the two sides of the same coin. If each element of the sample space denoted by S is pragmatically observed, then S is necessarily a finite set. It may be convenient to think of S as embedded in a larger and manageable space denoted by R . Thus, a continuum of real numbers such as an interval appears. Each element of S is an event or proposition. Each element of S can also be a set individuated by a proposition. A particular set that is individuated by a proposition is given by the interval [ a , a ] = { a } , with a R .

5.3. The Prevision of a Discrete Random Quantity Is a Scalar Product

The prevision of a discrete random quantity is an algebraic operation that takes two equal-length sequences of real numbers and returns a real number ([25]). Two sequences of real numbers are respectively the possible values for a random quantity denoted by x and their probabilities denoted by p . The algebraic structure where this operation happens is a finite dimensional linear space over R ([26,27,28]). In particular, finite dimensional linear spaces over R can have a Euclidean nature, so the scalar product of two ordered sequences of real numbers is the dot product of their Cartesian coordinates. Such a product is independent from the choice of a particular Cartesian coordinate system. Such a product is therefore independent from the choice of a specific orthonormal basis. The scalar product of two ordered sequences of real numbers satisfies the following properties:
x , p = p , x ,
where x and p belong to the same finite dimensional linear space over R ;
p , x + y = p , x + p , y ,
where p , x , and y belong to the same finite dimensional linear space over R . Furthermore, p represents the same probabilities associated with the possible values for X and the possible values for Y in such a way that one writes P ( X + Y ) = P ( X ) + P ( Y ) ;
a p , x = a p , x , a n d p , a x = a p , x ,
with a R . In addition, if all possible values ​​for two or more random quantities become impossible at a later time, then the condition according to which the scalar product of two ordered sequences of real numbers is said to be non-degenerate is satisfied. Since the scalar product of two ordered sequences of real numbers is independent from the choice of a specific orthonormal basis, there exists a fundamental intrinsic property. Such a property can be extended whenever two or more logically independent random quantities are studied. If m 2 logically independent random quantities are taken into account together with their marginal distributions remaining fixed, then m 2 joint distributions having a fundamental invariance property can appear. This invariance property makes clear that the weights of some joint distributions can change unlike the ones of the corresponding marginal distributions. In this way, multilinear relationships between variables can be studied ([29,30,31]). Furthermore, particular exchangeability relationships between variables can be shown. On the other hand, many researchers dealt with the study of the notion of exchangeability ([32,33,34]). The set of events is here embedded in the finite dimensional linear space over R of random quantities. It is possible to increase the dimension of a linear space over R just as it is possible to reduce it. Everything can be represented in a linear form provided one considers an appropriate number of dimensions. In particular, a reduction of dimension is observed when the two ordered sequences of real numbers, which are respectively the possible values for a random quantity and their probabilities, are put on the real line. This line is isomorphic to R . This line is therefore a one-dimensional linear space over R .

5.4. Repeated Samples Are Invariant with Respect to Geometric Translations

The deviations or errors from a fixed value are invariant with respect to geometric translations. If s denotes the initial random sample of n observations and v denotes a vector that is known to be as the translation vector, then the translation function denoted by T v will be expressed by
T v ( s ) = s + v ,
where s and v are two equal-length ordered sequences of real numbers. Here, whenever we pass from a geometric translation to another one, s remains unchanged unlike v . As the number of samples increases indefinitely, the deviations or errors from a fixed value are invariant with respect to a geometric translation, which is expressed by v every time. The components of v are all equal, so v is every time an ordered sequence of real numbers that are all equal. As the number of samples increases indefinitely, the properties of D remain unchanged. They are such that one has E ( D ) = 0 and var ( D ) = σ D 2 = k σ 2 .

6. Conclusions

In this paper, a reinterpretation of the central limit theorem is shown. Since there exists an infinite number of probability laws that are formally admissible, it is possible to consider an infinite number of weighted averages. Thus, the way of understanding the statistical model to which a specific and pragmatic distribution is compared is not a functional scheme in the continuum, but it is itself a specific and pragmatic distribution. The deviations or errors from a fixed value are calculated. They are normally distributed. They are invariant with respect to geometric translations. It is possible to enlarge the reasoning, so developments and future perspectives that underlie the reinterpretation of the central limit theorem are discussed. In general, this research paper shows that one cannot have blind faith in the analogical arguments that suggest the use of particular analytical tools in the application of the calculus of probability to specific problems. All criteria based on measure theory are totally illusory whenever they lead to specifying the conclusion more precisely than the problem itself allows. If there are problems whose values are intrinsically indeterminate between given limits, then to use methods that authorize the derivation of a uniquely determined answer to them, via the introduction of arbitrary mathematical conventions, is unjustified and inadmissible. It is wrong that formally admissible answers to these problems are systematically excluded. It is wrong to contemplate an infinite number of alternatives, whether this number is intrinsically illusory, just to obtain a mathematical result that is shown, without any risk, as a sure prediction. What is said in this research paper implies that events and random quantities can be embedded inside finite dimensional linear spaces over R , where interesting invariance properties are discovered and made clear. Such properties allow to connect probability to multilinear issues that can innovatively be treated in different fields of science.
  • This study was not funded
  • The authors declare that they have no conflict of interest
  • This study does not contain any studies with human participants or animals performed by any of the authors
  • For this type of study formal consent is not required
  • Authors can confirm that all relevant data are included in the article

References

  1. de Finetti, B. Probabilism: a critical essay on the theory of probability and on the value of science. Erkenntnis 1989, 31, 169–223. [Google Scholar] [CrossRef]
  2. de Finetti, B. Logical foundations and measurement of subjective probability. Acta Psychologica 1970, 34, 129–145. [Google Scholar] [CrossRef]
  3. Carnap, R. The two concepts of probability: the problem of probability. Philosophy and Phenomenological Research 1945, 5, 513–532. [Google Scholar] [CrossRef]
  4. Doob, J.L. Probability as measure. The Annals of Mathematical Statistics 1941, 12, 206–214. [Google Scholar] [CrossRef]
  5. Halmos, P.R. The foundations of probability. The American Mathematical Monthly 1944, 51, 493–510. [Google Scholar] [CrossRef]
  6. Scott, A.J.; Wild, C.J. Fitting logistic models under case-control or choice based sampling. Journal of the Royal Statistical Society: Series B (Methodological) 1986, 48, 170–182. [Google Scholar] [CrossRef]
  7. Binder, D.A. On the variances of asymptotically normal estimators from complex surveys. International Statistical Review/Revue Internationale de Statistique 1983, 51, 279–292. [Google Scholar] [CrossRef]
  8. Hájek, J. Limiting distributions in simple random sampling from a finite population. Publications of the Mathematical Institute of the Hungarian Academy of Sciences: Series A 1960, 5, 361–374. [Google Scholar]
  9. Hájek, J. Asymptotic theory of rejective sampling with varying probabilities from a finite population. Annals of Mathematical Statistics 1964, 35, 1491–1523. [Google Scholar] [CrossRef]
  10. Madow, W.G. On the limiting distributions of estimates based on samples from finite universes. Annals of Mathematical Statistics 1948, 19, 535–545. [Google Scholar] [CrossRef]
  11. Davis, B.; McDonald, D. An elementary proof of the local central limit theorem. Journal of Theoretical Probability 1995, 8, 693–701. [Google Scholar] [CrossRef]
  12. Benoist, Y.; Quint, J.F. Central limit theorem for linear groups. The Annals of Probability 2016, 44, 1308–1340. [Google Scholar] [CrossRef]
  13. de Jong, P. A central limit theorem for generalized quadratic forms. Probability Theory and Related Fields 1987, 75, 261–277. [Google Scholar] [CrossRef]
  14. Hall, P. Central limit theorem for integrated square error of multivariate nonparametric density estimators. Journal of Multivariate analysis 1984, 14, 1–16. [Google Scholar] [CrossRef]
  15. Klartag, B. A central limit theorem for convex sets. Inventiones mathematicae 2007, 168, 91–131. [Google Scholar] [CrossRef]
  16. Peligrad, M.; Utev, S. Central limit theorem for linear processes. The Annals of Probability 1997, 25, 443–456. [Google Scholar] [CrossRef]
  17. Barron, A.R. Entropy and the central limit theorem. The Annals of Probability 1986, 14, 336–342. [Google Scholar] [CrossRef]
  18. Esseen, C.G. A moment inequality with an application to the central limit theorem. Scandinavian Actuarial Journal 1956, 1956, 160–170. [Google Scholar] [CrossRef]
  19. Von Bahr, B. On the convergence of moments in the central limit theorem. The Annals of Mathematical Statistics 1965, 36, 808–818. [Google Scholar] [CrossRef]
  20. Edwards, W.; Lindman, H.; Savage, L.J. Bayesian statistical inference for psychological research. Psychological Review 1963, 70, 193–242. [Google Scholar] [CrossRef]
  21. Sanfilippo, G.; Gilio, A.; Over, D.E.; Pfeifer, N. Probabilities of conditionals and previsions of iterated conditionals. International Journal of Approximate Reasoning 2020, 121, 150–173. [Google Scholar] [CrossRef]
  22. Coletti, G.; Petturiti, D.; Vantaggi, B. Possibilistic and probabilistic likelihood functions and their extensions: common features and specific characteristics. Fuzzy Sets and Systems 2014, 250, 25–51. [Google Scholar] [CrossRef]
  23. Berti, P.; Rigo, P. Finitely additive mixtures of probability measures. Journal of Mathematical Analysis and Applications 2021, 500, 125114. [Google Scholar] [CrossRef]
  24. Angelini, P. Financial decisions based on zero-sum games: new conceptual and mathematical outcomes. International Journal of Financial Studies 2024, 12, 56. [Google Scholar] [CrossRef]
  25. Angelini, P. Invariance of the mathematical expectation of a random quantity and its consequences. Risks 2024, 12, 14. [Google Scholar] [CrossRef]
  26. Jordan, P.; Neumann, J.V. On inner products in linear, metric spaces. Annals of Mathematics 1935, 36, 719–723. [Google Scholar] [CrossRef]
  27. Ficken, F.A. Note on the existence of scalar products in normed linear spaces. Annals of Mathematics 1944, 45, 362–366. [Google Scholar] [CrossRef]
  28. Falkner, N. A characterization of inner product spaces. The American Mathematical Monthly 1993, 100, 246–249. [Google Scholar] [CrossRef]
  29. Angelini, P.; Maturo, F. Tensors Associated with Mean Quadratic Differences Explaining the Riskiness of Portfolios of Financial Assets. Journal of Risk and Financial Management 2023, 16, 369. [Google Scholar] [CrossRef]
  30. Angelini, P. Extended Least Squares Making Evident Nonlinear Relationships between Variables: Portfolios of Financial Assets. Journal of Risk and Financial Management 2024, 17, 336. [Google Scholar] [CrossRef]
  31. Angelini, P. Comparisons between frequency distributions based on Gini’s approach: principal component analysis addressed to time series. Econometrics 2025, 13, 32. [Google Scholar] [CrossRef]
  32. Diaconis, P. Finite forms of de Finetti’s theorem on exchangeability. Synthese 1977, 36, 271–281. [Google Scholar] [CrossRef]
  33. Diaconis, P.; Freedman, D. Finite exchangeable sequences. The Annals of Probability 1980, 8, 745–764. [Google Scholar] [CrossRef]
  34. Spizzichino, F. A concept of duality for multivariate exchangeable survival models. Fuzzy Sets and Systems 2009, 160, 325–333. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated