Submitted:
09 March 2023
Posted:
09 March 2023
Read the latest preprint version here
Abstract
Keywords:
0. Background
1. Preliminaries
1.1. Motivation
1.2. Preliminary Definitions
- (1)
- (2)
- For all , exists (if A is countable infinite then for every , must be a finite set since is a discrete uniform distribution of ; otherwise, if A is uncountable, then is the normalized Lebesgue measure or another uniform measure on (e.g. [7]) such for every the Lebesgue measure or some other uniform measure on exists and is finite. [2, p.32-37].
- (3)
- For all , is positive and finite where is intrinsic. (For countably infinite A, is the counting measure where is positive and finite since is finite. For uncountable A, is the Lebesgue or radon-nikodym derivative on some other uniform measure on (e.g. [7]) where either of the measures on are positive and finite.)
- The element
- The set is arbitrary and uncountable.
- The element
- The set is arbitrary and uncountable.
- (1)
- Arrange the x-value of the points in the sample of uniform ε coverings from least to greatest. This is defined as:
- (2)
- Take the multi-set of the absolute differences between all consecutive pairs of elements in (1). This is defined as:
- (3)
- Normalize (2) into a probability distribution, where for multi-set X, we have as the cardinality of all elements in the multi-set, including repeated ones. This is defined as:
- (4)
- Take the entropy of (3), (for further reading, see [8, p.61-95]]). This is defined as:
- (a)
- We take the cardinality of the sample of the uniform ε coverings of (def. 5) divided by the smallest cardinality of the sample of the uniform ε coverings of (def. 5), where the entropy on the sample of uniform coverings on (def. 6) is larger than the entropy on the sample of uniform coverings on (def. 6). In other words, if:then the ratio at the beginning of the paragraph is defined (using 1.2.8) as
- (b)
-
We take the cardinality of the sample of uniform ε covering of (def. 5) divided by the largest cardinality of the sample of the uniform ε covering of (def. 5), where the entropy on the sample of uniform coverings on (def. 6) is smaller then the entropy on the sample of uniform coverings on (def. 6). In other words if:then the ratio at the beginning of the paragraph is defined (using 1.2.10) as
- (1)
- If using equations 1.2.9 and 1.2.11 we have that:we say converges uniformly to A at a superlinear rate to that of .
- (2)
-
If using equations 1.2.9 and 1.2.11 we have that:we say converges uniformly to A at a linear rate to that of .
- (3)
-
If using equations 1.2.9 and 1.2.11 we have that:we say converges uniformly to A at a sublinear rate to that of .
1.3. Question on Preliminary Definitions
- (1)
- Are there “simpler" alternatives to either of the preliminary definitions? (Keep this in mind as we continue reading).
2. Main Question
- (1)
- The expected value of f on each term of the pre-structure is finite
- (2)
- The pre-structure converges uniformly to A
- (3)
- The pre-structure converges uniformly to A at a linear or superlinear rate to that of other non-equivalent pre-structures of A which satisfies (1) and (2).
- (4)
- The generalized expected value of f on the pre-structure (i.e. an extension of def. 3 to answer the full question) satisfies (1), (2), and (3) while having an unique & finite value.
- (5)
- A choice function is defined that chooses a pre-structure from A such the following satisfies (1), (2), (3), and (4) for the largest possible subset of .
- (6)
- If there is more than one choice function that satisfies (1), (2), (3), (4) and (5), we choose the choice function with the “simplest form", meaning for a general pre-structure of A, when each choice function is fully expanded, we take the choice function with the fewest variables/numbers (excluding those with quantifiers).
3. Informal Attempt to Answer Main Question
3.1. Choice Function
3.2. Questions on Choice Function
- (1)
-
What unique pre-structure would contain (if it exists) for:
- (a)
- where if , we want
- (b)
- where if , we want
- (c)
- where we’re not sure what would be in this case. What would be if it’s unique?
3.3. Generalized Expected Values
- (1)
- For a worst-case f defined on countably infinite A (e.g. countably infinite "pseudo-random points" non-uniformly scattered across the real plane), one might typically use from equation 3.3.1 (since countable sets might need just one iteration of C for the generalized expected value to be unique); otherwise, one may use from equation 3.3.2 for finite iterations of C.
- (2)
- For a worst-case f defined on uncountable A, we might have to use from equation 3.3.3 as the function is so difficult to analyze. We can imagine this function as an uncountable number of "pseudo-random" points non-uniformly generated on a subset of the real plane (see 4.1 for a visualization.)
3.4. Questions Regarding The Answer
- (1)
-
Using prevalence and shyness [3,4], can we say the set of f where , , or , from equations 3.3.1, 3.3.2 and 3.3.3 respectively, have unique and finite values that form either a prevalent or neither prevalent nor shy subset of ? (If the subset is prevalent, this implies that either of the generalized expected values can be unique and finite for a “large" subset of ; however, if the subset is neither prevalent nor shy we need a more precise definition of “size" which takes “an exact probability that the expected values are unique & finite"—some examples (which are shown in this answer [9]) being:
- (a)
- Fractal Dimension notions
- (b)
- Kolmogorov Entropy
- (c)
- Baire Category and Porosity
- (2)
- There might be a total of 120 variables in the choice function C (excluding quantifiers). Is there a choice function with fewer variables (ignoring quantifiers) which answers criteria (1), (2), (3) & (4) of the main question in §2 for a "larger" subset of ? (This might be impossible to answer since such a solution cannot be shown with prevalence or shyness [3,4])—therefore, we need a more precise version of “size" with some examples, again, shown in [9].
- (3)
- If question (2) is correct, what is the choice function C using the most generalized expected value that fully answers the question in §2?
- (4)
- Can either , , or from equations 3.3.1, 3.3.2 and 3.3.3 respectively (when A is the set of all Liouville numbers [10] and ) give a finite value? What would the value be?
- (5)
- Similar to how definition 11 in §4 approximates the expected value in definition 1, how do approximate , , and from equations 3.3.1, 3.3.2 and 3.3.3 respectively?
- (6)
- Can programming be used to estimate , , and from equations 3.3.1, 3.3.2 and 3.3.3 respectively (if an unique/finite result of either of the expected values exist)?
3.5. Applications
- (1)
-
In Quanta magazine [11], Wood writes on Feynman Path Integrals: “No known mathematical procedure can meaningfully average1 an infinite number of objects covering an infinite expanse of space in general. The path integral is more of a physics philosophy than an exact mathematical recipe."—despite Wood’s statement, mathematicians Bottazzi E. and Eskew M. [12] found a constructive solution to the statement using integrals defined on filters over families of finite sets; however, the solution was not unique as one has to choose a value in a partially ordered ring of infinite and infinitesimal elements. In addition, although there were ways of preventing the use of the axiom of choice (within their integral), the axiom was still required for certain cases.
- (a)
- Perhaps, if Botazzi’s and Eskew’s Filter integral [12] is not enough to solve Wood’s statement, could we replace the path integral with expected values , , and from equations 3.3.1, 3.3.2 and 3.3.3 respectively? (See, again, §4.1 for a visualization of Wood’s statement.)
- (2)
-
As stated in § 1.1, “when the Lebesgue measure of A, measurable in the Caratheodory sense, has zero or infinite volume (or undefined measure), there may be multiple, conflicting ways of defining a "natural" uniform measure on A." This is an example of Bertand’s Paradox which shows, "the principle of indifference (that allows equal probability among all possible outcomes when no other information is given) may not produce definite, well-defined results for probabilities if applied uncritically, when the domain of possibilities (i.e. the sample space) is infinite [13].Using 3.2, perhaps if we take (from def. 3.1.9):then for , if we want and we get the following:Then might serve as a solution to Bertand’s Paradox (unless there’s a simpler and which helps solve the main question in ).Now consider the following:
- (a)
-
How do we apply (or a simpler solution) to the usual example which demonstrates the Bertand’s Paradox as follows: for an equilateral triangle (inscribed in a circle), suppose a chord of the circle is chosen at random—what is the probability that the chord is longer than a side of the triangle? [14] (According to Bertand’s Paradox there are three arguments which correctly use the principle of indifference yet give different solutions to this problem [14]:
- i
- The “random endpoints" method: Choose two random points on the circumference of the circle and draw the chord joining them. To calculate the probability in question imagine the triangle rotated so its vertex coincides with one of the chord endpoints. Observe that if the other chord endpoint lies on the arc between the endpoints of the triangle side opposite the first point, the chord is longer than a side of the triangle. The length of the arc is one-third of the circumference of the circle, therefore the probability that a random chord is longer than a side of the inscribed triangle is .
- ii
- The "random radial point" method: Choose a radius of the circle, choose a point on the radius, and construct the chord through this point and perpendicular to the radius. To calculate the probability in question imagine the triangle rotated so a side is perpendicular to the radius. The chord is longer than a side of the triangle if the chosen point is nearer the center of the circle than the point where the side of the triangle intersects the radius. The side of the triangle bisects the radius, therefore the probability a random chord is longer than a side of the inscribed triangle is .
- iii
- The "random midpoint" method: Choose a point anywhere within the circle and construct a chord with the chosen point as its midpoint. The chord is longer than a side of the inscribed triangle if the chosen point falls within a concentric circle of radius the radius of the larger circle. The area of the smaller circle is one-fourth the area of the larger circle, therefore the probability a random chord is longer than a side of the inscribed triangle is .
4. Glossary
4.1. Example of Case (2) of Worst Case Functions
4.2. Question Regarding Section 4.1
4.3. Approximating the Expected Value
- (1)
- We need to know when point x is in set A or not
- (2)
- We need to be able to generate points from a density g that is on a support that covers A but is not too much bigger than A
- (3)
- We have to be able to compute and for each point
References
- (https://mathoverflow.net/users/35357/michael greinecker), M.G. (https://mathoverflow.net/users/35357/michael greinecker), M.G. Demystifying the Caratheodory Approach to Measurability. MathOverflow, [https://mathoverflow.net/q/34007]. URL:https://mathoverflow.net/q/34007 (version: 2010-07-31).
- T., L.; E., R. T., L.; E., R. The maximum entropy of a metric space. https://arxiv.org/pdf/1908.11184.pdf.
- Ott, W.; Yorke, J.A. Prevelance. Bulletin of the American Mathematical Society 2005, 42, 263–290. [Google Scholar] [CrossRef]
- Hunt, B.R. Prevalence: a translation-invariant “almost every” on infinite-dimensional spaces 1992. https://arxiv.org/abs/math/9210220. [CrossRef]
- (https://stats.stackexchange.com/users/173082/ben), B. (https://stats.stackexchange.com/users/173082/ben), B. In statistics how does one find the mean of a function w.r.t the uniform probability measure? Cross Validated, [https://stats.stackexchange.com/q/602939]. https://stats.stackexchange.com/q/602939 (version: 2023-01-24).
- B. , P. 3 ed.; John Wiley & Sons: New York, 1995; pp. 419–427. [Google Scholar]
- (https://mathoverflow.net/users/46214/mark mcclure), M.M. (https://mathoverflow.net/users/46214/mark mcclure), M.M. Integral over the Cantor set Hausdorff dimension. MathOverflow, [https://mathoverflow.net/q/235609]. https://mathoverflow.net/q/235609 (version: 2016-04-07).
- M., G. M., G. 2 ed.; Springer New York: New York [America];, 2011; pp. 61–95. https://ee.stanford.edu/~gray/it. [CrossRef]
- (https://math.stackexchange.com/users/13130/dave-l renfro), D.L.R. (https://math.stackexchange.com/users/13130/dave-l renfro), D.L.R. Proof that neither “almost none” nor “almost all” functions which are Lebesgue measurable are non-integrable. Mathematics Stack Exchange, [https://math.stackexchange.com/q/4623168]. https://math.stackexchange.com/q/4623168 (version: 2023-01-21).
- Grabowski, A.; Kornilowicz, A. Introduction to Liouville Numbers. Formalized Mathematics 2017, 25. [Google Scholar] [CrossRef]
- C., W. C., W. Mathematicians Prove 2D Version of Quantum Gravity Really Works. Quanta Magazine, 2021. [Google Scholar]
- E., B.; M., E. E., B.; M., E. Integration with Filters. https://arxiv.org/pdf/2004.09103.pdf.
- Shackel, N. Bertrand’s Paradox and the Principle of Indifference. Philosophy of Science 2007, 74, 150–175. [Google Scholar] [CrossRef]
- Drory, A. Failure and Uses of Jaynes’ Principle of Transformation Groups. Foundations of Physics 2015, 45, 439–460. [Google Scholar] [CrossRef]
- B., K. B., K. Visualization of Uncountable Number of Psuedo-random Points Generate on Subset of the Real Plane, 2023. https://www.wolframcloud.com/obj/4e78f594-1578-402a-a163-ebb16319ada2.
- Maimon O., R. L. 2 ed.; Springer New York: New York [America];, 2010; pp. 851–852. [CrossRef]
| 1 | Meaningful Average—The measure inside the average is canonical when the measure is derived from the radon-nikodym derivative of the uniform probability measure [2, p. 32-37] |
| 2 | Wood wrote on Feynman Path Integrals: “No known mathematical procedure can meaningfully average 1 an infinite number of objects covering an infinite expanse of space in general." |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
