Preprint
Article

This version is not peer-reviewed.

Towards a Measure Theory of Semantic Information

Submitted:

21 March 2025

Posted:

24 March 2025

You are already at the latest version

Abstract
The classic account of the quantification of semantic information is that of Bar Hillel and Carnap. However, their approach assigns the maximum informativeness to a contradiction. This situation is what Floridi refers to as the Bar Hillel-Carnap paradox, and he developed a novel theory designed to remove this paradox. Unfortunately his approach does not succeed in that aim. In this paper I critique Floridi’s theory of strongly semantic information on its own terms and show where it succeeds and fails. I then present a novel approach based on methods from communication theory and quantum probability that meets all the requirements stipulated by Floridi and removes the paradox.
Keywords: 
;  ;  ;  

1. Introduction

There is a growing interest in Information and its philosophy, particularly with the current wave of agitation (positive and negative) surrounding the successes of AI across a range of fields. Associated with this is some interest in measuring information, and informativeness, the most famous of which is Bar-Hillel and Carnap’s quantification of the semantic content of a message.1 More recently Luciano Floridi identified in their work a problem which he referred to as the Bar-Hillel-Carnap paradox (BCP). He went on to propose a solution to this that would, he hoped, give a robust foundation to the quantification of semantic information, which he referred to as the “Theory of Strongly Semantic Information” (TSSI).[1] There have been a number of engagements with, and critiques of, TSSI.2 However each of these has assumed that Floridi’s construction and analysis is correct on its own terms. The criticism mainly focussed on what followed therefrom. In this paper I shall review TSSI as it is presented by Floridi. We shall find that there are a number of issues that reduce its usefulness with regard to the task for which it was designed. Nonetheless, it does provide a step in the right direction, and does suggest a way forward. Having identified the problems with TSSI on its own terms, in the rest of the paper I propose an improved version of TSSI that overcomes the identified deficiencies and provides a more general means to measure the informativeness of a message, which I shall refer to as the Measure Theory of Semantic Information, (MTSI). In particular it meets all of Floridi’s criteria for the quantification of information and removes the BCP.

2. Information and the “Bar-Hillel Carnap” Paradox

Floridi, following Dretske, [6] adheres to the Veridicality Thesis (VT). That is, in order to count as information a message must be true ([7] chap. 4). The idea of “false information” is in the same boat as the term “false friend”: a useful and evocative description, but with the clear meaning of not being a friend at all [6]. This view is not universally accepted, but for the purposes of this paper it will, without prejudice, be taken as read that VT holds, so that we can focus on Floridi’s arguments for the quantitation of information that he develops.
Floridi borrows from situation logic the term infon (symbolised as σ ) [8,9] “to refer to discrete items of factual semantic information qualifiable in principle as true or false, irrespective of their semiotic code and physical implementation”3. Using this, the General Definition of Information is given as a quad-partite definition:
G D I σ (an infon) is an instance of semantic information iff:
  • GDI.1:
    σ consists of ndata (d) for n 1
    GDI.2:
    the data are well-formed ( w f d )
    GDI.3:
    the wfd are meaningful ( m w f d = δ )
    GDI.4:
    the δ are truthful.
This definition makes use of a prior term data, or datum which is ultimately a lack of uniformity[11, p 23]. Which leads to a description of information as “a distinction that makes a difference.”4 The definition of a datum is then:[11, p 23]
Dd:
datum = d e f x being distinct from y, where x and y are two uninterpreted variables and relation of `being distinct’ as well as the domain, are left open to further interpretation.
Floridi uses these to distinguish between the Bar Hillel and Carnap approach, that he calls the “Theory of Weakly Semantic Information” (TWSI), in which truth merely supervenes on information, and gives rise to a paradox, and the “Theory of Strongly Semantic Information” (TSSI), his own formulation which he believes overcomes that paradox. The former is defined by GDI.1 – GDI.3, whereas the latter uses all four in the definition of information.
Floridi provides an analysis of where TWSI leads,[7] in particular as described by Bar-Hillel and Carnap [13] as a grounding for his TSSI. The context for Floridi’s approach is as a response to the definition of semantic information put forward by Bar-Hill and Carnap in the 1960s. They identified the amount of information in, or the semantic content (CONT) of, a proposition as being inversely related to the prior probability of the proposition (referred to as the Inverse Relationship Principle, IRP).
C O N T ( p ) = 1 P ( p )
Here P ( p ) is the prior probability of p and C O N T could be, for example a set of possible worlds, or a set of propositions, and is a measure of the likelihood of p not being the case. One upshot of this is that while the semantic content of a tautology is zero, as one would expect, C O N T ( ) , is 1. That is, a contradiction is maximally informative. This is rather surprising. As Bar-Hillel and Carnap put it:
“It might perhaps, at first, seem strange that a self-contradictory sentence, hence one which no ideal receiver would accept, is regarded as carrying with it the most inclusive information. It should, however, be emphasized that semantic information is here not meant as implying truth. A false sentence which happens to say much is thereby highly informative in our sense. Whether the information it carries is true or false, scientifically valuable or not, and so forth, does not concern us. A self-contradictory sentence asserts too much; it is too informative to be true.” [13, p 229].
That is, Bar-Hillel and Carnap reject the VT and see truth as supervening on information. It is this feature that gives rise to what Floridi calls the Bar-Hillel Carnap paradox (BCP), wherein the impossible situation of a contradiction is the most conveys that maximum information.

3. The Theory of Strongly Semantic Information

In order to understand the criticism that will follow in the next section we need to present an outline of the substance of Floridi’s TSSI. Floridi proposes that the best place to start is with “an analysis of the quantity of semantic information in σ including a reference to its alethic value. This is TSSI” ([7] p 117). To this end he identifies three desiderata for a theory of semantic information; it should: ([7] p 117)
D.1:
avoid any counterintuitive inequality comparable to BCP;
D.2:
treat the alethic value of σ not as a supervenient but as a necessary feature of semantic information, relevant to the quantitative analysis;
D.3:
extend a quantitative analysis to the whole family of information-related concepts: semantic information vacuity and inaccuracy, informativeness, misinformation (what is ordinarily called `false information’), disinformation.
These three desiderata are fundamental. The first two are unproblematic; the third is unproblematic in the abstract, but the way it is unpacked may give rise to some issues (see Section 4). It this unpacking that Floridi proceeds to develop. Throughout the discussion Floridi makes use of an example universe, E, consisting of all possible worlds, or states, arising from the conjunction of a set of basic infons.5
As well as the statement of BCP as an inverse relation:
C O N T ( σ ) = 1 P ( σ )
based on what an infon excludes, he also highlights the relation between the informativeness, ι , of an infon and its semantic content, stated as:
ι ( σ ) = C O N T ( σ )
The particular quantitative measure that Floridi chooses for TSSI is not probability but the discrepancy, ϑ , from the actual state of affairs.6
As a step towards formulating a relation to enable the calculation of informativeness for an infon, Floridi identified five criteria that he considered any method of quantifiction must meet (which he labelled M1-M5). These criteria are uncontroversial given the aims of TSSI, but, as we shall see in the next section Floridi’s chosen means of implementation does not actually meet all these criteria.
The criteria are as follows: the true state will have zero discrepancy [M1]; both a tautology [M2] and a contradiction [M3] will have the maximum discrepancy; contingently true [M4] and contingently false [M5] infons will have discrepancies that lie strictly in the range between zero and the maximum discrepancy 7

3.1. Degrees of Inaccuracy

The model universe is maximal for states8 (i.e., no further conjunctive propositions may be added without introducing a contradiction) and so it must contain the true state. Since the infons are conjunctions, any infon, σ , other than that representing the actual state of affairs will be false. However, different infons will contain greater or fewer components (conjuncts) that are false, and this gives rise the idea of degree of falsity or inaccuracy. It is straightforward then to create a metric of the distance from the true state of affairs as the ratio of the number of false components to the length of the infon (that is, the total number of conjuncts in the infon). More formally:
ϑ ( σ ) = e ( σ ) / l ( σ )
Here l is the length of the infon (which in the case of Floridii’s example is 6), e is the number of erroneous conjuncts in the infon, and ϑ is the distance of the infon from the true state of affairs (the negative sign refers to the fact that it deals with degrees of error); and spans the range 0 (matching the actual situation) to -1 (the infon contains no truth).

3.2. Degrees of Vacuity

The other aspect of distance from the actual state of affairs arises when the infon is true but more abstract than the actual (true) state. The most extreme example of this is a tautology, which includes the actual situation, but is uninformative due to its being true in all circumstances. This has a distance of +1. In order to fill in the gap between these two extremes Floridi introduced the semi-dual.9 By this means a set of classes are identified which contain all the infons with the same number of disjunctions. Each member of the class will have the same number of ways of being true given that some components of the infon are false. So in this case the distance is the ratio of the number of ways of being true, n, to the size of the universe10 s l (where l is the length of the infon and s is the number of values the infon can take: 2 in this case). More formally:
ϑ ( σ ) = n / s l

3.3. Degrees of Informativeness

Now that a putatively suitable distance metric has been identified for these two situations Floridi is able to use it to provide a means of quantifying the degree of informativeness, ι , of an infon. He proposes that the distance be viewed as spanning the range [-1, +1], with the actual state of affairs at the origin, the LHS being the degree of inaccuracy (hence the negative value) and the RHS the degree of vacuity.
The fact that, as Floridi sees it, the BCP arises, at least in part, from the inverse relation between information content and probability in that formulation suggests that a relation that identifies the actual state of affairs as having maximum informativeness and the two extremæ (tautology and contradiction), zero informativeness will serve better as a metric.
He proposes a quadratic relation as being the simplest relation that meets a number of criteria that he considers mandatory for a function to quantify the degree of informativeness. Most of these are unobjectionable, but a couple appear more problematic and less than optimal. (We will describe thee and the issues surrounding them in Section 4.1 below). The precise formulation used by Floridi is:
ι ( σ ) = 1 ϑ 2 ( σ )
see Figure 1.
The motivation, in part at least, for this quadratic is the claim that it meets a number of other conditions that he believes follow directly from M1- M5. These conditions are: that when the discrepancy ϑ , is zero the informativeness of the infon, ι , is 1 [E1]; that the integral of ι . must be bounded in the interval between minimum and, maximum values (i.e., it must be a proper integral) [E2]; when the discrepancy is at a maximum, the informativeness must be zero [E3]; when the discrepancy is in the range between minimum and maximum, then the informativeness lies in the range 0 1 [E4]; a small variation in discrepancy yields a substantial change in informativeness; [E5]; and finally [E6]: the marginal information function (which is the derivation of the equation for informativeness) is a linear function. This last is meant to approximate the ideal that “all atomic messages ought to be assigned the same potential degree of informativeness”. It can be agreed that these are, for the most part, suitable criteria and conditions for any function that aims to quantify informativeness to possess. Unfortunately, as we shall see in the next section TSSI fails to meet several of them. Most importantly it does not yield an informativeness of zero for a contradiction, and therefore fails in the central aim of removing the BCP.

4. Some Problems with TSSI

It is clear that Floridi has taken a step forward in helping to understand semantic information and the various desiderata associated therewith. However, there are a few problems in the current formulation that limit its applicability.

4.1. The Status of Contingency and Contradiction

A major reason, if not the reason, for the development of TSSI was to provide a solution to the BCP. That being the case, having a clear and coherent means of calculating the informativeness of a contradiction is crucial, and it is this that the relation in Equation 3 seeks to provide.
Floridi [7] gave formal statement of each of the criteria; I state the problematic ones here:
( M . 3 ) ¬ w σ f ( σ ) = 1
That is, “if (it is estimated that) σ is false and conforms to no possible situation, then σ is a contradiction and it is assigned the maximum degree of negative discrepancy.” ([7] p 120) This is as it should be and so the calculation of inaccuracy should naturally give this answer for a contradiction, but, as we shall see, that is not what happens.
The next criterion is stated as:
( M . 4 ) w ϑ σ ( 0 > f ( σ ) > 1 )
meaning “if (it is estimated that) σ is contingently false, then it is assigned a degree of discrepancy with a value less than 0 but greater than –1 (degrees of semantic inaccuracy).” ([7] p 120)
Unfortunately for any state that is a conjunction of atomic infons, the distance metric given in equation 1, is in violation of both these criteria. For example if the worlds/ situations are conjunctions of natomic infons giving a universe, E, of 2 n possible worlds/situations.11 The distance from the actual world, w, is calculated from equation 1.
In the case where all the atomic infons in the state are false, the degree of inaccuracy (i.e., the distance from w), ϑ is n n , which is equal to -1. Since this is a contingent falsehood it is in violation of (M.4).
On the other hand, if we apply Equation 1 to a contradiction then we will find that the inaccuracy is not equal to -1. This is because any contradiction must contain at least one conjunct that is true, and so its degree of inaccuracy can never be -1 (in violation of (M.3))! In fact for a contradictory infon containing n conjuncts the number of true conjuncts will lie in the range from 1 to n 1 , and so ϑ will lie in the range 1 n to n 1 n .12
In the light of both these issues we will need to find a more appropriate means to represent the inaccuracy and vacuity of an infon. I address this in Section 5.

4.2. The Information Space

Another issue with this representation arises if we wish to use the informativeness relation to represent the relationship between different information statements, or shifts between infons (e.g., by means of abstraction). As an illustration consider the situation where the infon (length six as in the original example) contains one false atomic statement. In that case the informativeness would be -0.972 (from equation 3). If we now abstract the infon, by means of a single disjunction, it becomes true with an informativeness of +0.953. This is discontinuous in the most basic sense: the single operation makes the informativeness jump from negative to positive without passing through zero (and to jump between different metrics). And this jumping back and forth would continue as more falsehoods and disjunctions are gradually introduced into the infon. This suggests that it is worth exploring to see if there is a representation and relationship that better captures what Floridi is aiming for here.
Another problem is that it is not clear what purpose [E5] serves. It is not unique to a quadratic (it can be equally well met by a circle), and it arguably conflicts with [E6]. Therefore it seems best to drop it as a criterion, and I shall ignore it for the rest of this paper.
There have been a number of critiques and criticisms of Floridi’s approach [2,3,5,14,15]. Some have made interesting points, but often do not hit the mark with respect to the space. Several of these have been dealt with by Sequoia-Grayson [16].

4.3. Some Conclusions Regarding TSSI

In this section we have highlighted a number of issues, and criticisms of Floridi’s approach, in particular, it does not achieve the main aim of resolving the BCP. At this point it is worthwhile to stop and reflect on the positive contributions and shortcomings of TSSI. There can be summarised as follows:
On the positive side:
  • Floridi identifies the weakness of TWSI, and the undesirable nature of the relation that makes a contradiction the most informative of messages.
  • He also provides a set of desiderata that provide, in a very clear way, the ideals for a theory of semantic information.
  • The quantitative relation that he suggest is a step forward in that it points towards a means of calculating the informativeness of an item without having to assign non-zero informativeness to a contradiction (while not quite achieving that goal).
However there are a number of issues which highlight that, despite these contributions, TSSI is not fit for purpose as it stands
  • The RHS and LHS of the space do not have the same metric, which is not desirable for a single straight line axis.
  • A contingent message which has only false atoms will have an informativeness of zero, in violation of M3.
  • A contradiction, because by definition it will have at least one true atom must have a non-zero informativeness. This is the main problem because it violates both M3, and the main motivation for the work, and therefore does not fully solve the problem of BCP.
Given the fundamental nature of these issues there is scope to explore another representation for a space in order to comply with the positive contribution while addressing the problems.

5. A More General Representation of Semantic Information

In this section I shall propose a different representation for the Information Space (IS) that will provide the means to meet the desiderata identified by Floridi and, most importantly, remove the BCP. This can be done utilising methods that have been used very successfully in other domains. In the Mathematical Theory of Communication, and related areas (such as Control Theory and Signal Theory) the Unit Circle (UC) is often used as a convenient means of representing signals and analysing stability. This is not surprising since in these domains sinusoids and trigonometric relations are commonplace, and they are defined in relation to a circle. In addition, in developing Quantum Probability (QP), based on Hilbert spaces, von Neumann utilised similar methods.[17] This being the case I argue that, given the problems identified in the previous section, the UC is worth exploring in this context as a better means of representing the relations of quantification of information.
It is a commonplace of such domains that incompatible variables or ranges be represented by orthogonal axes, whether it be Real and Imaginary numbers in communication theory or `up’ and `down’ in quantum theory. Two key features of the UC that are relevant here are the ease with which one can calculate quantitative relations by means of the cosine rule, and related to that, the fact that pure True infons and pure False infons are orthogonal to each other (and hence with no mapping since the cosine is zero). As such, True and False can be used as the axes of the space with respect to which the deviations are measured. In this case the true state-of-affairs is again the origin. Messages which contain both true and false atoms will be distributed in the space between the axes.
The first things that need to be done then are to change the space from that shown in Figure 1 (2-D) to that shown in Figure 2 (3-D, with the third dimension, representing ι ( σ ) , coming out of the page). Then we need to decide on what the best, and most coherent, means to calculate the distance from the true state-of-affairs is. In deciding what scales to utilise in measuring the distance from the true state, we need to consider two things: the distance along each axis, and the position in the plane. Before we settle this we should look at what messages can be created. This is best done by means of an example. Floridi used a universe consisting of three atoms and two predicates. which gave a space of 64 states (or worlds). However, as D’Alfonso[3] has pointed out, the actual number of possible messages is 2 64 ( 10 19 ) for this universe. We wish our space to represent all these messages and hence will need a simpler example to demonstrate things. In fact the only manageable possibility for these purposes is a single predicate with two atoms. This yields a universe of two atomic infons, four worlds (or states), and sixteen messages.13 Table 1 shows the complete set of messages for the two atomic infons x and y.14 From Table 1 we see that the number of worlds in which any message may be true (or false) ranges between 0 and 4. In particular note that the tautology is true in 4 worlds (and the contradiction is false in 4 worlds).
Again we wish to quantify the deviation ϑ of each infon from the actual state. Let us assume that the actual state of affairs is, for example, x y . That state is true in one world, so in order to calculate the deviation we first identify the difference in the number of worlds between the state of interest and the true state, call that k, (e.g., the difference between x y and T is 3 for this example). The deviation from the true state (for the horizontal axis) is then the ratio of k to the total number of worlds, call that m. This is expressed formally as:
ϑ T ( σ ) = k m
This is similar to the metric used by Floridi for the RHS of his space. We now require two variables to represent the deviation from the true state of affairs; call these ϑ T ( σ ) and ϑ F ( σ ) for the True and False axes respectiely. We keep ι ( σ ) as the informativeness metric (which is the third dimension in the space, and would be coming out of the page).
We have already seen that on the LHS he used the ratio of the number of false atoms to the length of the state as a metric. That the metrics were different was, as mentioned previously, odd for a straight line scale, but for the current representation it makes sense for those messages that may be true but cannot lie on the horizontal axis because these contain a mixture of true and false atoms. In this case, the number of false atoms can be used as a metric for how far round towards the vertical (false) axis they lie. For the current example, there can be zero, one, or two false atoms, which suggest a suitable set of steps for the angle of each message relative to the horizontal axis. This is depicted in Figure 2. In the general case, for an infon with n atoms, the angle between each ray would be π / ( 4 n 2 ) rads.
So far we have only dealt with messages that are true, and we now need to look at how the false infons should be placed in the space. As it happens this is straightforward and follows directly from the relation between the two axes. Being orthogonal they represent contradictory infons: where all the infon lying on the horizontal axis are true and contain only true atoms, those lying on the vertical axis are false and contain only false atoms. In fact they form a “mirror image” reflected through a 45 line in the space (the dotted line in Figure 2). In this case the distance from the origin is given by the number of worlds in which the infon is false (relative to the true state-of-affairs). Hence the same formula can be used to calculate the distance from the origin. Now we have that the two halves of the space are symmetric around the 45 line with the contradictory infons being located at the same relative positions, and therefore having the same distance from the origin (true state).
We are now in a position to put forward a suitable metric to replace that proposed by Floridi. The equation of a UC is:
x 2 + y 2 = r 2 = 1
which for the information metric space becomes:
ϑ T 2 ( σ ) + ϑ T 2 ( σ ) + ι 2 ( σ ) = 1
from which we get the formula for the informativeness metric:
ι ( σ ) = 1 ( ϑ T 2 ( σ ) + ϑ F 2 ( σ ) )
At this point we have a clear means whereby we can calculate the deviation of an infon from the true state affairs, and a metric to identify the informativeness of any infon. However, things do not end there. Floridi utilised his metric to quantify the degree of inaccuracy and degree of vacuity of infons. His approach is very straightforward for inaccuracy, but it is quite complex for the amount of vacuity (requiring the integral of the distance). It is not difficult to see from Equation 7 that it would be even more so for our space, requiring the double integration of what could be a complex shape.15 Fortunately we can simplify things significantly by following the approach taken by von Neumann in the development of QP.
In QP the basic metric is a wave amplitude, which is a periodic (sinusoidal) displacement: the maximum absolute amplitude is 1 and the minimum is 0, ranging around a UC. That is if one makes the square of the amplitude as the probability of an electron being within that amplitude then one has
P ( x ) + P ( y ) = 1
which is a measure rather than a metric.
By analogy, if we square the metric values ϑ T , ϑ F and ι we will (as with von Neumann) obtain measure values that generate a new space. The axis obtained from ϑ T gives a measure of vacuity or the degree to which one is uninformed (or ignorant).16 Let us label that Φ u . The other axis then is a measure of misinformation, Φ m say.17 And finally, the third axis gives a measure of the informativeness of the infon (and would come out of the page in Figure 3), call that axis Φ i .
Now we have:
Φ i + Φ m + Φ u = 1
Or
Φ i = 1 ( Φ m + Φ u ) = 1 Φ R
this is MTSI.
As things are set up in the IS for MTSI a contradiction has informativeness of zero (as does a tautology) as required to avoid BCP. However, for reasons presented previously, neither a contradiction nor a tautology lies on the axes of the space since they both contain both true and false atomic infons. Strictly speaking there is no infon on either axis that has a value of zero. On the other hand because (as mentioned previously) a contradiction (or tautology) can contain as many true or false atoms as there are atoms in the messages, each ray not on the axes will contain a version of a contradiction (or tautology) at its zero point, as shown by the M 0 and M 15 labels in Figure 3. As the number of atomic infons increases the points representing a contradiction or tautology will asymptotically approach their respective axes.
The final issue is that according to our measure space all infons and their contradictories have the same value for informativeness. This starts, as we have just seen, with the contradiction and tautology: they are contradictories and both have value zero. As another example consider the atomic infon x; it is on the Φ u axis and will be true in 50% of cases. Its contradictory, x , is located on the Φ m axis and is false in 50% of cases. They are images of one another through the 45 ( π / 4 ) line and have the same value of informativeness. In one sense it is not surprising that true and false infons have the same value, after all that was also the case for TSSI, but it did not capture this relation for contradictories. But what is surprising is that this also means that for the true state of affairs, its contradictory will also have a value of 1. That is, for the example used, both x y and ( x + y ) yield Φ i = 1 .
The simplest way to show why this is the case is again by means of an example. Consider a familiar problem scenario: You are confronted by two doors. Behind one is a million dollars, behind the other is instant death. The doors are protected by two guards, one of whom only tells lies and the other of whom only tell the truth. You have to choose one door, and, obviously, you want to pick the money not the demise.
There are two situations we can explore to make things clear: one in which you know which.guard is which, and one in which you don’t. To help you make your decision you are allowed to ask one guard one question. In the first case you can simply ask either guard directly which door you should choose. If you ask Guard 1, then you choose the other door from the one they tell you (because they will misinform you), and if you ask Guard 2, then you choose the door they tell you to take. Simple. Here we see that regardless of which guard you ask, you become equally well informed as to what action to take, because there is a reliable process (logic) that leads you to the correct answer.
In the second situation, which is the one usually used in the puzzle, you can still ask a single question and regardless of which guard you ask, become equally well informed as to which door to choose. In this case you ask either guard what answer the other guard would give. If you ask Guard 1 they will falsely tell you (i.e., misinform you) that Guard 2 would say Door 2. and if you ask Guard 2 they will truthfully tell you that Guard I would say Door 2. So you are now informed that Door 2 is the door you don’t want and you can safely choose Door 1. Again, this is because propositional logic provides a robust and reliable means of leading you to the desired result.
We are now in a position to consider how this new approach to semantic information matches up to the desiderata, the BCP and the specified conditions put forward by Floridi (which I believe are correct).18 In Section 4 we saw that TSSI did not meet all the stated conditions for various reasons. Lets look at each one in turn and see how MTSI compares.
  • Clearly M1- M3 are satisfied in a straightforward manner by the new version of the distance metric, and overcome the problem that M3 poses for TSSI.
  • The spirit of M4 and M5 is also met straightforwardly, but is adapted to conform to the 2D situation rather than the single dimension of TSSI. For M4 and M5, again MTSI achieves what TSSI aimed for but did not provide (i.e., 0 < Φ R < 1 ). So MTSI gives a robust measure of inaccuracy without the problems highlighted for TSSI.
  • Then also E1, E3 and E4 have their direct counterpart in MTSI. The fact that MTSI is measure based means that the calculation of quantity of information is simplified with respect to the integral needed for E2.
  • Floridi states that “all atomic messages ought to be assigned the same potential degree of informativeness and therefore, although [E6] indicates that the graph of the model has variable gradient, the rate at which ι ( σ ) changes with respect to change in ϑ ( σ ) should be assumed to be uniform, continuous and linear.” Here again, since the MTSI measure is affine, the aims of E6 are met exactly, whereas they are only approximated with the quadratic.
  • The information space of MTSI includes all possible messages explicitly, and not just the conjunctive states/ worlds as in TSSI.
  • Floridi points out that the space he proposed would allow for continuous model to be analysed as well. The same applies here. For the boolean case, as one increases the length of the infon, the space becomes denser but the relative positions of the messages in the space remains the same (e.g., atomic infons will still be true in half the cases etc).
There have been a number of challenges to VT and Floridi’s TSSI recently. It is beyond the scope of this paper to address these in detail, but a couple of comments are in order. The situation is not helped by the fact that the criticisms deal with different versions, or statements, of VT which makes it harder to address the criticisms. Nonetheless, the paper by Fresco et al[19] seems to be based on having a reliable means of shifts from a false to true status. It should be obvious from the foregoing description, and particularly the example of the two doors, that this is not a problem for MTSI. On the other hand, Ferguson[4] attacks VT by means of carefully constructed paradoxes whereby VT gives rise to statements that both “convey information” and do not “convey information.”’19 However one does not need to go to these lengths because Floridi has already provided examples of statements that are information and not information, depending on the accessibility of the relevant propositions.[11, p 50] The example Floridi uses is of a man phoning a garage to get a repair done, stating: “My wife left the car lights on overnight, and now the battery is flat.” This statement is false (and hence not information) since it was the man who left the light on. But the relevant part for the mechanic is only that the battery is flat, which is true. One might also suggest in passing that if one were to follow Ferguson here, and reject VT, then unless the term “misinformation” is evacuated of all meaning, any statement would both “convey information” and “convey-misinformation”. which does not seem to be an improvement. No doubt the debate on VT will continue, but there is nothing in the current criticism that demands jettisoning it quite yet, and a lot to suggest its continued value.
Finally, and as the observant reader will note, MTSI has the form of an inverse relation. This indicates that it is not the IRP in and of itself that leads to, or entails, the BCP, but the fact that the original version only utilised a single variable. That is, it was at the wrong Level of Abstraction.[20] TWSI is the same formula as Popper suggested for “verisimilitude” ([21] p 233f), and Floridi proposed that “likeness” should be used to describe the proximity of messages to the state of interest. This is precisely what MTSI, at least potentially, provides, and complements Cevolani’s contention that “truthlikeness” rather than “informativeness” should be the focus of attention.[10] Hence the use of MTSI for verisimilitude is one possible direction for future work. Another question that arises from this analysis is: could the approach taken here be used to explore the distribution of messages in the information space when the datum of interest is more abstract, or representative of a different Level of Abstraction? ([7] ch. 3 )

Acknowledgments

With thanks to Patrick Allo, David Lusseau, and Ulrich Stegmann who gave helpful comments on an earlier version of this paper.

References

  1. Floridi, L. Outline of a Theory of Strongly Semantic Information. Minds and Machines 2004, 14, 197–222. [Google Scholar]
  2. Cevolani, G. Strongly semantic information and information about the truth. Trends in Logic 2014, 41, 59–74. [Google Scholar]
  3. D’Alfonso, S. On quantifying semantic information. Information 2011, 2, 61–101. [Google Scholar] [CrossRef]
  4. Ferguson, T.M. Two paradoxes of semantic information. Synthese 2015, 192, 3719–3730. [Google Scholar] [CrossRef]
  5. Wolf, M.J. Analysis, clarification and extension of the Theory of Strongly Semantic Information. Ethics & Politics 2011, 13, 246–254. [Google Scholar]
  6. Dretske, F.I. Information and the Flow of Knowledge; CSLI, 1999.
  7. Floridi, L. The Philosophy of Information; OUP, 2011.
  8. Devlin, K.J. Logic and Information; CUP, 1991.
  9. Israel, D.; Perry, J. What is Information? In Information, Language and Cognition; Hanson, P.P., Ed.; University of British Columbia Press, Vancouver, 1990; pp. 1–28.
  10. Cevolani, G. Verisimilitude and strongly semantic information. Ethics & Politics 2011, 13, 159–179. [Google Scholar]
  11. Floridi, L. Information: A Very Short Introduction; Oxford: OUP, 2010. [Google Scholar]
  12. MacKay, D.M. Information, Mechanism and Meaning; MIT Press, Cambridge, MA, 1969.
  13. Bar-Hillel, Y.; Carnap, R. An outline of a theory of semantic information. In Language and Information; Bar-Hillel, Y., Ed.; Addison-Wesley, 1964; pp. 221–274.
  14. D’Agostino, M. Semantic Information and the trivialization of logic: Floridi on the Scandal of Deduction. Information 2013, 4, 33–59. [Google Scholar] [CrossRef]
  15. Dodig-Crukovic, G. Empirical modelling and information semantics. Mind and Society 2008, 7, 1157–166. [Google Scholar]
  16. Sequoiah-Grayson, S. The Metaphilosophy of Information. Minds and Machines 2007, 17, 331–344. [Google Scholar]
  17. von Neumann, J. Mathematical Foundations of Quantum Mechanics; Princeton: Princeton University Press, 1955 (1932).
  18. Coghill, G.M. Notes on the Floridian Theory of Strongly Semantic Information. Technical report, University of Aberdeen, 2016.
  19. Fresco, N.; McGivern, P.; Ghose, A. Information, veridicality and inferential knowledge. American Philosophical Quarterly 2017, 54, 61–75. [Google Scholar]
  20. Floridi, L. The method of levels of abstraction. Minds and Machines 2008, 18, 303–329. [Google Scholar]
  21. Popper, K. Conjectures and Refutations; Routledge, 1963.
1
One could include Shannon Information, or the work by the British scientist Donald MacKay, but they were not focussed on semantics. In fact Shannon explicitly stated that his was a syntactic metric. Therefore I shall follow Floridi and refer to that as the “Mathematical Theory of Communication” (MTC).
2
For example: [2,3,4,5]
3
For the purposes of this paper these may be considered to be propositions as has been done by [10] and [3]. However, while keeping that in mind I shall continue to use the term “infon” throughout this paper for continuity with Floridi.
4
This description is attributed, by Floridi and others, to [12]. However neither I nor anyone else, including MacKay’s widow has been able to confirm that he is the originator of the description, nor where it might be found. However, insofar as it serves as a the contradictory to the common description of being uninformative as “a distinction without a difference” it remains an accurate description and should continue to be used.
5
Floridi presents it as containing all possible messages, but as D’Alfonso [3] has pointed out, this is not the case.
6
Floridi gives a fuller set of comparisons to justify the development of TSSI, but this summary is sufficient for the purposes of this paper.
7
For details of the formal expression of these five criteria see [7].
8
Though, as noted, not for messages.
9
A semi-dual is an infon in which the operators are changed from conjunction to disjunction, but the components are not negated. For example a contradiction is the semi-dual of a tautology.
10
Here s represents the number of values an infon can take (in this case two: {T, F}), and l is the length of the infon
11
Floridi uses states consisting of six atomic infons, giving a universe of 64 worlds/situations.
12
This also gives rise to an issue highlighted by [3] relating to inconsistency. He identifies that inconsistent infons that have greater or fewer true constituents would have higher or lower informativeness. Unlike Floridi, and the position proposed in this paper, he thinks that is intuitively as it should be.
13
Even three atomic infons would give eight worlds and 256 messages; which is too unwieldy for illustrative purposes. To put this in perspective, consider a high quality print of this page with, say, 600 dpi. That would give around 28 million dots on the page: several orders of magnitude fewer than would be required to represent all the possible messages in Floridi’s original example. In fact, about a decade ago it was estimated that around 10 19 pieces of information had been generated since the invention of writing.
14
A note about nomenclature. In this paper, I am using the symbolic representation from digital systems and communication for ease of presentation That is, conjunction is represented by a dot (though it is usually omitted where no confusion will ensue). Disjunction is represented by a plus sign and negation by a prime stroke. This gives rise to the Sum-of-Products (SOP), equivalent to The Conjunctive Normal Form (CNF) of propositional logic.
15
See [18] for further details of this.
16
Recall that Floridi calculated vacuity by calculating the area under the curve of his relation: squaring a distance has the same `units’ as an area.
17
Or ’Disinformation’, but we may ignore that and assume no malicious intent.
18
Floridi states them as conditions for an adequate metric, but here it is more appropriate to consider them with respect to the proposed information measure.
19
It is not clear why Ferguson utilises the term “convey information” rather than simply “information”. It seems to lead to ambiguity since he says it is synonymous with “semantic information” but also refers to it as occurring “to some degree”.
Figure 1. The degree of informativeness (after Floridi [7]). Here, ι ( σ ) = 1 ϑ 2 ( σ ) .
Figure 1. The degree of informativeness (after Floridi [7]). Here, ι ( σ ) = 1 ϑ 2 ( σ ) .
Preprints 153109 g001
Figure 2. The structure of the information metric space.
Figure 2. The structure of the information metric space.
Preprints 153109 g002
Figure 3. The distribution of messages in the information measure space.
Figure 3. The distribution of messages in the information measure space.
Preprints 153109 g003
Table 1. The complete set of messages.
Table 1. The complete set of messages.
Message Message No. of worlds in No. of true atoms
Number which it is True
M0 x x 0 1
M1 x y 1 2
M2 x y 1 1
M3 x y 1 1
M4 x y 1 0
M5 x 2 1
M6 y 2 1
M7 x y + x y 2 2
M8 x y + x y 2 2
M9 y 2 0
M10 x 2 0
M11 x + y 3 2
M12 x + y 3 1
M13 x + y 3 1
M14 x + y 3 0
M15 x + x 4 1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated