Preprint
Article

This version is not peer-reviewed.

No Small Observer Can Verify a Black Hole Firewall

Submitted:

03 March 2026

Posted:

04 March 2026

You are already at the latest version

Abstract
Quantum gravity is often probed by thought experiments that assume unlimited classical control over quantum systems. The firewall paradox is one such thought experiment. It asks if one observer can decode early Hawking radiation, and then later compare with degrees of freedom behind the horizon. Here I show that this protocol is constrained by an instruction budget. If the correct decoding depends on the black hole microstate, a hidden internal configuration, then the observer must specify which decoder to use. The number of decoders needed to cover all microstates grows exponentially with the black hole entropy, so the decoder specification requires a number of bits that scales with that entropy. A covariant entropy bound limits how much information can be stored within a bounded causal patch. This enforces a horizon-scale lower bound on the region required for single-observer verification, even when computational complexity is ignored. More broadly, the result highlights a principle for quantum gravity experiments. Control information is physical, and bounds on storage can decide what is testable.
Keywords: 
;  ;  

1. Introduction

Black hole complementarity is the idea that two observers can each use a consistent description, even if those descriptions cannot be combined into one single measurement record, because no single observer can access both sets of data [1,2]. One description has a smooth horizon for infallers. The other treats evaporation as unitary, meaning information is not destroyed and in principle can be reconstructed from the outgoing Hawking radiation.
I’ll begin with a little background on these terms. A quantum state is a rule for predicting measurement outcomes. A pure state is the special case where that rule comes from one wavefunction1. A mixed state is what you get when you have ordinary uncertainty about the preparation, or when you look at only part of an entangled system. A purifier is an extra system on top of all this. A purifier together with the mixed system forms a pure joint state.
Meanwhile a black hole horizon is not a surface, but a point of no return where the gravity is too strong for anything to get back out2. If you fall freely in a gravitational field, you feel like you are floating in deep space, so it is believed that the horizon is “smooth” in the sense that a person passing through it will not detect any difference. Black hole evaporation is widely believed to be unitary. Unitary means information is preserved (transformed but never destroyed), at least in principle. Evaporation refers to Hawking radiation, the process by which a black hole radiates away its mass and energy. Unitarity then says the information about what fell in is encoded in correlations of the outgoing radiation rather than being destroyed.
So now you have some background, assume two people named Alice and Bob. Alice stays outside the black hole, and Bob falls in. Alice collects the Hawking radiation and confirms it is unitary, but never sees the inside. Bob passes through the horizon and confirms it is smooth, but can never send a message back. Subjectively, physics holds up for both of them, but objectively their realities contradict one another. Complementarity holds that one should discard the objective view because no single measurement can access both sides [1].
The firewall paradox, often called AMPS after Almheiri, Marolf, Polchinski and Sully, attacks this by proposing a single-observer test [3]. If evaporation is unitary, then the early radiation R should contain enough information to predict a late outgoing mode B. If the horizon is smooth, then B should also be maximally entangled with an interior partner A. Maximally entangled means A and B share the strongest possible quantum correlations; each subsystem alone looks completely random, but together they behave like a perfectly correlated pair. Quantum mechanics does not allow one system to be maximally entangled with two independent systems. Breaking the entanglement between A and B “snaps” the vacuum state3, creating a firewall of high-energy particles that blasts anything crossing the horizon. But a firewall contradicts the smooth horizon. This is the monogamy problem: B cannot be entangled with bothA and R.
AMPS proposed that one observer could collect the early radiation outside, distil the information from it, then jump into the black hole and compare with the interior mode. A single subjective observer would then witness the contradiction. To prevent this, the universe would have to “burn” them with a firewall before they cross the horizon, which removes the smooth horizon. This expresses a tension between quantum theory and gravity. The structural connection to observer-dependent paradoxes has also been noted through reframings as an extended Wigner’s friend scenario [4].
However, if no single observer can perform the needed joint experiment, then the AMPS firewall paradox is not a problem. Instead of a paradox, there would be a constraint on which measurements are physically meaningful in quantum gravity.

1.1. What This Paper Asks

This paper asks whether the AMPS verification protocol can be implemented inside one bounded causal patch once I include the control information needed to choose the correct decoder. I model the decoder choice as physical control information, meaning the distinguishable control states that select which decoding procedure is applied [5,6]. Treating the covariant entropy bound as a memory ceiling limits how many distinct decoder choices any bounded patch can hold [7]. Separately, the alpha-bit analysis of Hayden and Penington bounds how large a code subspace can share one fixed, state-independent decoder at a given accuracy, even for mixtures and for states entangled with an external reference system [8]. In evaporating setups this tension appears as state dependence of interior reconstructions for mixed states [9,10].
The implication is that some firewall contradictions may be untestable by any observer who only accesses a subhorizon patch. That shifts the target for quantum gravity. It is not only about what exists, but what can be certified.

1.2. The Protocol

Hawking radiation is the outgoing quantum radiation emitted by a black hole as it evaporates. I label the early Hawking radiation as R. It is the radiation already emitted to infinity. An old black hole is one that has emitted enough Hawking radiation that R can in principle predict properties of the late radiation. I label a late outgoing mode as B. This is one of the later quanta emitted after the black hole is old. I label the would-be interior partner of B as A. If the horizon is smooth, then A and B should be strongly entangled.
A decoder is a physical procedure, a map or circuit, that acts on R and outputs a system that purifies B. Purifies means that the output and B together form a pure state, meaning the joint system can be described by one wavefunction rather than a probabilistic mixture. A causal patch is the region of spacetime whose degrees of freedom can still influence the infaller before the comparison is completed.
The AMPS-style verification task can be stated as five steps.
1.
Collect the early radiation R outside the horizon.
2.
Wait until the black hole emits a fresh outgoing mode B.
3.
Choose a decoder and apply it to R to produce a candidate purifier of B.
4.
Cross the horizon and access the interior partner mode A of the same B.
5.
Compare the candidate purifier with A and B inside one causal patch.
The first three steps happen outside. The last two steps happen inside. The word verify means a single observer can do all five steps using only the degrees of freedom that fit in their patch.

1.3. Where The Obstruction

If one fixed decoder worked uniformly for all black hole microstates, the story would be simple. You would bring that one decoder and run it. Holographic reconstruction bounds that demand uniformity on mixtures and on states entangled with an external reference system imply this uniformity fails in general. Instead, one fixed decoder can only be trusted on a limited code subspace. Different microstates require different decoders.
This paper just points out the obvious. The decoder choice itself is physical. At verification time the infaller must carry a control register K that selects which decoder is being applied. If there are many decoders, then K must have many reliably distinguishable internal states. The covariant entropy bound acts as a ceiling on how many such internal states can fit in a bounded patch.

1.4. Why Waiting Longer Doesn’t Help

It is tempting to think the protocol could be saved by taking more time. In Stack Theory, the operational notion of time at a layer is the count of changes that the layer can represent [6]. If nothing that your vocabulary can express changes, then no time passes at that layer. This means many step descriptions can be rephrased as bigger state descriptions by storing intermediate results [11].
The firewall verification task is different. It is a co-instantiation task. Co-instantiation means all required ingredients are true at the same objective instant for one observer. The observer needs one moment in which the late mode B, its interior partner A, the decoded purifier, and the decoder selector K are all physically present inside one causal patch. Seeing each ingredient at some point along a long worldline does not create one joint measurement record. Section S3 of the supplement formalises this gap between ingredient-wise occurrence and co-instantiation.
This is not the Harlow–Hayden time barrier, which is a claim about how long the distillation computation takes [12]. Even if an oracle, meaning a hypothetical device that performs computation for free, performed the computation instantly, the physical control state that selects the right decoding operation still has to fit in the patch [13,14].

1.5. The Theorem

Theorem 1 turns the above story into one inequality. It uses three numbers.
S BH is the black hole horizon entropy in bits. It is log 2 | Φ micro | where Φ micro is the set of mutually exclusive microstates compatible with the same exterior. ρ 2 is the fraction of the horizon area that bounds the infaller patch. α * is a number between 0 and 1 that controls how large a code subspace can share one fixed decoder at fixed error tolerance. It is defined precisely in Box 3.
The theorem shows that a verifier confined to a patch with boundary area fraction ρ 2 can only succeed on all microstates if
ρ 2 1 α * c ε S BH .
For large black holes the correction term is negligible, so the punchline is ρ 2 1 α * . In a representative case where one decoder works uniformly on a code subspace carrying 20 % of the black hole microstate entropy, the decoder label alone costs about 80 % of the entropy in control bits and demands access to about 80 % of the horizon area.
Figure 1 plots the large entropy form of the theorem. It tells you the minimum patch size fraction ρ min required for each reconstruction fraction α * . The curve is ρ min = 1 α * when subleading constants are neglected.
The Extended Data figures in the supplement are toy computations. They do not simulate black holes. They illustrate the two counting steps: how join capacity can exceed a patch ceiling, and how decoder coverage scales with a code subspace size bound.
If you want to reproduce these toy plots, see Section 7. The exact toy parameters are listed there and are also recorded in the CSV source data files. For a plain English description of each figure and its source data, see figures/README.md and source_data/README.md. One warning helps avoid confusion. The toy exponent parameter called α in Extended Data Figure 2 is not the same object as the theorem parameter α * . The theorem parameter α * is the geometric reconstruction fraction defined in Box 3. The toy parameter α is just a dial that sets a subset size d max = 2 α n in a finite set cover model.

1.6. Contributions

Many firewall discussions focus on computation time, asking whether the distillation circuit is too slow to run before the infaller reaches the singularity, the region deep inside where curvature becomes extreme and semiclassical descriptions break down. This paper isolates a different bottleneck: the physical memory needed to specify which distillation circuit to run when the correct circuit depends on hidden microscopic details of the black hole state. In this model, the instructions for the right reconstruction can be the scarce resource before the data.
Concretely, I make the following contributions. First, I treat the decoder label as a physical control state that must fit inside the same causal patch as the final comparison. Second, using alpha-bit reconstruction bounds, I translate state dependence into a lower bound on how many distinct decoders are needed to cover all microstates at fixed error tolerance. Third, I translate that decoder count into a lower bound on control bits, using the Stack Theory capacity definition in Box 2. Fourth, I combine that control bit requirement with the covariant entropy bound, treated as a patch memory ceiling, to obtain an explicit lower bound on the required horizon area fraction in Theorem 1. Fifth, I include a deterministic toy code script, source data, and plots that illustrate the counting steps without requiring any quantum gravity simulation.
The principle that control information is a physical resource in holographic settings is shared with Kubicki, May and Pérez-García [15]. They apply it to general bulk computation. The present paper applies this lens specifically to AMPS verification and combines it with alpha-bit coverage bounds to produce the horizon area fraction inequality in Theorem 1.

1.7. Outline

If you want the formal definitions, go to Section 2. If you want the main result, go to Theorem 1 and Figure 1. If you want to reproduce the plots, see Section 7.

2. Glossary and Notation

This paper aims to be readable by physicists, computer scientists, and information theorists. It uses a small vocabulary from quantum gravity and from quantum information. I define those terms here and then stick to them. The main result is Theorem 1. If you want to reproduce the figures, see Section 7.

2.1. Information Capacity

I use a simple capacity measure that counts how many distinct internal states a physical system can reliably represent. I write I ( v ) for the number of bits of distinction available to a vocabulary v. Box 2 gives the formal Stack Theory definitions. In plain English, if a patch can only support 2 10 reliably distinct internal states, then it has at most 10 bits of usable capacity no matter how clever the computation is.

2.2. Time as Change

Stack Theory treats time at a vocabulary as the count of changes that the vocabulary can represent [6]. Given an underlying trajectory of world states, the induced layer trajectory deletes repeated frames that look identical at that vocabulary. If the encoding stops changing, the layer clock stops too.
If nothing you can notice changes, then no time passes for you at that descriptive level. This matters because waiting longer is only helpful if it creates new distinguishable states. It cannot increase the number of distinct states that fit inside one causal patch at one moment. The firewall verification task is a one-moment comparison, so it is governed by co-instantiation capacity rather than by elapsed time.

2.3. Basic Mathematical Conventions

Preprints 201243 i001

Bits and base-two logarithms.

All logarithms are base two, so information is measured in bits. I use log 2 so that one bit corresponds to a factor of two in the number of distinguishable states. If a physical system can be in N reliably distinguishable states, then it can carry at most log 2 N bits of information.

Fundamental constants.

I write c for the speed of light, G for Newton’s gravitational constant, and for the reduced Planck constant. They are shown explicitly only to make dimensions clear. If you prefer natural units you can set c = G = = 1 everywhere without changing any dimensionless inequality in this paper.

Reliably distinguishable states.

I say two internal states are reliably distinguishable if there exists some measurement that can tell which one you have with a small error probability. This is the operational notion behind capacity bounds like the covariant entropy bound.

Operationally distinguishable states and procedures.

I say two states or procedures are operationally distinguishable if there exists some experiment that can tell them apart. For states, this means there is some measurement whose outcome statistics differ for the two states. For procedures, such as two decoder channels, this means there is some input state and some measurement on the output that produce different statistics depending on which procedure was applied. In plain English, if no possible test can notice a difference, then the two procedures are the same for all physical purposes.
In this paper I adopt the weakest possible counting convention. I only distinguish decoders to the extent that they differ on the input domain that the verifier can actually supply, at the tolerated error level. If two decoding procedures act identically on every protocol-relevant input state, then I treat them as the same decoder and they do not require two different control settings. This makes the decoder family size | D | as small as it can be, which makes the no-go theorem harder to prove.

Order one fractions

When I say an order-one fraction of the horizon area, I mean a constant fraction like 20 % or 80 % that does not shrink as the entropy grows.
Preprints 201243 i002
An environment Φ is the set of mutually exclusive ways the world could be. It can be infinite. A program p Φ is one yes or no question about the world. The answer is yes exactly on the states in p. A vocabulary v is the list of programs the observer can actually implement. In this paper it is finite (in related work this is justified using the Bekenstein bound [16]).
A statement l v is a set of programs that can all be true at once. Its truth set T ( l ) is the set of world states where every program in l answers yes. The layer L v is the list of statements that are satisfiable, meaning they have a nonempty truth set.
The encoding Enc v ( ϕ ) is the list of programs in v that answer yes on the state ϕ . Two different world states are distinguishable by v only if they induce different encodings. The realised encoding set Im ( Enc v ) is the set of distinct encodings that actually occur as ϕ ranges over Φ . The encoding capacity is I ( v ) = log 2 Im ( Enc v ) . It is the number of bits of distinction that v can represent about the world.
Even if Φ is infinite, a finite vocabulary can realise only finitely many distinct encodings. In fact Im ( Enc v ) 2 | v | , so I ( v ) | v | . So a finite list of questions carves an unbounded world into finitely many operational cases.
A bounded patch R can only hold finitely many reliably distinguishable internal states. I write I max ( R ) for the corresponding bit ceiling. If a protocol needs more distinctions than that ceiling, no amount of extra computation can make it fit in one causal patch.
A supposed paradox becomes operational only if the relevant joint vocabulary fits inside one causal patch. If it does not, the paradox is a request for an unphysical measurement.
For the main theorem, the only quantity I will use from Box 2 is the key capacity I ( v K ) . It equals log 2 of the number of reliably distinguishable control states of the decoder selector register. The full vocabulary structure is used in Supplementary Section S2 and in Extended Data Figure 1 where I discuss capacity complementarity from joins.

2.4. Toy Example

Take Φ = { a , b , c , d } . Let p 1 = { a , b } and p 2 = { a , c } . Let v = { p 1 , p 2 } . The encoding of each state is
World state ϕ Encoding Enc v ( ϕ )
a { p 1 , p 2 }
b { p 1 }
c { p 2 }
d
So the realised encoding set is Im ( Enc v ) = { { p 1 , p 2 } , { p 1 } , { p 2 } , } . It has Im ( Enc v ) = 4 elements, so I ( v ) = log 2 4 = 2 bits.
Now split v into two smaller vocabularies v 1 = { p 1 } and v 2 = { p 2 } . Each one has only two realised encodings, so I ( v 1 ) = I ( v 2 ) = 1 bit. But their join is v 1 v 2 = { p 1 , p 2 } , which is the original v, and so I ( v 1 v 2 ) = 2 bits. If a patch had a ceiling between 1 and 2 bits, then each description would fit on its own, but the joint description would not fit. This is the finite toy analogue of complementarity as a join overflow. Extended Data Figure 1 repeats this pattern at larger scale using random programs.

3. Single Patch Versus Distributed

AMPS refers to the firewall argument of Almheiri, Marolf, Polchinski and Sully [3]. It is often told using a distillation step outside the horizon and a comparison step inside. That is a distributed protocol. The firewall contradiction is instead a statement about what one observer can certify. The observer is supposed to know that B is purified by R and also know that B is entangled with an interior partner A.
I therefore distinguish two notions.
Definition 1
(Single patch verifier). A single patch verifier is a protocol whose decoding choice and verification outcome are both implemented by degrees of freedom confined to one infaller causal patch at the moment of verification. External preprocessing is allowed only insofar as its output is carried into the patch as physical state.
A protocol counts as single patch only if the choice of decoder and the final yes or no verdict are both carried by physical degrees of freedom inside the same infaller patch at the moment of comparison. If some outside agent does work and then merely tells the infaller which decoder to use, that instruction still has to be stored inside the patch as a physical control state. If it does not fit, the protocol is distributed rather than single patch.
Remark 1.
If you allow a distributed protocol, then you have already granted complementarity. You have accepted that no single patch contains the joint refinement of the exterior and interior descriptions.
The next section proves a no-go theorem for single patch verifiers in a controlled reconstruction model.

4. No-go Theorem From Control and Reconstruction Limits

Preprints 201243 i003
The assumptions are listed in Box 3. From assumption 1 (patch capacity cieling) I have a hard memory ceiling for anything that must sit inside one causal patch. From assumption 2 (the microstate coverage requirement) I have a well defined requirement for successful decoding. The verifier must work for every microstate in Φ micro , not just for a lucky subset. When I say a decoder works on a subset of microstates I mean it works uniformly on every state in the code subspace they span, including mixed and reference-entangled states. Assumption 3 is the holographic input. It says that if the reconstruction is state independent, one fixed decoder can only cover an exponentially small code subspace of microstates unless the geometric fraction parameter α * is close to one. Assumption 4 is the operational bridge. Different decoders require different physical control settings. I count decoders in the most forgiving way. Only differences that matter on protocol-relevant inputs are counted as different control settings. The theorem below is what you get when you combine these ceilings and then ask one patch to hold the decoder choice.

4.1. A Covariant Capacity Ceiling for a Subhorizon Patch

For a nonrotating, uncharged black hole of mass M, the horizon radius is the Schwarzschild radius R s : = 2 G M / c 2 and the horizon area is A 0 : = 4 π R s 2 . Let 𝓁 p 2 : = G / c 3 be the Planck area. The horizon entropy in bits is given by [20,21]
S BH : = A 0 4 𝓁 p 2 ln 2 .
The factor ln 2 converts the usual natural logarithm convention into base two so the unit is bits.
I model the infaller as confined to a bounded causal patch R . Let A ( R ) denote the maximal area of a codimension-two cross-section of the boundary of R . I adopt Assumption 1. I treat the covariant entropy bound as a patch capacity ceiling. In bits [7],
I max ( R ) A ( R ) 4 𝓁 p 2 ln 2 .
Suppose the patch boundary area satisfies A ( R ) ρ 2 A 0 for some ρ ( 0 , 1 ] . This defines ρ 2 as the accessible area fraction of the horizon.
Lemma 1
(Covariant subhorizon ceiling). For any causal patch R with A ( R ) ρ 2 A 0 ,
I max ( R ) ρ 2 S BH .
Proof. 
Combine (2) with A ( R ) ρ 2 A 0 and the definition (1).    □
The inequality above counts only the control information needed to select a decoder. It does not budget for storing the early radiation data, the distilled output, or the final comparison record. Any such storage competes for the same finite patch capacity, so including those costs would only make the no-go stronger.
Note also that Lemma 1 avoids any appeal to weak-gravity assumptions about the verifier. It is an area-based ceiling for any information that must fit inside one infaller patch.

4.2. Decoder Means Physical Key

The AMPS verifier must select a decoder that distils a purifier of a late mode B from the early radiation R. Let D be the set of decoder channels that are distinct for this verification task. Two candidate procedures count as the same decoder if they are operationally indistinguishable on every protocol-relevant input state at the tolerated error level. This convention minimises | D | and therefore makes the key lower bound as weak as it can be.
It also means the argument is not about where the decoder description sits. The key can be explicit, like a written label. It can be implicit, like the internal configuration of a programmable device after it has interacted with R. At the moment the device applies the decoding map, it must be in one of | D | operationally distinguishable configurations. That is the physical key.
Selecting among | D | operationally distinct decoders is a control problem that requires physical state inside the patch. I formalise this as control information in the Stack Theory sense [5,6]. A selector register that can implement any D D needs at least | D | reliably distinguishable settings. So its key vocabulary must satisfy I ( v K ) log 2 | D | by the capacity definition in Box 2. The next subsection turns that informal point into a single counting inequality that also accounts for limited reconstruction coverage.

Quantum Programming Does Not Evade the Key

In this paragraph the word program is used in the programmable quantum processor sense. It refers to a control register state that selects an operation. This is not the same as a Stack Theory program, which is a subset of an environment. If one models the selector as a quantum program register, the conclusion is strengthened. For exact deterministic programming of a finite set of distinct unitaries, the program states must be mutually orthogonal, so the program register dimension is at least the number of implemented unitaries [13]. Approximate programming relaxes orthogonality, but quantitative results still impose nontrivial resource costs as the target family grows and the tolerated error shrinks. Kubicki, Palazuelos and Pérez-García provide quantitative lower bounds for approximate programmable devices. These bounds show that the required program register dimension grows at least polynomially in the target system dimension and in 1 / ε for natural families of measurements and channels. I use this only to support the qualitative point that allowing a small error does not make the physical program register free [14]. The inequality in Theorem 1 uses exact distinguishability of the control states. In the approximate regime the required program register dimension can exceed | D | , which would only strengthen the bound. A related use of programmable processor bounds together with gravitational entropy bounds to constrain bulk computation in AdS/CFT appears in Kubicki, May and Pérez-García [15]. My argument targets firewall verification and combines this programming requirement with alpha-bit coverage bounds to obtain an explicit horizon area fraction requirement.
The decoder key bound below is the step that closes the common loophole. A short classical description of a decoder is not a physical key unless that description is instantiated inside the patch as a distinguishable control state.

4.3. Reconstruction Bounds Limit Coverage

I now use a mainstream model class in which the state dependence of interior reconstruction is controlled. Concretely, I work in evaporating black hole models where the black hole is coupled to a non-gravitating bath [9,10]. The bath is an ordinary quantum system that stores Hawking radiation and can be measured without gravitational backreaction. The early radiation lives in bath degrees of freedom that I call R. The remaining black hole is described by the gravitational region or, in AdS/CFT language, by a boundary CFT. Interior reconstruction is described by entanglement wedge reconstruction. In plain English, this means that some interior operators can be represented as operators acting on exterior degrees of freedom, but only on a limited code subspace of microstates. The size of that code subspace is controlled by a geometric quantity that appears in Proposition 1.
Definition 2
(Error metric). For density matrices ρ and σ on the same Hilbert space, define the trace distance
Δ ( ρ , σ ) : = 1 2 ρ σ 1 .
Here X 1 : = Tr X X is the trace norm. An error tolerance ε means I accept trace distance at most ε between the decoded state and the target state. Throughout, an “error at most ε” means the same thing.
The trace distance is a standard way to say how distinguishable two quantum states are. It is zero only when the states are identical. It is at most one. If the trace distance is at most ε , then no measurement can tell the two states apart with advantage larger than about ε .
Definition 3
(Decoder microstate coverage). Fix an error tolerance ε ( 0 , 1 ) . Fix the microstate set Φ micro from Box 3. For a decoder channel D D , I define d ε ( D ) to be themaximum cardinality of a subset Φ micro Φ micro such that D reconstructs the purifier of B with error at most ε uniformly on all states supported on the code subspace spanned by Φ micro 4. Uniformly includes mixtures of those microstates and states entangled with an external reference. Let
d ε max : = max D D d ε ( D ) .
Think of d ε ( D ) as the number of mutually exclusive microstate cases that one fixed decoder D can handle5. Uniformly means the decoder works not just for each microstate case but also for arbitrary mixtures and even when the microstate is entangled with an external reference system. This is the regime relevant for the holographic reconstruction bounds I use later. The geometric parameter α * below comes from a competition between two quantum extremal surfaces.
A quantum extremal surface is a surface that extremizes the generalized entropy in semiclassical gravity. The area gap A 2 A 1 acts like an entanglement budget that controls how large a code subspace can share one fixed decoder.
Proposition 1
(Hayden–Penington microstate coverage ceiling). Work in the holographic reconstruction setting analysed by Hayden and Penington [8]. Let A be the boundary region available to the verifier. Let A 1 and A 2 be the areas of the two competing quantum extremal surfaces anchored to A , with A 2 > A 1 . Let A 0 be the horizon area so that S BH = A 0 / ( 4 𝓁 p 2 ln 2 ) . Define
α * : = A 2 A 1 A 0 ( 0 , 1 ) .
Fix an error tolerance ε ( 0 , 1 ) measured in trace distance. There exists a constant c ε that does not scale with S BH such that the following holds.
Consider any single, state-independent reconstruction channel D acting on the verifier-accessible region A. Then
d ε ( D ) 2 α * S BH + c ε and hence d ε max 2 α * S BH + c ε .
Proof 
(Proof sketch). Hayden and Penington analyse entanglement wedge reconstruction for a boundary region A in a black hole code space entangled with an external reference system R [8]. Two quantum extremal surfaces with areas A 1 and A 2 compete. In their derivation, decoding the intermediate region between the two surfaces from A for all states in a chosen subspace reduces, at leading order in G N , to requiring that the entropy S ( R ) never exceed the area gap. More precisely, their decoding condition is equivalent to the condition discussed around their equation (4.12) in Section 4.
4 G N S ( R ) ψ < A 2 A 1
uniformly over those states ψ . Here G N is Newton’s constant. Up to unit conventions, an area term divided by 4 G N is the Bekenstein–Hawking entropy in natural logarithm units. In this paper I express S BH in bits via A 0 / ( 4 𝓁 p 2 ln 2 ) , so the scaling matches after dividing by ln 2 . If R purifies a maximally mixed state on a d-dimensional subspace, then S ( R ) ln d . Requiring uniform success on all mixed and reference-entangled states therefore forces
ln d A 2 A 1 4 G N + O ( 1 ) .
Hayden and Penington use natural logarithms. Dividing by ln 2 converts to bits and yields log 2 d α * S BH + O ( 1 ) . In my notation I can take d to be d ε ( D ) from Definition 3, since any covered set of d ε ( D ) orthonormal microstates spans a d ε ( D ) -dimensional code subspace on which D works uniformly. I absorb fixed accuracy and finite-size corrections into the additive constant c ε .    □
Lemma 2
(Decoder keys are hidden entropy). Let M be afinitemicrostate6 set with | M | = 2 S . Let D be a finite family of operationally distinct decoders.
For each D D , let M ( D ) M be the set of microstates on which D succeeds uniformly at some fixed accuracy. Assume a uniform coverage bound
| M ( D ) | d max for every D D .
If the protocol must succeed for every microstate, meaning D D M ( D ) = M , then
| D | | M | d max = 2 S d max .
Therefore any control register that must select among D at verification time must support at least | D | reliably distinguishable control states. Equivalently any key vocabulary v K representing that register must satisfy
I ( v K ) log 2 | D | S log 2 d max .
Proof. 
The covering condition gives
| M | D D | M ( D ) | | D | d max .
Rearrange to get the lower bound on | D | . A selector register that can implement any D D needs at least one reliably distinguishable control state per operationally distinct decoder. So the register has at least | D | internal states and any representing vocabulary satisfies I ( v K ) log 2 | D | .    □
Interpretation. If each decoder label leaves at most d max microstates unresolved, then the label must carry the rest of the micro information. So the instruction budget is a hidden entropy term.

4.4. No Subhorizon Firewall Verifier Theorem

I can now state the main no-go theorem. It is a single inequality.
Theorem 1
(No subhorizon AMPS verifier under reconstruction limits). Assume an old black hole with horizon entropy S BH in bits. Fix the microstate set Φ micro from Box 3 and assume | Φ micro | = 2 S BH . Fix an error tolerance ε ( 0 , 1 ) .
Consider a single patch verifier whose entire implementation at verification time is confined to a causal patch R with A ( R ) ρ 2 A 0 for some fixed ρ ( 0 , 1 ] . If the verifier must succeed for every microstate ϕ Φ micro with error at most ε, then any such protocol requires a key vocabulary with
I ( v K ) S BH log 2 d ε max .
If Proposition 1 holds, then
I ( v K ) ( 1 α * ) S BH c ε .
Since the key must be physically instantiated inside R , success requires
I max ( R ) I ( v K ) .
By Lemma 1, this is possible only if
ρ 2 S BH S BH log 2 d ε max .
Equivalently
ρ 2 1 log 2 d ε max S BH .
In particular, inserting Proposition 1 yields the necessary condition
ρ 2 1 α * c ε S BH .
For large S BH , this approaches ρ 2 1 α * .
The theorem is a capacity ledger. If one fixed decoder works uniformly on at most d ε max microstates, then covering all microstates in Φ micro , which has size 2 S BH , requires at least 2 S BH / d ε max different decoders. Picking one decoder out of that family costs at least log 2 ( 2 S BH / d ε max ) = S BH log 2 d ε max control bits. The covariant entropy bound says a subhorizon patch with area fraction ρ 2 can store at most about ρ 2 S BH bits. So the patch can hold the decoder choice only if ρ 2 is large enough.
Proof. 
Apply Lemma 2 with M = Φ micro so that | M | = 2 S BH . For each operational decoder D D , take M ( D ) to be a uniform success set at accuracy ε . By definition of d ε max I have | M ( D ) | d ε max for every D. The single-observer requirement means the protocol must succeed for every microstate, so the success sets cover Φ micro . Lemma 2 then gives the key budget lower bound
I ( v K ) S BH log 2 d ε max .
Lemma 1 gives the patch ceiling I max ( R ) ρ 2 S BH . Since the key must be physically instantiated inside R , I need I max ( R ) I ( v K ) . Combine the inequalities and divide by S BH to get
ρ 2 1 log 2 d ε max S BH .
Finally, insert Proposition 1 to obtain the specialised form with α * and c ε .    □
Corollary 1
(It takes horizon-scale area to verify a horizon). Fix any ρ ( 0 , 1 ) and fix ε ( 0 , 1 ) . Under the assumptions of Theorem 1, if
α * < 1 ρ 2
then no single patch verifier confined to A ( R ) ρ 2 A 0 can succeed on all microstates for sufficiently large S BH . Equivalently, if α * is bounded away from one, any successful verifier must control a patch whose boundary area is an order-one fraction of the horizon area.
Proof. 
Theorem 1 gives the necessary condition ρ 2 1 α * c ε / S BH . If α * < 1 ρ 2 , then for sufficiently large S BH this condition fails.    □
As a concrete example, take α * = 0.2 and neglect the subleading term c ε / S BH . A single fixed decoder then works uniformly on at most 2 0.2 S BH microstates, that is a code subspace with 20 % of the black hole entropy in bits.
If you prefer a concrete count, take S BH = 10 bits. Then there are 2 10 = 1024 microstates and one decoder can cover at most 2 0.2 × 10 = 4 of them. Covering all microstates therefore needs at least 1024 / 4 = 256 decoders and selecting among them costs log 2 256 = 8 control bits.
In general, selecting the correct decoder costs about 0.8 S BH control bits. Theorem 1 gives the requirement ρ 2 0.8 and hence ρ 0.8 0.89 . In other words, a successful verifier must control a patch whose boundary area is about 80 % of the horizon area. Conversely, a patch that accesses only a quarter of the horizon area, meaning ρ 2 = 1 / 4 , could only work if α * 0.75 .
In black hole plus bath models that realise the Page curve using islands, the two surfaces with areas A 1 and A 2 are the competing quantum extremal surfaces that appear in the generalized entropy minimisation problem. For a given geometry and time slice, A 1 and A 2 can be computed semiclassically, so α * = ( A 2 A 1 ) / A 0 can in principle be extracted as a function of time. Penington and Almheiri et al compute these competing surfaces in explicit models as part of entanglement wedge reconstruction and the Hawking radiation entropy calculation [9,10]. In this paper I treat α * as an input parameter because its numerical value is model-dependent.

4.5. Summary Table and Tradeoff Curve

The table below collects the variables.
Quantity Meaning
S BH horizon capacity in bits, S BH = A 0 4 𝓁 p 2 ln 2
d ε max maximal code subspace dimension per fixed decoder at error ε
α * reconstruction fraction, α * = ( A 2 A 1 ) / A 0
c ε error term that depends on ε and does not scale with S BH
| D | decoder family size, | D | 2 S BH / d ε max 2 ( 1 α * ) S BH c ε
I ( v K ) key bits, I ( v K ) log 2 | D | S BH log 2 d ε max ( 1 α * ) S BH c ε
ρ patch size fraction, A ( R ) ρ 2 A 0
No-go condition     ρ 2 1 log 2 d ε max S BH 1 α * c ε S BH
Figure 1 plots the same inequality as a tradeoff curve between α * and the required patch fraction ρ .

5. What the Theorem Does and Does Not Show

Theorem 1 does not prove that firewalls exist. It also does not prove that firewalls do not exist. It proves something narrower and more operational. Under controlled reconstruction assumptions, the standard firewall verification protocol does not fit inside any subhorizon causal patch unless log 2 d ε max is close to S BH , which corresponds to reconstruction that is close to state independent.
There are of course some potential escape hatches if we violate or make very specific assumptions.
1.
Exponential state independence. If α * is extremely close to one, then one decoder can succeed uniformly on code subspaces whose size in bits is close to the full black hole entropy S BH . That is a strong form of state independence.
2.
Horizon-scale control. If α * is bounded away from one, then Theorem 1 forces ρ 2 to be bounded below by a fixed constant that does not shrink as S BH grows. The verifier must control a patch whose boundary area is an order-one fraction of the horizon area. That is not a small infaller experiment (hence the word “small” in the title of this paper).
3.
Give up microstate coverage. If one only demands success on an exponentially small fraction of microcases, the key size can shrink by an amount that scales with S BH . Succeeding on any fixed nonzero fraction still costs order S BH control bits, up to an additive constant. This is not the AMPS claim.
This is not the Harlow–Hayden complexity barrier [12]. Even an oracle for the computation does not remove the need to physically specify which channel is being applied. The obstruction is representational and covariant. It is also not removed by stretching the protocol over more time. In Stack Theory terms, that replaces a demand for co-instantiation at one instant by a weaker demand that each ingredient shows up somewhere in a time window [6,11]. The AMPS paradox is a single-observer contradiction. It only bites if the observer can hold one joint record in which the ingredients are simultaneously present. If the patch ceiling forbids that co-instantiation, then no amount of time sharing can restore the missing joint record.
More broadly, whenever the correct operation depends on hidden microscopic details, the program register can become the scarce resource before the data register does.

6. Conclusions

All I have really done in this paper is translate established generalisation-optimal learning results from Stack Theory into physical language and apply them to the black hole firewall paradox [22], implementing what was already alluded to but not formally established in a conference paper [16].
I have shown that a single observer confined to a subhorizon causal patch faces a capacity bottleneck when attempting the AMPS firewall verification protocol. The core inequality ρ 2 1 α * c ε / S BH says that if the geometric reconstruction fraction α * is bounded away from one, then the observer must control a patch whose boundary area is an order-one fraction of the horizon area. The bottleneck is not computational complexity but representational capacity: the physical control register that selects the correct decoder must fit inside the same patch as the verification outcome.
The result is conditional on the assumptions stated in Box 3. The covariant entropy bound is treated as a memory ceiling, which is an interpretation step beyond the formal bound on light-sheet entropy. The state-dependence input relies on holographic reconstruction bounds from evaporating black hole plus bath models. If either assumption is weakened, the quantitative bound changes, but the qualitative point survives whenever one fixed decoder cannot cover all microstates.
Three directions for future work look both feasible and falsifiable. First, island-style reconstruction thresholds can be recast as explicit join constraints, expressing the usual entanglement wedge story as an inequality over capacities rather than a narrative about observers. Second, the universal success requirement in the theorem can be relaxed to a tunable success fraction and matched to known code subspace constructions, turning the key bound into a family of tradeoff curves that can be compared directly to explicit state-dependent reconstruction schemes. Third, the same separation between computation and control should apply beyond black holes. If the covariant entropy bound is a genuine memory ceiling, then similar capacity ledger constraints should appear wherever the correct operation depends on hidden microscopic details, including cosmological horizons and any scenario where the bound is tight.
Finally, none of these steps require Stack Theory to be a theory of everything. It is merely the road I took to get to this conclusion. Whatever quantum gravity is and whatever formalism we use, quantum gravity should respect this covariant information accounting.

7. Methods and Supplementary Information

All results in the main text are analytic. All the experiments, code, this paper’s Supplementary Information (SI) and the supporting Stack Theory community of work is available in the online GitHub repository [6]. Figure 1 is a plot of a closed-form inequality from Theorem 1. Extended Data Figure 1 and 2 are toy computations that illustrate the finite capacity bookkeeping behind Box 2 and Lemma 2. They do not model quantum gravity, holographic dynamics, or black hole evaporation. They are included only for intuition and can be omitted without changing any claim in the main text.

7.1. Supplementary Code 1 in Plain English

Supplementary Code 1 is the Python file stack_physics_suite_experiments.py. It produces the PNG figures under figures and the CSV source data files under source_data. For a plain English description of each PNG and each CSV file, see figures/README.md and source_data/README.md. The script fixes its random seeds and it strips variable PNG metadata. If you rerun it, you should reproduce the numerical CSV outputs exactly. The PNG files should match on the same plotting stack, but different Matplotlib versions can produce minor byte-level differences.

7.2. How to Reproduce the Toy Computations

Unzip the bundle and run python -B stack_physics_suite_experiments.py from the bundle root. The script rewrites three PNG figures and three CSV source data files. The PNG files live under figures and the CSV files live under source_data. The parameter values used for each plot are stored in the corresponding CSV file and are summarised here.
Output What it is Key parameters
Figure 1 theorem curve ρ min = 1 α * 400 grid points for α * [ 0 , 1 ]
Extended Data Fig. 1 toy complementarity histogram env_n = 12 , m = 6 , p = 0.2 , samples = 200 , seed = 2
Extended Data Fig. 2 toy decoder cover scaling α = 0.6 , n { 6 , 8 , 10 , 12 } , factor = 12 , trials = 2 , seed = 2
What the toy parameters mean. The parameter env_n is the number of bits used to index the toy world states. So the environment has 2 env _ n mutually exclusive states. The parameter m is the number of yes or no questions in each vocabulary. The parameter p is the probability that a given world state is included in a given random question. The parameter n is the number of microstate bits in the decoder coverage toy model. The toy microstate set has size | H | = 2 n . The parameter alpha sets the per-decoder coverage size d max = 2 α n . The parameter factor controls how many random candidate decoders I generate before greedy cover. The parameter trials is how many independent random covers I run at each n. The parameter seed is the fixed random seed used for reproducibility.

7.3. How to Read the Plots

This bundle includes one analytic curve and two toy plots. All three are about counting distinctions, meaning how many different internal cases a verifier must be able to tell apart.

Figure 1.

This is the theorem in picture form. Pick a value of α * on the horizontal axis. That value says how large the best-case decoder code subspace can be, as a fraction of the full microstate entropy. The vertical axis tells you the smallest patch fraction ρ that can hold the decoder choice. If α * is far below one, then ρ is close to one, meaning the verifier needs access to most of the horizon area.

Extended Data Fig. 1.

This histogram shows a join overflow. Each vocabulary by itself fits under the ceiling. But their join often exceeds the same ceiling. That is the toy analogue of complementarity as a failure of joint realisability under a capacity bound.

Extended Data Fig. 2.

This plot shows why state dependence forces key bits. If one decoder only covers a fraction 2 α n of 2 n microstates, then you need about 2 ( 1 α ) n decoders to cover everything. Taking log 2 turns that into a line of slope 1 α . The greedy set cover curve sits above the lower bound because random decoders overlap.

7.4. Figure 1 in the Main Text

This figure is not a simulation. It plots the large entropy limit of the theorem. In the simplified regime where log 2 d ε max α * S BH and subleading constants are neglected, the theorem gives
ρ min = 1 α * .
I sample 400 evenly spaced values of α * between 0 and 1 and plot the resulting curve.
Algorithm translated to English.
1.
Sample α * uniformly on a grid between 0 and 1.
2.
Compute ρ min = 1 α * at each grid point and plot the curve.

7.5. Extended Data Figure 1

The point of this plot is to show complementarity as a capacity overflow in a finite toy model. I build a toy world with | Φ | = 2 12 mutually exclusive states. I sample two vocabularies v 1 and v 2 . Each vocabulary contains m = 6 random programs. Each program is a random subset of Φ obtained by including each state independently with probability p = 0.2 . For each sampled pair, I form the join vocabulary v 1 v 2 by taking the union of their programs. I compute I ( v 1 v 2 ) and plot a histogram over 200 samples. The illustrative ceiling in the figure is 9.8 bits. The random seed is fixed to 2.
Algorithm translated to English. A histogram is a bar chart of counts grouped into bins.
1.
Fix the toy parameters env _ n , m, p, the number of samples, and the random seed.
2.
For each sample, draw two random vocabularies v 1 and v 2 on the same finite environment Φ .
3.
Form the join vocabulary v 1 v 2 by taking the union of their programs.
4.
Compute the join encoding capacity I ( v 1 v 2 ) = log 2 Im ( Enc v 1 v 2 ) .
5.
Plot the histogram of these values and draw the ceiling line.
The horizontal axis in the histogram is the join capacity I ( v 1 v 2 ) in bits. The vertical axis is the number of sampled vocabulary pairs that fall in each bin. The vertical reference line is the illustrative ceiling.

7.6. Extended Data Figure 2

Warning about notation. The toy exponent parameter α used in this section is not the same object as the theorem parameter α * . The theorem parameter α * comes from competing quantum extremal surfaces. The toy parameter α only controls the integer subset size d max = 2 α n in a finite set cover model.
The point of this plot is just to illustrate the decoder counting step in a finite toy model. I model the microstate basis as a finite set H with | H | = 2 n . I assume one decoder can succeed on at most d max = 2 α n microstates. Covering all microstates then requires at least | H | / d max decoders. I plot log 2 | H | / d max as a function of n.
I also plot a random overlap model. In that model, each decoder is a random subset of size d max and I apply a greedy set cover algorithm. Set cover means choosing a collection of decoders whose union contains every microstate in H. Greedy means I repeatedly pick the decoder that covers the largest number of microstates that are still uncovered. Because random subsets overlap, the greedy cover typically needs more decoders than the lower bound. The parameter values in Supplementary Code 1 are α = 0.6 , n { 6 , 8 , 10 , 12 } , factor 12, trials 2, and random seed 2. Factor sets how many random decoders I generate in the toy model, relative to the lower bound. Specifically I generate M = factor × | H | / d max random subsets before running greedy cover.
Algorithm in plain English.
1.
Choose α , a list of microstate bit sizes n, and the random seed.
2.
For each n, set | H | = 2 n .
3.
Set the per-decoder coverage to d max = 2 α n .
4.
Compute the no-overlap lower bound L = | H | / d max and record log 2 L .
5.
Generate M = factor × L random candidate decoders, each as a subset of H of size d max .
6.
Run greedy set cover to select a collection whose union is all of H. Record the cover size and average it over the requested number of trials.
The Supplementary Information contains a proof of the capacity complementarity inequality, a formal separation between single patch and distributed realisation following the Temporal Gap non-commutation result, Theorem 3 in [11], and robustness checks for the AMPS ledger. It also documents the toy computations in more detail and explains how each plot connects back to the inequalities used in the main theorem.

Provenance

The central technical result is Theorem 1 and its corollaries. It packages the decoder selection problem into a simple covariant capacity ledger inequality. The argument combines a covariant area-based capacity ceiling [7] with a state-dependence bound for interior reconstruction [8] and standard limits on programmable quantum operations [13,14]. Related uses of programmable processor bounds together with gravitational entropy bounds to constrain bulk computation in AdS/CFT appear in [15]. The present paper applies this control information lens to firewall verification and combines it with alpha-bit coverage bounds to obtain an explicit horizon area fraction requirement. The Stack Theory primitives used to formalise capacity and control information follow [5,6,16,23]. All plotted data are generated from closed-form expressions or from synthetic toy models in Supplementary Code 1.   

Supplementary Materials

The following supporting information can be downloaded at the website of this paper posted on Preprints.org.

Data Availability Statement

No empirical datasets were generated or analysed. Source data for Figure 1 and Extended Data Figures 1 and 2 are provided as CSV files in the bundle under source_data. These CSV files are generated by Supplementary Code 1.

Code Availability

As stated earlier the code is in the folder named after this paper in the GitHub repository [6]. Supplementary Code 1 (stack_physics_suite_experiments.py) reproduces Figure 1 and Extended Data Figure 1 and 2 and writes the corresponding source data CSV files. It depends on numpy and matplotlib. A vendored copy of stacktheory_suite (version 0.8.4) is included in the top-level directory of the GitHub, which you’ll need to re-run the experiment code.

AI-Assisted Technologies

In the preparation of this manuscript I used ChatGPT, Claude, Grok and Gemini for language editing, refactoring code, checking for errors, generating feedback / simulated reviews, and for detailed searches into less familiar topics. The actual substance of this research was from a human, as evidenced by discussion of these ideas in an earlier conference paper on which this paper substantially depends [16]. All I’ve really done here is translate my existing work into a new context. All theoretical claims, derivations, and final wording were reviewed and approved by me. I take full responsibility for the content.

Author Contributions

I (M.T.B.) conceived the study, developed the theory, and wrote the manuscript.

Conflicts of Interest

I declare no competing interests.

Correspondence and Requests for Materials

Correspondence and requests for materials should be addressed to M.T.B. (m@michaeltimothybennett.com).

References

  1. Susskind, L.; Thorlacius, L.; Uglum, J. The stretched horizon and black hole complementarity. Physical Review D 1993, 48, 3743–3761. [Google Scholar] [CrossRef] [PubMed]
  2. ’t Hooft, G. On the quantum structure of a black hole. Nuclear Physics B 1985, 256, 727–745. [Google Scholar] [CrossRef]
  3. Almheiri, A.; Marolf, D.; Polchinski, J.; Sully, J. Black holes, complementarity, or firewalls. Journal of High Energy Physics 2013, arXiv:hep2013, 062. [Google Scholar] [CrossRef]
  4. Hausmann, L.; Renner, R. The firewall paradox is Wigner’s friend paradox. arXiv 2025, arXiv:quant. [Google Scholar] [CrossRef]
  5. Bennett, M.T. How To Build Conscious Machines. PhD thesis, The Australian National University, 2025. [Google Scholar] [CrossRef]
  6. Bennett, M.T. Technical Appendices, 2025. Archived release on Zenodo. Available online: https://github.com/ViscousLemming/Technical-Appendices. [CrossRef]
  7. Bousso, R. The holographic principle. Reviews of Modern Physics 2002, 74, 825–874, [hep-th/0203101. [Google Scholar] [CrossRef]
  8. Hayden, P.; Penington, G. Learning the Alpha-bits of Black Holes. Journal of High Energy Physics 2019, arXiv:hep2019, 007. [Google Scholar] [CrossRef]
  9. Penington, G. Entanglement wedge reconstruction and the information paradox. Journal of High Energy Physics 2020, arXiv:hep2020, 002. [Google Scholar] [CrossRef]
  10. Almheiri, A.; Hartman, T.; Maldacena, J.; Shaghoulian, E.; Tajdini, A. The entropy of Hawking radiation. Reviews of Modern Physics 2021, arXiv:hep93, 035002. [Google Scholar] [CrossRef]
  11. Bennett, M.T. A Mind Cannot Be Smeared Across Time. In Proceedings of the Forthcoming in Proceedings of the AAAI 2026 Spring Symposium on Machine Consciousness: Integrating Theory, Technology, and Philosophy, Burlingame, California, USA, April 2026. [Google Scholar]
  12. Harlow, D.; Hayden, P. Quantum computation versus firewalls. Journal of High Energy Physics 2013, arXiv:hep2013, 085. [Google Scholar] [CrossRef]
  13. Nielsen, M.A.; Chuang, I.L. Programmable quantum gate arrays. Physical Review Letters 1997, 79, 321–324. [Google Scholar] [CrossRef]
  14. Kubicki, A.M.; Palazuelos, C.; Pérez-García, D. Resource Quantification for the No-Programming Theorem. Physical Review Letters 2019, arXiv:quant122, 080505. [Google Scholar] [CrossRef] [PubMed]
  15. Kubicki, A.M.; May, A.; Pérez-García, D. Constraints on physical computers in holographic spacetimes. SciPost Phys. 2024, arXiv:quant16, 24. [Google Scholar] [CrossRef]
  16. Bennett, M.T. Is Complexity an Illusion? In Proceedings of the 17th International Conference on Artificial General Intelligence; Springer, 2024. Lecture Notes in Computer Science. [Google Scholar] [CrossRef]
  17. Bekenstein, J.D. Universal upper bound on the entropy-to-energy ratio for bounded systems. Phys. Rev. D 1981, 23, 287–298. [Google Scholar] [CrossRef]
  18. Bennett, M.T. Computational Dualism and Objective Superintelligence. In Proceedings of the 17th International Conference on Artificial General Intelligence; Springer, 2024. Lecture Notes in Computer Science. [Google Scholar] [CrossRef]
  19. Bennett, M.T. Are Biological Systems More Intelligent Than Artificial Intelligence? Philosophical Transactions of the Royal Society B: Biological Sciences. Special issue on Hybrid agencies: crossing borders between biological and artificial worlds 2024, arXiv:cs. [Google Scholar] [CrossRef]
  20. Bekenstein, J.D. Black holes and entropy. Physical Review D 1973, 7, 2333–2346. [Google Scholar] [CrossRef]
  21. Hawking, S.W. Particle creation by black holes. Communications in Mathematical Physics 1975, 43, 199–220. [Google Scholar] [CrossRef]
  22. Bennett, M.T. Optimal Policy Is Weakest Policy. In Proceedings of the Artificial General Intelligence. Springer; Lecture Notes in Computer Science; 2025; Vol. 16057, pp. 43–56. [Google Scholar] [CrossRef]
  23. Bennett, M.T. The Optimal Choice of Hypothesis Is the Weakest, Not the Shortest. In Proceedings of the 16th International Conference on Artificial General Intelligence; Springer, 2023; Lecture Notes in Computer Science; pp. 42–51. [Google Scholar] [CrossRef]
1
Mathematically it can be written as a unit vector | ψ , or as a rank-one density matrix ρ = | ψ ψ | .
2
Since light can’t get out, might look like a pitch black surface.
3
In quantum field theory, “empty space” is not nothing. It is a pattern of entanglement between adjacent regions. In the Rindler and Unruh picture this entanglement is what makes the vacuum look thermal to accelerated observers. Breaking this pattern releases energy.
4
In Stack Theory terms d ε ( D ) can be understood as weakness of the weakest correct policy, but in raw microstates rather than a higher level embodied language.
5
For those familiar with Stack Theory, the number of mutually exclusive microstate cases a decoder can handle would be proportional to Stack Theory’s weakness measure, if one interprets the decoder as an embodied policy as in [16].
6
M is finite meaning it is not a complete description of the universe microstate like ϕ Φ would be.
Figure 1. The large entropy limit of Theorem 1 gives a simple tradeoff curve between the reconstruction fraction α * and the required patch size fraction ρ . In the regime d ε max 2 α * S BH with subleading constants neglected, the theorem reduces to ρ 2 1 α * , so ρ min = 1 α * . The curve is evaluated on 400 evenly spaced values of α * . The horizontal axis is α * . The vertical axis is ρ min . Source data are provided in source_data/source_data_Fig1_rho_min_vs_alpha.csv. When α * is not close to one, ρ min is an order-one fraction of the horizon, meaning a small infaller patch cannot hold the decoder choice.
Figure 1. The large entropy limit of Theorem 1 gives a simple tradeoff curve between the reconstruction fraction α * and the required patch size fraction ρ . In the regime d ε max 2 α * S BH with subleading constants neglected, the theorem reduces to ρ 2 1 α * , so ρ min = 1 α * . The curve is evaluated on 400 evenly spaced values of α * . The horizontal axis is α * . The vertical axis is ρ min . Source data are provided in source_data/source_data_Fig1_rho_min_vs_alpha.csv. When α * is not close to one, ρ min is an order-one fraction of the horizon, meaning a small infaller patch cannot hold the decoder choice.
Preprints 201243 g001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated