Black Holes and Complexity via Constructible Universe

: The relation of randomness and classical algorithmic computational complexity is a vast and deep subject by itself. However, already, 1-randomness sequences call for quantum mechanics in their realization. Thus, we propose to approach black hole’s quantum computational complexity by classical computational classes and randomness classes. The model of a general black hole is proposed based on formal tools from Zermelo–Fraenkel set theory like random forcing or minimal countable constructible model L α . The Bekenstein–Hawking proportionality rule is shown to hold up to a multiplicative constant. Higher degrees of randomness and algorithmic computational complexity are derived in the model. Directions for further studies are also formulated. The model is designed for exploring deep quantum regime of spacetime.


Introduction and Motivations
Susskind et al. (e.g., [1][2][3]) in a series of papers have approached the complexity of quantum systems. Their findings are particularly interesting in the case of black holes (BH) and the horizons which are formed in spacetime. This program can be seen as the effort toward understanding the regime where quantum mechanics (QM) and spacetime overlap each other as deeply involved physical structures. Thus far, we are left with separated formalisms of two fundamental physical theories i.e., QM and general relativity (GR) where this last certainly describes spacetime as dynamical entity. The celebrated Penrose-Hawking singularity theorems indicate that there has to exist a regime of spacetime where description by GR is not applicable any longer and where presumably certain theory of quantum gravity will have to find its place. This is where black holes are formed and where singularities transcending the usual GR description have to appear. Understanding BH leads in fact to understanding the quantum regime of spacetime and gravity (QG). However, we are facing here the crucial difficulty and we know neither the correct mathematical structure of QG nor even the more or less successful guess of it. Thus, the attitude to QG from the perspective of BHs should be based on the careful recognition of mathematics involved in this area keeping at the back of a head the possibility for a proper guess which would eventually emerge from the formalism. Historically, the Penrose-Hawking singularity theorems resulted with the careful studying of GR's mathematical structure and finding its limitations. Now, we try to find proper indications coming from the quantum side (QM and BHs), rather than GR that could lead to understanding of QG. In particular, the classical world can be seen now as defining limitations for the QM formalism.
Besides the phenomenon of BHs, there exist even more aspects of QM that recently received much attention from researchers. One is what can be called quantum information processing and, in particular, the 'mutual' search for an effectively working version of a quantum computer. Anyway, 'quantum computing' in a general form is realized in nature in various ways as evolution of entangled or interfered states, which eventually undergo unpredictable wave function collapse, and the process can be used to characterize physical systems. This was an idea of Susskind et al. to work out intrinsic characteristics of BHs in terms of quantum computational complexity. In their analysis, quantum circuits appear where the minimal number of gates indicates the complexity of the quantum process (see below in this section). They argued that BHs are intrinsically realizing such quantum computational complexity and many physical properties of BHs like entropy are derivable from it. Another important aspect of QM recently under intense investigation is the problem of inherent randomness of QM. In mathematics, there exists a vast domain devoted to defining and studying various concepts of randomness and their interrelation with classical computational complexity (see, e.g., [4], Appendix B). Whether QM is actually random and what degree of randomness it realizes is a deep problem in general (see, e.g., [5,6]). We consider the domain of algorithmic complexity based on classical Turing machines as a classical one. From that point of view, the computational complexity referred to in the papers by Suskind et al. belongs to the quantum domain. The general premise we follow in this work is that the classical algorithmic computational complexity delineates limitations for the quantum one (similarly to the classical realm of spacetime setting limitations for quantum BH). However, the classical and quantum computational complexities overlap in a specific sense and this overlapping can be used to characterize quantum BHs and classical spacetime-and presumably gives insights about the mathematics of QG. In this paper, we address the question whether, and under what conditions, the quantum complexity of BHs can be seen as classically originated. Even confirmed to some degree, it would indicate that the QM realm of BHs and the classical realm of spacetime can be seen as overlapping domains from the complexity point of view (where the complexity links them both). Such approach is based on two important observations. Firstly, BHs are inherently characterized by their entropy (as the Beckenstein-Hawking area law) and the algorithmic complexity we use here is also related to entropy. Secondly, the binary sequences of 'successes' and 'fails' as resulted in the process of infinite fair coin tossing (with the precise probability 1 2 / 1 2 at each toss) cannot be realized by a deterministic classical process [6]. It follows that randomness realized by such sequences has to be quantum even though the entire process seems to be describable classically. It seems to be, but actually it is not. In this way, randomness can be seen as another aspect of the overlapping domains of QM and spacetime realms. Finally, we carefully approach the mathematics suitable for the description of both realms. It seems that the most suitable mathematical formalism is one based on models of axiomatic set theory and this fact follows from the structure of QM lattice of projections L. The basic reason behind such statement is 'Solovay genericity', the phenomenon known from random forcing in set theory and from algorithmic computational complexity and randomness where it is called 'weak genericity'. Precisely, this kind of genericity is realized within a QM lattice of projections on infinite dimensional Hilbert spaces.
The entire project is embedded in formal mathematics, and various mathematical concepts appear that are not in direct everyday use by most of the physicists, which by itself may 'hide' the simplicity of basic ideas. There are notions like randomness, models of Zermelo-Fraenkel set theory, forcing or even axiomatic formal systems (that already require some familiarization with the usual mathematical practice). However, when one adopts a slight extension of accessible mathematical tools, the initially involving and logically distant ideas become rather direct and easily expressible. To facilitate access to the results in the paper, we have attached appendices where the reader can find indications and guidelines regarding basic constructions in QM and computational complexity. As already mentioned above, the important clues regarding mathematics of QG should come from the unifying description of BHs by formal set theory methods (e.g., [7,8]). Now let us be more specific. Suskind et al. have proposed the measure of intrinsic computational complexity of states in a Hilbert space of dimension 2 K , where K is the number of qubits. A way to the complexity went through the relational complexity C(U, V) between two unaries, i.e., unitary complex matrices U, V ∈ SU(2 K ) where this last space is much bigger than the state space CP(2 K − 1)-the complex projective space of dimension 2 K − 1. Let G be a set of quantum gates considered as fundamental.
C(U, V) = the minimal number of gates g i ∈ G ⊂ SU(2 K ) such that U = g 1 g 2 . . . g k V .
Then, the complexity of each V ∈ SU(2 K ) is defined as In fact, C(U, V) is a right-invariant metric on SU(2 K ) which differs essentially from the usual inner product metric [1] d(U, V) = arccos Tr|UV † | . Then, the task is to find the structural ingredients of QM responsible for the complexity of quantum systems and BHs in particular. The quantum gates and circuits involved in the definitions of the complexity distance above are based on unitary evolution of states and, as such, do not refer to the collapse of a wave function taking place during a quantum measurement. In this case, we would live in a world of quantum computations that permits entanglement and maintains it without classical reduction. However, when we allow measurements on entangled systems, we still have much room for computational complexity in the resulting classical realm. In addition, the measurement in QM that enforces classicality is the source of randomness and unpredictability in QM [5,6]. This algorithmic complexity in classical contexts (based on classical Turing machines-see Appendix B) belongs to classical complexity theory. However, we expect that the degree of randomness it carries is high enough for giving an alternative to quantum computing point of view in the context of BHs. As we mentioned already, such expectation is supported by the result that classical 1-random behavior of the coin-tossing generating infinite binary sequence cannot be realized as a deterministic process and, in fact, it requires QM (e.g., [6]). Thus, 'a kind of' classical randomness in principle can be used to mimic quantum phenomena.
Our approach is based on the method of formalization, as developed in e.g., [8,9]. It relies on using models of axiomatic set theory (Zermelo Fraenkel set theory possibly with the axiom of choice-ZFC) to quantum mechanics and gravity especially for the overlapping domain of QM and 'classical' spacetime. The reason for using varying models of ZFC is the inherent structure of the lattice L of QM where the set-theoretical random forcing is generated by atomless Boolean algebras of projections (the measure algebras). We present the construction, along with building the BH model, in Section 3. It is shown there that local Boolean contexts in L are atomless Boolean algebras and that it leads to the non-trivial forcing (see also Appendix A). This random forcing governs the change of the model according to L α → L α [r], where r is certain random infinite binary sequence. Next, we follow the observation that the variety of models can play the role of additional internal degrees of freedom (DoF) in physical theories [8]. Applying these forcings and such generated DoF to the BH model built in Section 3, the rule of proportionality of entropy to the surface of the horizon of a BH has been derived in Section 3.1. The second main result of this paper also relies on forcing and ZFC models. Namely, the randomness of infinite binary sequences-r's (of possible outcomes of QM measurements added by forcing extensions to L α ) leads to the characterization of the inherent randomness assigned to BHs. This is our Theorem 1 which states that given any class 0 (n) , n = 0, 1, 2 . . . known from the algorithmic computability theory (see Appendix B), one finds that the relative to 0 (n) 1-randomness (i.e., in fact n-randomness) is realized by generic BHs. Thus, again, the source of the results above lies in the structure of the QM lattice L that governs the change of the ZFC models by random forcing. We expect that this very fact should survive, in this or another form, in a final version of QG. Such statement is a hypothesis; however, uncovering formalism substantiating, it can be considered as the third main result of the paper.
The work is organized as follows: in the next section, we will briefly review the ZFC models techniques and explain (following [10,11]) that mathematics of QM can be formulated in the minimal standard countable transitive model (CTM) of ZFC, L α . However, the statistical outcomes of QM are not L α -expressible. In the subsequent section, we will discuss the randomness and computational complexity from the perspective of L α and QM in it. Having developed the general setup for complexity in QM (based on random forcing and models of ZFC), we will apply it to derive BH horizon area law in Section 3.1 and to characterize the computational complexity of BH in Section 3.2 (Theorem 1). These results are based on the lattice L that supports the random forcing, and this fact is presented in Section 3 and Appendix A. More thorough and complete treatment of the relation of QM and randomness will be the topic of a separate publication.

Mathematical Formalism of QM in the Minimal Constructive Model of ZFC
While using mathematics in physics, its formal foundations are mostly of no concern and are usually understood intuitively. Actually, even most mathematicians in their research do not need to use directly methods of model theory. This situation makes the fact that QM formalism has deep connections with axiomatic set theory even more interesting.
Set is a basic concept of most of mathematics-even objects in physical theories like QM are formally sets, e.g., operators or Hilbert spaces. Set theory is the basis for analysis, algebra, topology, etc.-simply for mathematics as it is used in theoretical reasoning by physicists. It should be described axiomatically via methods of formal logic to avoid technical problems like Russell's paradox. The most common is Zermelo-Fraenkel set theory with the axiom of choice, abbreviated as ZFC, and it is recognized as foundation of (almost all of) mathematics. A model of ZFC is a mathematical structure M = (M, ) (an interpretation) that satisfies all axioms and derived in ZFC formulas. Model M is said to be standard when is the usual membership relation ∈ and M is called transitive if M is a transitive class (each of its elements is also its subclass). Referring to standard transitive models is typical and one should treat such references as technical in the next sections. The important thing to realize here is that each theorem of ZFC must be true in all its models. Hence, the usual mathematical relationships stated as theorems should not cause any problems when expressing formalism and postulates of QM within a model of ZFC.
There exist many non-isomorphic models of ZFC. An important concept of 'absoluteness' helps us to grasp some of the differences between them (it is related to 'seeing' formal objects as from inside or outside of a model). For example, a property ψ(x) = 'x is a real number' is absolute. This means that any set satisfying ψ does it whether we relate it to some standard transitive model or not. On the other hand, χ(x) = 'x is the set of real numbers' is not absolute. When we consider a collection of all sets satisfying ψ, then we will obtain the class R of all reals. We know that there exist (and are relevant for this paper) standard CTMs of ZFC. Let M = (M, ∈) be such a model. We can denote the set from this model that satisfies χ as R M . Since, from the outside, M is countable, then R M is countable too and hence it cannot be equivalent to R (in fact, the relation between the two-from the outside M perspective-can be stated as R M = R ∩ M). However, inside model M, we do not know about any 'missing' reals and R M is actually uncountable from the inside M point of view (so-called Skolem's paradox). Moreover, we have formal tools to extend a model of ZFC, among other things possibly adding some of the 'missing' reals (in non-trivial cases). The method for extending a ZFC model is called 'forcing' and there are various versions of it (differentiated also by what 'kind' of real numbers are added). When a standard CTM M is extended via appropriate forcing to some M[G] (extension by addition of so-called generic ultrafilter G), then their sets of reals are related in a similar manner, namely R M ⊂ R M [G] (when seen from the outside of M/inside of M[G]). Since the real numbers can be seen as infinite binary sequences-and these can be related to repetitions of a given quantum measurement-this fact is of special interest for us. Gödels's constructible universe (a particular model of ZFC) is a class of sets that can be fully described by simpler sets-starting with an empty set, we construct more and more complicated sets by applying basic set operations to the sets from previous stage. The construction is ordered by ordinal numbers-transitive sets whose all elements also are transitive sets (the finite ones are exactly natural numbers). Namely, for each β ∈ Ord (class of all ordinals), we obtain the stage in so-called constructible hierarchy where P(A) is a power set of A and I(A) is the smallest set containing A and closed under the following operations: creating ordered and unordered pairs, set difference, Cartesian product, performing permutation of an ordered triple, and taking the domain of a binary relation. The constructible universe is defined as the union of all L β , and it is a standard transitive model of ZFC. However, to obtain a minimal constructible mathematical universe to work in, we do not need to sum up all the stages. In particular, the countable set defined as L α = δ<α L δ , where α is the smallest ordinal for which there exists a standard CTM of ZFC that does not contain it, is the standard minimal CTM of ZFC.
For a comprehensive exposition along with the discussion on formal intricacies of all the topics mentioned above, we refer an interested reader to classical manuscripts, like [12]. Now, we can address the issue of expressing the mathematics of QM in L α . According to the discussion above, mathematical theorems of functional analysis, etc. that are used in QM calculations are true in any ZFC model and hence will not cause problems here. What is important-and needs additional commentary-if we want L α to be the carrier for mathematics of QM is the representation of postulates of QM. What interferes with relating QM to the model are non-absolute defining properties of basic structures, e.g., if H is a Hilbert space inside L α , it may not be the Hilbert space when seen from outside of it (or in another standard transitive model that contains L α as a submodel).
We should indicate here that, for further considerations, it is only needed that one can express in L α the appropriate mathematical objects from the postulates and that, after extending the model via forcing, one can appropriately relate them to their counterparts in the new model. The existence of sequences representing outcomes of (possibly infinite) repetitions of a given measurement (when seen from the outside of the model) is a separate problem (see the discussion on reals in models above) and by itself was discussed more extensively elsewhere (e.g., in [13]). To address absoluteness issues, one can proceed as follows. Following Benioff's technical assumptions, we will consider only observables that are questions (yes/no outcome of a measurement), but one can carry on analogically without this restriction. Inside the model L α , each quantum system is studied over a complex Hilbert space H L α and states are represented by density operators from the set B(H L α ) of bounded linear operators over H L α . We want these points to be true: (a) Outside the model L α , there exists a Hilbert space H and an isometric monomorphism U from (c) Let S and Q be collections of state preparation procedures and question measuring procedures, respectively. There exist maps f : S → B(H) and g : Q → B(H) such that, for any s ∈ S and q ∈ Q, the values f (s) and g(q) are respectively the density and projection operators (outside L α ). There also exist analogical maps f L α and g L α such that appropriate values are density and projection operators from B(H L α ) (inside the model L α ). Moreover, in the respective domains of f L α and g L α , the mean values of these projections in the set of equally prepared states by these density operators are equal outside L α , i.e., Formal constructions showing how one could always find appropriate maps and spaces from these points can be found in Benioff's original work (parts B, C, and D of Sec. III, [10]). All considerations were expressed generally as interplay between the perspective of L α and outside of this model, but one can proceed analogically in the case of any standard transitive model M of ZFC that extends the minimal one, L α ⊂ M. Note also that the remarks of this section regarding QM in L α might be applied basically to any standard transitive model of ZFC.

Models of ZFC and Computational Complexity of BH's State Space
In this section, we consider the situation where, during the formation of a BH in spacetime, as a result of the gravitational collapse of a star, the state where QM is in L α (as in the previous section) is attainable-and then we draw conclusions from such scenario. To realize this, we need to construct a mathematical model of BH based on the varying models of ZFC. While performing this task, we present step by step formal ingredients of a general mathematical setup. Thus, in points 1-5 below, together with Appendix A, elements of QM and operator algebras appear which intertwine with constructions from formal ZFC. Finally, point 5 explains how random forcing arises in QM from L which leads to a change of the ZFC model. This last observation opens the way toward algorithmic computational complexity in the context of BHs via random forcing extensions (see Theorem 1).
Let us consider the increasing density of energy in spacetime due to the gravitational collapse of a sufficiently massive body and corresponding increase of the curvature of spacetime as basic starting points. In principle, the corresponding expressions are unlimited and hence divergent. Consider a 4-submanifold of spacetime N ⊂ M 4 with increasing 4-curvature and let it be a part (say open submanifold) of 4-sphere S 4 with decreasing radius r, i.e., k ∼ 1/r 2 and N ⊂ S 4 ∩ M 4 , where k is the Ricci scalar curvature. General relativity (GR) tells us that any such region in spacetime can be covered by local, diffeomorphic to the flat R 4 , patches (with the Minkowski metric). When r → 0, the diameters of the patches are also decreasing to zero showing the singularity of GR in such limit. Let us instead assume the scale of density of energy ρ s and the radius r s such that, for each density ρ > ρ s and radius r < r s , the local patches of the manifold have finite and positive diameters. However, they are not any longer the local patches of a spacetime submanifold. In this way, we have dismounted a highly curved 4-submanifold into the set of local patches of finite diameters not exceeding δ > 0 and the submanifold N with the 4-curvature bounded from below (bigger than 1/r 2 s ). Then, the set of R 4 will take part in the quantum description while the submanifold N, with its smooth atlas of arbitrary small size local patches, will be associated with the still singular classical GR description.
To understand this via set theory, the crucial element is that, during the collapse, the model of ZFC we work in can be weakened or changed into another model of ZFC so that the model changes when BH is forming. More precisely, behind the horizon quantum description requires the CTM L α which differs from the model M before the horizon. We can think of M as a model of ZFC with a 'more complete' real line R M compared to R L α ⊂ R M . One might consider R M as R and hence R L α is its countable dense subset. There are few comments regarding this point. First, neither M nor L α can prove that they are models of ZFC nor that there exists a ZFC model at all. If any of them could, this would prove the consistency of ZFC, which is known to be unprovable in this way (Gödel's incompleteness results). In fact, we do not so far have any proof or disproof of the consistency of ZFC. Moreover, when L α is not a submodel of M (and certainly M is not a submodel of L α ), then M does not 'see' some sets of L α . This means that any observer using tools available in M has no (or at best limited) access to sets in L α . This situation represents set-theoretic counterpart of the horizon in physics of BHs which we make use of. Furthermore, given the two models as above on both sides of the horizon from the point of view of spacetime in which BH was formed, the model L α resides in a, bounded by horizon of the BH, singular region of spacetime, i.e., it is constructed in the bounded by two-dimensional horizon three-dimensional space (+internal time) such that the singularity is described by QM in L α . QM singularity and spacetime horizon are linked by random forcings generated by the lattice of projections on the infinite dimensional separable complex Hilbert space states (see [8]). More precisely, the set {U i } i∈I of local open frames (with finite diameters) spans both, the local structure of the horizon and the quantum state space of BH, and the set emerges from QM expressed in L α . This set replaces GR-singular open cover of the infinitely small (part of) spacetime manifold crushed to a singular region and develops horizon.
In the remaining part of this section, we will collect basic mathematical facts in favor of the above scenario and then draw general conclusions regarding BHs and complexity. 3. The importance of B ⊂ L comes from the fact that a way from deep quantum regime to the classical 2-valued window goes through such Boolean local context B. In particular, any reduction which is assigned to a measurement in QM also factors through such B: any projection to an eigenspace, containing an eigenstate of an observable a, belongs to certain MASA B from L, which is picked with probability given by Born's probability rule.  A well-known fact that quantum observables like position or momentum require infinite dimensional Hilbert space of states shows that the atomless Boolean algebras involved always emerge here where spacetime overlaps with QM in a non-trivial way. This is the case of the description of BHs presented in this paper.
5. As we have already noticed, B-being a Boolean context for the QM lattice L-is the unavoidable step toward reaching 2-valued stage of classical world. This can be implemented as a homomorphism h : B → 2 where 2 is the 2-valued classical Boolean algebra {0, 1, ∧, ∨}, but we require that the completeness of B is somehow maintained by such h. More precisely, h : B → 2 should be completely additive in a sense that for any subfamily S ⊂ B of B in L the following holds h( S) = {h(s)|s ∈ S} . (4) 6. Now, let us work with B supposing that QM is in L α as in Section 2. It follows that Bs are also internal algebras, i.e., measure algebras Bor([0, 1])/ N in L α and, as such, they are still atomless Boolean algebras in the model. Then, the homomorphism h : B → 2 carries very important information when considered internally to L α . First, one cannot fulfill the condition (4) in L α . The reason is the fact that, when S ⊂ B and B as above is in L α , then h −1 (1), 1 ∈ 2 = {0, 1} is so-called generic ultrafilter in B (in L α ), e.g., [8] However, it is known that, whenever B is atomless in L α , the generic ultrafilter U / ∈ L α [12]. This means that, for any measurement, if QM is in L α , then h (giving rise to classical results) does not exist in L α . A natural question that follows would be if there exists any model M of ZFC containing L α ⊂ M such that h ∈ M ? The answer is positive and there indeed exists such a minimal extension of L α , denoted L α [h], and it contains the homomorphism h. The model L α [h] is known in set theory as the random forcing extension of L α [12] and, in the context of this paper, it describes the measurement, along with its classical outcome, consistently. Thus, we have reached the scenario where QM world is explored by classical windows and the process factors through the set-theoretic forcing extensions of a certain model of ZFC. Now, we are ready to draw further conclusions regarding BHs.

BH Horizon
The following picture emerges: given the model L α describing the region containing the singularity of BH, one is directed toward the horizon via forcings extensions L α [h] as in the previous subsection. The family of such extensions can not be reduced to a single extension and the variety of them corresponds to the family of local coordinate frames defined on the horizon [8]. The important point of this approach is the effective reduction of the Hilbert space dimension from infinite to a certain finite value to be determined shortly. The reduction takes place due to the presence of the horizon which effectively bounds the dimensionality and also explains the statistical QM origins of entropy. The BH entropy originates in QM regime and is not determined by the entropy of matter and energy creating a BH in the collapse. It is rather the structural formal property of the presented model.
Forcing extensions starting in L α and ending at the horizon with {L α [h i ], i ∈ I}, correspond to the projections of 4-dimensional local coordinate frames onto two-dimensional ones on the horizon. Projecting on the horizon means leaving with the formal description by the i-th local extension L α [h i ] but also relate the purely QM regime with the classical geometry of the horizon. What is the role of the number of forcing extensions? It certainly relates to the number of local two-dimensional 'patches' of the horizon. Let the area of the horizon be A h . Each (geometrical) point x ∈ Hor of the horizon belongs to certain open two-dimensional domain U (2) x in L α [h x ] and this last one is the projection of a 4-dimensional domain U x -if the diameters {δ x |x ∈ Hor} of each U (2) x are such that the minimal length l bounds them from below, i.e., ∃ l>0 ∀ x∈Hor δ x > l (this agrees with the discussion at the beginning of Section 3). Thus, the minimal number N of local domains generated by QM in L α is finite where J min. is finite since the area A h is finite. Thus, we have minimally N classical patches covering the horizon that come from a kind of classical reduction of quantum superposed states-on the QM side of L α , there are superposed states that undergo reduction to classical patches. The minimal such superpositions are given by bipartite states which makes this similar to the case of N qubits. Then, we ask about the entire dimensionality of a Hilbert space allowing for such generalized quantum computing processes that would be the minimal dimension of the Hilbert space in our model of BH. Thus, supposing that each horizon's patch emerges from the reduction of a superposed bipartite state, we can estimate the minimal dimensionality of the Hilbert space of states H BH of BH (e.g., [16]) as Hence, the minimal state space of BH is spanned on 2 N independent states. We can also approximate the entropy of BH in this approach. The entropy reads S BH ∼ log 2 2 N = N and thus taking the proportionality coefficient A it is S BH = A · N. In terms of the BH horizon area A h , we obtain which is the usual Bekenstein-Hawking formula up to the coefficient A. Within these general considerations, we do not determine it. One could wonder how it is possible to have 4-dimensional regions bounded by a two-dimensional horizon. The answer is related to the expressible power of models of ZFC and is also responsible for BH complexity as we will see in the next subsection. In any ZFC model, one can define not only any space, like R n , n ∈ N, but also any manifold or even any set of arbitrarily high cardinality, not only ω, ω 1 , but even higher. Given the model confined in the behind horizon domain, we have formally all objects definable in it. This special feature is the consequence of that that a model of ZFC changes when crossing the horizon. Thus, the model confined to the behind horizon region is different than the outside model. QM and GR are expressible in the models. More generally, everything that can be proved formally in ZFC remains provable literally in the same way in every model. There are, however, specific sentences independent on the axioms which still can be proved in one forcing extension and its negation in another, and they can make difference between the extended models. This phenomenon is related to randomness in QM-we do not investigate this topic here any further (for more details see [17]), but we do apply it when analyzing computational complexity below. On the set-theoretical side, this is connected with relative consistency and provability by forcings in the extended models. It is fair to say that there is a great variety of forcing procedures known for mathematicians but precisely one-the so-called Solovay forcing, which is responsible for randomness, is realized in QM [8,18].

Computational Complexity and BHs
Thus, behind the horizon, the model L α is changing toward the family L α [h i ], i ∈ I, and this generates the local descriptions of the horizon. This process can also be responsible for the computational complexity in the context of BHs. Randomness and computational algorithmic complexity are in general very wide domains of intense mathematical enquiries, e.g., [4,19,20], and, to manage it on the reasonable number of pages, we develop here the main line of argumentation while leaving more formal definitions and properties to Appendix B. However, for more complete and systematic picture, we refer readers to excellent monographs as above.
The phenomena of random behaviour, randomness, and related computational complexity can be expressed mathematically by properties of infinite binary sequences. Due to use of ZFC models, this presentation is directly applicable to our model of BHs. To better understand this, let us take closer look at certain aspects of randomness of finite and infinite sequences and their relation to the computational complexity. The sequences can model the outcomes of the series of QM measurements, but this is not mandatory, which means that, even in the formalism of QM, there is a room for such random behavior without interpreting the infinite sequences as outcomes. The essential examples are given by the completely additive homomorphisms h i which are modeled by a random infinite sequences, i.e., those which are added by forcings as in L α [h i ] (see [5]). These infinite sequences are represented formally by random real numbers which are random binary {0, 1} sequences (e.g., ([12], p. 243), [18]). The sequences are random with respect to L α since they fulfill the following defining property [RAND] of randomness [18] [RAND] A binary infinite sequence r ∈ R M[r] is random with respect to the model M if r omits all Borel sets of measure zero coded in M.
Now we see that, due to expressing QM in L α and taking it as M above, we are able to confine randomness of infinite sequences within the behind horizon region of BH. Next, we are going to express the computational complexity intrinsically characterizing BHs by relating it to randomness as above. The interplay between randomness and computational complexity is a vast subject by itself (see, e.g., [4,19]) and here we only touch a few aspects relevant in the BH context. First, let us approach randomness of infinite binary sequences (reals) from the Cantor space 2 ω via Turing computability (as we noted above, these reals correspond to h i 's in our description of BHs). In Appendix B, some useful facts are presented about Turing classes and arithmetical classes of sets needed here.
Randomness of infinite sets and binary sequences can be defined in various ways. However, four main attitudes exist: patternless structures (the measure-theoretic paradigm), incompressibility property (the computational paradigm), unpredictability and genericity-that appear to be equivalent to the so-called 1-random case (except genericity). Even though they diverge in higher random classes, there are still plenty of cross properties-that have been developed within the years-that characterize one of them by the others [4,6,19]. Genericity refers to the relation of forcing notions to randomness, and we have distinguished it as a valid approach to randomness. In fact, generic sets are frequently and deeply present in all other three classes above but also develops a specific field in which complexity of BHs is naturally expressible. Our generic class of random phenomena contains in particular Solovay (random) forcing but also Cohen forcing (and some others). Cohen forcing usually is understood when world genericity appears in random phenomena [4]. The great exposition of genericity in this wider context (containing also Solovay forcing) can be found in [21].
The [RAND] property above is the example of patternlessness within the genericity class (Solovay forcing). The definition by Martin-Löf (see Appendix B), Solovay, or Schnorr are instances of patternlessness and are based on the omitting of the measure zero sets from suitable computability class. Depending on the computational strength of the class of sets of measure zero, one obtains corresponding strength of randomness. It is complemented by the requirement that random sequences belong to all full measure 1 unspecific sets.
The computational complexity of strings refers to incompressibility that we will discuss briefly below. There exist functions that measure complexity of binary sequences-aiming at expressing their randomness. One of the historically first such objects are plain Kolmogorov complexity C and Kolmogorov complexity K. Let σ ∈ 2 <ω , i.e., σ is a finite binary string, and let f : 2 <ω → 2 <ω be a partial computable function. Then, the Kolmogorov complexity of σ with respect to f reads Then, the finite sequence σ is random relative to f when C f (σ) ≥ σ (one cannot perform any algorithm of 'length' related to f shorter than σ to describe or compute σ). The dependence on f is unnecessary since there exists the universal partial computable function U such that, for each f as above, there exists a finite binary sequence ν f that f (σ) = U(ν f σ) for all partial computable f : 2 <ω → 2 <ω .
To deal with infinite sequences and study their algorithmic randomness, one defines so-called prefix-free Turing machines. A prefix-free subset A ⊂ 2 <ω contains such σ ∈ 2 <ω that any proper extension τ of σ does not belong to A. A prefix-free function f : 2 <ω → 2 <ω is such function with prefix-free domain. A prefix free Turing machine is T with domain (a set of sequences on which T operates) being a prefix-free subset of 2 <ω . A function U : 2 <ω → 2 <ω is a universal partial computable prefix-free function if, for each partial computable prefix-free function f : 2 <ω → 2 <ω , there exists a string ν f such that f (σ) = U(σν f ) for all σ ∈ 2 <ω .
Then, fixing U, we can define the Kolmogorov's prefix-free complexity of σ, K(σ) as The definition of n-random sequences σ, given in Appendix B, refers directly to the arithmetical complexity classes ∑ 0 n (see Appendix B) which means that randomness is deeply involved in computational complexity (and conversely). Below, one finds some examples of how it can go.
(c) A set A is ∞-often C-random if there are infinitely many n such that C(A n) ≥ n − O(1).

Then, it holds that
A is 2-random ⇐⇒ A is ∞-often C-random.
(d) A is 1-random ⇐⇒ ∑ n∈N 2 n−K(A n) < ∞ (and there are many other results of this kind about randomness and computational complexity).
Our concern now is to find a degree of algorithmic computational complexity assigned to BHs as formed in spacetime. We already noticed that random forcing (Solovay) is apparently present as the building block of the presented model of BHs (see [RAND] above). This is why we are going to analyze the complexity in the model via generic randomness (e.g., [21]). The crucial observation is the following result: Theorem 1. The computational complexity realized in the forcing model of BH based on L α is at least that of 1-random sequence relative to 0 (n) for all n ∈ N.
It remains to note that the [RAND] property means Solovay ω-genericity (relative to L α ) holding for all n ∈ N and taking place for BHs in our model. From Theorem 2, weak n-randomness holds for all n ≥ 1 and, from Theorem 3, one infers that n-randomness for all n ≥ 1 is attained in the presented model of BH. However, from the property (a) above, the statement of the theorem holds true.
Note that the arithmetical hierarchy (see Appendix B), on which the above results are based, comprises sets which are definable without parameters in the fragment of ZFC namely in hereditary finite model of sets, (HF, ∈), where the axiom of infinity does not hold ( [12], p. 480). L α is a model of all axioms of ZFC, and this shows that the results are independent in this respect on the model L α . The thorough discussion of this point will be performed in the separate publication.

Discussion
We presented a model of BH complexity based on classical notions of computational complexity. However, we allowed dynamical pattern of models of ZFC such that the working universe of sets undergoes sudden change while crossing the horizon to another model of ZFC. In this respect, this is the continuation of our previous recent works [8,9]. Physically, the change of a model depends on the scale of density of energy and we do not specify precisely the values but expect they should be found in the Planckian scales. Instead, we tested the entire possibility of the model-changing in the context of BHs. Formally, working in certain model of ZFC is advantageous since we can have a simplified universe where all theorems of ZFC remain true and the properties like randomness are efficiently defined with respect to the universe. There are in general plenty of possibilities for the choice of a model describing the singular interior of BH; however, one is particularly distinguished. This is the countable minimal model L α which has been shown by Benioff [10,11] to be strong enough to carry all the mathematics of QM and, at the same time, randomness of the sequences of outcomes of QM measurements should be found outside. We showed that they are in the random forcing extensions L α [h i ]'s and the derivation was based on the formal structure of QM.
Applying these findings to BHs and, assuming the change of the set universe to L α containing the singularity, we also tested (heuristically) the model against the Beckenstein-Hawking entropy relation and found that it is retrieved in the model up to a (undetermined so far) constant. Next, we analyzed the computational complexity realized in this model of BHs. The recent approach by Susskind et al. (e.g., [1,3]) estimates the computational complexity of the state spaces of BHs via quantum computing circuits and the measure of complexity are related to the entropy of BHs. We found the complexity based on the classical concepts of computing and randomness (see Section 3.2 and Theorem 1). However, the two approaches cannot be disjoint. The reason is that any conceivable construction of even the weakest 1-random infinite sequence requires QM and cannot be realized within the deterministic classical realm (e.g., [6]). This is why-when we showed that BHs in our approach require n-randomness classes, n ≥ 1, by 'classical' means-the quantum realm comes back through the back doors as it should be when one thinks about BHs at the fundamental level. There exists a vast amount of literature regarding randomness and QM (see, e.g., [5]). However, the proposed by us, ZFC countable models point of view seems to be a promising alternative. It is worth emphasizing that a ZFC-based approach to QM has been already present in physics before (e.g., [10,11,22,23]) and-in a sense-our research continues these works.
Much remains to be done. Every fundamental version of QG should reproduce finiteness of the entropy and 1/4 factor in the Beckenstein-Hawking law (B-H). For example, loop quantum gravity confirms finiteness and proportionality of entropy to the area of the horizon [24] and superstring theory supports the B-H law for some class of BHs, e.g., [25]. When we expect that the presented approach aims to be fundamental in some respect for QG (forcing extensions from the lattice L should survive in a successful version of QG), we should address the B-H precise formula in the model of BH explored here. We do not know how to achieve it now, though certain attempts are being performed. The relation of randomness and QM understood from the point of view of varying models of set theory can be pushed further-in particular toward better understanding of the measurement problem in QM and resulting randomness envisioned by the Born's rule. Next, we included the recognition of the regime of quantum gravity based on the dynamical ZFC environment and the fundamental role assigned to gravitons and their ZFC counterparts. In addition, the points closer to those not resolved in this paper, like whether we can be sure that QM is in the model of ZFC-is this mandatory or just a formal possibility? What else in physics emerges from such eventual necessity (one would be the presented model of BHs)? The varying models of ZFC extend our toolkit toward studying the overlapping region of QM and classical spacetime touched here (and more extensively in [8])-the approach should be developed further. Connected with this is the modification of the smooth spacetime structure due to such overlapping with QM which also was signaled in [8,9]. The issue of independence of the results in the previous section on the choice of a model of ZFC and the role of infinite dimensional Hilbert space of states in such fundamental approaches should be further explained. Finally, is QM random, 1-random, 2-random, or ω-random in a sense that it could be tested experimentally? Infinite random binary sequences are (or are not) random but, even mathematically, to prove that a single such sequence is random belongs to formally undecidable problems--let alone how to recognize from finite parts of random sequences their randomness as an infinite ones by experimental means. The experimental grasping, along with theoretical understanding, of these kinds of problems within QM is important for designing and manipulating the true random sequences in practice (e.g., [26]). Most of the issues raised above will be addressed in our forthcoming publication [17].
Then, one relates essentially arithmetical classes with the Turing computational classes via jump operator as is given by the Post's results A ∈ ∑ 0 n ⇐⇒ A = C-c.e. where C is a set in ∑ 0 n−1 ∏ 0 n−1 A ∈ ∑ 0 n ⇐⇒ A = 0 (n−1) -c.e. (A is computably enumerable from the set 0 (n−1) ).

(A4)
Arithmetical and computational classes and jump operators are also important tools characterizing higher classes of randomness of sets and sequences.
Given an infinite binary sequence σ ∈ 2 ω and µ the product measure on (the Cantor space) 2 ω , the randomness of σ can be defined as (Martin-Löf e.g., [4]): 1. ML test: A sequence {A n , n ∈ N} of uniformly c.e. (i.e., together with the set of its indices ( [4], p.11)) of ∑ 0 1 classes (∑ 0 1 subsets of sequences from 2 ω ) such that µ(A n ) < 2 −n , ∀ n∈N . 2. A ⊂ 2 ω is ML-null when there exists an ML test {A n , n ∈ N} such that A ⊆ n∈N A n . 3. σ ∈ 2 ω is ML-random if {σ} is not ML-null (for each ML test). 4. A ML test {A n , n ∈ N} is universal when n∈N B n ⊂ n∈N A n for all ML-tests {B n , n ∈ N}. Lemma A2. There exists a universal ML test.
ML-random sequence σ ∈ 2 ω is known to be 1-random. The following modification of the above definition of ML randomness explains the hierarchy of n-random sets, for all n ≥ 1.
i. ML n test: A sequence {A k , k ∈ N} of uniformly c.e. of ∑ 0 n classes (∑ 0 n subsets of sequences from 2 ω ) such that µ(A k ) < 2 −k , ∀ k∈N .
ii. A ⊂ 2 ω is ML n -null when there exists a ML n test {A k , k ∈ N} such that A ⊆ k∈N A k .
iii. σ ∈ 2 ω is n-random if {σ} is not ML n -null (for each ML n test).