1. Introduction
Time is one of the deepest-rooted experiences in adult human beings. As such, it has been a fundamental object of study in philosophy and physics from the antiquity to the present. Naively, we can try to define
1 time in a minimalist way as a concept that enables an observer to describe and measure the perceived physical fact that “the world changes”. However, in this definition there is an echo of circularity as for most, if not all, definitions of time.
In Newtonian physics, time is an absolute, primary concept. However, relativity theory demolished the notions of absolute time and absolute simultaneity, and, as a consequence, the flow of time, which shapes the first-person experience of an observer, appeared as an illusion. This paved the way to explore the possibility that time itself were not a primordial concept, but a derived one, and contributed to the impulse of mathematizing and constructing physics [
4,
5,
6,
7,
8] in terms of more abstract primary concepts
2.
Quantum mechanics went beyond relativity concerning the difficulties in grasping the observer-independent reality of nature. The Copenhagen interpretation of quantum mechanics highlighted the absence of meaning in describing a system without referring to the results of a measuring apparatus. Thus, the reality of a system, i.e., the definite value of the intrinsic properties that characterized such a system, could be affirmed only in connection with the actually performed observations by a classical meter.
Quantum mechanics also raised doubts about the validity of the classical principle of causality, i.e., that every event has an antecedent determined cause, and pointed at a possible intrinsic randomness as a fundamental feature of nature [
10].
In addition, there was a central tension between quantum mechanics and relativity about the principle of locality
3, i.e., the existence of a limiting velocity (the speed of light) for the propagation of an interaction through space, as was especially manifest in the experimental observations of non-classical correlations between entangled particles or Bell inequality violations [
15].
Time seems to play a crucial role for the coexistence of both theories. For instance, backward in time causation [
16,
17,
18] could explain Bell inequality violations preserving locality (see also Refs. [
19,
20]). Even in certain circumstances, quantum mechanics can alleviate some of the problems that the theory of relativity brings about regarding the nature of time. One of these cases is the solution to the difficulties and mysteries in the concept of time supplied by assuming the many-worlds interpretation (MWI) of quantum mechanics [
21] as analyzed by Deutsch [
22]. Another example is the resolution in the framework of quantum mechanics of the so-called
grandfather paradox that appears in general relativity in those regions of spacetime for which closed timelike curves (CTCs) are possible mathematical solutions. The
grandfather paradox reflects the logical inconsistency of those closed trajectories in which a time traveler could kill his own grandfather, therefore making impossible his own existence. The resolution of the paradox in a classical framework [
23] is generally considered unsatisfactory [
24,
25] because it essentially requires suppressing by fiat the initial conditions of trajectories driving to the inconsistent solutions (enunciating a global principle of consistency that would prohibit the particular local solutions leading to inconsistency). However, a quantum model devised by Deutsch in order to analyze the closed timelike curves (D-CTCs) in terms of quantum information flows guarantees the existence of a consistent solution for all initial conditions.
The interest of studying CTCs in the framework of quantum mechanics resides not only in the solution of paradoxical results in this specific milieu, but also in the analysis of the interplay between relativity theory and quantum mechanics, and the deep implications on the nature, foundations, and coexistence of both theories.
The real, physical existence of CTCs is still a central, open question. However, simulations of CTCs, and also open timelike curves (OTCs), on photonic circuits have already been performed on the laboratory [
26,
27].
A set of theoretical studies have also been developed on the special properties that quantum mechanical systems would present near a D-CTC
4. One of the most astounding results, obtained by Aaronson and Watrous [
30], is the computational equivalence of a classical and a quantum computer in the presence of a CTC. This result highlights the adequacy of the CTC scenario as testing ground to study the relationship between quantumness and classicality.
Mirroring the successful scheme in relativity, a traditional area of study in the foundations of quantum mechanics is the search of fundamental principles from which quantum mechanics might be derived. Different inspiring approaches have been considered, e.g., that of Zeilinger [
31] in which information plays a central role. In spite of the tremendous efforts of leading experts in quantum foundations, complete, satisfactory derivation of quantum mechanics has not been achieved yet. Therefore, it remains an open question in the foundations of quantum mechanics. What will be the final solution of this problem (if it can be found at all)? For the moment, it is difficult to forecast. Therefore, a variety of approaches should be tested. We proceed within the information framework unifying a batch of models (see, e.g., [
32,
33,
34,
35,
36]).
In the present article, the possibility of simulating D-CTCs in the framework of an information-theoretic, Darwinian approach to quantum mechanics (DAQM) [
37,
38,
39,
40,
41,
42] and its implication for the progress in the understanding of the foundations of quantum mechanics are analyzed. There are two specific concepts in this approach that play a central role in the theory. First, time is assumed to be, all the way down, a real fundamental magnitude in nature, in the sense that accepting the reality of time in its deepest meaning implies the possible change of the physical laws of nature, as extensively analyzed by Unger and Smolin [
9].
The second specific, crucial concept in DAQM is anticipation. If the previous anticipation by an agent of a future event or algorithm outcome actually coincides with such event or outcome in the future, then from the point of view of information, this situation might be operationally equated to backward in time causation. Anticipation is built on the capacity of a system to process information. This capacity is assumed in DAQM as a fundamental property of matter. If information is central to physics, then it is natural to explore the possibility that matter possesses the capacity of processing information. Supplementing every physical system with an information space in which a methodological probabilistic classical Turing machine processes the information that arrives at the system turns physical systems into information-theoretic Darwinian systems, susceptible of variation, selection, and retention [
43], on which natural selection might act as a metalaw, driving the evolution of natural laws or regularities—the algorithms stored on the information space of every physical system that govern their behavior—and plausibly, as will be analyzed within the article, generating quantum mechanical behavior as result of an optimized information strategy that might determine the direction of evolution in the long term.
Darwinism has been successfully applied to a long list of disciplines by several overarching schemes as universal Darwinism [
44] or generalized Darwinism [
43]. Darwinian processes have also been identified in physics, describing adequately intricate problems ranging from cosmology [
45] to the emergence of classical states from quantum interactions [
46]. This ubiquity of the Darwinian mechanisms is a reason in favor of examining the implications of DAQM.
The aim of this article is to explore the possibility of inducing quantum behavior for a physical system endowed with a probabilistic classical Turing machine through the simulation of a D-CTC on its information space, exploiting the theoretical equivalence between a classical and a quantum computer both with access to a D-CTC that has been demonstrated by Aaronson and Watrous [
30]. In
Section 2, a brief description of the D-CTC model is outlined. The characterization of a fundamental physical system in DAQM is analyzed in
Section 3. The simulation of a D-CTC on the information space of a physical system in the framework of DAQM is studied in
Section 4. The plausible emergence of quantum mechanical behavior from a sketchy mathematization of the dynamics of fundamental physical systems modelled as information-theoretic Darwinian systems in the framework of DAQM is discussed in
Section 5. The conclusions are drawn in
Section 6.
Finally, notice that, in the present study, the existence of real CTCs in the physical spacetime is not discussed, but only the implications of the possible simulation of a CTC in the information space of a DAQM particle.
2. Deutsch Quantum Model for a CTC
The conventional studies of CTCs are framed in the geometrical analysis of these solutions of Einstein’s field equations of general relativity for classical systems traveling along the spacetime [
47]. Deutsch [
24], adopting a different perspective, disregards the dynamics of the motion in the spacetime, assuming that the trajectories are classical and given, and considers traveling systems endowed with internal quantum mechanical degrees of freedom. Then, the physics of CTCs is analyzed in terms of the information flows within a physically equivalent quantum computational circuit that models the interaction process. It is assumed that the interactions between systems are localized in the quantum gates. Therefore, the states of the systems stay unchanged between gates.
One of the meaningful elements suggesting the exploration of a quantum mechanical description as a possibility to solve CTCs paradoxes is that, by considering quantum mechanical systems, the space of states of the traveling systems is enhanced, including linear superpositions and mixtures [
48]. Deutsch [
24] then proceeds to demonstrate (Deutsch’s fixed point theorem) that in quantum mechanics is always possible to find a consistent solution for any possible initial condition in the presence of a CTC. In this way, the D-CTC model overcomes the unsatisfactory limitation imposed by the classical resolutions (e.g., see Lewis [
23] and Novikov [
49]) of the
grandfather paradox in which the way out from the paradox basically consists in claiming a philosophical global consistency criterion that impedes by fiat the initial conditions that, although allowed by local physical laws, would conduct to physical contradictions.
The quantum circuit that represents in the model of Deutsch for a CTC the process for a particle that approaches the CTC is drawn in
Figure 1. Let us assume that the particle is a qubit. The particle that enters the CTC (see
Figure 1a) in the quantum state ρ(1) interacts in the quantum gate
U with an older version of itself that has come out of the past mouth of the wormhole in the state ρ(3). After the interaction, the younger version of the particle is in the state ρ(2) and enters the future mouth of the wormhole as the older version of the particle in the state ρ(4) leaves the CTC region traveling towards the unambiguous future. The process can be described in a “pseudo-time” narrative
5 [
25] from the intrinsic perspective of the particle following the timelike curve for which the proper time of the particle is increasing [
47], with the number between parentheses for every state indicating the proper time order along the line. The model requires as consistent condition that the state of the younger system leaving the gate, ρ(2), be the same as the state of the older system entering the gate, ρ(3). That is to say, ρ(2) = ρ(3).
The circuit represented in the
Figure 1a can be conveniently transformed into a simpler denotationally equivalent circuit [
24], as shown
Figure 1b, for which the outputs of the transformed circuit are the same function of the inputs as in the original circuit, as defined by Deutsch [
24], by including a
gate after the original gate
, so that the interaction in the transformed circuit is given by
. In this way the one-particle circuit is replaced with an equivalent two-particle circuit [
26,
50] constituted by a chronology-respecting particle (CR) that interacts with another particle confined in a closed timelike curve (CTC). Now, the consistent condition in this equivalent circuit, shown in
Figure 1b, is obtained by requiring that the quantum state (
) of the chronology-violating particle that emerges from the past mouth of the wormhole in the closed timelike curve be the same as the state that enters the future mouth (
):
where it has been considered that the chronology-respecting particle enters the gate in the pure state
(i.e.,
) and the partial trace is over the Hilbert space of the chronology-respecting (CR) particle. Deutsch [
24] showed (fixed-point theorem) that in the framework of quantum mechanics for a general interaction
there is always at least a solution
of Equation (1) for any initial state
. This solves the
grandfather paradox. An intriguing question of the model is the way in which nature has to manage in order to find the solution to Equation (1). This problem is usually named the
knowledge paradox, since the solution to Equation (1) must be known at the same time as the state
enters the gate, i.e., before the interaction occurs. Deutsch [
24] gives an answer to the
knowledge paradox6 in the framework of the many-worlds interpretation (MWI) of quantum mechanics. In our model, in the framework of DAQM, the
knowledge paradox would be solved through the action of Darwinian evolution. This point is discussed in
Section 5.
On the other hand, the state of the chronology-respecting particle that abandons the CTC region is given by the following expression:
where the partial trace now refers to the Hilbert space of the particle trapped in the closed timelike curve.
The dependence of the output state
on the input state
is non-unitary, since mixed states can occur on the CTC as solutions (
) of Equation (1), and nonlinear. This nonlinearity induced by the self-consistent condition (Equation (1)) is the source [
30] of the stunning computational efficiency that enables a classical computer in the presence of a CTC to reach the same computational performance as a quantum computer with access to a CTC.
4. Simulation of a CTC on the Information Space of a Particle in DAQM
Let us consider the information space of a particle
in DAQM. According to the model, the new, refreshed information about the surrounding systems that has arrived at the particle transported by the information carriers, which also convey momentum and energy, has been coded by the algorithm
into a probability distribution function
9 on the surrounding particles phase space. This function reflects the epistemic state of the particle about the locations occupied by the surrounding systems and their dynamics at time
. The wave function
10 of the particle can be constructed from the probability distribution function
by the algorithm
. That probability distribution function
enters the classical network
, which is schematically represented in
Figure 3, interacting with the second input
that has come out from the anticipation module
of the particle. After the interaction, the output of the first channel is
, the anticipated probability distribution function of the surrounding systems at time
, and the output of the second channel is
which must equal
11 the input
.
Note that is in general a statistical mixture. Therefore, the interaction with , the input in the first channel of the circuit , may cause modifications in different components of the mixture, provided that the modifications statistically compensate each other so that the mixture itself comes out unchanged from the second channel of the circuit.
As will be discussed in the next section, the anticipation module is a particular contextual solution that only works properly at the concrete system in which it has been developed and in the specific environment in which the system is immersed.
The program , the anticipation module and the network are all of them classical elements that perform their task on a classical Turing machine as initially assumed. This means that, first, the program contains and applies the principles of quantum mechanics but on a classical processor, second, the classical anticipation module is able to find out the function that stays stationary after coming out from the second channel of the classical network (i.e., it is able to find out the fixed point for the second channel of the network) and, finally, this circuit correctly calculates as output of the first channel the function that allows the program to compute (on other module) the wave function of the system at time .
Notice that all the elements and functions in the circuit of
Figure 3 are classical. Therefore, it seems that this classical circuit would not be able to simulate a D-CTC as represented in
Figure 2, given that a quantum circuit with quantum states is apparently required. However, as shown by Tolksdorf and Verch [
58] (see also Aaronson and Watrous [
30]), the D-CTC model can be implemented on a classical network acting on classical statistical mixtures, since in order to ensure the existence of a fixed point for a map (the core of the model), the key mathematical properties are of statistical nature. In particular, the state space must be convex and complete [
58]. In consequence, exclusive quantum properties are not necessary to carry out a D-CTC circuit.
Therefore, the classical circuit shown in
Figure 3 could simulate a D-CTC on the information space of particle
, provided that a procedure to obtain the objects
,
and
, whose characteristics have been defined in this section, be specified. However, if the classical Turing machine has access to a D-CTC, then according to the result shown by Aaronson and Watrous [
30], this classical information processor would be computationally equivalent to a quantum computer and might induce in real time quantum behavior on the particle.
5. Generation of the Anticipation Module and the Program in the Information Space of a Particle in DAQM
The characterization of a fundamental particle in DAQM [
37,
38,
39,
40,
41,
42] potentially endows these particles with the defining properties of information-theoretic Darwinian systems whose populations are then susceptible of evolution under natural selection
12. It is assumed that at time
there is a distribution of particles that are exclusively governed by their respective randomizers. As time goes on, algorithms that progressively take control of the particles’ behavior are randomly developed as a consequence of the arrival of information at every particle. Different procedures
13 on the information space of particles mimicking those encountered in biological genetic systems might enable the variation of the control algorithms and their posterior selection and retention or inheritance through the populations dynamics of DAQM particles in physical space.
The main mechanism of variation is associated with the random errors in the write/read operations in the Turing machine. Then, those populations of particles with variations that induce better adapted traits have more chances to prevail by the action of natural selection. The variations are stored in the memory of the Turing machine.
DAQM particles could disappear through two basic processes satisfying the conservation of energy: emitting all their energy or being absorbed by another particle. These processes would be extremely frequent when the algorithms controlling the particles were still simple and prequantum. Relics of this prequantum world might still exist in certain regions of the universe. The detection of these relics would be a possible experimental test of DAQM.
Unequivocally, fitness is the central magnitude in evolution. However, an open question in the Darwinian evolutionary theory is whether there is a direction for evolution, i.e., whether evolution is predictable in the long run. The successive mathematizations of evolution (Fisher’s synthesis [
59], Price equation [
60], evolutionary game theory [
61], Chaitin’s algorithmic information theory applied to evolution [
62], citing some of the main landmarks) have proven crucial to progress in the understanding of the concepts and processes of the theory.
In addition, different hypotheses (maximum entropy production principle [
63], free energy principle [
64], increasing complexity [
65], etc.) have been proposed in order to gain insight into the main mechanisms that would characterize Darwinian evolution in the long run (and also into the processes that would drive to the emergence of life [
66]).
In a broad sense, complexity seems to occupy a strategic place in the toolkit of long-term evolution. However, there is no general consensus [
65,
67] on the definition of complexity
14 in evolution. The challenge, therefore, is to find out a characterization of complexity that captures the essence of the underlying mechanisms of evolution in the long term. In addition, this concept should be general enough to encompass all systems in all environments, but at the same time specific enough to have predictive potentiality. Therefore, measures of the detailed structural and functional complexity of a system should better be discarded, since these characteristics of a system are deeply connected to the properties of the niche or environment in which the system is immersed and to the particular role played by the system in such environment.
Information seems to play a central part in the appearance of life. Perhaps the informational aspect of life is the key property [
68]. Walker and Davies [
68] describe the emergence of life as a kind of physical phase transition in which algorithmic information would gain context-dependent causal efficacy over the matter in which the information was stored [
68]. According to Walker and Davies [
68], the crucial magnitude would not just be the amount of information, as measured for example by Shannon’s information, but a quantity measuring that type of contextual causal efficacy possessed by biological information [
68].
Among the multiple variations in the representation of complexity in biology, there is a contextual definition of “physical complexity”, as it is named by Adami et al. [
67], that characterizes the genomic complexity of a biological system in terms of the information about the environment that is stored in the system for a fixed environment. It is then demonstrated [
67] that, assuming this definition, according to the performed simulations, complexity must always increase.
In the present article another contextual definition, which is named “survival information complexity”, is introduced. One of the most characteristic features of complex systems is the difficulty in predicting [
69] their future behavior. The idea of the new definition is to assign a high degree of complexity to a system that both is able to compute reliably the future behavior of its surrounding systems and, in turn, its outflow of information is minimized (this second characteristic implies that, for the surrounding systems, it would be difficult to predict the behavior of the system). The essentials of the definition are then summarized in that informationally the optimal strategy of any system at any environment would be maximizing the anticipation capacities of the system and minimizing the information sent outside by the system. Thus, survival information complexity measures the capability of a system for optimizing its information flows against survival. This optimization of the information flows for survival would plausibly be a common satisfactory strategy for all systems independently of their specific structures, functions, characteristics or environments. Therefore, if this definition of complexity admits quantification, then it seems an interesting candidate to explore the predictability of evolution in the long run.
Survival information complexity defined in this way does not automatically ensure a higher fitness in the short term. Other specific structural and functional traits or particular strategies may turn out more advantageous for survival in a concrete environment. However, in the long run, assuming an evolving environment, those populations that are informationally more complex, in the sense of the definition we have adopted, would have, in the end, a higher probability of adapting to the changing environment, therefore increasing their survival expectations, and, consequently, suggesting to identify the increase of “survival information complexity” as one of the main properties that would determine the direction of evolution in the long run.
Therefore, a system with information processing capabilities is then considered to be informationally more complex when, on the one side, it is able to anticipate the positions of its surrounding systems with greater precision, and, on the other side, it is also able to minimize the outflow of information.
Let us then define mathematically the survival information complexity
of a DAQM system
whose location is
at time
by introducing two terms:
The first one,
, measuring the capacity at time
of system
for anticipating the behaviour of the surrounding systems at a future time
is:
where
is the number of systems that surround system
;
(
) are the positions that will really be occupied at time
by the surrounding systems of system
;
is the position that will be occupied at time
by the surrounding system
of system
as calculated by system
at time
on the Turing machine of its information space; and, finally,
is the squared distance between the positions that will be actually occupied by system
and system
at time
(see
Figure 4 for a visualization of an example for the definitions of real and calculated systems locations).
Notice that in DAQM the location of a system is the physical property that is defined at any time, given that the trajectory is continuous.
In the definition of survival information complexity
, the second term,
, evaluates the degree of optimization of the system’s information outflow by measuring the capacity of stealth of the system
, i.e., the lack of precision of the anticipation of its position by its surrounding systems at time
in a certain environment:
where
, as for the first term, is the number of systems that surround system
, with
being the position that will be occupied by system
at time
;
is again the squared distance between the positions that will be actually occupied by system
and system
at time
; and, finally,
is the position that will be occupied at time
by system
as calculated by system
at time
on the Turing machine of the information space of system
.
More sophisticated definitions of may be envisioned (e.g., including the integration over a meaningful interval of time or introducing a kind of anticipation depth or considering the response time of the information processor of the system) and certain technicalities should be further discussed (e.g., avoiding singularities by defining a minimum value for denominators; discussing the inaccessibility of certain magnitudes for system at time , what implies a delayed calculation of the quantity; the inherent difficulty of calculating the survival information complexity, except for models), but the given definition captures the central elements of the concept.
This mathematical characterization of survival information complexity measures the capability of a system for anticipating the behavior of its environment and for dynamically minimizing the options of its surrounding systems to foresee its location. This magnitude then takes higher values for those systems that predict more accurately the dynamics of their surrounding systems and whose own behavior, on the contrary, is more difficult to be predicted by their surrounding systems.
Therefore, this magnitude might plausibly describe the information processing complexity of biological systems correctly, independently of the structure or ecological niche of the considered biological system, including higher forms of organisms. For example, a higher value of seems to identify a better adapted organism for survival in the long run for both a prey and a predator. Thus, it seems a reasonable candidate to explore its adequacy as a quantity that could identify the trends of evolution in the long term. Is this magnitude also suitable to describe the evolution in the long run of information-theoretic DAQM physical particles?
To answer this question, let us schematically examine the way in which Darwinian evolution would imply the generation of the anticipation module and the program , which codes the mathematical quantum formalism, on the information space of a DAQM particle , assuming that the increase of the survival information complexity would determine the direction of evolution in the long term.
The maximization of the first term would directly induce the selection of those populations of particles that developed an anticipation module on the information space of the particles. Increasing values of would imply more reliable anticipations of the configuration of the surrounding systems, and therefore increasing fitness of the system to its environment in the long term. By developing an anticipation module, the information about the surroundings that had been captured by the system would have an added value for survival.
As mentioned in
Section 4, the anticipation module
, as a consequence of Darwinian evolution under natural selection, would be a contextual algorithm adapted to the specific evolutionary history of every population of particles. Is nature able to construct such a program? Let us discuss this question.
First, there are some results that point to the capability of Darwinian evolution to find out the paths to transform certain exponential time problems into polynomial time ones [
70,
71,
72]. This is a characteristic feature of quantum computation and the kind of improvement that the presence of an anticipation module could imply in computation problems. In the study of Chatterjee et al. [
70,
71], there is no connection with an anticipatory procedure. However, in some theoretical scenarios in which the expected time of evolution is exponential with the sequence length that undergoes adaptation in the model [
70,
71], a mechanism is proposed that is able to beat the exponential barrier and allows evolution to work in polynomial time, as demonstrated in computer simulations [
70,
71].
Second, a central element in biological complex systems is the brain. The results of neural network studies foster the perspective of defining and characterizing the brain as a predictive machine [
73,
74] that learns to conjecture the causes of incoming sensory information. Therefore, Darwinian evolution acting on biological systems supports the idea that anticipation elements are generated by natural selection in biological systems with information processing capabilities.
It is meaningful to remark on the different orders of magnitude in the time scale of evolution for biological systems on Earth (at least 3.5 billion years [
75]) and for physical systems in the universe, on which Darwinian evolution would be acting (as hypothesized by DAQM) from the Big Bang (around 13.8 billion years [
76]). This difference must be taken into account when establishing parallelisms between both realms.
Let us now consider the problem of generating in the information space of a particle the classical algorithm
that codes the quantum mechanical rules. Increasing values of the
term would imply the generation of the Hilbert space structure for the state space of the systems, since it has been analyzed [
77,
78,
79,
80,
81] that the Hilbert space structure optimizes the information retrieval and inference capabilities of an information system, therefore leading to the improvement of the response time of the system that would induce an increase in fitness in the long run.
On the other hand, as mentioned above, the optimal dynamics for a system, from a general informational point of view, would be the one that minimizes the information outflow of the system, therefore the one that minimizes the accuracy of the predictions that its surrounding systems calculate about the behavior of the system. Thus, the maximization of the second term,
, as implied by the minimization of the outflow of information from the particle, plausibly brings about [
37,
38,
39,
40,
41,
42] the dynamical postulates of quantum mechanics, i.e., the Schrödinger equation and, as in Bohm-like theories, the guiding equation for calculating the velocity of the particle from the wave function. In particular, the Schrödinger equation can be directly derived, following Frieden [
82], by applying the principle of minimum Fisher information on the probability distribution function for the position of a particle [
82,
83].
Notice that although DAQM can be considered a kind of generalization of Bohmian mechanics, DAQM is local, as discussed in Ref. [
42], in which Bell inequality violations in the framework of DAQM are analyzed using a natural model for characterizing entanglement in this theory.
Further work has to be done in order to give a detailed account of the mechanisms underlying the deduction
15 of the quantum mechanical postulates in DAQM. In particular, the role played by entanglement shaping the interaction between particles.
Finally, let us briefly discuss the role and characteristics of the randomizer
that is stored in the information space of every particle as an intrinsic property. At time
, the randomizer would completely control the self-interaction of the system. As time goes by, its weight in the dynamics of the successive particle populations would progressively diminish as the program
evolved. However, assuming that the evolution of the information-theoretic particles of DAQM could be adequately modelled by a zero-sum elementary game, then, according to game theory [
84], Darwinian evolution would plausibly preserve a complementary role in the control of the particle for
, since randomness seems to be an essential element to optimize the survival expectations of a system, given that it would protect the system against the risk of completely deciphering its strategy by an adversary.
An open question of the model is the intrinsic nature of the randomizer. Is randomness an authentic fundamental, non-derivable property of matter? Or is it only pseudo-randomness, a kind of classical algorithm that yields an appearance of randomness? The difference is crucial for the model, since while true randomness, in principle, cannot be cracked, on the contrary, algorithmic pseudo-randomness, however intricate might it be, could be theoretically decrypted using enough resources and adequate observational and computational methods, as modern machine learning procedures show [
85]. An intermediate road might be to construct the randomizer, although still classical, [
39] following the mathematical prescription of the “free will function” defined by Hossenfelder [
86] or including a computationally irreducible algorithm as defined by Wolfram [
87], which would guarantee its unpredictability even for its own system. Depending on the intrinsic nature of randomness in matter, it could act as a screening wall for the anticipation depth.
The plausible development of the modules and in the information space of DAQM fundamental particles subject to evolution under the main action of natural selection has been analyzed. The emergence of quantum behavior in such particles would be the result of Darwinian evolution acting on matter, “the survival of the fittest” in a universe of information-theoretic fundamental particles.
DAQM helps to explore the relationship between classicality and quantumness from another point of view, namely, the central role that information seems to play in the quantum theory. However, this point of view is common to most modern reconstructions of quantum mechanics [
88]. In many of these reconstructions [
88], quantum systems are fundamentally characterized as carriers and processors of information. If this is an adequate representation of the world, and Darwinian natural selection might operate, then the appearance of the capability of projecting possible future configurations of the environment, the appearance of anticipation, would just be a question of time.
The second central characteristic of DAQM systems is that their properties are calculated in-flight on their information spaces in response to the interactions with other systems or to measurements in experiments. However, this reflects Bohr’s complementarity [
11,
12,
13,
14], from another perspective, the concept of objective indefiniteness [
89], or in the saying of Peres [
90] ‘unperformed experiments have no results’.
Therefore, in-flight calculated properties and anticipation, the two main elements associated with possessing information processing capabilities, constitute, in addition to the intrinsic randomness supplied by the randomizer on the information space of every particle, the backbone of DAQM for explaining quantum behavior in a natural way.
DAQM also explains in a remarkably natural manner the special adaptation and high-performance level of quantum computation for solving optimization problems and efficiently simulating quantum systems. As it has been analyzed in this article, a fundamental particle in DAQM would be a classical-like system supplemented with information processing capabilities that would simulate quantum mechanical behavior in real time by optimizing its information flows through the action of Darwinian natural selection.