Superdeterminism Without Conspiracy

Superdeterminism - where the Measurement Independence assumption in Bell's Theorem is violated - is frequently assumed to imply implausibly conspiratorial correlations between properties $\lambda$ of particles being measured and measurement settings $x$ and $y$. But it doesn't have to be: a superdeterministic but non-conspiratorial locally causal model is developed where each pair of entangled particles has unique $\lambda$. The model is based on a specific but arbitrarily fine discretisation of complex Hilbert space, where $\lambda$ defines the information, over and above the freely chosen nominal settings $x$ and $y$, which fixes the exact measurement settings $X$ and $Y$ of a run of a Bell experiment. Pearlean interventions, needed to assess whether $x$ and $y$ are Bell-type free variables, are shown to be inconsistent with rational-number constraints on the discretised Hilbert states. These constraints limit the post-hoc freedom to vary $x$ keeping $\lambda$ and $y$ fixed but disappear with any coarse-graining of $\lambda$, $X$ and $Y$, rendering so-called drug-trial conspiracies irrelevant. Points in the discretised space can be realised as ensembles of symbolically labelled deterministic trajectories on an `all-at-once' fractal attractor. It is shown how quantum mechanics might be `gloriously explained and derived' as the singular continuum limit of the discretisation of Hilbert space; It is argued that the real message behind Bell's Theorem has less to do with locality, realism or freedom to choose, and more to do with the need to develop more explicitly holistic theories when attempting to synthesise quantum and gravitational physics.


Introduction
A deterministic hidden-variable model is said to be superdeterministic -not a word the author would have chosen -if the so-called Measurement Independence assumption (sometimes referred to as the Statistical Independence assumption or the λ-independence assumption), is violated [19] [44] [14] [22] [47].Here ρ is a probability density on a set of hidden variables λ, and x and y ∈ {0, 1} denote experimentally chosen measurement settings -for concreteness, nominally accurate polariser orientations.Without (1), it is impossible to show that model satisfies the CHSH version of Bell's inequality |C(x = 0, y = 0) − C(x = 0, y = 1) + C(x = 1, y = 0) + C(x = 1, y = 1)| ≤ 2 (2) where C denotes a correlation on Bell-experiment measurement outcomes over an ensemble of particle pairs prepared in the singlet state.
The argument that models which violate (1) are conspiratorial, originates in a paper by Shimony, Horne and Clauser, written in response to Bell's paper on local beables [7].Shimony et al write: In any scientific experiment in which two or more variables are supposed to be randomly selected, one can always conjecture that some factor in the overlap of the backwards light cones has controlled the presumably random choices.But, we maintain, skepticism of this sort will essentially dismiss all results of scientific experimentation.Unless we proceed under the assumption that hidden conspiracies of this sort do not occur, we have abandoned in advance the whole enterprise of discovering the laws of nature by experimentation.
The drug trial is often used to illustrate the contrived nature of such a conspiracy.For example [18]: . . .if you are performing a drug versus placebo clinical trial, then you have to select some group of patients to get the drug and some group of patients to get the placebo.The conclusions drawn from the study will necessarily depend on the assumption that the method of selection is independent of whatever characteristics those patients might have that might influence how they react to the drug.Related to this, superdeterminism is sometimes described as requiring exquisitely (and hence unrealistically) finely tuned initial conditions [4], or as negating experimenter freedom [51].A number of quantum foundations experts (e.g.[29] [49] [3] [1]) use one or more of these arguments to dismiss superdeterminism in derisive terms.
A new twist was added by Aaronson [1] who concluded his excoriating critique of superdeterminism with a challenge: Maxwell's equations were a clue to special relativity.The Hamiltonian and Lagrangian formulations of classical mechanics were clues to quantum mechanics.When has a great theory in physics ever been grudgingly accommodated by its successor theory in a horrifyingly ad-hoc way, rather than gloriously explained and derived?It would seem that developing superdetermistic models of quantum physics is a hopeless cause.However, the purpose of this paper is to attempt to show that superdeterminism has been badly misunderstood, to rebuff these criticisms and indeed suggest that violating (1) is perhaps the only sensible way to understand the experimental violation of Bell inequalities.Importantly, we show that whilst conspiracy would imply a violation of (1), the converse is not true.In Section 2, we define what we mean by a non-conspiratorial violation of (1) and how it differs from these more traditional conspiratorial violations.Motivated by this, and the unshieldable effects of gravity as described in Section 3, a non-conspiratorial superdeterministic model is described in Section 4, based on a specific discretisation of complex Hilbert Space [35].Using the homeomorphism between p-adic integers and fractal geometry, the model is linked to the invariant set postulate [33] -the universe is evolving precisely on some special dynamically invariant subset of state space.In Section 5 it is shown how this model violates (1) non-conspiratorially, and indeed violates (2) in exactly the same way as does quantum mechanics.In Section 5.3 we show that the superdeterministic model is locally causal.In Section 6 we discuss common objections to superdeterminism including fine tuning, free will and the so-called drug-trial analogy.Addressing Aaronson's challenge in Section 6.3, we show how the state space of quantum mechanics can be considered the singular continuum limit of the discretised Hilbert space of the superdeterministic model.A possible experimental test of the supderdeterministic model is discussed in Section 7.
Before embarking on this venture, one may ask answer the question: why bother?After all, quantum mechanics is an extremely well tested theory, and has never been found wanting.Why not just accept that quantum theory violates the concept of local realism -whatever that meansand get on with it?The author's principal motivation for pursuing theories of physics which violate (1) lies in the possibility of finding a theory of quantum physics that is not inconsistent with the locally causal nonlinear geometric determinism of general relativity theory.As discussed in Section 8, results from this paper suggest that instead of seeking a quantum theory of gravity ('quantum gravity') we should be seeking a strongly holistic gravitational theory of the quantum, from which the Euclidean geometry of space-time is emergent as a coarse-grained approximation to the p-adic geometry of state space.This, the author believes, is the real message behind the experimental violation of Bell inequalities.

Conspiratorial and Non-conspiratorial Interventions
There is no doubt that conspiratorial violations of Bell inequalities, of the type mentioned in the Introduction, imply a violation of (1).Here we are concerned with the converse question: does a violation of (1) imply the existence of a conspiratorial hidden variable theory?In preparing to answer this question, we quote from Bell's response 'Free Variables and Local Causality' [7] (FVLC) to Shimony et al [7].In FVLC, Bell writes: I would insist here on the distinction between analyzing various physical theories, on the one hand, and philosophising about the unique real world on the other hand.In this matter of causality it is a great inconvenience that the real world is given to us once only.We cannot know what would have happened if something had been different.We cannot repeat an experiment changing just one variable; the hands of the clock will have moved and the moons of Jupiter.Physical theories are more amenable in this respect.We can calculate the consequences of changing free elements in a theory, be they only initial conditions, and so can explore the causal structure of the theory.I insist that B [Bell's paper on the theory of local beables [7]] is primarily an analysis of certain kinds of physical theory.
To understand the significance of this quote, we base the analysis of this paper around the thought experiment devised by Bell in FVLC, where, by design, human free will plays no explicit role.Bell supposes x and y are determined by the outputs of two pseudo-random number generators (PRNGs).These outputs are sensitive to the parity P x and P y of the millionth digits of the PRNG inputs.Bell now makes what he calls a 'reasonable' assumption: But this peculiar piece of information [whether the parity of the millionth digit is odd or even] is unlikely to be the vital piece for any distinctively different purpose, i.e., it is otherwise rather useless . . .In this sense the output of such a [PRNG] device is indeed a sufficiently free variable for the purpose at hand.
It is important to note that, in this quote, Bell deflects discussion away from statistical properties of some ensemble of runs of an experiment where measurement settings are supposedly selected randomly (as per the Shimony et al quote above), and focusses on one individual run of an experiment.
There is an important reason for this.When discussing conspiratorial hidden-variable models of the Shimony et al type, it is assumed that in any large enough ensemble with common value of λ, there exist sub-ensembles for each of the four pairs of measurement settings (00, 01, 10 and 11).In this context, (1) implies that the four sub-ensembles are statistically equal.Conversely, in a conspiratorial violation of (1), the four sub-ensembles are statistically unequal.In such a situation, ρ(λ|xy) can be interpreted as a frequency of occurrence within each of the four sub-ensembles.It is worth noting that in such a frequency-based interpretation of ρ, the issue of counterfactual definiteness -a central issue below -never arises.This has led to a misconception that counterfactual definiteness plays no role in Bell's Theorem.
Importantly, hidden-variable models do not have to be like this.It is possible that the value of λ is unique to each run of a Bell experiment.The model described below has this property.In this situation, if ρ(λ|xy) were to define a frequency of occurrence, and a particle with value λ was measured with settings x and y and could only be measured once, then ρ(λ|xy) ̸ = 0 and ρ(λ|x ′ y) = ρ(λ|xy ′ ) = ρ(λ|x ′ y ′ ) = 0 where x ′ = 1 − x and y ′ = 1 − y.However, this does not itself imply a violation of (1) -it merely emphasises that ρ(λ|xy) is fundamentally not a frequency of occurrence in an ensemble but rather is a probability density defined on an individual particle with value λ.
With that in mind, let us continue to focus, as Bell does, on a single entangled particle pair.If x and y were not free variables, then P x or P y would not be vital for 'distinctively different' purposes.That is to say, we could vary P x or P y without having a vital impact on distinctly different systems.In the language of Pearl's causal inference theory [38], if x and y were not free variables, there would exist small so-called interventions which by design changed P x or P y , and by consequence had a vital impact on distinctly different systems.
The most important part of this paper is to draw attention to two possible ways this might happen.The first is the conventional way where the effect of the intervention propagates causally from its localised source in space-time, somehow vitally influencing distinctly different systems.It is hard to imagine how varying something as insignificant as the parity of the millionth digit of an input to a PRNG could have such a vital impact.For this reason, Bell argued, the PRNG output should indeed be considered a free variable.This, of course, is not unreasonable.
However, there is a second possibility -one that was not considered by Bell -that such interventions are simply inconsistent with physical theory.That is to say, the hypothetical state of the universe where P x or P y is perturbed but all other distinctly different elements of the universe are kept fixed, is inconsistent with the laws of physics.If such an intervention was hypothetically applied to a localised region of the universe, the ontological status of the whole universe would change; clearly a state of the universe as a whole only exists if all parts of it satisfy the laws of physics.Of course, we cannot perform an actual experiment to test directly this potential inconsistency: in changing the millionth digit, the hands of the clock and the moons of Jupiter will have moved, as Bell notes.Hence addressing the question of whether the PRNG output is a free variable in the sense of this paragraph requires studying the mathematical properties of physical theory.This is Bell's point in the first quote and it is the topic of this paper.
Below we develop a model where each particle pair has unique λ with the property As will be shown this can be interpreted as a locally causal non-conspiratorial violation of (1).It implies that an intervention on x, keeping λ and y fixed, leads to a state of the world which is inconsistent with the model postulates and therefore has zero probability.Clearly, this cannot be an intervention within space-time.Instead the intervention describes a hypothetical perturbation which takes a point in the state space of the universe (labelled by the triple (λ, x, y)), consistent with the model and hence with ρ(λ|xy) ̸ = 0, to a state (λ, x ′ , y) which is inconsistent with physical theory and hence has ρ(λ|x ′ y) = 0. Importantly, this means a non-conspiratorial interpretation of (3) implies that physical theory does not have the post-hoc property of counterfactual definiteness.
The essential nature of counterfactuals in Bell's Theorem was pointed out by Redhead [41] some years ago.This important point seems to have been lost in more recent discussions of Bell's Theorem.However, one needs to be careful to not throw the baby out with the bathwater.Counterfactual reasoning is both pervasive and important in physics.Indeed, it is central to the scientific method [28].The fact that we can express laws of physics mathematically gives us the power to estimate what would have happened had something been different.Such estimates can lead to predictions and the predictions can be verified by experiment.We clearly do not want to give up counterfactual reasoning entirely in our search for new theories of physics.We address this concern by noting that output from experiment and physical theory (particularly when the complexities of the real world are accounted for) involves some inherent coarse-graining.Experiments have some nominal accuracy and output from a computational model is typically truncated to a fixed number of significant digits.We can represent such coarse graining by integrating exact output of physical theory over small volumes V ϵ > 0 in state space, where V ϵ → 0 smoothly as ϵ → 0. In developing a superdeterministic model below, we will require that counterfactual definiteness holds generically when variables are coarse grained over volumes V ϵ , for suitably defined small values ϵ > 0. The model based on discretised Hilbert Space, described in Section 4 has this property.As discussed below, this renders the drug-trial analogy irrelevant.
These matters are subtle, and it seems Bell appreciated this.Instead of derisively rejecting theories where (1) is violated he concludes FVLC with the words: Of course it might be that these reasonable ideas about physical randomisers are just wrong -for the purposes at hand.Indeed, in his last paper 'La Nouvelle Cuisine', Bell writes: An essential element in the reasoning here is that [polariser settings] are free variables.... Perhaps such a theory could be both locally causal and in agreement with quantum mechanical predictions.However, I do not expect to see a serious theory of this kind.I would expect a serious theory to permit 'deterministic chaos' or 'pseudorandomness' . . .But I do not have a theorem about that.
The last sentence is insightful, because, as we discuss, there is no such theorem.Indeed the reverse: here we develop a serious non-classical model where polariser settings are not free variables, utilising geometric concepts in deterministic chaos.

The Andromedan Butterfly Effect
The purpose of this section is to note how the Principle of Equivalence makes the interaction of matter with gravity especially chaotic.Here we repeat a calculation first reported by Michael Berry [8] and further analysed in [45].It is well known that the flap of a butterfly's wings in Brazil can cause a tornado in Texas.But, could the flap of a butterfly's wings on a planet in the Andromedan galaxy cause a tornado in Texas?
We consider molecules in the atmosphere as hard spheres of radius R with mean free distance l between collisions.We wish to estimate the uncertainty ∆θ M in the angle of the M th collision of one of the spheres with other spheres, due to some very small uncertain external force.It is easily shown that ∆θ M grows exponentially with M .In particular With l ≈ 10 −7 m and R ≈ 10 −10 m, l/R ≈ 10 3 .After how many collisions M is the position of a molecule in Earth's atmosphere sensitive to the gravitational uncertainty in the uncertain flap of a butterfly's wing in the Andromeda galaxy?Let r denote the distance between Earth and Andromeda.The flap of a butterfly's wing through a distance ∆r will change the gravitational force on our target molecule by an amount ∆F = ∂F/∂r ∆r = (Gm mol m wing /r 3 ) ∆r. Uncertainty in the flap of an Andromedan butterfly's wing will therefore induce an uncertainty in the acceleration of a terrestrial atmospheric molecule by an amount ∆a = 2Gm wing r 3 ∆r (5) If τ denotes the mean time between molecular collisions, there is an uncertainty in the direction of the molecule ∆θ 1 ≈ τ 2 ∆a/R.Plugging in m wing = 10 −5 kg, ∆r = 10 −2 m, τ = 10 −9 s and r = 10 22 m, gives ∆θ 1 = 10 −90 .How large is M before ∆θ M ≈ 1? From the above, 10 3M ≈ 10 90 .Hence after about 30 collisions the direction the direction of travel of the terrestrial molecule has been rendered completely uncertain by the gravitational effect of the Andromedan butterfly.Indeed one can go further: the direction of the terrestrial molecule is rendered completely uncertain by the uncertain position of a single electron at the edge of the visible universe after about only 50 or so collisions (which below we will round up to an order of magnitude of 10 2 ).Once the molecules of the Earth's atmosphere have been disturbed in this way, it is only a matter of a couple of weeks before the nonlinearity of the Navier-Stokes equations leads to uncertainty in a large-scale weather pattern, such as a Texan tornado [27] [36].
Nowhere above did the mass of the molecule enter the calculation.The same calculation could just as well apply to the measurement process in quantum physics -where a particle interacts with atoms in some measuring device.Indeed the same calculation could apply to molecules in an experimenter's brain, affecting the decisions they make.
As a matter of principle, the direction of motion of a molecule after 100 collisions is for all practical purposes uncomputable.If a computation (say on a supercomputer) was attempted, then the direction of the molecule would depend on the motion of electrons in the chips of the supercomputer.Self-referentially, the computational software would have to include a representation of the computation.Technically, the Andromedan Butterfly Effect describes a computationally irreducible system [50] -one that cannot be computed with a computationally simpler system.This is the first step in our argument that (1) could be violated because the universe should be considered a rather holistic chaotic system.Gravity is what makes the universe holistic.Unlike the other forces of nature, the effects of gravity cannot be shielded by negative charges.However, by itself, this argument doesn't imply that the parity of the millionth digit, or the flap of a butterfly's wings in Andromeda is a vital piece of information for determining distinctly different systems.For that we need more from the theory of chaos.

A Superdeterministic Model 4.1 Nominal vs Exact Measurement Settings and Hidden Variables
We build on Bell's thought experiment whereby polariser orientations are determined by the parities P x and P y in Alice and Bob's PRNGs (supposing there is no particular reason why these would be odd or even).Manifestly, P x and P y only determine the polariser orientations to some nominal accuracy.That is, x, y ∈ {0, 1} determine small neighbourhoods of the 2-sphere of orientations, referred to as ϵ disks.None of the results below depend on the magnitude of ϵ as long as ϵ > 0 (as discussed in Section 6.3, the limit ϵ = 0 in the proposed model is singular).It will be assumed -consistent with our search for an underlying deterministic theory -that when measurements are made on particular particles, the measurement outcomes ±1 are associated with some exact, albeit unknown, polariser orientations X and Y. Here, X and Y denote unit vectors in physical 3-space from the centre of a unit ball, to the surface of the 2-sphere of orientations.The corresponding points on the 2-sphere to which the vectors point are written as X and Y .The nominal directions x and y refer to unit vectors pointing to the centroids of the ϵ disks.In the discussion below, we assume that the probabilistic nature of the quantum measurement problem arises because the measurement outcome is typically sensitive to the exact measurement settings within an ϵ disk (c.f.fractally intertwined basins of attraction [32]).Hence, any given nominal setting is consistent with an ensemble of possible exact settings.
With this in mind, we let λ describe all of the variables which, over and above P x and P y , determine the exact measurement settings X and Y.These variables include Andromedan butterflies and electrons at the end of the visible universe.When measuring a single qubit we can frame it like this: consider a spacelike hypersurface S in the past of some event E where the measurement outcome was known, and before the event where the PRNG output was known, then λ must include data on S, on and inside E's past light cone, with the exception of P x (or its determinants).In this way we can write X = E(λ, x), Y = E(λ, y).The extension of this causal picture for entangled qubits is discussed in Section 5.3.

Rational Quantum Mechanics: RaQM
Motivated by John Wheeler's plea to excise the continuum from physical theory [48] -see also [17] -a way to introduce non-conspiratorial superdeterminism into quantum physics is to discretise Hilbert space [11] [12] [35] [13].At the experimental level this is surely unexceptionable: all experiments which confirm quantum mechanics will necessarily confirm a model of quantum physics based on discretised Hilbert Space, providing the discretisation is fine enough.However, as discussed, such a discretisation has profound implications for the interpretation of quantum experiments

Single Qubits
Consider a qubit prepared in a state |1⟩, and written as with respect to some arbitrary basis {|1 ′ ⟩, | − 1 ′ ⟩}.We call this a 'proper basis' if where p is some large prime number and 0 ≤ m 1 , n 1 ≤ p.In a proper basis, it can be shown [37] that |ψ(θ, ϕ)⟩ has a representation as a bit string of p deterministic elements ±1 where cos 2 θ/2 equals the fraction of elements +1 in the bit string and ϕ denotes a cyclical permutation of elements of the bit string.Here we will assume p is a large prime.In RaQM the measurement basis is a proper basis, and the measurement output is a deterministically selected element of the bit string.In this way, each proper basis is a basis with respect to which a measurement could potentially be made by the experimenter.A measurement cannot be made with respect to a basis which is not proper.For each proper basis, the potential measurement outcome is associated with some exact measurement orientation (θ e , ϕ e ) in physical 3-space, which satisfies the rationality constraints written in base-p.In this way, (θ, ϕ) correspond to nominally accurate measurement orientations.
In this representation, measurement outcome probabilities (over ensembles of elements of the bit string) automatically satisfy Born's Rule.Hence Born's Rule is not a separate axiom in RaQM.
If the exact preparation setting is represented by X, and exact measurement setting Y, then the first of ( 8) can be written X • Y ∈ Q, where • denotes the scalar product of two vectors.Because of Niven's theorem angles that determine probabilities and angles that determine phases, c.f., (8), are relatively incommensurate, except at the precise values ϕ e = 0, π/2, π, 3π/2.One can assume that such precise values never occur in practice (any gravitational wave would disturb a system away from such a precise value), though they may be relevant for theoretical reasons (e.g. when one considers a measurement performed with a precisely opposite measurement direction).
An important corollary to Niven's Theorem -central to this paper -is what we refer to as the Impossible Triangle Corollary: Corollary.Let △XY Z be a triangle on the unit sphere with rational internal angles, not precisely equal to multiples of 45 Proof.Assume X • Z = cos θ XZ is rational, where θ XZ denotes the exact angular distance between X and Z on the unit sphere, etc.By the cosine rule for △XY Z, where ϕ Y is the exact internal angle of the triangle at the vertex Y .Since cos θ XY and cos θ Y Z are both rational, then from (9), sin ) cos 2 ϕ Y must be rational.Again, since cos θ XY and cos θ Y Z are both rational, cos 2 ϕ Y and hence cos 2ϕ Y must be rational.But this is impossible since ϕ Y is itself rational and ϕ Y is not a multiple of 45 • .Hence cos θ XZ must be irrational.
The Impossible Triangle Corollary is vital for explaining the notion of non-commutativity in RaQM [37].Consider a particle with spin prepared (with Stern-Gerlach device SG0) relative to some exact orientation X.It is passed through Stern-Gerlach device SG1 with exact orientation Y.The spin-up output beam of SG1 is passed through Stern-Gerlach device SG2 with exact orientation Z.By RaQM, X • Y and Y • Z must be rational.In a run of the experiment, a detector in one output channel of SG2 will register a particle.Consider a hypothetical counterfactual experiment on the same particle (same λ) where SG1 and SG2 are swapped (commuted).The measurement outcome from this hypothetical experiment is undefined: by the Impossible Triangle Corollary, if X • Y and Y • Z are rational, then X • Z is not.
We note in passing that it is straightforwardly shown [35] that the ensemble representation of the single qubit state in RaQM satisfies a uncertainty principle relationship, i.e.
Here ∆S x and ∆S y are associated with standard deviations of bit-strings, and ⟨S z ⟩ denotes a bit-string ensemble mean.

Multiple Qubits
Bell's inequality is based on measurements of pairs of entangled particles prepared in the quantum mechanical singlet state with correlations where σ denote Pauli matrices.
In RaQM, an n-qubit system is represented by a set of n bit strings.Hence in RaQM ( 11) is represented by two correlated bit strings (each with equal numbers of +1s and −1s).The bits represent measurement outputs from members of an ensemble, defined by deterministic laws.Corresponding to any one ensemble member (which can be labelled by λ) the exact measurement settings corresponding to x, y must satisfy i.e., cos θ XY is rational.Keeping X fixed and perturbing Y → Y ′ , then X and Y ′ are also permissible exact settings if X • Y ′ is rational, and the angle ϕ subtended at X between the two great circles XY and XY ′ satisfies Similar rational constraints apply, fixing Y and perturbing X → X ′ .
In any small neighbourhood of some point which does not satisfy the rationality conditions ( 13) and ( 14), there will, for large enough p, exist points which do satisfy the rational conditions.Hence it will be impossible to violate the rationality conditions in any coarse-graining no matter how fine, for large enough p.

The Invariant Set Postulate
A clear disadvantage of discretised Hilbert space is that the sum of two discretised Hilbert vectors is no longer guaranteed to be a Hilbert vector -indeed typically it is not by Niven's theorem.However, if discretised Hilbert vectors represent symbolic strings describing ensembles of deterministic worlds, one can speculate that arithmetic closure exists at some deeper deterministic level.But what type of deterministic system would be consistent with the rationality constraints described above?The discussion in Section 3 suggests a chaotic model may be relevant.However, we need something in addition to mere chaos to account for the rationality constraints, we need a chaotic system evolving on an invariant subset of state space.
Consider, for example, the chaotic model that Lorenz [26] discovered in his quest to understand the deterministic non-periodicity of weather.These equations describe a classical dynamical system and can be integrated from any triple of initial states (x L , y L , z L ) at t = 0.However, at t = ∞ this model has an emergent non-classical property: all states lie on a measure-zero, fractionally dimensioned, dynamically invariant subset I L of state space, known as the Lorenz attractor.That I L is fractionally dimensioned implies that I L has a non-Euclidean, fractal geometry.That I L is an invariant set implies that if a point lies on I L , its future evolution will continue to lie on I L for all time, and its past evolution has lain on I L for all time.That I L has measure zero implies that the probability that a randomly chosen point belongs to I L is equal to zero (random with respect to the uniform measure on the Euclidean state space R 3 spanned by (x L , y L , z L )).This is consistent with the notion that the rationals describe a set of measure zero in the continuum field of real numbers.Conversely, associated with I L is a fractal invariant measure ρ L (a Haar-type measure).Points which do not lie on I L have ρ L = 0.Such points define states which are inconsistent with a non-classical dynamical system where all states lie on I L by definition.Such a system is non-classical because I L is a non-computable subset of state space [10] [16] -non-computability being a post-quantum discovery of 20th Century mathematics .Now suppose we are given some timeseries from the output of the Lorenz equations in this nonclassical limit where the system is evolving on I L .No matter how long is the timeseries, we cannot estimate statistical quantities such as correlations or conditional frequencies more accurately than the accuracy to which the timeseries has been outputted (e.g. the number of significant figures of output variables).That is to say, we must treat all estimates of frequency (and correlation) as functions of coarse-grained variables defined from non-zero ϵ balls of volume V ϵ in state space.A key property of I L is that no matter how small is V ϵ , as long as it is non-zero, it is undecidable as to whether a point inside V ϵ has ρ L = 0.However, by the fractal nature of I L , we know such points exist.The results from Section 3 suggests the universe itself be considered a chaotic system.Consistent with the discussion above, we assume the universe is a deterministic chaotic dynamical system evolving precisely on its fractal invariant set I U , with invariant measure ρ U .This is referred to as the invariant set postulate [33].States q which do not lie on I U must be assigned a measure and hence a probability ρ U (q) = 0.It is worth noting that geometric properties of a system's invariant set (e.g. its non-integer dimension) provides a relativistically invariant description of chaos, in contrast with positivity of Lyapunov exponents [15].
In number theory, Ostrowsky's theorem states that there are only two inequivalent norm-induced metrics on the rational numbers Q: the Euclidean metric and the p-adic metric [25].It is well known that the set of p-adic integers is homeomorphic to a Cantor Set with p iterated pieces [42].We can therefore suppose that states in discretised Hilbert Space represent ensembles of trajectories on a fractal invariant set at some level of fractal iteration, where each trajectory at one level of iteration is associated with an ensemble of p trajectories at the next level of fractal iteration.In this way, a deterministic state which does not satisfy the rationality constraints of RaQM corresponds to a state of the world which does not lie on the invariant set I U .In this picture, the measurement process corresponds to a jump from one fractal iteration of I U to the next -that is to say a 'quantum jump' describes an increment in fractal iteration.This suggests a picture of the evolution of time similar to a fractal zoom [46].
Although in a true fractal, the depth of iteration is infinite, the ideas expressed here continue to hold if the depth of iteration of I U is in fact finite.An invariant set with finite depth of iteration corresponds to a strictly periodic limit cycle.Computational representations of fractal attractors are in fact periodic limit cycles.We will assume below that I U is in fact a periodic limit cycle.
As discussed above, although deterministic, such models are not classical.Classical models are associated with deterministic initial conditions and dynamical evolution equations expressed in terms of differential (or finite difference) equations on the reals or complex numbers.Typically one can vary initial conditions as one likes, and perturbed initial conditions can typically be integrated from the evolution equations without issue.In classical models, the ontology of states does not depend on their lying on invariant sets, nor on their having rational-number characteristics.This has consequences for our understanding of free will, discussed in Section 6.
Notice that the results discussed here do not depend on how large is p, as long as it is not infinite.Moreover, by writing ϵ ∼ 1/p, it can be seen that violations to the rationality constraints in RaQM can be completely eliminated by coarse graining over ϵ disks, no matter how small is ϵ.

The Bell (1964) Inequality
In this subsection, we focus on the original Bell inequality [5] For some specific run in an experiment to test (17), suppose Alice chooses the nominal orientation x 1 and Bob the nominal orientation x 2 .In keeping with the discussion above, λ, x 1 and x 2 fix a pair of exact measurement settings X 1 = E(λ, x 1 ) and X 2 = E(λ, x 2 ).To satisfy (13), Unlike the nominal settings, Alice and Bob have no control over exact settings, hence have no choice as to whether this rationality constraint is satisfied.There are many points in the two ϵ disks for x 1 and x 2 which satisfy this constraint (see Fig 1).
In order that a putative hidden-variable theory satisfies (17), it is necessary that, in addition to the real-world run where the particles were measured with nominal settings (x 1 , x 2 ), the same particles (same λ) could have been measured with nominal settings (x 1 , x 3 ) and (x 2 , x 3 ) with definite outcomes ±1.We will consider these hypothetical worlds in turn.
For the first, keeping λ fixed, Alice continues to choose the nominal setting x 1 , whilst Bob hypothetically chooses the nominal setting x 3 .Since X 1 = E(λ, x 1 ), then keeping λ and x 1 fixed, Figure 1: a) The three circles correspond to ϵ disks on the unit sphere associated with the three nominal measurement settings x 1 , x 2 and x 3 in Bell's 1964 inequality.The straight lines represent great circles on the unit sphere whose cosine of angular distance is rational.By the Impossible Triangle Corollary, it is impossible for the cosine of all three angular distances (i.e.X 1 • X 2 , X 1 • X 3 and X 2 • X 3 ) to be rational.Because of this, ( 1) is violated non-conspiratorially. b) In a more conventional model, there is no requirement for the exact settings to be held fixed when comparing real and hypothetical worlds with the same hidden variables and same nominal settings.In this model it is always possible to satisfy (1) and hence the violation of Bell inequalities must imply violation of local realism or some grotesque conspiracy.
X 1 is fixed.By contrast, keeping λ fixed but transforming x 2 to x 3 implies a hypothetical change in Bob's exact setting from X 2 to some (unknown) X 3 .This transformation is consistent with RaQM as long as the rationality conditions ( 13) and ( 14) hold.It is therefore necessary that X 1 • X 3 ∈ Q and that ϕ/2π ∈ Q where ϕ denotes the angle between the great circles X 1 X 2 and X 1 X 3 at the point X 1 of intersection.There are plenty of exact settings X 3 in the ϵ neighbourhood for x 3 for which these rationality constraints are satisfied.
But now, c.f. the third term in (17), we consider the hypothetical world where, keeping λ fixed, Alice chooses the nominal direction x 2 and Bob x 3 .With X 3 = E(λ, x 3 ), X 3 is fixed by its value for the first hypothetical world.We now invoke the key property of the singlet state: if the measurement outcome was +1 (say) when Bob's exact setting was X 2 , then the measurement outcome will be −1 in a hypothetical world where Alice's exact setting was X 2 .Hence, considering the exact settings corresponding to the third term in (17), X 2 and X 3 must both be held fixed at their previously determined values.However, appealing to the Impossible Triangle Corollary, if X 1 •X 2 and X 1 •X 3 are rational and if ϕ is rational and not precisely a multiple of 45 • (gravitational waves from Andromeda will help ensure that), then X 2 • X 3 cannot be rational.See Fig 1.
In essence, for each run in a Bell experiment, one of the two counterfactual runs is inconsistent with the rationality constraints of the hidden variable model and therefore must be assigned a probability ρ = 0. Put another way where, to emphasise, all orientations in (18) are nominal.If one configuration occurs in reality (so that its probability is not identically zero) then (18) implies that (1) is violated.

The CHSH Inequality
This is a straightforward extension of the argument above though we no longer use the singlet property that, with the same exact settings, Alice and Bob must have opposite measurement outputs.
Here x, x ′ , y and y ′ (with x ′ = 1 − x, y ′ = 1 − y as before) denote four ϵ disks (i.e.nominal settings) on the 2-sphere of orientations.In a given run where Alice chooses x and Bob y, we write the corresponding exact settings as X, Y where In order that a putative hidden-variable theory satisfies (2), it is necessary that, in addition to the real-world run, the same particles (with the same λ) could have been measured with nominal settings xy ′ , x ′ y and x ′ y ′ , with definite outcomes ±1.
As before, with By repeated application of the Impossible Triangle Corollary, we now show it is impossible to satisfy (20) and (21).
We do this again by contradiction.Suppose that ( 20) and ( 21) are satisfied and consider the two triangles △XX ′ Y ′ and △XX ′ Y .By the cosine rule for spherical triangles for each of the two triangles Subtracting these equations then must be rational.Writing where r 1 , r 2 are rational.However, by Niven's Theorem, providing γ and δ are not precise multiples of 45 • , A 2 1 and A 2 2 must be irrational, hence A 1 and A 2 must be irrational.Moreover they must be independently irrational since γ and δ can be varied independently of one another.Hence generically A 1 − A 2 = A must be irrational which is the contradiction we are looking for.
This in turn leads to the following general conclusion.In the situation where Alice chose x and Bob y, then, keeping the particles' hidden variables fixed, at least one of the three counterfactual choices must violate the rationality conditions: 1) Alice and Bob chose x and y ′ , Alice and Bob chose x ′ and y, or Alice and Bob chose x ′ and y ′ .Similar to ( 18) If ρ(λ|xy) ̸ = 0, i.e. one configuration occurs in reality then (25) implies that (1) is violated.As with Bell's 1964 inequality, the Impossible Triangle Corollary implies that (1) is violated without conspiracy.
It can be noted that since RaQM is based on an arbitrarily fine discretisation of complex Hilbert Space, by letting p be sufficiently large, RaQM must violate Bell's inequality as closely we like to the quantum mechanical violation of Bell's inequality.

Local Causality
Below we consider a locally causal violation of the rationality constraints (19) and (20).
Fig 3 illustrates a space-time diagram where a photonic source emits two entangled photons.S is a spacelike hypersurface through the event where the particles were emitted by the source.The photons are measured by Alice and Bob's detectors with nominal settings x and y and exact settings X and Y .The nominal settings are determined by two PRNGs shown in the figure.By the discussion above, the particle's hidden variables λ, together with the parities P x and P y , determine these exact settings.We suppose that, consistent with the counterfactual violation of ( 19) and ( 20), two components, λ A corresponding to information in the past light cone of Alice's measurement event, and λ B corresponding to information in the past light cone of Bob's measurement event.We write Alice and Bob's measurement outcomes (= ±1) in the form: Figure 3: A space-time diagram illustrating the locally causal nature of the proposed superdeterministic model based on RaQM.Suppose Alice's measurement outcome is +1.This measurement outcome could have been −1 if P x (the parity of the millionth digit) were different or if λ A were different.However, local causality demands that Alice's measurement outcome could not have been −1 if P y were different or λ B was different, keeping P x and λ A fixed.Importantly, this does not exclude the possibility that Alice's measurement outcome would be inconsistent with the laws of physics, and hence is undefined, if P y had been different or λ B had been different, keeping P x and λ A fixed.According to RaQM and the invariant set postulate, such undefinedness arises because the corresponding Pearlean interventions are inconsistent with rationality constraints and take the universe off its invariant set.Because of this, the event A, whilst determined by data on and inside the past light cone of A, does not depend on space-time events at or near B, consistent with Bell's notion of locality in La Nouvelle Cuisine.
Whatever the values of S A and S B in (26), it is never the case that where λ ′ A , λ ′ B describe perturbed data on S (see Fig 3).That is to say, a hypothetical intervention in space-time which alters either Bob's PRNG input or the hidden variable λ B in the past light cone of Bob's measurement event, keeping P x and λ A fixed, and changes Alice's measurement outcome, violates the rationality conditions (19) and (20).In simple words this implies that Alice's measurement outcome does not depend on Bob's measurement settings -the essence of locality.The space-time events at B are irrelevant for determining the event A because any intervention which changes B keeping the data on S A fixed is inconsistent with the putative model and hence does not define a putative event in space time.But this is precisely Bell's characterisation of local causality in La Nouvelle Cuisine -see Fig 4 in [6].Put another way, the event A is entirely determined by data on S A , the subset of S contained on and in the past lightcone of A -the essence of relativistic causality.

Objections to Superdeterminism
Below we address some of the objections that have been raised against superdeterminism with the RaQM/invariant set model in mind.

The Drug Trial
Each human in a drug trial is unique; having unique DNA for example.However, the characteristics used to allocate people randomly to the active drug or placebo groups are based on coarse-grain attributes.Are they young or old?Are they male or female?Are they black or white?We typically assume the two groups contain equal numbers of such coarse-grain attributes.In a conspiratorial drug trial, the selection process is manipulated so that the two groups do not contain equal numbers of coarse-grain attributes.
Although RaQM violates (1), it does not violate any version of (1) where coarse-grained hidden variables are used in place of hidden variables.For large enough p, it is always possible to find counterfactual worlds which satisfy the rationality constraints in a coarse-grained volume V ϵ of state space, no matter how small is ϵ.That is to say, in any sufficiently-large ensemble of individual runs, where λ denotes a coarse-grained value of λ and ρ f denotes a frequency of occurrence.In this sense, (28) does not imply (1) and the drug trial conspiracy is an irrelevance to RaQM.

Fine Tuning
The fine-tuned objection (e.g.[4]) rests on the notion that superdeterminism appears to require some special, atypical, initial conditions.Perhaps one might view an initial state lying on a fractal attractor as special and atypical -after all a seemingly tiny perturbation (changing P x keeping λ fixed) can take the state of the universe off its invariant set I U .Although the Euclidean metric accurately describes distances in space-time, the p-adic metric is more natural in describing distances in state space when the inherent geometry of state space is fractal [25].From the perspective of the p-adic metric, a fractal invariant set is not fine-tuned: a perturbation which takes a point off I U is a large perturbation (of magnitude at least p), even though it may appear very small from a Euclidean perspective.Conversely, perturbations which map points of I U to points of I U can be considered small amplitude perturbations.Similarly, we must ask with respect to what measure are states on I U deemed atypical.Although states on I U are atypical with respect to a uniform measure on the Euclidean space in which I U is embedded, they are manifestly typical with respect to the invariant measure of I U [20].
In claiming that a theory is fine tuned, one should first ask with respect to which metric/measure is the tuning deemed fine -and then ask whether this the natural metric/measure to assess fineness.

Singular Limits and Aaronson's Challenge
Aaronson's challenge (see the Introduction) raises a more general question: what is the relationship between a successor theory of physics and its predecessor theory?There is a subtle but important relationship brought out explicitly by Michael Berry [9] that is of relevance here.
Typically an old theory is a singular limit of a new theory, and not a smooth limit, as a parameter of the new theory is set equal to infinity or zero.A singular limit is one where some characteristic of the theory changes discontinuously at the limit, and not continuously as the limit is approached.Berry cites as examples the old theory of ray optics is explained from Maxwell theory, or the old theory of thermodynamics is explained from statistical mechanics.His claim is that old theories of physics are typically singular limits of new theories.
If quantum theory is a forerunner of some successor superdeterministic theory, and Berry's observation is correct, quantum mechanics is likely to be a singular limit of that superdeterministic theory.Here the state space of quantum theory arises at the limit p = ∞ of RaQM, but not before.For any finite p, no matter how big, the incompleteness property that led to the violation of (1) holds.However, it does not hold at p = ∞.From this point of view, quantum mechanics is indeed a singular limit of RaQM's discrete Hilbert space, at p = ∞.It is interesting to note that pure mathematicians often append the real numbers to sets ('adeles') of p-adic numbers, at p = ∞.However, the properties of p-adic numbers are quite different to those of the reals for any finite p no matter how big.Here the real-number continuum is the singular limit of the p-adics at p = ∞.The relationship between QM and RaQM is very similar.
In physics, it is commonplace to solve differential equations numerically, i.e. to treat discretisations as approximations of some continuum exact equations, such that when the discretisation is fine enough, the numerical results are as close as we require to the exact continuum solution.This is not a good analogy here.A better analogy is analytic number theory, considered as an approximation to say the exact theory of prime numbers.If one is interested in properties of primes for large primes, treating p as if it were a continuum variable can provide excellent results.However, here the continuum limit is the approximation and not the exact theory.
Contrary to Aaronson's statements, the singular relationship between a superdeterministic theory and quantum mechanics is exactly as one would expect from the history of science.

Free Will
Superdeterminism is often criticised as denying experimenter free will.Nobel Laureate Anton Zeilinger put it like this [51]: We always implicitly assume the freedom of the experimentalist...This fundamental assumption is essential to doing science.If this were not true, then, I suggest, it would make no sense at all to ask nature questions in an experiment, since then nature could determine what our questions are, and that could guide our questions such that we arrive at a false picture of nature.
A clear problem here is that the notion of free will is poorly understood [24] and therefore hard to define rigorously.It was for this reason that Bell introduced his PRNG gedanken experiment -to show that it was possible to discuss (1) and its potential violation without invoking free will.
Nevertheless, to avoid the charge of conspiracy, an experimenter must be able to choose in a way which is indistinguishable from a random choice.For the present purposes we can think of this as being consistent with free will.For this reason, experimenters have found increasingly whimsical ways of choosing measurement settings -such as bits from a movie, or the wavelength of light from a distant quasar -in an attempt to mimic randomness.
In It is well known that brains are low-power noisy systems [43] [36] where neurons can be stochastically triggered.In practice, the source of such stochasticity is thermal noise.However, such noise -arising from the collision of molecules -will have an irreducible component due to the Andromedan butterfly By the Andromedan butterfly effect, we can assume that the outputs of these brains is sensitive to all data on S in ALICE and BOB's past light cones and not just λ.We can use this to assert that Alice and Bob's brains cannot be corrupted by the values of the hidden variables.
effect.In such circumstances, as discussed in Section 3, this can render the action of the brain non-computational.In particular, if we were to construct a model of Alice and Bob's brains which are driven by the subset λ of data on S (or indeed any subset), this model will not provide reliable predictions of their brains' decisions.This surely provides evidence of our ability to choose in ways which are for all practical purposes random.
However, the Invariant Set Postulate provides insights which may help shed new light on the age-old dilemma of free will.Consider two possible invariant sets I U and I U ′ .Here, for given λ, we suppose I U permits the settings 00 and 11 in a Bell experiment, whilst, keeping λ fixed, I U ′ permits the settings 01 and 10.These invariant sets differ (very slightly) in terms of their geometries, i.e. in terms of the (underlying deterministic) laws of physics.
If Alice and Bob are free to choose their measurement settings, then, prior to their choosing, all observations of the universe must be consistent with the universe belonging to either of the I U and I U ′ .However, once Alice and Bob have chosen, not only is one of I U and I U ′ consistent with available observations, the other becomes inconsistent with the supposed deterministic laws of physics.
Suppose Alice and Bob chose 00 and let q denote a state of the universe prior to Alice and Bob choosing.The question then arises: was it always the case -even before Alice and Bob chose 00that q ∈ I U and not I U ′ ?Or did the very act of Alice and Bob choosing result in q ∈ I U rather than q ∈ I U ′ ?In thinking about these questions, it is important to distinguish determinism from predestination.In a conventional deterministic initial-value problem, initial conditions are specified independently from evolutionary laws and the evolved state is predestined from the initial state.By contrast, if one thinks of the geometry of the invariant set as primitive, then the choices Alice and Bob make are no more 'predestined' from q than q is predestined from their choices.Instead, all one can say is that, because of determinism, the earlier and later states must be dynamically consistent.Importantly, as a global state space geometry, I U is consistent with what Adlam calls an 'all at once' constraint [2].One could say that the geometric specification of I U , and hence whether q ∈ I U , depends as much on states on I U to the future of q as on states to the past of q.This future/past duality exists because neither the proposition q ∈ I U , nor Alice and Bob's choice given q, is computational (in our finite system it is computationally irreducible).
Hence, is it both true that q ∈ I U (rather than I U ′ ) before Alice and Bob chose, and also true that the act of choosing required q ∈ I U rather than q ∈ I U ′ .Importantly, as there is no notion of temporal causality in state space, it would be wrong to call this latter fact retrocausality.Simply, it is a consequence of I U being an all-at-once constraint.This type of analysis helps explain the so-called delayed choice paradoxes in quantum physics.
It is possible to conclude not only that experimenter choices are indeed freely made (Nobel Laureates can be assured that their brains are not being subverted), but that these choices can determine which states of the universe are consistent with the laws of physics and which not (surely a fitting role for Nobel Laureates).This has some significant implications for the role of intelligent life in the universe more generally, which the author will discuss elsewhere.

Experimental Tests
A key result from this paper is that we will not be able to detect non-conspiratorial superdeterministic violations of (1) by studying frequencies of measurement outcomes in a Bell experiment.We must look for other ways of testing such theories.
Of course, QM is exceptionally well tested and if a superdeterministic theory is to replace QM, it must clearly be consistent with results from all the experiments which support QM.Here RaQM has a free parameter p, which, if large enough, can replicate all existing experiments.This is because with large enough p, discretised Hilbert space is fine enough that it replicates to experimental accurace the probabilistic predictions of a theory based on continuum Hilbert space (and Born's Rule to interpret the squared modulus of a state as a probability -something automatically satisfied in RaQM).Conversely, however, if p is some finite albeit large number, then in principle an experiment with free parameter p exp can study situations where p exp > p where there might be some departure between reality and QM [21].
One conceivable test of RaQM vs QM probes the finite amount of information that can be contained in the quantum state vector |ψ⟩ (in RaQM n qubits are represented by n bit strings of length p).In RaQM, the finite information encoded in the quantum state vector will limit the power of a general purpose quantum computer, in the sense that RaQM predicts the exponential increase in quantum compute speed with qubit number for a range of quantum algorithms may generally max out at a finite number m of qubits.
The key question concerns the value of m.Could it be related to the number of collisions in the Andromeda butterfly effect?We will return to this elsewhere.

Conclusions
Attempts to develop models which violate (1) do not justify the derision from a number of researchers in the quantum foundations community over the years.Not least, Bell himself did not treat the possible violation of (1) with derision and accepted that seemingly reasonable ideas about the properties of physical randomisers might be wrong -for the purposes at hand. .
A superdeterministic (and hence deterministic) model has been proposed which is not conspiratorial, is locally causal, does not deny experimenter free choice and is not fine tuned with respect to natural metrics and measures.The model, based on a discretisation of Hilbert space, is not a classical hidden-variable model, i.e. it derives its properties from post-quantum-theory mathematical science (particularly that of non-computability and computational irreducibility).By considering the continuum of complex Hilbert Space as a singular limit of a superdeterministic discretisation of complex Hilbert Space, Aaronson's challenge to superdeterminists, to show how a superdeterministic model might gloriously explain quantum mechanics, can be met.
One of the most important conclusions of this paper is that we need to be extremely cautious when invoking the notion of an 'intervention' in space time, at least in the context of fundamental physics.Such interventions form the bedrock of Pearl's causal inference modelling [38], and causal inference has been used widely in the quantum foundations community to try to analyse the causal structure of quantum physics e.g.[52] [40].Here we distinguish between two types of intervention: one that is consistent with the laws of physics and one that is not.The effect of the former type of intervention, if it is initially contained within a localised region of space-time, must propagate causally in space-time, constrained by the Lorentzian metric of space time.By contrast, the latter type of intervention simply perturbs a state of the universe from a part of state space where the laws of physics hold, to a part of state space where the laws of physics do not hold.If this superdeterministic model is correct, theories of quantum physics based on causal inference models which adopt an uncritical acceptance of interventions will give misleading results.
The results of this paper suggest that the way gravity interacts with matter may be central to understanding the reasons why the universe can be considered a holistic dynamical system evolving on an invariant set, and hence why Hilbert space should be discretised.This suggests that instead of looking for a quantum theory of gravity, we should instead be looking for a gravitational theory of the quantum [34] [39].However, importantly, the results here suggest such a theory will not be found by probing smaller and smaller regions of space-time, ultimately the Planck scale.It will instead be found by incorporating into the fundamental laws of physics, the state-space geometry of the universe at its very largest scales [36].Planck-scale discontinuities in space-time may instead be an emergent property of such (top-down) geometric laws of physics.
In this regard, a recent proposal [31] for synthesising quantum and gravitational physics describes gravity as a classical stochastic field.The latter is consistent with our discussion of the Andromedan butterfly effect.However, it is not consistent with our discussion of Bell's inequality.On the other hand, if one simply acknowledges that a stochastic gravitational field is a 'for all practical purposes' representation of a chaotic system evolving on a fractal invariant set, then Oppenheim's model may become consistent with both the proposed superdeterministic violation of Bell's inequality and with realism and the relativistic notion of local causality.
In the author's opinion, this is the real message -not non-locality, indeterminism or unrealitybehind the violation of Bell's inequality.

Figure 2 :
Figure 2: Illustrating the CHSH experiment where x, x ′ , y and y ′ denote ϵ-disks associated with nominal measurement settings, under the control of the experimenters.X, X ′ , Y and Y ′ are points on the unit 2-sphere corresponding to exact measurement settings, and straight lines correspond to great circles joining these points.By the relationship X = E(λ, x) and Y = E(λ, y), fixing λ and a nominal measurement setting fixes an exact measurement setting.As discussed in the text, if Alice and Bob chose the nominal settings xy, then, by the Impossible Triangle Corollary with λ fixed, the exact settings associated with one of the three counterfactual nominal settings xy ′ , x ′ y or x ′ y ′ cannot satisfy the rationality conditions of RaQM.
Fig 4 we replace the two PRNGs in Fig 3 by Alice and Bob's brains.

Figure 4 :
Figure 4: Here the PRNGs of Fig 3 are replaced with the experimenters' brains.By the Andromedan butterfly effect, we can assume that the outputs of these brains is sensitive to all data on S in ALICE and BOB's past light cones and not just λ.We can use this to assert that Alice and Bob's brains cannot be corrupted by the values of the hidden variables.