Causality Is an Effect, II

Causality follows the thermodynamic arrow of time, where the latter is defined by the direction of entropy increase. After a brief review of an earlier version of this article, rooted in classical mechanics, we give a quantum generalization of the results. The quantum proofs are limited to a gas of Gaussian wave packets.


Introduction
The history of multiple time boundary conditions goes back-as far as I know-to Schottky [1,2], who, in 1921, considered a single slice of time inadequate for prediction or retrodiction (see Appendix A). There was later work of Watanabe [3] concerned with prediction and retrodiction. Then, Schulman [4] uses this as a conceptual way to eliminate "initial conditions" prejudice from Gold's [5] rationale for the arrow of time and Wheeler [6,7] discusses two time boundary conditions. Gell-Mann and Hartle also contributed to this subject [8] and include a review of some previous work. Finally, Aharonov et al. [9,10] proposed that this could solve the measurement problem of quantum mechanics, although this is disputed [11]. See also Appendix B.
This formalism has also allowed me to come to a conclusion: effect follows cause in the direction of entropy increase. This result is not unanticipated; most likely everything having to do with arrows of time has been anticipated. What is unusual, however, is the ability to prove the thesis mathematically.
The proof given [12] only applies to classical mechanics. The present work extends it to quantum mechanics. Moreover, since the previous paper was buried in conference proceedings, and the present work uses similar arguments, I will repeat some parts of that proof. Note though that the present work is extremely limited: it applies to particles in a gas with Gaussian wave functions. I think it should apply more generally, but that is not what I prove.
A remark is in order on the arrow of time. Often the sequence of cause and effect is taken as primary [13], but in the present context this is shown not to hold. Our argument depends on two conditions: the nature of perturbation and two-time boundary conditions. Both will be elaborated on. A future condition on quantum problems may make measurement deterministic but is hollow in that it does not lead (to my knowledge) to testable claims. There is the possibility of advanced effects ( [6,14]) but so far, this has not proved measurable.

Perturbation
A "cause" can be a change (or intervention, in the dynamics or in the initial conditions) or it can be a precursor ("pre" in the usual meaning of the term). In a change, you can compare the outcome of two similar but not identical "causes." The case of "precursor" is different. You can say one thing leads to another, A is a precursor of B, but you only consider A, not some variant, say A . I will adopt the meaning "change." This allows study of "effects", whereas I do not know how to examine ordinary time evolution. One should Note that both beginning and end are forced to have low entropy (by definition, assuming 0 and T are small). This is illustrated for the "cat map" in the first graph of Figure 1. See Appendix C for information on the "cat map" and on our procedure. Then, we do something different! Two times for perturbation are contemplated: the first during the rising of the entropy, the second during its fall. By detailed calculation (not reproduced here, but see below, Section 5) it is found that causality, the macroscopic change in behavior, follows the direction of entropy increase, that is in the falling entropy case, the macroscopic behavior changes earlier in the time parameter (which is neutral, i.e., has two different arrows and in between doesn't have an arrow).
Intuitively, this is simple. Let the perturbation be at t 0 and 0 < t 0 < τ < T − τ < T. Between 0 and t 0 (before the perturbation) both perturbed and unperturbed have the same boundary conditions, hence their macroscopic behavior is similar (but the microscopic behavior is, in general, different). After the perturbation there is no constraint, i.e., no effective boundary condition: both systems go to equilibrium, from which there is adequate time (recall T > 2τ) to reach (whether perturbed or not) a specific region within the unit square. There was however a perturbation and there are different microscopic and macroscopic paths. Similarly, for T − τ < t 0 < T the boundary conditions at T and t 0 (just after the perturbation in the neutral parameter t) are the same for perturbed and unperturbed systems, hence the macroscopic behavior is the same. However, they both go (in (−t)) to equilibrium; hence, the two systems (perturbed and unperturbed) have different macroscopic paths (N.b., distinguish carefully the words "macroscopic" and "microscopic."). This is illustrated in the case of the cat map, both for 0 < t 0 < τ and T − τ < t 0 < T. See Figure 1, second and third images.

Quantum Version
The first step is to show that with low entropy conditions at both (distantly separated) times, the entropy increases in between. To calculate entropy in subspaces of Hilbert space presence or absence in that subspace should be defined and numbers of states counted. This leads to having regions of 6-dimensional x-p space-to mimic the classical space-and that is possible. What we do is take a region of phase space and select a basis of states whose maximum value is in this region. Coherent states will do the job. For convenience, the value of the spread used to define those states might be the fixed point of [15], but that is not necessary. Finally the dimension of the Hilbert subspace will be the number of states in that region. What this means is that one can go back to the classical method of using (the logarithm of) volume in phase space as a measure of entropy.
Therefore, as was done previously, we take coarse grains that are regions of phase space-coordinates and momentum. The density matrix involves mainly the diagonal elements. Thus even for identical particles the separation of locale as well as separation of momenta makes the density matrix nearly diagonal. For more on this theme, see [16].
Even for identical particles this leads to cancellation. In one dimension the wave function for a pair of Gaussians is (not normalized) (The variables are x 1 and x 2 ; all the others are constants.) The diagonal elements of the density matrix (calculated from Equation (2)) already show signs of cancellation, as follows: As is evident, if x α is significantly different from x β or k α from k β , then there is already cancellation or rapid oscillation. With more particles the effect is stronger. This, by the way, is the reason that isolated systems can be analyzed without paying attention to symmetrization with respect to all identical particles in the universe.
To get low entropy at the beginning and the end confine the wave functions-at the beginning and end-to particular coarse grains. In between the two times considered, wave functions will spread. This sounds like the classical definitions, but once you have coarse grains the definitions are not all that different. However, to imitate [12] it is necessary that ψ final = U T ψ initial , where U T is the propagator from time-0 (zero is the initial time) to time-T (T is the final time). This imposes a significant constraint on the wave function, in particular, the wave function, under U T , should not spread. If spreading were to happen, the entire space available, not just the target in phase space, would be occupied and the entropy would not drop.
In a recent article [15], we found that Gaussians that scatter do not spread. (This was explored further in [17], but the principal application to Gaussians was done in [15].) The idea is that scattering provides localization. In [17], it is argued that wave functions become Gaussian (often) but that involves decoherence, which is human perception. Is everything a Gaussian? Obviously not; atomic wave functions can be complicated and even the hydrogen atom is a different function. Nevertheless, for the purpose of this demonstration a Gaussian will be adequate, at least for showing that sometimes causality is an effect.
The requirement that ψ final = U T ψ initial and that both be confined to a (small) region of phase space (at t = 0 and t = T) is severe. However, based on the results of [15], it can be done. It is possible that the entropy would not be strictly zero (due to the Gaussian's never vanishing) but it can be made small. The same holds in momentum space.
At this point we are back in classical mechanics and the proof is straightforward. Since non-standard definitions are used in [12] we repeat the proof (now using standard definitions).

Classical Proof
We present a précis of what has been done in our previous work ( [12]). On the phase space Ω let µ be a measure and µ(Ω) = 1. Let the dynamics be a measure-preserving map φ (t) on Ω, with φ (t) (ω) the time-t image of an initial point ω ∈ Ω. The coarse graining of Ω, providing a notion of "macroscopic," are sets with the following properties: {∆ α }, α = 1, . . . , G, with ∪ α ∆ α = Ω, ∆ α ∩ ∆ β = ∅ for α = β. Let χ α be the characteristic function of ∆ α and let v α = µ(∆ α ) = χ α (ω)dω (ω ∈ Ω). If f is a function on Ω, its coarse graining is defined as Let the system's distribution in Ω be described by a density function ρ(ω). The entropy to be used for studying irreversibility involves coarse graining and is defined as with ρ formed from ρ as in Equation (4). (In other notation, S = − Ω ρ(ω) log ρ(ω) dω.) The relative entropy (Kullback-Leibler divergence), to which S(ρ) is related, was given in [12] with a different sign from the usual. Moreover, the illustration given in Figure 1 uses a different definition of entropy. Turning to the system at hand, it is required to start (t = 0) in a subset 0 ⊂ Ω and end (t = T) in a subset T ⊂ Ω. (The fact that Gaussian wave functions necessarily do not vanish anywhere may lead to a small correction.) The points of Ω satisfying this two-time boundary condition are In [12,18,19] it is argued that for chaotic dynamics and for sufficiently long times T, = ∅. (Whether this carries over to quantum mechanics will be dealt with later.) We assume that there is a relaxation time τ, and that T τ. As a consequence (Recall that for mixing dynamics, φ (t) satisfies lim t→∞ µ A ∩ φ (t) (B) = µ(A)µ(B). This is true for the t → ∞ limit, but we assume that there is a time (τ) such that the decorrelation condition holds. Moreover, φ is measure preserving.) Under φ (t) , becomes To calculate the entropy, the density, which was ρ(0) = χ /µ( ) at time-0, must be coarse grained. The important quantity for the entropy calculation is If T − t > τ then the following will hold Using the measure-preserving property of φ (t) , the factors µ( T ) in both numerator and denominator of ρ α cancel, leading to This is precisely what one gets without future conditioning, so that all macroscopic quantities, and in particular the entropy, are indistinguishable from their unconditioned values. Working backward from time-T one obtains an analogous result. Define a variable s ≡ T − t and set˜ (s) ≡ (T − s). Theñ If s satisfies T − s > τ, then when the density associated with˜ (s) is calculated, its dependence on 0 will drop out. It follows that For a time-reversal invariant dynamics this will give the entropy the same time dependence coming back from T as going forward from 0. Even in the absence of T invariance, there should be roughly the same behavior because of the absence of any dissipative dynamics. Now we turn to perturbations. Call the unperturbed system A. The microstates are in the set (formerly called ). System B, the perturbed case, has an instantaneous transformation act on it at time-t 0 . Call this transformation ξ. (This transformation is called ψ in [12]. The letter ξ is chosen to avoid confusion with the wave function.) It is not dissipative-the arrow does not arise from such an asymmetry. ξ is invertible and measure preserving. Successful solutions must go from 0 to T under the transformation φ (T−t 0 ) ξφ (t 0 ) . The microstates for system B are therefore Again, there is the assumption that B is not empty, which will be taken up for quantum mechanics in Section 6. Clearly, A and B are different-at all times. But as will now be shown, for mixing dynamics and for sufficiently large T, the following hold: (1) for t 0 close to 0, the only differences in macroscopic behavior between A and B are for t > t 0 ; (2) for t 0 close to T, the only differences in macroscopic behavior between A and B are for t < t 0 .
The direction of causality follows the direction of entropy increase. The proof is nearly the same as before. Again we use a time τ such that the mixing decorrelation holds for time intervals longer than τ. First consider t 0 close to 0. The observable macroscopic quantities are the densities in grain-∆ α , which are, for t < t 0 , As before, the mixing property, for , which is the initial-value-only macroscopic time evolution. For ρ B α , the only difference is to add a step, ξ −1 . But this step is as measure preserving as φ itself, and therefore, as before, . Thus A and B have the same macrostates before t 0 . For t > t 0 , ρ A α (t) continues its behavior as before. For ρ B α (t) things are different: Now, I require T − t > τ. If this is satisfied the T dependence drops out and This shows that the effect of ξ is the usual initial-conditions-only phenomenon. If we repeat these arguments for t such that T − t is small, then just as we showed in Section 3, the effect of ξ will only be at times t less than t 0 .

The Set 0 ∩ φ −T T Is Not Empty under Quantum Evolution
The assertion, the set 0 ∩ φ −T T is not empty under quantum evolution, is sometimes true. What has been shown is that Gaussian wave functions do not spread indefinitely, but rather are localized each time they scatter [15]. Thus, our arguments hold for a gas of particles that are Gaussians and occasionally scatter. As found in [15,17], wave functions overlap, but scattering usually takes place at distances on the order of = mean free path (for scattering).
What happens in more general situations is not known. One theory of the quantum measurement process ( [19], namely, there is no such thing as a measurement) proposes that non-entanglement is usually the case. However, experiment has not yet ruled on this issue.

Conclusions
There are many defects to this quantum version of entropy increase determining the direction of causality. What is not clear is whether a quantum theory is at all needed. Causality as usually understood is a macroscopic phenomenon and not quantum mechanical. Nevertheless, we point out the deficiencies of the present work.
There is uncertainty over black holes. If one considers two time boundary conditions purely as a way to get rid of "initial conditions" prejudice, OK, but if there is any thought that the world is actually symmetric (approximately) then that too must be dealt with. Do black holes evaporate? Do they become white holes? Are there black holes at all? (See the blog by Cramer [20].) We will not deal with any of these questions and allow our results to depend on the answers.
Our arguments are phrased in terms of chaotic dynamics. Chaos is problematic for quantum mechanics. Our previous paper ( [12]) deals also with harmonic oscillators (which are not chaotic) but has other limitations. For the argument above to be meaningful all that is really required is a relaxation time. It is plausible that this exists in quantum theory.
Although we have briefly mentioned identity of particles, that has not been seriously dealt with.
Finally, there is the requirement that ψ final = U T ψ initial (where U T is the evolution operator for time T) and that both be low entropy, confined to a region of phase space. In some versions of quantum mechanics this is a natural requirement, but if that is wrong, there are severe limitations on the applicability of these results (see Section 6).
We mention in passing that our understanding of the "destiny" of Aharonov et al. does not fix the results of measurements (the "destiny" being a future value of the wave function). This is because general wave functions allow many outcomes to a given experiment.