1. Introduction
In [
1], we developed an axiomatic system for thermodynamics that incorporated information as a fundamental concept. This system was inspired by previous axiomatic approaches [
2,
3] and discussions of Maxwell’s demon [
4,
5]. The basic concept of our system is the
eidostate, which is a collection of possible states from the point of view of some agent. The axioms imply the existence of additive conserved quantities called
components of content and an entropy function
that identifies reversible and irreversible processes. The entropy includes both thermodynamic and information components.
One of the surprising things about this axiomatic system is that despite the absence of probabilistic ideas in the axioms, a concept of probability emerges from the entropy
. If state
e is an element of a uniform eidostate
E, then we can define
States in
E with higher entropy are assigned higher probability. As we will review below, this distribution has a uniquely simple relationship to the entropies of the individual states and the overall eidostate
E.
The emergence of an entropic probability distribution motivates us to ask several questions. Can this idea be extended beyond uniform eidostates? Can we interpret an arbitrary probability distribution over a set of states as an entropic distribution within a wider context? What does the entropic probability tell us about probabilistic processes affecting the states within our axiomatic system? In this paper, we will address these questions.
2. Axiomatic Information Thermodynamics
In physics, a formal axiomatic system can serve several purposes. It can serve to ensure consistent reasoning with unfamiliar or counter-intuitive mathematical objects. Furthermore, it may clarify the underlying concepts of a subject and their relations. This in turn can help to understand how widely, and under what circumstances, those concepts may apply to the physical world. Here, we review the system in [
1] and describe the ideas represented by the definitions and axioms.
Our system aims to capture how an agent with finite knowledge might describe states and processes in the external world. The set of states is denoted , but the agent’s knowledge is represented by an eidostate, which is a finite collection of possible states in . (The term eidostate comes from the Greek eidos, meaning to see.) Thus, the agent’s knowledge of the state may be incomplete. No probability distribution is assumed a priori for the eidostate; it is a simple enumeration of possibilities. The set of eidostates is called . (Without too much confusion, we may identify a state with the singleton eidostate so that can be regarded as a subset of .)
Eidostates may be combined by the operation +, which simply represents the Cartesian product of the sets. Thus, if A and B are distinct eidostates, the combination is not the same as . Two eidostates A and are similar () if they are formed by the same Cartesian factors perhaps put together in a different way. Our first axiom is
Axiom 1 (Eidostates). is a collection of sets called eidostates such that
- (a)
Every is a finite non-empty set with a finite prime Cartesian factorization.
- (b)
if and only if .
- (c)
Every non-empty subset of an eidostate is also an eidostate.
The agent may control state transformations, including those that involve the acquisition, use, or deletion of information. This produces a relation → on . We interpret to mean that there exists a dynamical evolution that transforms A into B, perhaps in the presence of some other “apparatus” states that undergo no net change. We assume that any simple rearrangement of an eidostate’s components is always possible, that transformations can be composed in succession, and so on:
Axiom 2 (Processes). Let eidostates , and .
- (a)
If , then .
- (b)
If and , then .
- (c)
If , then .
- (d)
If , then .
Formally, a process is a pair of eidostates, and it is said to be possible if either or . An eidostate A is uniform if, for all , is possible. Uniform eidostates, as we will see, have particularly tractable properties. The existence of non-uniform eidostates is neither forbidden nor guaranteed by the axioms, and there are models both with them and without them.
Consider a non-deterministic process in which initial state a might evolve into either or . We describe this as a deterministic process on eidostates (that is, ). The principle underlying the “Second Law” in our axiomatic system is simple to state. Given a collection A of possible states, no process can deterministically eliminate one or more of the elements of A without any other effect. Possibilities cannot simply be “deleted” by the agent. That is,
Axiom 3. If and B is a proper subset of A, then .
To borrow an example from the next section, suppose the agent does not know whether a coin is showing “heads” (h) or “tails” (t). This might be represented by the eidostate . The agent might examine the coin to discover its actual state, but this is not a deterministic process, and it might end up in either h or t. To guarantee that the coin is h, the agent will need to intervene in the coin’s state, which in principle may involve changes to other states.
On the other hand, it is reasonable to assume the existence of certain “conditional” processes:
Axiom 4 (Conditional processes).
- (a)
Suppose and . If and then .
- (b)
Suppose A and B are uniform eidostates that are each disjoint unions of eidostates: and . If and then .
We now introduce information concepts into the system. This is based on the idea of a record state. State r is a record state if there exists another state such that . That is, a particular record state may be freely created or deleted. An information state is an eidostate containing only record states; the set of these is called . A bit state is an information state with exactly two elements, and a bit process is a process of the form . Then, we have
Axiom 5 (Information). There exist a bit state and a possible bit process.
Devices such as Maxwell’s demon accomplish state transformations by acquiring information. We accommodate these into our theory via a new axiom:
Axiom 6 (Demons). Suppose and such that .
- (a)
There exists such that .
- (b)
For any , either or .
The first clause asserts that what one demon can do (transforming a into b by acquiring the information in J), another demon can undo (transforming b into a by acquiring information in I). The second clause envisions nearly reversible demons.
Two more axioms serve more technical purposes. The first regularizes the → relation by its asymptotic behavior:
Axiom 7 (Stability). Suppose and . If for arbitrarily large values of n, then .
The second posits the existence of states in which conserved quantities may be stored—e.g., the height of a weight used for storing energy. We have
Axiom 8 (Mechanical states). There exists a subset of mechanical states such that
- (a)
If , then .
- (b)
For , if then .
(This axiom is identical to one posited by Giles [
2], whose system was a model for our own.) Note that the axiom does not assert that
is non-empty; it is perfectly consistent to have a model for the axioms with no mechanical states.
We introduce one additional axiom that allows us to relate uniform eidostates to singleton states. Consider, for instance, a situation in which a one-particle “Szilard” gas is confined to a volume. This situation corresponds to a single state (as in the first part of Figure 2 below). Now, we introduce a partition so that the gas particle might be either in one side or the other (that is, state or ). If we remove the partition, the original state is restored. Thus, . Considering many such examples, we arrive at our final axiom:
Axiom 9 (State equivalence). If E is a uniform eidostate, then there exist states such that and .
The system yields some striking results. For instance, we can establish the existence of components of content, which are numerical quantities that are conserved in every possible process. A component of content Q is a real-valued additive function on the set of states . (Additive in this context means that .) In a uniform eidostate E, every element has the same values of all components of content, so we can without ambiguity refer to the value . The set of uniform eidostates is denoted . This set includes all singleton states in , all information states in and so forth, and it is closed under the + operation.
Perhaps the most far-reaching theorem established by the axioms is the following (Theorem 8 in [
1]):
Theorem 1. There exist an entropy function and a set of components of content Q on with the following properties:
- (a)
For any , .
- (b)
For any and component of content Q, .
- (c)
For any , if and only if and for every component of content Q.
- (d)
for all .
On uniform eidostates, the entropy function is completely determined (up to a non-mechanical component of content) by the → relations among the eidostates.
3. Coin-and-Box Model
Now that we have described our system in some detail, we will introduce a very simple model of the axioms. None of our later results depend on this model, but a definite example will be convenient for explanatory purposes. Our theory deals with configurations of coins and boxes; the states are arrangements of coins, memory records, and closed boxes containing coins. We have the following:
Coin states, which can be either h (heads) or t (tails) or combinations of these. It is also convenient to define a stack state to be a particular combination of n coin states h: . The coin value Q of a compound of coin states is just the total number of coins involved. A finite set K of coin states is said to be Q-uniform if every element has the same Q-value.
Record states r. As the name suggests, these should be interpreted as specific values in some available memory register. The combination of two record states is another record state. Thus, r, , , etc., are all distinct record states. Record states are not coins, so .
Box states. For any Q-uniform set of coin states C, there is a sequence of box states . Intuitively, this represents a kind of closed box containing coins so that . If , then we denote the corresponding “basic” box states by .
We now must define the relation → among eidostates.
We assume that similar eidostates can be transformed into each other in accordance with our axioms. As far as the → relation is concerned, we can freely rearrange the “pieces” in a compound eidostate.
For coin states, .
If r is a record state, for any a. In a similar way, for an empty box state, .
If K is a Q-uniform eidostate of coin states, .
Now, we add some rules that allow us to extend these to more complex situations and satisfy the axioms. In what follows, A, , B, etc., are eidostates, and s is a state.
- Transitivity.
If and , then .
- Augmentation.
If , then .
- Cancelation.
If , then .
- Subset.
If and , then .
- Disjoint union.
If A and B are both disjoint unions and , and both and , then .
Using these rules, we can prove a lot of → relations. For example, for a basic box state, we have
. From the subset rule, we have
(but not the reverse). Then, we can say
from which we can conclude (via transitivity and cancelation) that
. The use of a basic box allows us to “randomize” the state of one coin.
Or consider two coin states and distinct record states
and
. Then
from which can show that
. That is, we can set an unknown coin state to
h if we also make a record of which state it is. A pretty similar argument establishes the following:
The bit state
can be deleted at the cost of a coin absorbed by the basic box. The basic box is a coin-operated deletion device, and since each step above is reversible, we can also use it to dispense a coin together with a bit state (that is, an unknown bit in a memory register).
These examples help us to clarify an important distinction. What is the difference between the box state
and the eidostate
? Could we simply replace all box states
with a simple combination
of possible coin eidostates? We cannot, because such a replacement would preclude us from using the subset rule to obtain Equation (
2). The whole point of the box state is that the detailed state of its contents is
entirely inaccessible for determining possible processes. From the point of view of the agent, putting a coin in a box effectively randomizes it.
It is not difficult to show that our model satisfies all of the axioms presented in the the last section with the mechanical states in
identified as coin states. The key idea in the proof is that we can reversibly reduce any eidostate to one with a special form:
where
is a stack state of
coins and
is an
information state containing
k possible record states. (This is true because all box states can be reversibly “emptied”, and any
Q-uniform eidostate of coin states can be reversibly transformed into a stack state and an information state, as in Equation (
5)). Relations between eidostates are thus reduced to relations between states of this form. We note that the coin value
q is conserved in every → relation, and no relation allows us to decrease the value of
k. In our model, there is just one independent component of content (
Q itself), and the entropy function is
. (We use base-2 logarithms throughout).
4. The Entropy Formula and Entropic Probability
Now, let us return to the general axiomatic system. One general result of [
1] is a formula for computing the entropy of a uniform eidostate
E in terms of the entropies of its elements
e. This is
It is this equation that motivates our definition of the entropic probability of
e within the eidostate
E:
Then,
and the probabilities sum over
E to 1. As we have mentioned, the entropy function
may not be quite unique; nevertheless, two different admissible entropy functions lead to the
same entropic probability distribution. Even better, our definition gives us a very suggestive formula for the entropy of
E:
where the mean
is taken with respect to the entropic probability, and
is the Shannon entropy of the distribution
P [
6,
7].
Equation (
9) is very special. If we choose an arbitrary distribution (say
), then with respect to this probability, we find
with equality if and only if
is the entropic distribution [
7]. Therefore, we might
define the entropic probability to be the distribution that maximizes the sum of average state entropy and Shannon entropy—a kind of “maximum entropy” characterization.
How is the entropic probability related to other familiar probability concepts? To quote [
1],
Every formal basis for probability emphasizes a distinct idea about it. In Kolmogorov’s axioms [
8], probability is simply a measure on a sample space. High-measure subsets are more probable. In the Bayesian approach of Cox [
9], probability is a rational measure of confidence in a proposition. Propositions in which a rational agent is more confident are also more probable. Laplace’s early discussion [
10] is based on symmetry. Symmetrically equivalent events—two different orderings of a shuffled deck of cards, for instance—are equally probable. (Zurek [
11] has used a similar principle of “envariance” to discuss the origin of quantum probabilities.) In algorithmic information theory [
7], the algorithmic probability of a bit string is related to its complexity. Simpler bit strings are more probable.
We may remark further that axiom systems like those of Kolmogorov or Cox do not assign actual probabilities in any situation; they simply enforce rules that any such assignment must satisfy. Entropic probability is an actual assignment of a particular distribution, which is determined by objective facts about state transformations.
5. Uniformization
A unique entropic probability rule arises from our → relations among eidostates, which in the real world might summarize empirical data about possible state transformations. But so far, this entropic probability distribution is only defined within a uniform eidostate E.
In part, this makes sense. An eidostate represents represents the knowledge of an agent—i.e., that the state must be one of those included in the set. This is the knowledge upon which the agent will assign probabilities, which is why we have indicated the eidostate E as the condition for the distribution. Furthermore, these might be the only eidostates, since the axioms themselves do not guarantee that any non-uniform eidostates exist. (Some models of the axioms have them, and some do not.) But can we generalize the probabilities to distributions over non-uniform collections of states?
Suppose
is a finite set of states that is possibly not uniform. Then, we say that
A is
uniformizable if there exists a uniform eidostate
, where the states
are mechanical states in
. The idea is that the states in
A, which vary in their components of content, can be extended by mechanical states that “even out” these variations. Since
is uniform, then
for any
. The abstract process
is said to be
adiabatically possible [
2]. Mechanical states have
, so the entropy of the extended
is just
which is independent of our choice of the uniformizing mechanical states.
What is the significance of this entropy? Suppose A and B are not themselves uniform, but their union is uniformizable. Then, we may construct uniform eidostates and such that either or , depending on whether or the reverse. In short, the entropies of the extended eidostates determine whether the set of states A can be turned into the set B if we imagine that these states can be augmented by mechanical states, embedding them in a larger, uniform context.
Given the entropy of the extended state, we can define
This extends the entropic probability to the uniformizable set
A.
Let us consider an example from our coin-and-box model. We start out with the non-uniform set
. These two basic box states have different numbers of coins. But we can uniformize this set by adding stack states, so that
is a uniform eidostate. The entropy of a basic box state is
, so we have
The entropic probabilities are thus
6. Reservoir States
So far, we have uniformized a non-uniform set A by augmenting its elements with mechanical states, which act as a sort of “reservoir” of components of content. These mechanical states have no entropy of their own. But we can also consider a procedure in which the augmenting states act more like the states of a thermal reservoir in conventional thermodynamics.
We begin with a mechanical state , and posit a sequence of reservoir states , which have the following properties.
For any
n and
m, we have that
, so that
is a non-negative constant for the particular sequence of reservoir states. This sequence
is characterized by the state
and the entropy increment
. Note that we can write
, where
.
For example, in our coin-and-box model, the basic box states act as a sequence of reservoir states with a mechanical (coin) state and an entropy increment . The more general box states form a reservoir state sequence with and , where and k is the number of states in K. For each of these box-state reservoir sequences, .
One particular type of reservoir is a
mechanical reservoir consisting of the states
,
,
, etc. We denote the
nth such state by
. For the
reservoir states,
. If we have a finite set of states
that can be uniformized by the addition of the
states, they can also be uniformized by a corresponding set of non-mechanical reservoir states
:
As before, we can find the entropy of this uniform eidostate and define entropic probabilities. But the
reservoir states now contribute to the entropy and affect the probabilities.
First, the entropy:
The entropic probability—which now depends on the choice of reservoir states—is
The reservoir states affect the relative probabilities of the states. For example, suppose
for a pair of states in
A. We might naively think that these states would end up with the same entropic probability, as they would if we uniformized
A by mechanical states. But since we are uniformizing using the
reservoir states, it may be that
and
have different entropies. Then, the ratio of the probabilities is
which may be very different from 1.
Again, let us consider our coin-and-box model. We begin with the non-uniform set
. Each of these states has the same entropy
, that is, zero. We choose to uniformize using basic box states
. For instance, we might have
Recalling that
, the entropy is
This yields probabilities
As an illustration of these ideas, consider the version of Maxwell’s demon shown in
Figure 1.
The demon is a reversible computer with an initial memory state . It is equipped with a reversible battery for storing energy, initially in mechanical state . The demon interacts with a one-particle “Szilard” gas, in which the single particle can move freely within its volume (state ). The gas is maintained in thermal equilibrium with a heat reservoir whose initial state is . We might denote the overall initial state by .
Now, the demon introduces a partition into the gas, separating the enclosure into unequal subvolumes, as in
Figure 2. The two resulting states are
and
, which are not equally probable. The probabilities here are entropic probabilities due to the difference in entropy of
and
.
Now, the demon records the location of the particle in its memory and uses this to control the isothermal expansion of the one-particle gas. The work is stored in the battery. At the end of this process, the demon retains its memory record, and the battery is in one of two mechanical states and . The gas is again in state . But different amounts of heat have been extracted from the reservoir during the expansion, so the reservoir has two different states and .
The overall final eidostate might be represented as
The states of the demon and the gas,
and
, have different energies and the same entropy. It is the reservoir states
and
that (1) make
F uniform (constant energy) and (2) introduce the entropy differences leading to different entropic probabilities for the two states.
A conventional view would suppose that the unequal probabilities for the two final demon states comes from their history—that is, that the probabilities are inherited from the unequal partition of the gas. In the entropic view, the unequal probabilities are due to differences in the environment of the demon, which is represented by the different reservoir states and . The environment, in effect, serves as the “memory” of the history of the process.
7. Context States
When we uniformize a non-uniform A by means of a sequence of reservoir states, the reservoir states affect the entropic probabilities. We can use this idea more generally.
For example, in our coin-and-box model, suppose we flip a coin but do not know how it lands. This might be represented by the eidostate
. Without further information, we would assign the coin states equal probability 1/2, which is the simple entropic probability. But suppose we have additional information about the situation that would lead us to assign probabilities 1/3 and 2/3 to the coin states. This additional information—this
context—must be reflected in the eidostate. The example in Equation (
14) tells us that this does the job:
The extended coin-flip state
includes extra context so that the entropic probability reflects our additional information.
In general, we can adjust our entropic probabilities by incorporating
context states. Suppose we have a uniform eidostate
, but we wish to
specify a particular non-entropic distribution
over these states. Then, for each
, we introduce eidostates
, leading to an extended eidostate
which we assume is uniform. The
values are the context states. Our challenge is to find a set of context states so that the entropic probability in
equals the desired distribution
.
We cannot always do this exactly, but we can always approximate it as closely as we like. First, we note that we can always choose our context eidostates to be information states. The information state
containing
n record states has entropy
. Now, for each
k, we closely approximate the ratio
by a rational number; and since there are finitely many of these numbers, we can represent them using a common denominator. In our approximation,
Now, choose
for each
k. The entropy of
becomes
From this, we find that the entropic probability is
as desired.
We find, therefore, that the introduction of context states allows us to “tune” the entropic probability to approximate any distribution that we like. This is more than a trick. The distribution represents additional implicit information (beyond the mere list of states ), and such additional information must have a physical representation. The context states are that representation.
8. Free Energy
The tools we have developed can lead to some interesting places. Suppose we have two sets of states, and , endowed with a priori probability distributions and , respectively. We wish to know when the states in A can be turned into the states in B, which are perhaps augmented by reservoir states. That is, we wish to know when .
We suppose we have a mechanical state , leading to a ladder of mechanical reservoir states . The mechanical state is non-trivial in the the sense that for any s. This means that there is a component of content Q such that . The set can be uniformized by augmenting the and states by mechanical reservoir states.
However, we still need to realize the
and
probabilities. We do this by introducing as context states a corresponding ladder of reservoir states
such that
is very small. Essentially, we assume that the reservoir states are “fine-grained” enough that we can approximate any positive number by
for some positive or negative integer
n. Then, if we augment the
and
states by combinations of
and
states, we can uniformize
and also tune the entropic probabilities to match the a priori
and
. The final overall uniform eidostate is
for integers
,
,
and
. The uniformized
and
eidostates are subsets of this and thus are themselves uniform eidostates. The entropic probabilities have been adjusted so that
We now choose a component of content
Q such that
. Since the overall state is is uniform, it must be true that
for all choices of
. Of course, if all of these values are the same, we can average them together and obtain
We can write the average change in the
Q-value of the mechanical state as
Since all of the states lie within the same uniform eidostate,
if and only if
—that is,
From this, it follows that
If we substitute this inequality into Equation (
33), we obtain
We can obtain insight into this expression as follows. Given the process ,
is the average increase in Q-value of the mechanical state, which we can call . Intuitively, this might be regarded as the “work” stored in the process.
We can denote the change in the Shannon entropy of the probabilities by . Since each or state could be augmented by a corresponding record state, this is the change in the information entropy of the stored record.
For each state
a, we can define the
free energy . We call this free “energy”, even though
Q does not necessarily represent energy, because of the analogy with the familiar expression
for the Helmholtz free energy in conventional thermodynamics. The average change in the free energy
F is
The free energy F depends on the particular reservoir states only via the ratio . Given this value, depends only on the and states, together with their a priori probabilities.
To return to our coin-and-box example, suppose we use the basic box states as reservoir states , and we choose the coin number Q as our component of content. Then, and , so that the free energy function . (If we use different box states as reservoir states, the ratio is different.)
With these definitions, Equation (
36) becomes
Many useful inferences can be drawn from this. For example, the erasure
Q-cost of one bit of information in the presence of the
-reservoir is
. This cost can be paid from either the mechanical
Q-reservoir state, the average free energy, or from a combination of these. This amounts to a very general version of Landauer’s principle [
12]: one that involves any type of mechanical component of content.