Next Article in Journal
The Quantum Relative Entropy of the Schwarzschild Black Hole and the Area Law
Next Article in Special Issue
Near-Horizon Carnot Engines Beyond Schwarzschild: Exploring Black Brane Thermodynamics
Previous Article in Journal
Sustainability of a Three-Species Predator–Prey Model in Tumor-Immune Dynamics with Periodic Treatment
Previous Article in Special Issue
Three Optima of Thermoelectric Conversion: Insights from the Constant Property Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rethinking Economic Measurement Using Statistical Ensembles

Signal Power and Light, Inc., Cordova, AL 35550, USA
Entropy 2025, 27(3), 265; https://doi.org/10.3390/e27030265
Submission received: 22 December 2024 / Revised: 18 February 2025 / Accepted: 27 February 2025 / Published: 3 March 2025

Abstract

:
The axiomatic framework of quantum game theory gives us a new platform for exploring economics by resolving the foundational problems that have long plagued the expected utility hypothesis. This platform gives us a previously unrecognized tool in economics, the statistical ensemble, which we apply across three distinct economic spheres. We examine choice under uncertainty and find that the Allais paradox disappears. For over seventy years, this paradox has acted as a barrier to investigating human choice by masking actual choice heuristics. We discover a powerful connection between the canonical ensemble and neoclassical economics and demonstrate this connection’s predictive capability by examining income distributions in the United States over 24 years. This model is an astonishingly accurate predictor of economic behavior, using just the income distribution and the total exergy input into the economy. Finally, we examine the ideas of equality of outcome versus equality of opportunity. We show how to formally consider equality of outcome as a Bose–Einstein condensate and how its achievement leads to a corresponding collapse in economic activity. We call this new platform ‘statistical economics’ due to its reliance on statistical ensembles.
JEL Classification:
C01; C51; C70; D01; D63; D81; E01; E17

1. Introduction

Recent work reformulating the axiomatic foundation of game theory [1] gives us a new footing for exploring economics. This work drastically restructures our economic thinking, forcing us to discard entire fields of study as being irrelevant or in direct opposition to a fundamental law of nature, the second law of thermodynamics. Moreover, it shows that we have not been using the correct metrics. While this approach discards many things, it strengthens some familiar conclusions. It does so by giving them a new, more meaningful context and significance, such as formally connecting neoclassical economics to game theory.
Our issue is not the data or tools we do not have. All of the topics in this paper could have been developed under the existing neoclassical and game theoretic frameworks. The data that we employ in our analysis are well-studied and by no means novel. The difference is in how we apply our tools and interpret the data.
The only new concept that we introduce to an economist is that of the statistical ensemble. It is not that this concept was not available; it was, but we did not recognize it. All it took was generalizing game theory and its axioms to lift the scales. Welcome to statistical economics.

1.1. Isomorphisms

The cross-pollination of ideas in economics and the physical sciences has a rich and significant history. This history is a testament to the shared goal of understanding both the worlds of people and things. This concurrent development comes from sharing the same fundamental mathematical structures, isomorphisms, which differ in only superficial properties [2].
Game theory’s new axiomatic basis [1] leverages such an isomorphism. In the 1960s, Pfanzagl [3] developed an axiomatic theory of measurement, a framework that provides a systematic way to assign numbers to objects, cardinality. Pfanzagl used this theory to show the equivalence of Savage’s subjective expected utility to von Neumann–Morgenstern utility. When we look at the axiomatic formulation of quantum mechanics, we see that quantum theory is fundamentally a theory of measurement.
Following Pfanzagl’s methods, QGT proved the functional equivalence between von Neumann entropy and vNM utility [1]. This work follows a long-established trend outside of mainstream economics, attempting to identify entropy’s role in economics (Jakimowicz [2] presents a meticulous summary of the historical prior work.) This effort is motivated by the second law’s attractiveness as an absolute law shaping dynamical behavior.
While an economic entropic isomorphism is clearly needed, it must be precise because there are many different entropies. There is only one second law and only one associated entropy with that law: the ensemble’s entropy (the ensemble entropy is the von Neumann entropy in the quantum context and the Gibbs entropy classically).
Previous applications of entropy in economics did not make this subtle and essential distinction, rendering them fundamentally flawed. Economists’ reluctance to apply isomorphisms without a proper theoretical basis is not incorrect, as prior entropy theories led to erroneous conclusions.
However, this hesitance is not entirely justified either; historical isomorphisms, such as production functions, applied out of empirical necessity were not only insightful but also later explained theoretically. This experience validates the physical science motif of observation first, followed by advancing theory. Experiential primacy is also true neurobiologically [4]; it behooves us not to place the cart before the horse.

1.2. Background of Ensembles in Economics

The formal application of ensembles in econophysics is a relatively recent advancement [5]; however, their application in mainstream economics remains limited due to the absence of an axiomatic justification [2]. Ensembles, or rather some of their metrics, have appeared in the economic literature [6,7], but these have not had any significant development.
Econophysicists seem content with developing isomorphisms, while economists are equally content to ignore these developments [8]. Until QGT proved utility to be the ensemble’s entropy [1], there was no theoretical framework to overcome this impasse. This proof revealed that many had mistakenly identified utility as being isomorphic with energy (This misconception began with Irving Fisher and was cited and even propagated by Jakimowicz [2]. I also held the same misconception until the proof was complete, and even then, I was still skeptical).
This misidentification limited economists’ and econophysicists’ ability to identify and adequately conceptually frame utility into a meaningful metric. Furthermore, entropy’s historical characterization of disorder is not helpful. A better conception of entropy is as a measure of complexity or, as we will find out shortly, potential.
We need to consider both open and closed systems because the behavior in each is opposite regarding the minimization or maximization of entropy, respectively. We can think of the closed thermodynamic system as the world around us or as a competitive game. It is a world where entropy is always maximized. This entropy maximization is the second law of thermodynamics; it is inexorable and unavoidable. An open system extracts the potential of the world around us and creates pockets of lower entropy through work expenditure, which increases global entropy (The behavior of an open system is the same as Maxwell’s daemon. Szilard presented a mechanism of how this daemon acts to increase global entropy, preserving the second law [9]). The open thermodynamic system describes the fundamental aspect of life and, by extension, economics.
Life transforms the chaotic potential of the world into low-entropy things. We must work to create and maintain these pockets of lower entropy (Work, as referred to here, is, by definition, the expenditure of exergy). This conception of entropy is directly contrary to the formulation of entropy economics by Georgescu-Roegen, which postulated that the action of human beings was to take low entropy resources from the environment and transform them into high entropy activities [2]. As a direct result, entropy economics suggests that there are material limits to growth. What we are presenting is that the limits to growth are solely dependent upon our use of energy to maintain the desired state of the world we create.
In recorded history, this pattern of behavior has played out time and time again: the Industrial Revolution, where we harnessed energy long stored in the ground and created means of converting it into a mechanism capable of supporting a population far more extensive than the conception or even belief of Malthus and our increasing ability to extract order from the potential around us, which is why Paul Erlich lost his bet with Julian Simon. Open systems must create and exploit an entropy gradient in order to perpetuate themselves.
If these systems allow themselves to reintegrate with the increasing entropy of chaos, we have, by definition, their death and dissolution. Our open economic ensemble or any other open ensemble (e.g., species) will reach a limit defined by its ability to access energy. Unfortunately, economists have not recognized and/or refused to acknowledge the centrality that energy plays in our lives and life in general [10]. The application of energy in society determines the “extent of the market” ([11] Book 1, Ch. 3), and the measured entropy quantifies the “division of labor”.
The pattern of reducing entropy and then increasing entropy by adding energy is essential for the function of any heat engine. In the economic context, we, as individual participants, work to reduce the potential of the world around us and, through the addition of energy, increase the complexity of our ensemble/society.
We witness this correctness even beyond empirical and theoretical framing. We see it in our societies’ oldest stories, such as the Enūma Eliš or the Book of Genesis. In both stories, the Creator makes Creation out of the chaos of the primordial soup, converting the potential of the world into something structured: in this case, life ([12] pp. 2–3). Genesis goes a step further; God enjoins a member of His Creation to name everything and assert dominion and care, analogous to God’s act of Creation. We can interpret this as God tasking humanity with the stewardship of life and an invitation/commandment to emulate His work of transforming the potential into the useful ([12] p. 7).
Touching on religion is unavoidable because we, as human beings, base our actions and choices on value (The canonical distribution, Equation (3), shows this concept of value-directing action through the Hamiltonian, H ^ , determining choice, ρ ^ ). Language even provides a value structure, restricting and directing our focus from the innumerable chaos of potential to a set of “things” with which we can use ([12] pp. 13–17). We see the world through a story, and the story we choose defines the world and our possible interactions ([12] p. 11). This use of the word value simultaneously contains the numeric representation and the ineffable subjective. What is “good”? Genesis outlines a framework where the continued act of Creation and the continuation of life are“very good”.
So, what direction do we point to as being where we aim in policy and philosophy? The choice is stark; we can choose life or death. Humans can be considered either a flawed embodiment of the good or a cancer on the planet that must be eliminated because of their consumptive tendencies. However, if we take a step back, all life is consumptive; thus, all of life and Creation, in that mindset, should be destroyed. Again, what is “good”?

1.3. What Changed to Now Allow Using Ensembles?

The fingerprints of ensembles have always been with us. Only the current development of econophysics leads the effort to apply ensembles empirically in economics [2,5]. It is interesting to note that it is entirely possible to develop and apply economic ensembles using a purely classical approach, just like Gibbs did in physics. However, economists did not develop ensembles; they practically ignored expected utility, which was a necessary tool in their development.
In physics, ensembles manifest an extensive property known as entropy, S, which, as Boltzmann pointed out, is closely related to the number of possible arrangements or microstates of a system that result in the same macroscopic properties (multiplicity), W (If k = 1 , entropy is dimensionless in its natural unit).
S = k log W .
However, multiplicity is never really considered in economics. Daniel Bernoulli mentioned it about 150 years before Boltzmann [6], but it was never rigorously pursued. Jakimowicz’s extensive review of entropy in economics neither mentions multiplicity nor ensemble entropy [2]. It presents some adjacent notions but does not fully delineate them. This paucity of literature is a significant indication that all previous expositions of entropy in economics do not use the ensemble’s entropy, confusing any subsequent interpretations.
vNM utility presented an avenue to develop the concept of ensembles formally. However, economists effectively ignored/limited the application of classical game theory because of the perceived flaws in the vNM formulation of utility. Morgenstern even noted that economists such as Allais were being short-sighted and that a different axiomatic formulation would render Allais’s and others’ arguments moot [13].
The axiomatic formulation of QGT, the proof of classical game theory as a special case of QGT, and the proof of the equivalence of von Neumann entropy and vNM utility, provided the necessary and sufficient foundation to discard the earlier critiques of game theory and EU [1]. By removing the independence axiom as a foundation of game theory, the Allais paradox has nothing to contradict. Because this paradox is so famous, we will use it as an example of how to show earlier critiques of game theory as irrelevant.

2. Materials and Methods

2.1. What Is an Ensemble?

We are using the word ensemble in what seems like a vague “suitcase” word that has no specific meaning because it can mean anything. This vagueness is due to how we measure/define the ensemble, which determines its behavior. The definition of an ensemble is, however, straightforward; an ensemble is a collection of things on which we can make some set of observations. In the quantum context, the ensemble average of an observable is the trace of the product of the potential observable outcomes, M ^ , multiplied by the probability observable, known as the density matrix, ρ ^ (See reference [1] (Def. 2.9) for the formal definition of a density matrix).
M = error ρ ^ M ^ .
These observables that scale with the ensemble’s size are called extensive (Some observables do not scale with the size of the ensemble. Price is an example of such an observable. This class of observable is called intensive). In the classical context, we observe some distribution of a quantity and then compute the average of that observation.
Some examples of ensembles are those of an individual across time or of a group of individuals at a specific time. Time and value are complementary measures in economics, as are their isomorphs, time and energy, in physics. Due to their complementarity, they obey the uncertainty principle. If we are to know something’s value, we cannot place the time; if we know the specific time, we cannot identify the value.
Some will react that this is an entirely subjective framework. It is; we can only measure and observe in finite time. As individuals, we cannot exist as each possible incarnation of the potential that we embody. We exist in finite time and finite locality. However, when we look at even just moderately sized populations, the ensemble behaves as if it is composed of indistinguishable entities behaving ergodically. For this reason, the assumption of ergodicity has some interesting consequences that may make those of a frequentist persuasion subjectively Bayesian; we as individuals are acting as if extensions of some quantity of “humanness” that is indistinguishable from one another and indistinguishable from other people across time. This interpretation will undoubtedly result in some interesting debate.
One of the more important properties of an ensemble is its distribution, which is straightforwardly derived from QGT’s axioms [1],
ρ ^ = e β H ^ Z [ β ] .
The term H ^ represents the observable of an individual’s or group’s values. The term β is a Lagrangian multiplier used to maximize the ensemble’s entropy/utility given the Hamiltonian. In a game theory setting, β represents a measure of the risk preference of the individual. We generally usually use its inverse, T, which has units of value (energy) and represents a measure of the economic activity of the ensemble. The term Z β is the partition function, the normalization constant, and is a pure functional of β (The partition function is very important; its logarithm represents a thermodynamic potential, but we restrict its use here, as this discussion would expand the scope of this paper. An interested reader should look at [1] (§ 2.e(iii)) or [14] for more detail).
There are many paths to deriving Equation (3), but the most clear and direct is through the integration of the axiom relating the time evolution of states with the Hamiltonian; if we add up the sum of the small choices based on our values, we arrive at the distribution of our actions as individuals and as groups of individuals. This concept also intuitively explains the uncertainty principle and the complementarity of time and value.
We can express different distributions of different ensembles, but that is not necessary for what we need here, although the concept is similar. The eigenvalues of the density matrix are the probabilities of classical game theory and, strictly speaking, represent our observations.
As Jaynes notes, the canonical distribution is a maximum entropy distribution [14]. Equation (3) represents any possible distribution (determined by the Hamiltonian), finite or continuous. Matsoukas had previously generalized and extended this concept classically [15].

2.2. The Distinction Between Classical and Quantum

One of the isomorphisms that quantum mechanics borrowed from economics is that the collapse of the wave function can be thought of as if the quantum object “decides” to be in one state or another based on the Hamiltonian. The quantum superposition represents an “undecided” state before a choice. This is a difficult anthropomorphism to ascribe to perceptibly inanimate objects. However, when we apply this analogy to clearly thinking and reasoning entities, we clearly distinguish between the interpretation of the quantum and classical.
The quantum state represents the undecided state, where we hold each of the possibilities of an outcome in a superimposed state. When we make a decision and subsequently demonstrate, make observable, that decision through action, we provide direct, measurable evidence of the collapsed choice. We can see why Rothbard insisted so stringently on the primacy of demonstrated preference.
This restriction severely limits us as experimentalists. We can only observe what we can measure; we only know the people we watch by the fruits of their choices. Thus, just like in quantum mechanics, we can only infer the structure of the Hamiltonian, the values of a person. Their entanglements influence their decisions and, thus, correspondingly, what we observe.
A way of thinking about entanglements is as relationships developed through repeated past interactions. One example of this would be a husband and a wife acting predictably as individuals with shared values of the couple. The Tit-for-Tat solution to the prisoner’s dilemma is another example of how such entanglements can form in competitive games. In a business setting, a contract represents an agreed-upon structure of entanglement that excludes Nash equilibria to hopefully some de minimus probability.
In a society, evolved norms provide predictability to the behavior of its members and a means of identifying the group as a whole from other groups that have a different set of values. Suppose we mix two ensembles with drastically different values. In that case, they are either compatible and work in concert or incompatible, where one takes advantage of the other based on the differences in the value operators.

2.3. Thermodynamics

The direct consequence of using an ensemble approach is that we describe the time evolution of the ensemble’s distribution through a set of differential equations or directly through the Euler equation. This methodology is no different than that of neoclassical economics. What is made explicit is that these representations describe a distribution of the observed parameter and that without the inclusion of the distribution’s entropy, those relationships are fundamentally incomplete.
There is no fundamental difference in the equations that we derive from an ensemble compared to those long familiar with neoclassical economics. What is different is what we are measuring and the explicit inclusion of the ensemble’s entropy.
In mainstream economics, the concern is over attempting to describe an individual’s utility function. Utility, while maximized, is not what motivates people; their values motivate them. Entropy/utility maximization drives the process; it is the “Invisible Hand” (Yakovenko illustrates exactly the process of entropy maximization that Smith is describing as the function of the invisible hand [5] (III.C)):
They are led by an Invisible Hand to make nearly the same distribution of the necessaries of life, which would have been made, had the earth been divided into equal portions among all its inhabitants [16].
(Part 4, Ch. 1)
Furthermore, the open system maximization of each individual’s values leads to the wealth of nations.
By preferring the support of domestic to that of foreign industry, he intends only his own security; and by directing that industry in such a manner as its produce may be of the greatest value, he intends only his own gain, and he is in this, as in many other cases, led by an Invisible Hand to promote an end which was no part of his intention [11].
(Book 4, Ch. 2)
We have known these phenomena for a long time but have been unaware of the underlying mathematical structure. Value, not utility, motivates choice; utility is the exploratory force that drives choice. Thus, our primary concern is attempting, however imperfectly, to discern others’ values, their Hamiltonian.

3. Results

Now that we have our tools, we can start to work on their application.

3.1. The Allais Paradox

List and Haigh conducted an interesting experiment looking at how professional risk takers, traders on the Chicago Board of Trade, evaluated risk compared to that of a control cohort of college students [17]. This difference in experience is a fortunate distinction for us to consider because we can evaluate the ability of each cohort to discern risk and select the optimal (for them) outcome. In the simple games, the students and traders behaved consistently, following nearly identical strategies. There was a difference in risk evaluation in the more complex composite game. The traders correctly identified the joint game as a composite game, whereas the students naively assumed that the games were independent.
We use the conventional ket, | i (Paul Dirac developed a compact and powerful notation to express the vectors and operators in quantum theory. Readers should familiarize themselves with their application, as we will be using Dirac’s notation throughout this paper.), with a numeral, i, to represent each orthonormal choice, called a basis, that List and Haigh presented to the students and traders:
  • | 1 Win $7 with certainty.
  • | 2 Win $7 with 75% chance, $10 with 20% chance, and $0 with 5% chance.
  • | 3 Win $7 with 25% chance and $0 with 75% chance.
  • | 4 Win $10 with 20% chance and $0 with 80% chance.
The first game, A, was the choice between | 1 and | 2 . The second game, B, was the choice between | 3 and | 4 . There was a final compound game, A B , of having to choose one outcome from each game and be paid the outcome of a sub-lottery between those two games with even probability. The compound game had the following choice sets | 1 , 3 , | 1 , 4 , | 2 , 3 , and | 2 , 4 .
We assume the choices of each game are independent of each other. Thus, the payout matrix, Hamiltonian for each game is H ^ A , H ^ B , H ^ A B for the first, second, and joint games, respectively:
H ^ A = 7 7.25 , H ^ B = 1.75 2 , and H ^ A B = 4.375 4.5 4.5 4.625 .
The observed data for the outcomes of each of the games for each cohort are in Table 1.
Using the Hamiltonians, the empirical density matrices, and the number of players N, we find the empirical entropies s , expected payout e , and through a simple regression β for each game as in Table 2 (We are following the convention of the total entropy as being S = N s , where s is the specific entropy of the individual; similarly, we have E = N e . Also, note that the units of β are the inverse units of the expected payout).
The “Canonical” for game A B in Table 2 represents the situation where the traders acted optimally for their estimated risk preference, β . The absence of the β for the students is that their strategy resulted in a much more complex Hamiltonian than can be expressed here. The student strategy in the compound game was
ρ ^ A B = ρ ^ A ρ ^ B
with a goodness of fit of χ 2 ( 3 , N = 30 ) = 1.20 , p = 0.78 using the empirical density matrices of the first cohort of students. This model suggests that the students treated the games as being distinct and not as an integrated composite. Furthermore, we can also conclude that the traders, who evaluate risk professionally, were able to consider the composite game as a composite and get close to an optimal outcome for their observed risk preference.
If we wanted, we could investigate why people prefer the “sure thing” slightly more than the optimal. This sure thing preference is significantly more telling for the students than the traders, but the effect clearly remains for traders, even showing up in the composite game. More work is needed to explore the Hamiltonians’ off-diagonal components and explain the “sure thing” preference. The preference of the “sure thing” also appears in neuroscience, where people act to limit choice (entropy) to be settled and that too much choice can induce anxiety [18].
Because we are dealing with a single group of people for each cohort, their risk preference should be the same from game to game. Because it is not, we have some other factor, likely the “sure thing”, modifying their preference. It is interesting to note that the risk preferences for the traders for Game B and A B are very close, giving credence to this line of thought.
We clearly do not have a paradox when using statistical economics. Instead, we uncovered that
1.
Novices use simple heuristics when confronted with complex games.
2.
People have a quantifiable preference for the “sure thing”.
How long have we been ignorant of being able to quantify these effects, and to what other insights have we made ourselves blind due to a frivolous and pedantic argument?

3.2. Econometric Analysis

Next, we turn to applying our theory to the macro scale and look at large populations outside of carefully constructed experiments, people acting in the wild, so to speak. Before we begin, let us take a step back and frame what we are going to be doing in the context of existing economic thought. As we noted earlier, Daniel Bernoulli [6] long ago proposed that the utility functional was logarithmic. In macroeconomics, there is a power law relationship known as the Cobb–Douglass production function,
Y = A L β K α .
where Y is the total economic output, A is the total factor productivity, also referred to as the Solow residual, L is labor input, K is the capital input, and α and β are Lagrangian multipliers. The connection between the Cobb–Douglas production function and Bernoulli’s utility is that they both consider total productivity as being proportional to utility.
In physics, we see a very similar equation as the fundamental equation of state describing an ideal gas [19] (§3-4). This connection is beyond mere coincidence. An ideal gas and people share the same underlying stochastic process; they are canonically distributed, as in Equation (3). The continuous distribution with a finite mean and the highest entropy is the exponential distribution, which is canonical. Energy in an ideal gas is distributed exponentially amongst the component atoms, just like income is distributed amongst people [20]. The canonical distribution of income does not just occur in the United States; it also holds, when studied, in 67 other countries [21]. Thus, it is clear that the correct interpretation of income distributions is as statistical ensembles.

3.2.1. Data

The next task is to determine which parameters to use in our model. We begin with the income data from the Internal Revenue Service from the 24-year period 1996–2019, as reported in Publication 1304, Table 1.1 [22]. The IRS data directly give us the total income and the number of taxpayers (There are a number of different metrics for income. We will use Adjusted Gross Income as our measure of income. AGI allows us to isolate the business activity of Limited Liability Companies, partnerships, or privately held companies with income reported on an individual’s K-1. The residual profit from these activities will be considered residual income of ownership, which AGI represents). The income distribution has income bins of different income levels over the years. To estimate the income distribution f ( m ) , we will use the mixed exponential–Pareto distribution of Banerjee and Yakovenko [20],
f ( m ; T , m 0 , α ) = e m 0 T arctan m m 0 Z 1 + m m 0 2 α .
Equation (7) is then integrated over the bin width to estimate the hyper-parameters: T —the monetary temperature of the thermal portion, Z—the partition function, m 0 —the thermal–epithermal crossover income, and α —the Pareto exponent using HMC with NUTS in Stan in the R statistical analysis framework. Since this is a relatively straightforward parameter estimation and the results differ little from previous work [20], we will not cover the specifics of the model or its development. The hyperparameters of Equation (7) are in Table A1. The ensemble’s entropy is then computed numerically for each year’s income distribution.
Ayers and Warr performed some extensive econometric work studying the impact of energy on the economy, specifically the Solow residual [10]. They showed that the Solow residual (which represents roughly 80% of GDP growth) could be explained entirely by the exergetic input into the economy (Exergy is a technical term that describes useful work extracted from a heat engine. Since all of our work comes from heat engines, exergy represents the measure of work done in/by the economy). Their work seems to have fallen on deaf ears in the mainstream economic community, as economists still treat energy as just another commodity instead of being central to economic activity [10] (p. 224). The entirely predictable economic collapse of Germany due to Energiewende is directly attributable to their artificial restrictions on energy supply. As Ayers and Warr note, making energy expensive reduces energy consumption, and all goods and services that depend on energy are reduced [10] (pp. xxi, 217–218).
To obtain the needed data to estimate the total exergy consumed each year, we used the Energy Information Agency’s Open Data API [23]. The data sets used in the analysis were the total energy consumed in a given year for the Residential (TERCBUS), Industrial (TEICBUS), Commercial (TECCBUS), and Transportation (TEACBUS) sectors. We multiplied the sectors’ consumed energy by the sector’s conversion efficiency: 65%, 49%, 65%, and 21%, respectively, to determine the exergy consumed by each sector [24].

3.2.2. Model

The extensive parameters for the ensemble of tax-paying individuals are the size of the ensemble N from the total number of taxpayers, the total income M, the ensemble’s specific entropy s, the partition function Z, and E, the total exergy input into the economy. It is paramount for us to understand the metrics we have at our disposal, as our understanding determines how we interpret them and their implications. By employing analogs to the equations of state for an ideal gas, we create a clear connection between income distributions and thermodynamic concepts. It is worth noting that while some of these relationships may be familiar to an economist, the novelty lies in their derivations.
The first hurdle we must overcome is grasping how to view the income distribution, Equation (7), and understanding its representation in the thermodynamic concept. Taking our income distribution, our first estimation of its Euler equation is (The use of the word estimation is because the selection functional (see [15] (§3.8.1)) was not formally determined; thus, the terms in Equation (8) do not have an exact correspondence to those in Equation (7). They have a conceptual equivalence but are not mathematically precise.),
S = β M M + N log Z .
The physics envy that economists have undeniably had with such accusations of “determinism” undoubtedly stems from a lack of understanding of what the physical sciences do. Equation (8) does not represent a physical process per se; it represents, completely, the distribution of some measurable ensemble quantity. Equation (8) is an equivalent mathematical representation of the observed income distribution. Thermodynamics, entropy, and the second law are not physical quantities, per se; they are stochastic properties that have physical manifestations and consequences.
Because we have not formally determined the selection functional for Equation (7), T T m , where from Equation (8), β m 1 / T m . While not precisely matching the actual partition function, we will take the regressor Z from Equation (7) as being sufficiently representative of the actual partition function. We can then rearrange Equation (8) as
T m = m s log Z .
where m is the average income and s is the numerically estimated Shannon entropy of Equation (7) and Table A1.
Next, we need to consider the exergy available to the ensemble. Because exergy is closely related to energy, we will place it in the same context as energy in statistical physics. This connection gives us an isomorphism between value and exergy. Because we must expend exergy and time in everything that we do, the measure of the time and exergy that we spend gives us a measure of our value of anything. Here, we present a purely subjective argument of value, but one that is balanced by an objective measure. For this reason, we consider our ensemble a single-commodity economy, money, where the total activity in the economy is given by exergy. Let us think about the model more conventionally. In this case, the exponential part of the distribution represents the payments to Labor, L of the Cobb–Douglas production function. The Pareto portion similarly represents the payments to Capital, K. These traditional economic components are still in the model but are incorporated drastically differently into the income distribution.
There is one item left to consider: What is the thermodynamic analog to money? Money is analogous to volume in thermodynamics. Money is the space in which economic activity occurs. We make the distinction here that it is not the total money supply that matters but rather the money available to an individual from their income. People ultimately own all companies. Thus, the individual realizes any profit that accrues to the company as income. As a result, we ignore companies’ paper value (stock price) and only consider them based on their dividends (realized gains).
Furthermore, we ignore individuals’ ownership of companies outside of the tax base and their realization of gains. This restriction balances the income earned from ownership of foreign companies (non-dollar-based). As a result, we consider this unknown error small and neglect it. Individual income represents the entire product of the economy realized by its constituent members. We can add other commodities to our model and treat them similarly to money; they increase the dimensionality of our analysis’s scalar space.
Since entropy is utility [1], we will consider our model of the economy as following Bernoulli’s logarithmic utility and be of a similar form to the Cobb–Douglas production function and that of an ideal (canonically distributed like income) gas [19] (eq 3.38),
s = s 0 + c m R log e e 0 + R log m m 0 .
We can use Equation (10) to derive two important equations of state (see [19] (§3-4)): the ideal money equation,
P m = R T ;
and a description of the internal action of an economy,
e = c m R T .
where P is the marginal value of money, m is the average individual income, R is the ideal money constant, T is the economic temperature (a measure of economic activity), e is the average exergy used by an individual, c m is the specific exergy capacity for a constant money supply.
Recognizing that β m = P / T and P = T / T m , we can express Equation (11) as
m = R T m .
which, when we perform a regression using the T m computed from Equation (9), we find the ideal money constant as 1.138 [ 1 / person ] ± 0.002 with χ 2 ( 1 , N = 23 ) = 559.9 ,   p < 2.2 × 10 16 from a simple linear regression. Figure 1 shows the average income versus the computed monetary temperature using Equation (9). The fitted model is in red.
Ideally, we would have some measure of the economic activity, its temperature, and perform a similar regression as the ideal money constant with Equation (12). However, this is not the situation that we face. Therefore, we use R and subtract the contribution of money to the ensemble’s utility,
s s 0 R log m m 0 = c m R log e e 0 .
From Equation (14), c m = 0.98 ± 0.08 , with significance of χ 2 ( 1 , N = 23 ) = 11.93 , p = 2.5 × 10 11 using s 0 = s 1996 , e 0 = e 1996 , and m 0 = m 1996 from Table A2. With parameters for Equation (10) determined, we can compute the remaining economic parameters; see Table A2.
We can test this theory against the data in an interesting manner. We recall from thermodynamics that we can define the ratio of the specific heats of an ideal gas as being
γ 1 + R c m .
For our ensemble, γ = 2.16 . There are a number of different ways that the volume of a system can be expanded or shrunk. In general, we can model these through what is known as a polytropic process. We can describe this process as being
P m n = C
where n is the polytropic coefficient and C is a constant. We find that the US economy over the study period underwent a polytropic expansion, with n = 1.31 with significance of χ 2 ( 1 , N = 22 ) = 36.66 , p < 2.2 × 10 16 . Figure 2 shows the data points, and the red line is the fit of Equation (16), the dollar’s demand curve, with elasticity n .
Based on 1 < n < γ and the fact that the income is generally expanding, we conclude that the economy has work extracted from it through the expansion of the money supply and that there is a net energy inflow into the economy. Where the extracted work is going remains an open question. It is entirely possible that because money creation is relegated to a small group, expanding the money supply allows them to purchase additional assets before the value has been removed from the money. These asset purchases then inflate prices. As the money diffuses farther into the economy, its value decreases until all currency has a marginally lower value.

4. Discussion

For this final part, we focus on the economic issue of income inequality, the nature of income distributions, and the concept of what is “fair”. For this discussion and acknowledgment of the centrality of energy in economic activity in the previous section, we represent the Hamiltonian as being the available exergy, E ^ , modifying Equation (3),
ρ ^ = e β E ^ Z [ β ] .
There are two general notions of equality:
1.
Equality of opportunity.
2.
Equality of outcome.

4.1. Equality of Opportunity

Addressing the equality of opportunity first, for a given state, each member of the ensemble has a 1 / N probability of being in a given state. Because the available energy for the ensemble is finite, the resulting maximum entropy distribution is the canonical distribution, Equation (17), where the eigenvalues of the Hamiltonian equate with the value of each state. This distribution is the exponential distribution in the continuous case. It is universal. In any case, restrictions are placed on the occupancy of the outcomes (policy entanglements), or money is taken from one and given to another, reducing the ensemble’s entropy (utility). A good way of thinking about entropy/utility is as complexity. In this context, entropy represents a quantitative measure of liberty or freedom.
As demonstrated by our society, we are allowed to own (control) property; we can accumulate wealth. Thus, the amount of wealth we have accumulated determines the amount of income (rent) we have. At its heart, rental income is a Markovian process in which each step determines the probability of the subsequent steps. The maximum entropy distribution for a continuous Markov process is the Pareto distribution, explaining why the upper portion of the income distribution is Pareto. Thus, even in this case, we still have equality of opportunity.
For a pure thermal society (a society in thermal equilibrium), the Gini coefficient will be at the low end, 0.5. For a pure epithermal society (a society where everyone is part of the Pareto distribution), if we had a Pareto exponent of 1.3, this would be a Gini of 0.625. In a society that is in the thermal equilibrium of the equality of opportunity, the minimum Gini is 0.5 and has the maximal distribution of wealth possible.

4.2. Equality of Outcome

If we look at the construction of the Gini Index, the line of perfect equality occurs when all the individuals in society have the same outcome. The line of equality has long been argued for as a desirable state, primarily from socialism, with a wide variety of different policies to enact the ideal state of wealth distribution. What does the mathematical conception of this ideal state look like?
First, since we are dealing with human beings living in a reality governed by the maximization of entropy, wealth is always canonically distributed. As previously discussed, restrictions on wealth (policy entanglements) will only lower the ensemble’s entropy. We cannot overlook the fact that the ensemble’s entropy always seeks its maximum subject to its constraints, even policy constraints. These structural impositions can lower the Gini coefficient. Any Gini < 0.5 has such policy entanglements; there will always be some level of inequality regardless of the policy restrictions.
Let us then ask if there is a theoretical situation or condition that could lead to the ideal of perfect equality. Fortunately for us, there is such a theoretical limit. Because there are quantum effects in our interactions, there is a theoretical state called a Bose–Einstein condensate (The members of a Bose–Einstein condensates can each occupy the same state (outcome). The Bose–Einstein statistics are usually derived from the Grand Canonical ensemble as a direct result because these ensembles exhibit symmetric behavior through permutation, P ^ i j | ψ = | ψ , where P ^ i j is the permutation operator and | ψ is a pure strategy. If the permutation is antisymmetric, P ^ i j | ψ = | ψ , then Fermi–Dirac statistics are followed. Games that would exhibit antisymmetric behavior are games where one can only win or loose and not share the same outcome, like sporting events). This condensate only occurs at very low energy states where the β of Equation (17) becomes very very large. In this situation, people will all occupy the same or very few states and have perfect equality of outcome. The unfortunate consequence of this is that there is little, if any, economic activity. Remember, T is a measure of economic activity; the hotter the temperature, the more activity there is. Conversely, if T is lower, the economy is colder with less activity; as β = 1 / T , a smaller T is a larger β .
Thus, theoretically and in practice, any policy that shifts from equality of opportunity to equality of outcome always reduces economic activity, and people at the societal level are always slightly worse off in proportion to the degree of the restriction. Furthermore, such outcomes only occur at extremely low levels of economic activity, e.g., in hunter-gatherer societies where there is minimal variation in roles (See the earlier discussion on the division of labor). If it is claimed that such equality can be achieved without extremely low economic activity, this would be contrary to the second law of thermodynamics. As Arthur Eddington famously said,
The law that entropy always increases holds, I think, the supreme position among the laws of Nature. If someone points out to you that your pet theory of the universe is in disagreement with Maxwell’s equations—then so much the worse for Maxwell’s equations. If it is found to be contradicted by observation—well, these experimentalists do bungle things sometimes. But if your theory is found to be against the Second Law of Thermodynamics, I can give you no hope; there is nothing for it to collapse in deepest humiliation [25].

5. Conclusions

By using the statistical economic framework, we have gained insights into both individual human choice and collective human action. The canonical ensemble has proved to be a handy and powerful tool for simplifying and clarifying our understanding of our interactions and our behavior. Hopefully, the two empirical examples provided show, in sufficient detail, how to formally use statistical economics as an analytical framework and how to think about and approach ensembles. We have even been able to resolve a centuries-long debate about income inequality just by studying the canonical distribution with a simple thought experiment and study of it.
This paper will hopefully prompt a more critical look into statistical economics, one that is capable of a substantive critique. Theories do not become better without being challenged. That is why this paper is a challenge to rethink our approach to economics.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data and all computer code are available publicly on GitHub: 1. For the US macroeconomic analysis see, https://github.com/crabel99/US-Entropic-Analysis, accessed on 26 February 2025. 2. For the analysis of the List and Haigh data see, https://github.com/crabel99/Allais-Analysis, accessed on 26 February 2025.

Conflicts of Interest

Author Cal Abel is employed by Signal Power and Light, Inc. and declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AGIAdjusted Gross Income
APIApplication Programming Interface
EIAEnergy Information Agency
EUExpected Utility Hypothesis
HMCHamiltonian Monte Carlo
IRSInternal Revenue Service
LLCLimited Liability Company
NUTSNo U-Turn Sampling
QGTQuantum Game Theory
SEUSubjective Expected Utility
USUnited States
vNMvon Neumann—Morgenstern

Appendix A

Table A1. Regression coefficients of the United States income tax distribution for the mixed exponential–Pareto distribution for 1996–2019.
Table A1. Regression coefficients of the United States income tax distribution for the mixed exponential–Pareto distribution for 1996–2019.
Year T [ $ ] m 0 [ $ ] α Z
199633,424.95120,921.41.19778932,783.57
199735,533.71127,040.71.13722234,999.70
199842,384.76118,780.61.27233140,485.83
199939,053.37135,461.61.10525738,521.05
200039,552.88129,135.11.19541838,577.32
200140,889.55131,838.81.30968339,434.20
200241,338.96133,736.01.36343839,692.01
200339,862.09123,672.21.34652738,181.91
200441,206.12126,851.51.27380639,714.94
200542,661.85128,592.31.22706441,236.77
200644,281.24130,389.21.22310342,733.78
200745,653.34135,186.91.20742144,153.72
200846,083.54142,474.81.25160344,528.59
200945,865.92146,023.11.33245444,095.15
201046,247.19147,527.91.27767444,690.98
201147,286.28144,684.11.30261445,422.53
201248,434.39144,688.41.23166346,758.26
201350,636.73149,014.31.32291848,356.20
201452,275.46150,729.51.28978949,987.84
201553,713.46156,673.11.27811051,487.21
201655,070.56159,914.61.31023552,585.58
201757,066.31164,395.71.28226854,608.08
201859,700.10170,770.01.27997257,100.77
201961,345.95169,178.01.32151458,167.35
Table A2. Economic variables from the canonical production function that was derived from the income distribution and estimates of exergy usage in the United States for 1996–2019.
Table A2. Economic variables from the canonical production function that was derived from the income distribution and estimates of exergy usage in the United States for 1996–2019.
YearP [GJ/k$]m [k$/person]N [people]e [GJ/person]T [GJ]s [1/person]
199610.46714637.68948120,351,210392.1858352.344411.51733
19979.68738440.59687122,421,993387.1879347.854111.59368
19989.05667343.40742124,770,661380.8353342.146911.75771
19998.46394646.07878127,075,147380.2339341.606611.70065
20007.95691849.20155129,373,502382.9225344.022111.69841
20017.86912747.37317130,255,240369.9594332.375911.70397
20028.11158946.38492130,076,442376.23338.009411.70206
20037.97344847.5919130,423,630376.7065338.437611.67136
20047.50880251.34242132,226,043378.2977339.867111.72381
20056.94543455.23813134,372,680372.5771334.727711.77325
20066.35821258.02852138,394,756356.8393320.588611.81362
20076.00493560.76228142,978,808352.2324316.449711.84845
20086.14904758.0051142,450,569349.0134313.557811.8414
20096.23869254.28291140,494,129336.5923302.398511.814
20106.16297656.61016142,892,054344.2103309.242611.83573
20115.92543257.60562145,370,240336.1274301.980811.85409
20125.36313862.7905144,928,473326.8532293.648711.89954
20135.47427361.71394147,351,299332.075298.340111.91875
20145.20273765.75103148,606,578333.6133299.722111.96088
20154.90771567.84563150,493,262324.588291.613711.9909
20164.85074868.04946150,272,156323.2772290.43612.00673
20174.53038572.00567152,903,232317.7014285.426712.05083
20184.47387475.71772153,774,296329.4796296.008312.09697
20194.29129675.83724157,796,805317.3104285.075412.11267

References

  1. Abel, C. The Quantum Foundations of Utility and Value. Phil. Trans. R. Soc. A 2023, 381, 20220286. [Google Scholar] [CrossRef] [PubMed]
  2. Jakimowicz, A. The Role of Entropy in the Development of Economics. Entropy 2020, 22, 452. [Google Scholar] [CrossRef] [PubMed]
  3. Pfanzagl, J. Theory of Measurement; John Wiley and Sons: New York, NY, USA, 1968. [Google Scholar]
  4. McGilchrist, I. The Master and His Emissary: The Divided Brain and the Making of the Western World; Yale University Press: New Haven, CT, USA, 2009. [Google Scholar]
  5. Yakovenko, V.M. Econophysics, Statistical Mechanics Approach to. In Encyclopedia of Complexity and Systems Science; Meyers, R.A., Ed.; Springer: New York, NY, USA, 2009; pp. 2800–2826. [Google Scholar] [CrossRef]
  6. Bernoulli, D. Exposition of a New Theory on the Measurement of Risk. Econometrica 1954, 22, 23–36. [Google Scholar] [CrossRef]
  7. Savage, L.J. The Foundations of Statistics, 2nd ed.; Dover Publications: New York, NY, USA, 1954. [Google Scholar]
  8. Stanley, E.H. Interview with Eugene H. Stanley. IIM Kozhikode Soc. Manag. Rev. 2013, 2, 73–78. [Google Scholar] [CrossRef]
  9. Szilard, L. On entropy reduction in a thermodynamic system by interference by intelligent subjects [NASA TT F-16723]. Zhurnal Phys. 1976, 53, 840–856. [Google Scholar]
  10. Ayers, R.U.; Warr, B. The Economic Growth Engine: How Energy and Work Drive Material Prosperity; International Institute for Applied Systems Analysis: Northhampton, MA, USA, 2009. [Google Scholar]
  11. Smith, A. The Wealth of Nations; W. Strahan and T. Cadell: London, UK, 1776. [Google Scholar]
  12. Peterson, J.B. We Who Wrestle with God: Perceptions of the Divine; Penguin: London, UK, 2024. [Google Scholar]
  13. Morgenstern, O. Some Reflections on Utility. In Expected Utility Hypothesis and the Allais Paradox: Contemporary Discussions of Decisions Under Uncertainty with Allais’ Rejoinder.; D. Reidel Publishing Company: Boston, MA, USA, 1979; pp. 175–183. [Google Scholar]
  14. Jaynes, E.T. Information Theory and Statistical Mechanics. Phys. Rev. 1957, 106, 620–630. [Google Scholar] [CrossRef]
  15. Matsoukas, T. Generalized Statistical Thermodynamics; Springer: Cham, Switzerland, 2018. [Google Scholar] [CrossRef]
  16. Smith, A. The Theory of Moral Sentiments; Andrew Millar: London, UK, 1759. [Google Scholar]
  17. List, J.A.; Haigh, M.S. A simple test of expected utility theory using professional traders. Proc. Natl. Acad. Sci. USA 2005, 102, 945–948. [Google Scholar] [CrossRef] [PubMed]
  18. Friston, K.; Schwartenbeck, P.; FitzGerald, T.; Moutoussis, M.; Behrens, T.; Dolan, R.J. The anatomy of choice: Active inference and agency. Front. Hum. Neurosci. 2013, 7, 598. [Google Scholar] [CrossRef] [PubMed]
  19. Callen, H.B. Thermodynamics and an Introduction to Thermostatistics, 2nd ed.; John Wiley & Sons: New York, NY, USA, 1985. [Google Scholar]
  20. Banerjee, A.; Yakovenko, V.M. Universal patterns of inequality. New J. Phys. 2010, 12, 1–25. [Google Scholar] [CrossRef]
  21. Tao, Y.; Wu, X.; Zhou, T.; Yan, W.; Huang, Y.; Yu, H.; Mondal, B.; Yakovenko, V.M. Exponential structure of income inequality: Evidence from 67 countries. J. Econ. Interact. Coord. 2019, 14, 345–376. [Google Scholar] [CrossRef]
  22. Internal Revenue Service. IRS SOI Tax Stats—Individual Statistical Tables by Size of Adjusted Gross Income, Individual Complete Report (Publication 1304), Table 1.1. United States Treasury Department 1996–2019. Available online: https://www.irs.gov/statistics/soi-tax-stats-individual-statistical-tables-by-size-of-adjusted-gross-income (accessed on 14 March 2023).
  23. Energy Information Agency. Open Data API. 2022. Available online: https://api.eia.gov/ (accessed on 14 March 2023).
  24. Lawrence Livermore National Laboratory. Energy Flow Charts: United States. 2022. Available online: https://flowcharts.llnl.gov/commodities/energy (accessed on 14 March 2023).
  25. Eddington, A.S. The Nature of the Physical World: Gifford Lectures; Cambridge University Press: Cambridge, UK, 1927; Volume 74. [Google Scholar]
Figure 1. Plot of the average income versus the computed monetary temperature.
Figure 1. Plot of the average income versus the computed monetary temperature.
Entropy 27 00265 g001
Figure 2. Polytropic expansion of the United States economy 1996–2019.
Figure 2. Polytropic expansion of the United States economy 1996–2019.
Entropy 27 00265 g002
Table 1. The reformatted results of List and Haigh [17].
Table 1. The reformatted results of List and Haigh [17].
GameOccupancyBasisCohort
A15, 17 | 1 , | 2 Student
A8, 19 | 1 , | 2 Trader
B8, 24 | 3 , | 4 Student
B3, 24 | 3 , | 4 Trader
A B 4, 13, 3, 10 | 1 , 3 , | 1 , 4 , | 2 , 3 , | 2 , 4 Student
A B 8, 7, 9, 30 | 1 , 3 , | 1 , 4 , | 2 , 3 , | 2 , 4 Trader
Table 2. Ensemble specific parameters calculated from List and Haigh [17].
Table 2. Ensemble specific parameters calculated from List and Haigh [17].
GameCohortN s e β [ $ 1 ]
AStudent32 0.691 $ 7.13 0.500
ATrader27 0.608 $ 7.18 3.46
BStudent32 0.562 $ 1.94 4.39
BTrader27 0.349 $ 1.97 8.32
A B Student30 1.23 $ 4.53
A B Trader54 1.17 $ 4.55 8.56
A B Canonical54 1.14 $ 4.56 8.56
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Abel, C. Rethinking Economic Measurement Using Statistical Ensembles. Entropy 2025, 27, 265. https://doi.org/10.3390/e27030265

AMA Style

Abel C. Rethinking Economic Measurement Using Statistical Ensembles. Entropy. 2025; 27(3):265. https://doi.org/10.3390/e27030265

Chicago/Turabian Style

Abel, Cal. 2025. "Rethinking Economic Measurement Using Statistical Ensembles" Entropy 27, no. 3: 265. https://doi.org/10.3390/e27030265

APA Style

Abel, C. (2025). Rethinking Economic Measurement Using Statistical Ensembles. Entropy, 27(3), 265. https://doi.org/10.3390/e27030265

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop