Open Problems within Nonextensive Statistical Mechanics

Nonextensive statistical mechanics has developed into an important framework for modeling the thermodynamics of complex systems and the information of complex signals. To mark the 80th birthday of the field’s founder, Constantino Tsallis, a review of open problems that can stimulate future research is provided. Over the thirty-year development of NSM, a variety of criticisms have been published ranging from questions about the justification for generalizing the entropy function to the interpretation of the generalizing parameter q. While these criticisms have been addressed in the past and the breadth of applications has demonstrated the utility of the NSM methodologies, this review provides insights into how the field can continue to improve the understanding and application of complex system models. The review starts by grounding q-statistics within scale-shape distributions and then frames a series of open problems for investigation. The open problems include using the degrees of freedom to quantify the difference between entropy and its generalization, clarifying the physical interpretation of the parameter q, improving the definition of the generalized product using multidimensional analysis, defining a generalized Fourier transform applicable to signal processing applications, and re-examining the normalization of nonextensive entropy. This review concludes with a proposal that the shape parameter is a candidate for defining the statistical complexity of a system.


Introduction
Nonextensive Statistical Mechanics (NSM) [1][2][3] has developed into an important framework for modeling the thermodynamics of complex systems [4][5][6] and the information of complex signals [7][8][9].The methodology ties together heavy-tailed statistics derived from a generalized entropy function and the resultant analysis, modeling, and design methods for systems impacted by nonlinear dynamics.Upon the 80 th birthday of the field's founder, Constantino Tsallis, I reflect on open problems that will stimulate future investigation and development of NSM.While there is much to celebrate in the applications of NSM, a review of open problems requires examination of some of the criticism [10][11][12][13][14][15] the field has received over its thirty-year development [16].The criticism ranges from questions about the interpretation of the generalizing parameter q to the justification for modifying the entropy function.In this paper, I will carefully examine several key concerns with a perspective of motivating further improvement and applicability of NSM.
Inspection of a few applications of NSM introduces the challenges of characterizing the properties of complex systems using q-statistics.Table 1 lists seven examples in which a theoretical underpinning is available to explain experimental observations.However, in each case, the mapping between the physical phenomena and the parameter q requires unexplained constants that detract from the ability of NSM to describe the physics of those systems.The relationship  = 1 +  in which x (or its inverse) is a physical property is common since x often defines a property of the system that induces nonlinearity.This property goes to zero (or infinity for inverse) when  = 1.Another typical relationship is  = 2 −  since this is the reflection about  = 1.If NSM was defined in terms of the physical property x, the reflection about 0 would simply be  = −.However, the full review of open problems in NSM will show that this simple translation is not adequate to account for multidimensional systems and the effects of other nonlinear elements.
Table 1: Applications of Nonextensive Statistical Mechanics.A variety of complex systems, such as atomic gases, space plasma velocities, financial volatility, cellular mobility, wavelets, and heat baths can be modeled using NSM.In each case, the mapping between the physical property and the parameter q requires numerical constants that diminish the ability of q-statistics to describe the physical phenomena.

Applications Physical Property
Relation to q Entropy of Hydrogen Atoms [17] M, number of atoms  = 1 + 1  * Space Plasma Velocities [18,19]  = , spectral index  = 1 + 1  * Volatility of Financial Markets [7,20] , nonlinear Fokker-Plank  = 2 −  Hydra-Cell Velocity [21,22] , nonlinear Kramers Equ  = 2 −  Wavelet Analysis [23] i, wavelet scale index  = 1 − 2 Heat Bath Thermodynamics [24,25] n, degrees of freedom,  = !# $ , d dimensions, N particles  =   − 1 Superstatistic Fluctuations [5] n, Chi-square deg. of freedom The analysis in this review is grounded in the role heavy-tailed statistics plays in modeling the nonlinear dynamics of complex systems.It will be shown that by decomposing the NSM parameter q into more direct physical properties, interpretations of NSM are clarified and the connections with the tail shape of distributions, such as the generalized Pareto distribution and the Student's t distribution, are simplified.This approach has been called Nonlinear Statistical Coupling (NSC).Here, I refer to the theory as NSM and reserve NSC or simply the coupling for the shape parameter, which may also be a candidate for quantifying statistical complexity.For simplification, I'll assume that distributions have a location of zero throughout.Also not included in the discussion are distributions, such as the Weibull distribution, which introduce modifications to the skew of the distribution.
Each section addresses a fundamental question and defines an open problem.In some cases, a comment will be provided suggesting directions for investigation.Solutions are specifically not provided because although the author has in some cases previously recommended a solution, the future direction of NSM is ultimately a community decision made by the investigators pushing the field forward.Section 2 reviews how q relates to the traditional parameters of heavy-tailed and compact-support distributions.Section 3 discusses the question of generalizing entropy.Section 4 examines the difference between mathematical fits and physical theories, and the role of independent random variables in clarifying the physical property of q.Section 5 highlights some inconsistencies in how the qproduct is defined and applied.Section 6 explains the limitations in the use of the q-Fourier transform as a physical model.Section 7 considers three different normalizations of the nonextensive entropy.Finally, section 8 asks whether a definition for statistical complexity is possible.
2 How is q-statistics related to traditional definitions of heavy-tailed distributions?
NSM began [16,26] with a proposal to generalize Boltzmann-Gibbs statistics by examining the properties of systems with a distribution of states modified by the power of q.This modified distribution with elements  ( ) is referred to as the escort distribution; however, it's unfortunate that the NSM literature has not been explicit that this expression necessarily defines q to be a real number of independent random variables sharing the same state.If q is an integer n elementary probability theory establishes that  % is the probability of  independent random variables each with probability .Fractional random variables are discussed further in Section 4. From this start, the Tsallis entropy and its maximizing distribution were derived to be: Warm-Up Problem: The first problem is not so much open, as it is a warm-up to ground the discussion of the other problems.How does the NSM parameter  relate to the shape of a distribution?And how does the Lagrange multiplier  relate to the scale of a distribution?I'll address this question via examination of the generalized Pareto and Student's t distributions.Both the probability density function (pdf) and the survival function (sf) are provided since the sf will provide insights into the definition of a generalized exponential function.To unify the discussion both distributions will be defined in terms of the shape parameter , though the Student's t is traditionally defined in terms of its reciprocal, the degrees of freedom,  = 1  * .The shape parameter is also referred to as the nonlinear statistical coupling or coupling due to its connection with nonlinear; further, the final problem will consider whether it is a candidate for quantifying statistical complexity.The distributions have three domains: Compact − Support −1 <  < 0 Exponential  = 0 Heavy − Tail  > 0 (3)

Definition 1 Generalized Pareto Distribution
The survival function (cf) is one minus the cumulative distribution function (cdf),  P = 1 − .The Pareto Type IV with a location of zero is defined in terms of a scale, , and two shape parameters,  and .

𝐹 P (𝑥; 𝜎, 𝜅, 𝛼
The probability distribution function (pdf) is the derivative of the cdf ;  ≥ 0, ,  > 0 (5) For Pareto Type II  = 1 and the cf and pdf reduce to ;  ≥ 0,  > 0 (7) Comment on Definition 1: The definition for Type IV is modified from the traditional approach to clearly distinguish between the decay of the tail in the limit as  goes to infinity, , and the raising of the variable to the power .Thus, the outer exponent is + !"# so that + !# is the asymptotic power.Nevertheless, the emphasis here will be on questions about NSM and connections to the long-standing traditions in statistical analysis.

Definition 2 Generalized Student's t Distribution
The Student's t distribution is traditionally defined in terms of the degrees of freedom, , however, to unify the discussion, the reciprocal shape parameter,  = 1  * , is used.The survival function of the generalized Student's t distribution, which depends on the Gauss hypergeometric function, 2F1, and the Beta function, B, is  P (; , , ) ' $ (, ; ; ) = j The Student's t probability density function is: Warm-Up Solution The exponent of the Pareto ( = 1) and Student's t ( = 2) distribution determines the relationship between q and the shape : From this relationship, the escort probability or density can be defined in terms of the shape The multiplicative term of the variable determines the relationship between the Lagrange multiplier  and the scale   = Open Problem 1: Notice that the Pareto Type II survival function is in the form of the generalized exponential function, This leads to a question regarding the definitions for the generalized algebra of NSM.In the development of NSM, the generalization of the exponential function has been applied to the pdf; however, would the sf be the more natural function to generalize?If so, then the shape parameter rather than q becomes the fundamental parameter of the NSM generalization of statistical mechanics.We'll see that this modification leads to a clearer definition of the multivariate distributions and to more direct physical interpretations.A related issue is that exp 4 (−) ≠ exp 4 +' () for  ≠ 0. This is important in the definition of distributions since it's the reciprocal of the exponential function rather than the negative of the argument that is important.Before continuing, the inverse of the generalized exponential function is the generalized logarithm, 3 Is a generalization of entropy necessary?
One of the challenges of statistical mechanics is that it's quite difficult even for seasoned experts to formulate an intuitive framework for its foundational concept, entropy.To address the question of the need for a generalized entropy we'll describe the issue in terms of a distribution's average density (or probability for non-continuous distributions).While most concepts in statistics are framed in terms of densities/probabilities (y-axis of distribution) and estimates of the random variable (x-axis of distribution), entropy is based on the logarithm of the probabilities.This transformation,  → log , is essential to providing an additive scale, so that the arithmetic average is the central tendency of the uncertainty, leading to the definition of entropy,  = − ∑  ( log ( .This is the informational entropy, which will be used in this paper, while the physical entropy includes multiplication by the Boltzmann constant.Notice, however, that the logarithm can be separated from the aggregation of the probabilities using the weighted geometric mean  = log ∏  ( +E / # (-' . For continuous distributions, the entropy is  = − ∫ () log ()  .∈Aand the equivalent of the weighted geometric mean of the density is exp(−), called the logaverage.Therefore, the weighted geometric mean can be used to examine average density or probability without resorting to the logarithmic transformation.Using equations (12)(13)(14)(15) the log-average is generalized to a function I'll refer to as the coupled log-average where the factor − 0 '&4 and its inverse is determined by the exponent of the distribution f.For discrete functions (9) reduces to the generalized mean as derived in Definition 3 of [27]  FGH (; , ) ≡ ~j  ( (0,4)  ( Figure 1 shows the Gaussian,  = 0, and three heavy-tailed Coupled Gaussians  = {0.5, 1, 2}.The distributions are normalized by their couple average density which is highlighted in the figure by a horizontal line.The couped average density is computed for each density with the matching coupling value, k.Furthermore, the matching coupled average of the density is always equal to the density at  =  ± .As the coupling or shape increases, the tail becomes heavier, and the log-average ( = 0), shown as dashed horizontal lines, approaches zero.Thus, the entropy, which is the logarithm of the average density, approaches infinity.Nevertheless, the Student's t distribution has a structure that is quite different from the variance of the Gaussian going to infinity.Something has clearly been lost in summarizing the uncertainty of the Student's t with just the entropy.In particular, the scale s of the Student's t distribution, which generalizes the standard deviation of the Gaussian and is referred to as the q-standard deviation in NSM, remains finite.The analysis shows that the generalized mean can be used to separate the effect that the shape and the scale have on measures of the uncertainty.
Open Problem 2: Given that the average generalized density is equal to the density at the mean plus/minus the scale for the Coupled Gaussian and the location plus the scale for the Coupled Exponential, can the relationship between the average generalized density and the average density be quantified in a manner that strengths the explanation of how the generalized entropy complements the entropy function in describing the uncertainty of a system.For instance, given that entropy is a measure of the degrees of freedom of a system, and the coupling is the inverse of the degrees of freedom, can the difference between the coupled entropy and the entropy be quantified in terms of the degrees of freedom?
Comment on Problem 2: An important aspect of the investigation of statistical degrees of freedom is its relationship with the thermodynamic degrees of freedom.As noted in Table 1 and described in [25], q is determined by the degrees of freedom, n, of a temperature bath.Substituting (11), the shape, which is the reciprocal of the statistical degrees of freedom, is related by taking  = 1, given that distribution is based on the energy.d is the dimensions of

2.
translational degrees of freedom, though rotational and vibrational could also be considered.N is the number of molecules.

NSM: Mathematical Fit or Physical Theory?
A common criticism of the NSM has been that it is merely a mathematical fit to physical phenomena given a free parameter rather than a physical theory that provides an explanatory description of complex systems [10,12,14,28].While this has been refuted by investigators in the NSM community [2,[29][30][31], let's take a moment to consider what distinguishes a physical theory from a mathematical fit.First, mathematical theories build from assumed axioms and deductively prove derivative theories.Physical theories are a subset of mathematical theories that are constrained by physical measurements of the world.So, demonstrating a fit between a mathematical theory and physical measurement is a crucial step toward a physical theory.But is a fit sufficient to qualify a relationship as a physical theory?In physics, we are seeking models that provide explanatory power in describing a system.As such, each term (variables and constants) in a physical theory must have a clear definition of its role in the physical model, otherwise, the model loses its ability to be explicative.
Further, an effective model must fulfill the requirement of being the simplest representation of a phenomenon.Occam's razor [32,33] was one of the first articulations of this principle and Bayes' Theorem quantifies this property by specifying the uncertainty created by the overfitting of more complex models (see Ch. 28 of [34]).In the context of NSM, these criteria establish a requirement that its defining parameter q have a clear physical definition and that this property provides a simpler explanation of the statistics of complex systems than the shape or degrees of freedom parameters that it seeks to replace.For even in the case where the equations of NSM can be derived from first principles [35,36], if the defining parameter does not have a physical definition, then the derivation still lacks a physical interpretation.
Furthermore, as noted in the introduction, q does in fact have a straightforward interpretation based on the escort distribution.The originating motivation of q-statistics was the consideration of systems defined by an escort distribution with probabilities The quantity  ( ) defines the probability of q random variables that occupy the same state i.
Thus, the necessary starting point for defining a physical property of q is the number of independent random variables sharing the same state.For continuous random variables, one can consider an approximate threshold to discretize the limit.The relevance to complex systems is that the independent components of a multivariate heavy-tailed distribution are nevertheless correlated (or conversely, if linearly uncorrelated the components are dependent).Because of the nonlinear dependence between the dimensions of a heavytailed distribution, there is a higher occurrence of discrete variables that are equal or continuous variables that are approximately equal than would occur for distributions with exponential decay.This property has recently been explored as an approach to filtering heavy-tailed samples to facilitate the estimation of their distribution [37].Nevertheless, the question remains whether the property of equal-valued independent random variables is central to describing the statistics of complex systems.Several investigators have suggested other interpretations, but close examination shows that the descriptions are equal to  − 1 or another function of q, rather than q itself.For instance, Wilk and Włodarczyk [38,39] show how the fluctuations (relative variance) of an inverse scale parameter 6 ! 3 7 are equal to  − 1.The problem is this does not provide an interpretation of q's statistical property, rather it shows that q is misaligned by -1 with a possible interpretation.The relative variance is indeed a useful property, and thus a variable  =  − 1 is a candidate for an approach to defining nonextensive statistical mechanics.But as we'll see in the following section, multidimensional analysis shows that neither q nor  − 1 are fundamental.
Open Problem 3: Define and provide evidence for a physical definition of the parameter q.Include in this definition an explanation of the role of the number of independent random variables via the expression  ) .Demonstrate that this physical property simplifies and/or improves the description of the statistics of complex systems in comparison to the shape or the shape's inverse, the degrees of freedom.

The q-product does not construct the multivariate distributions
Borges [26] initiated and other investigators [40][41][42] further developed a q-algebra to encapsulate the core functions of nonextensive statistical mechanics.The foundational functions are a generalization of addition and multiplication, though the two do not form a generalized distributive property.The lack of distribution property was in part because the q-sum was primarily relevant to the combining of q-logarithms while the q-product was primarily relevant to the combining of q-exponentials.Here are the definitions in those contexts, While these constructions provide a useful shorthand for some of the complex relationships in nonextensive statistical mechanics, when applied to statistical analysis a significant shortcoming is evident.A bedrock principle of probability theory, which was discussed in the last section, is that independent probabilities multiply to form the joint probability.Thus, a natural question arises regarding the properties of the q-product of probabilities.Putting aside for a moment, the normalization of the q-exponential and q-Gaussian distributions, which add a further complication, how does the q-product of their distributions relate to the multivariate forms of these distributions?From the definition of the q-product we have where  is one for the q-exponential distribution and two for the q-Gaussian.Unfortunately, the expression on the right has very little to do with the multivariate form of these distributions when no cross-terms  ( 0  , 0 exist.This is because the exponents of the distribution include both a dimensional term and .Even for just the one-dimensional case, this led investigators to define 1 −  ≡ 2(1 − ) to account for the distinctions.The multivariate form of these distributions [43] is proportional to From this expression, it is evident that trying to force the multiplicative term inside the brackets and the exponent to be 1 −  and ' '+) respectively, results in several distortions of the physical properties.First, in NSM the physical scale of the distributions s is typically buried in a parameter referred to as the generalized inverse temperature,  ) = 4 ('+))/ " .And from the exponent, q is defined by the relationship To address the multivariate case, Umarov and Tsallis [44] formulated the following definitions for the multivariate Gaussian case ( = 2), Far from illuminating the multivariate statistics of complex systems, these types of expressions provide evidence that q is not aligned with the statistical properties of complex systems.Again, to interpret such an expression it is not enough to understand that q is the number of equal-valued independent random variables, one also needs to explain the physical role of each term in the right-hand expression.Without these explanations, the relationship fulfills a mathematical fit but falls short of a physical theory.
Open Problem 4: While the q-product is often referenced regarding its role in defining qindependence, the form does not lead to the structure of the multivariate heavy-tailed distributions in the manner that the product of distributions equates with the multivariate distribution of independent variables.A definition for the generalized product for NSM is needed that is based on the properties of the multivariate distributions, as was proposed in [43].
6 Does the q-Fourier transform model the properties of complex signals?
One of the celebrated results of NSM is the proof of a generalized central limit theorem (q-CLT) [45] that converges to q-Gaussians for random variables defined to have a property of q-independence.The nonlinear dependence described by q-independence relies on a generalization of the Fourier transform that maps q-Gaussians to a q*-Gaussians.Given a more general form of the Fourier transforms, a natural application would be the design of filters for signals with long-range correlations and/or fluctuations, the tell-tale characteristic of signals from a complex system.And yet, to date, there appear to be no applications of the q-Fourier transform to signal processing.Related to the lack of applications is the lack of a symmetric inverse [46][47][48], one of the key properties that has made the Fourier transform the foundation of signal processing.To frame this problem, I'll examine the Fourier transform in the context of the symmetry between the compactsupport and heavy-tailed functions of NSM [49].
The gap, in what should be a straightforward application of NSM, results from the disconnect between the mathematical relationships for the q-CLT and the physics of signal processing.Recall that the Fourier transform takes a function as an input (called the signal) and outputs another function (called the image) that preserves all the information about the original function.The process can be inverted with a function that has the same structural form.The image has been proven to represent the sinusoidal frequencies of the original signal and is used throughout engineering and science to craft filters for noise reduction, match filtering, and countless other purposes.As the name "image" implies, the FT is a kind of mirror.When applied to probability distributions, the FT mirror has the property of transforming wide, high-entropy distributions into narrow, low-entropy image functions.The Gaussian turns out to be the symmetrical function of this process, whereby the FT of a Gaussian is also Gaussian (though no longer normalized to integrate to one).And the variance of the image is proportional to the inverse of the signal's variance.
Unfortunately, as currently defined the q-FT violates this basic relationship between a signal and its image.The q-FT transforms both the tail shape and the scale of the distribution.Focusing on the tail shape, the transformed value of q and its translation into the shape parameter are determined from the definition of q-FT to be Table 2 shows how different domains of the heavy-tailed q-Gaussian distributions are transformed by the q-FT into wider-tailed images.The Cauchy distribution ( = 1,  = 2) highlights the difficulties in applying the q-FT to signal processing since the image function is an impulse function ( = ∞,  = 3).This suggests that the Cauchy distribution is the limit of physically realizable distributions.There are systems such as the Standard Map in which the Cauchy does act as a limiting distribution [50,51].At the same time, the coherent noise model [52] and Erhenfest's dog-flea model [53] have been measured to have q-Gaussians with shape/q values ( = 1.53,  = 2.21) and ( = 2.08,  = 2.35), respectively.And yet, distributions in this domain of very slow tail decay (1 <  < ∞, 2 <  < 3) have a q-FT image function with a divergent integral (−∞ <  < −1, 3 <  < ∞) suggesting that these would not arise in physically realizable systems.
In contrast to the q-FT, the Fourier transform maps heavy-tailed q-Gaussians into functions that are a product of a power-law term and a modified Bessel function of the second-order, which has an exponential tail decay: The power-law term is rising but sharply dampened by and in the limit as  → ∞ dominated by the exponential decay term.It is noteworthy, that the exponent of the powerlaw term ' $4 turns out to be a conjugate mapping between the exponents of the heavy-tailed and compact-support q-Gaussians.That is for  > 0, the heavy-tail and compact-support domains are related by: Heavy − tailed Compact − support In [49] I proposed a symmetrical generalization of the Fourier transform that maps the q-Gaussians between their compact-support and heavy-tailed domains.However; the transform included a mapping of the q parameter that did not generalize to other functions.
A requirement for a complete definition is a mathematically rigorous mapping between the infinite domain of the heavy-tailed distributions and the finite-domain compact-support functions.
Table 2: Description of the q-Gaussian domains and their associated q-FT image.The q-FT transforms functions such as the q-FT to images with slower decaying tails.So for example, the last row is the domain of distributions with undefined mean, which has an image with a divergent integral.
q-Gaussian Domain q-FT Image Description Shape, k q Description Shape q Finite Mean Finite Var.
[0, ' Open Problem 5: As currently defined, the q-Fourier transform of NSM has limited physical applications, since a) the inverse is not symmetric and b) the image function has slower decaying tails.Can these limitations be validated by limits within physical applications of heavy-tailed distributions or can a symmetric generalization of the Fourier transform be defined?A candidate for a symmetric Fourier transform maps q-Gaussians between their heavy-tailed and compact-support domains but currently lacks a general mapping between these domains.Can a mathematically rigorous mapping between the compact-support and heavy-tailed domains be defined that would qualify as a generalization of the Fourier transform? )  ( ( was investigated.The normalized Tsallis entropy was found to not satisfy the Lesche stability requirement [56][57][58] and has since been dismissed in favor of the original Tsallis entropy form.
However given the insight regarding the distinction between the power and normalization for the generalized exponential and logarithms discussed in Section 5 another normalization can be considered.The coupled entropy [27,59] is defined as In [27] it was found that the coupled and Tsallis entropy are (constant, asymptotically constant) as a function of the coupling for the generalized Pareto distribution the matching coupling value and a scale equal to (one, non-one), respectively.However, for the matched coupled Gaussian the coupled entropy rose in value while the Tsallis entropy decayed with increasing coupling value, i.e. the more heavy-tailed distributions.
The issue of normalization for a generalized entropy comes into sharper focus when considering the role of a generalized sum in defining the nonlinear combination of entropies.Substituting for q the relationship between the coupled, normalized, and Tsallis entropies is: While the q-sum of q-entropies has been defined as (using the coupling notation) ) Notably, the B-G-S entropy scales with the degrees of freedom and the nonextensive entropies modify this scaling [35,60].Given that  is the inverse of the statistical degrees of freedom, the coupled sum of the coupled entropies directly expresses this modification.
Open Problem 6: What is the proper normalization of a generalized entropy and how does the normalization impact the relationship between a generalized entropy and the statistical degrees of freedom?Stability issues caused a rejection of the normalized Tsallis entropy; however, neither the normalized nor unnormalized Tsallis entropy considers how the derivatives of a cdf into a pdf impacts the relationship between the definition of the generalized exponential and logarithmic functions and the structure of a pdf and its generalized entropy.When this is accounted for the nonlinear term combining a nonextensive entropy (coupled entropy) is precisely the inverse of the statistical degrees of freedom.Does this suggest a criterion for the normalization?
Comment on Problem 6: There are a variety of applications that may be impacted by the normalization of the NSM entropy.For instance, the robustness of machine learning algorithms have been improved using both q-entropy [61,62] and coupled entropy [59] generalizations.A careful analysis of whether the difference in normalization impacts the performance improvements would contribute to determining the importance of the normalization.Entropic analysis has been shown to be an effective measure of financial market volatility but greater detail is needed to determine the relative advantages of different forms of generalized entropy [63,64].

A measure of complexity?
The derivation of nonextensive entropy began with the investigation of systems with a modified distribution in which the probability of a state is raised to the power q.As discussed in sections 2 and 4, this necessitates a physical interpretation of q as the number of independent random variables sharing the same state.Unfortunately, clarifying this interpretation raises questions as to whether q is a fundamental or secondary property of complex systems.The more fundamental question is how should the statistical complexity of a system be quantified.The mismatch between q's physical interpretation and the fundamental properties of complex systems may be why the field has avoided addressing the issue.
Nevertheless, an approach to quantifying the statistical complexity of a system may be quite simple.The property Nonlinear Statistical Coupling was first introduced with the candidate 1 − , which did fulfill the need for the linear domain to have a value of zero; however, multidimensional analysis exposed that isolating the nonlinear properties required decomposition.As shown in equation (23), q is dependent on three properties, the dimension d, the nonlinearity of the random variable , and the redefined nonlinear statistical coupling .The coupling term is not new, in fact, it has a long tradition within statistical analysis as the shape parameter defining the deviation from exponential decay and it is the inverse of the degrees of freedom used to define the Student's t distribution.And so, the final open problem for the reader to consider is whether the coupling or shape parameter is an appropriate measure of a system's statistical complexity.
Open Problem 7: Does the shape parameter also referred to as the nonlinear statistical coupling provide a quantification of a system's statistical complexity?Can this definition of statistical complexity be related to other forms of complexity, such as algorithmic complexity?Explain the statistical complexity in terms of its inverse the statistical degrees of freedom.For instance, given samples from which to determine a model, does the nonlinearity of the function define the deterministic complexity of the model?And do the statistical degrees of freedom (samples minus model parameters) determine the inverse of the statistical complexity of the model?

Conclusion
While NSM has advanced the modeling of uncertainty within complex systems there remain many open problems worthy of investigation.In this paper, issues arising from the use of the parameter q as a focal point for modeling complex systems are examined.These issues are framed in terms of a set of open problems, including: 1. Should the generalized exponential function, originally proposed by (Borges, 2004), be applied to the survival function rather than the probability density functions?2. Can the difference between generalized entropy and BGS entropy be explained in terms of the degrees of freedom and its inverse the nonlinear statistical coupling?3.For NSM to be a complete physical theory, a clear physical interpretation of q is required.Determine whether the number of independent random variables sharing the same state is the appropriate interpretation of q. 4. Define the q-product using the properties of the multivariate distributions of qstatistics.5.The q-Fourier transform does not seem to model the physical image of a heavytailed signal.For example, the Cauchy distribution is transformed into a delta function, which could not be used for real-world signal processing.Can a generalization of the Fourier transform be defined which utilizes the complementary properties of the compact-support and heavy-tailed domains?6.The normalization of the coupled entropy differs from both the normalized and unnormalized Tsallis entropy.Is there a criterion that would clarify a preference between these three normalizations of the generalized entropy for complex systems?7. Define a measure of statistical complexity.
The author has proposed that the nonlinear statistical coupling, which is equal to the shape parameter and the inverse of the degrees of freedom, is a measure of statistical complexity.It is left to the reader to examine this set of open problems, determine satisfactory solutions, and consider whether a reframing of nonextensive statistical mechanics leads to a focus on the fundamental properties of complex systems.

Figure 1 :
Figure 1: Contrast between Average Density and Average Generalized Density

7
How should nonextensive entropy be normalized?During the early investigations of nonextensive entropy, a question arose regarding the proper probability to weight the generalized entropy.Tsallis entropy,  ) * ≡ '+∑ E / 0 / )+' = − ∑  ( )  )  ( ( (Tsallis, 2009), is weighted by  ( ) , however, this form does not make use of the escort probability defining the constraints for the generalized maximum entropy formalism.For this reason the normalized Tsallis entropy [54,55]  )