Abstract
This paper presents a new formula for the entropy of a distribution, that is conceived having in mind the Liouville fractional derivative. For illustrating the new concept, the proposed definition is applied to the Dow Jones Industrial Average. Moreover, the Jensen-Shannon divergence is also generalized and its variation with the fractional order is tested for the time series.
1. Introduction
The quest of generalizing the Boltzmann–Gibbs entropy has become an active field of research in the past 30 years. Indeed, many formulations appeared in the literature extending the well-known formula (see e.g., [1,2,3,4]):
The (theoretical) approaches to generalize Equation (1) may vary considerably (see e.g., [1,2,3]). In this work we are particularly interested in the method firstly proposed by Abe in [1], which consists of the basic idea of rewriting Equation (1) as
Concretely, we substitute the differential operator in Equation (2) by a suitable fractional one (see Section 2 for the details) and then, after some calculations, we obtain a novel (at least to the best of our knowledge) formula, which depends on a parameter .
The paper is structured as follows. Section 2 introduces and discusses the motivation for the new entropy formulation. Section 3 analyses the Dow Jones Industrial Average. Additionally, The Jensen-Shannon divergence is also adopted in conjunction with the hierarchical clustering technique for analyzing regularities embedded in the time series. Finally, Section 4 outlines the conclusions.
2. Motivation
Let us introduce the following entropy function:
where is the gamma function.
We define the quantity as the Liouville information (See Figure 1).
Figure 1.
Liouville information .
Our motivation to define the entropy function given by Equation (3) is essentially due to the works of Abe [1] and Ubriaco [4]. Indeed, in Section 3 of [4], the author notes (based on Abe’s work [1]) that the Boltzmann-Gibbs and the Tsallis entropies may be obtained by
and
respectively. From this, he substitutes the above differential operator by a Liouville fractional derivative (see Section 2.3 in [5]) and then he defines a fractional entropy (see (19) in [4]). With this in mind and taking into account the generalization of the Liouville fractional derivative given by the “fractional derivative of a function with respect to another function” (see Section 2.5 in [5]) we consider using it in order to define a novel entropy. The Liouville fractional derivative of a function f with respect to another function g (with ) is defined by [5,6],
It is important to keep in mind that our goal is to obtain an explicit formula for the entropy. Therefore, we can think that a “good” candidate for g is the exponential function, due to the fact that and also the structure of Equation (4). We chose . Let us then calculate Equation (4) with and . We obtain
It follows that,
and after using the property , we finally get
We have, therefore, motivated the definition provided in Equation (3).
Remark 1.
3. An Example of Application
The Dow Jones Industrial Average (DJIA) is an index based on the value of 30 large companies from the United States traded in the stock market during time. The DJIA and other financial indices reveal a fractal nature and has been the topic of many studies using distinct mathematical and computational tools [7,8]. In this section we apply the previous concepts in the study of the DJIA in order to verify the variation of the new expressions with the fractional order. Therefore, we start by analyzing the evolution of daily closing values of the DJIA from January 1, 1985, to April 5, 2019, in the perspective of Equation (3). All weeks include five days and missing values corresponding to special days are interpolated between adjacent values. For calculating the entropy we consider time windows of 149 days performing a total of years.
Figure 3.
Evolution of the Dow Jones Industrial Average (DJIA) versus time t from January 1, 1985, to April 5, 2019.
Figure 4.
Evolution of the versus time t from January 1, 1985, to April 5, 2019, and .
We verify that has a smooth evolution with that plays the role of a parameters for adjusting the sensitivity of the entropy index.
The Jensen-Shannon divergence () measures the similarity between two probability distributions and is given by
where , and and represent the Kullback-Leibler divergence between distributions P and M, and P and Q, respectively.
For the classical Shannon information the can be calculated as:
In the case of the Liouville information the can be calculated as:
Obviously, for we obtain the Shannon formulation.
For processing the data produced by the we adopt hierarchical clustering (HC). The main objective of the HC is to group together (or to place far apart) objects that are similar (or different) [9,10,11,12]. The HC receives as input a symmetrical matrix D of distances (e.g., the ) between the n items under analysis and produces as output a graph, in the form of a dendogram or a tree, where the length of the links represents the distance between data objects. We have two alternative algorithms, namely the agglomerative and the divisive clustering. In the first, each object starts in its own singleton cluster and, at each iteration of the HC scheme, the two most similar (in some sense) clusters are merged. The iterations stop when there is a single cluster containing all objects. In the second, all objects start in a single cluster and at each step, the HC removes the ‘outsiders’ from the least cohesive cluster. The iterations stop when each object is in its own singleton cluster. Both iterative schemes are achieved using an appropriate metric (a measure of the distance between pairs of objects) and a linkage criterion, which defines the dissimilarity between clusters as a function of the pairwise distance between objects.
In our case the objects correspond to the years, from January 1, 1985, to December 31, 2018, that are compared using the and resulting matrix D (with dimension ) processed by means of HC.
Figure 5, Figure 6 and Figure 7 show the trees generated by the hierarchical clustering for the Shannon and the Liouville Jensen-Shannon divergence measures (with ), respectively. The 2-digit labels of the ‘leafs’ of the trees denote the years.
Figure 5.
Tree generated by the hierarchical clustering for the Shannon .
Figure 6.
Tree generated by the hierarchical clustering for the Liouville and .
Figure 7.
Tree generated by the hierarchical clustering for the Liouville and .
We note that our goal is not to characterize the dynamics of the DJIA time evolution since it is outside the scope of this paper. In fact, we adopt the DJIA simply as a prototype data series for assessing the effect of changing the value of in the and consequently in the HC generated tree. We verify that in general there is a strong similarity of the DJIA between consecutive years. In what concerns the use of the Shannon versus the Liouville , we observe that for close to 1 both entropies lead to identical results, while for close to 0 the Liouville produces a distinct tree. Therefore, we conclude that we can adjust the clustering performance by a proper tuning of the parameter .
4. Conclusions
This paper presented a new formulation for entropy based on the (generalized) Liouville definition of fractional derivative. The generalization leads not only to a new entropy index, but also to novel expressions for fractional information and Jensen-Shannon divergence. The sensitivity of the proposed expression to variations of the fractional order is tested for the DJIA time series.
Author Contributions
Conceptualization, R.A.C.F.; Data curation, J.T.M.; Formal analysis, R.A.C.F.; Methodology, R.A.C.F.; Software, J.T.M.; Visualization, J.T.M.; Writing—original draft, R.A.C.F. and J.T.M.
Funding
Fundação para a Ciência e a Tecnologia (FCT): IF/01345/2014.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Abe, S. A note on the q-deformation-theoretic aspect of the generalized entropies in nonextensive physics. Phys. Lett. A 1997, 224, 326–330. [Google Scholar] [CrossRef]
- Machado, J.A.T. Fractional Order Generalized Information. Entropy 2014, 16, 2350–2361. [Google Scholar] [CrossRef]
- Tsallis, C. Possible generalization of Boltzmann-Gibbs statistics. J. Stat. Phys. 1988, 52, 479–487. [Google Scholar] [CrossRef]
- Ubriaco, M.R. Entropies based on fractional calculus. Phys. Lett. A 2009, 373, 2516–2519. [Google Scholar] [CrossRef]
- Kilbas, A.; Srivastava, H.; Trujillo, J. Theory and Applications of Fractional Differential Equations; Elsevier: Amsterdam, The Netherlands, 2006. [Google Scholar]
- Samko, S.; Kilbas, A.; Marichev, O. Fractional Integrals and Derivatives: Theory and Applications; Gordon and Breach Science Publishers: Yverdon-les-Bains, Switzerland, 1993. [Google Scholar]
- Machado, J.A.T. Complex Dynamics of Financial Indices. Nonlinear Dyn. 2013, 74, 287–296. [Google Scholar] [CrossRef]
- Machado, J.A.T. Relativistic Time Effects in Financial Dynamics. Nonlinear Dyn. 2014, 75, 735–744. [Google Scholar] [CrossRef]
- Hartigan, J.A. Clustering Algorithms; John Wiley & Sons: New York, NY, USA, 1975. [Google Scholar]
- Sokal, R.R.; Rohlf, F.J. The comparison of dendrograms by objective methods. Taxon 1962, 11, 33–40. [Google Scholar] [CrossRef]
- Kaufman, L.; Rousseeuw, P.J. Finding Groups in Data - An Introduction to Cluster Analysis; Wiley-Interscience: Hoboken, NY, USA, 2005. [Google Scholar]
- Maimon, O.; Rokach, L. Data Mining and Knowledge Discovery Handbook; Springer: Berlin/Heidelberg, Germany, 2005. [Google Scholar]
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).