The Homological Nature of Entropy†
AbstractWe propose that entropy is a universal co-homological class in a theory associated to a family of observable quantities and a family of probability distributions. Three cases are presented: (1) classical probabilities and random variables; (2) quantum probabilities and observable operators; (3) dynamic probabilities and observation trees. This gives rise to a new kind of topology for information processes, that accounts for the main information functions: entropy, mutual-informations at all orders, and Kullback–Leibler divergence and generalizes them in several ways. The article is divided into two parts, that can be read independently. In the first part, the introduction, we provide an overview of the results, some open questions, future results and lines of research, and discuss briefly the application to complex data. In the second part we give the complete definitions and proofs of the theorems A, C and E in the introduction, which show why entropy is the first homological invariant of a structure of information in four contexts: static classical or quantum probability, dynamics of classical or quantum strategies of observation of a finite system. View Full-Text
Share & Cite This Article
Baudot, P.; Bennequin, D. The Homological Nature of Entropy. Entropy 2015, 17, 3253-3318.
Baudot P, Bennequin D. The Homological Nature of Entropy. Entropy. 2015; 17(5):3253-3318.Chicago/Turabian Style
Baudot, Pierre; Bennequin, Daniel. 2015. "The Homological Nature of Entropy." Entropy 17, no. 5: 3253-3318.