Entropy 2007, 9(1), 1-26; doi:10.3390/e9010001

A Utility-Based Approach to Some Information Measures

Standard & Poor’s, 55 Water Street, 46th Floor, New York, NY 10041, USA
* Author to whom correspondence should be addressed.
Received: 23 May 2006; Accepted: 5 January 2007 / Published: 20 January 2007
PDF Full-text Download PDF Full-Text [197 KB, uploaded 16 September 2008 11:01 CEST]
Abstract: We review a decision theoretic, i.e., utility-based, motivation for entropy and Kullback-Leibler relative entropy, the natural generalizations that follow, and various properties of thesegeneralized quantities. We then consider these generalized quantities in an easily interpreted spe-cial case. We show that the resulting quantities, share many of the properties of entropy andrelative entropy, such as the data processing inequality and the second law of thermodynamics.We formulate an important statistical learning problem – probability estimation – in terms of ageneralized relative entropy. The solution of this problem reflects general risk preferences via theutility function; moreover, the solution is optimal in a sense of robust absolute performance.
Keywords: Generalized Entropy; Generalized Kullback-Leibler Relative Entropy; Decision The- ory; Expected Utility; Horse Race; Tsallis Entropy; Statistical Learning; Probability Estimation; Risk Neutral Pricing Measure

Article Statistics

Load and display the download statistics.

Citations to this Article

Cite This Article

MDPI and ACS Style

Friedman, C.; Huang, J.; Sandow, S. A Utility-Based Approach to Some Information Measures. Entropy 2007, 9, 1-26.

AMA Style

Friedman C, Huang J, Sandow S. A Utility-Based Approach to Some Information Measures. Entropy. 2007; 9(1):1-26.

Chicago/Turabian Style

Friedman, Craig; Huang, Jinggang; Sandow, Sven. 2007. "A Utility-Based Approach to Some Information Measures." Entropy 9, no. 1: 1-26.

Entropy EISSN 1099-4300 Published by MDPI AG, Basel, Switzerland RSS E-Mail Table of Contents Alert