Estimation Bias in Maximum Entropy Models
AbstractMaximum entropy models have become popular statistical models in neuroscience and other areas in biology and can be useful tools for obtaining estimates of mutual information in biological systems. However, maximum entropy models fit to small data sets can be subject to sampling bias; i.e., the true entropy of the data can be severely underestimated. Here, we study the sampling properties of estimates of the entropy obtained from maximum entropy models. We focus on pairwise binary models, which are used extensively to model neural population activity. We show that if the data is well described by a pairwise model, the bias is equal to the number of parameters divided by twice the number of observations. If, however, the higher order correlations in the data deviate from those predicted by the model, the bias can be larger. Using a phenomenological model of neural population recordings, we find that this additional bias is highest for small firing probabilities, strong correlations and large population sizes—for the parameters we tested, a factor of about four higher. We derive guidelines for how long a neurophysiological experiment needs to be in order to ensure that the bias is less than a specified criterion. Finally, we show how a modified plug-in estimate of the entropy can be used for bias correction. View Full-Text
Share & Cite This Article
Macke, J.H.; Murray, I.; Latham, P.E. Estimation Bias in Maximum Entropy Models. Entropy 2013, 15, 3109-3129.
Macke JH, Murray I, Latham PE. Estimation Bias in Maximum Entropy Models. Entropy. 2013; 15(8):3109-3129.Chicago/Turabian Style
Macke, Jakob H.; Murray, Iain; Latham, Peter E. 2013. "Estimation Bias in Maximum Entropy Models." Entropy 15, no. 8: 3109-3129.