Estimation Bias in Maximum Entropy Models

Jakob H. Macke, Iain Murray, Peter E. Latham

Research output: Contribution to journalArticlepeer-review

Abstract

Maximum entropy models have become popular statistical models in neuroscience and other areas in biology and can be useful tools for obtaining estimates of mutual information in biological systems. However, maximum entropy models fit to small data sets can be subject to sampling bias; i.e., the true entropy of the data can be severely underestimated. Here, we study the sampling properties of estimates of the entropy obtained from maximum entropy models. We focus on pairwise binary models, which are used extensively to model neural population activity. We show that if the data is well described by a pairwise model, the bias is equal to the number of parameters divided by twice the number of observations. If, however, the higher order correlations in the data deviate from those predicted by the model, the bias can be larger. Using a phenomenological model of neural population recordings, we find that this additional bias is highest for small firing probabilities, strong correlations and large population sizes—for the parameters we tested, a factor of about four higher. We derive guidelines for how long a neurophysiological experiment needs to be in order to ensure that the bias is less than a specified criterion. Finally, we show how a modified plug-in estimate of the entropy can be used for bias correction.
Original languageEnglish
Pages (from-to)3109-3129
Number of pages21
JournalEntropy
Volume15
Issue number8
DOIs
Publication statusPublished - Aug 2013

Fingerprint

Dive into the research topics of 'Estimation Bias in Maximum Entropy Models'. Together they form a unique fingerprint.

Cite this