From Yaroslav's blog, this is of interest to the Experimental Probabilistic Hypersurface approach which computes the probability distribution.that maximizes entropy :for difficult to compute models (read too long to run on a computer). Here is the paper: How biased are maximum entropy models? by Jakob H. Macke, Iain Murray, Peter E. Latham. The abstract reads:
Maximum entropy models have become popular statistical models in neuroscience and other areas in biology, and can be useful tools for obtaining estimates of mutual information in biological systems. However, maximum entropy models fit to small data sets can be subject to sampling bias; i.e. the true entropy of the data can be severely underestimated. Here we study the sampling properties of estimates of the entropy obtained from maximum entropy models. We show that if the data is generated by a distribution that lies in the model class, the bias is equal to the number of parameters divided by twice the number of observations. However, in practice, the true distribution is usually outside the model class, and we show here that this misspecification can lead to much larger bias. We provide a perturbative approximation of the maximally expected bias when the true model is out of model class, and we illustrate our results using numerical simulations of an Ising model; i.e. the second-order maximum entropy distribution on binary data.