Nonadditive Entropies Yield Probability Distributions with Biases not Warranted by the Data
Abstract
Different quantities that go by the name of entropy are used in variational principles to infer probability distributions from limited data. Shore and Johnson showed that maximizing the Boltzmann-Gibbs form of the entropy ensures that probability distributions inferred satisfy the multiplication rule of probability for independent events in the absence of data coupling such events. Other types of entropies that violate the Shore and Johnson axioms, including nonadditive entropies such as the Tsallis entropy, violate this basic consistency requirement. Here we use the axiomatic framework of Shore and Johnson to show how such nonadditive entropy functions generate biases in probability distributions that are not warranted by the underlying data.
- Publication:
-
Physical Review Letters
- Pub Date:
- November 2013
- DOI:
- 10.1103/PhysRevLett.111.180604
- arXiv:
- arXiv:1312.1186
- Bibcode:
- 2013PhRvL.111r0604P
- Keywords:
-
- 05.20.-y;
- 02.50.Tt;
- 89.70.Cf;
- Classical statistical mechanics;
- Inference methods;
- Entropy and other measures of information;
- Condensed Matter - Statistical Mechanics;
- Physics - Data Analysis;
- Statistics and Probability
- E-Print:
- Phys. Rev. Lett., 111, 180604 (2013)