Beyond Bayes: On the Need for a Unified and Jaynesian Definition of Probability and Information within Neuroscience
Open Access
- 19 April 2012
- journal article
- research article
- Published by MDPI AG in Information
- Vol. 3 (2), 175-203
- https://doi.org/10.3390/info3020175
Abstract
It has been proposed that the general function of the brain is inference, which corresponds quantitatively to the minimization of uncertainty (or the maximization of information). However, there has been a lack of clarity about exactly what this means. Efforts to quantify information have been in agreement that it depends on probabilities (through Shannon entropy), but there has long been a dispute about the definition of probabilities themselves. The “frequentist” view is that probabilities are (or can be) essentially equivalent to frequencies, and that they are therefore properties of a physical system, independent of any observer of the system. E.T. Jaynes developed the alternate “Bayesian” definition, in which probabilities are always conditional on a state of knowledge through the rules of logic, as expressed in the maximum entropy principle. In doing so, Jaynes and others provided the objective means for deriving probabilities, as well as a unified account of information and logic (knowledge and reason). However, neuroscience literature virtually never specifies any definition of probability, nor does it acknowledge any dispute concerning the definition. Although there has recently been tremendous interest in Bayesian approaches to the brain, even in the Bayesian literature it is common to find probabilities that are purported to come directly and unconditionally from frequencies. As a result, scientists have mistakenly attributed their own information to the neural systems they study. Here I argue that the adoption of a strictly Jaynesian approach will prevent such errors and will provide us with the philosophical and mathematical framework that is needed to understand the general function of the brain. Accordingly, our challenge becomes the identification of the biophysical basis of Jaynesian information and logic. I begin to address this issue by suggesting how we might identify a probability distribution over states of one physical system (an “object”) conditional only on the biophysical state of another physical system (an “observer”). The primary purpose in doing so is not to characterize information and inference in exquisite, quantitative detail, but to be as clear and precise as possible about what it means to perform inference and how the biophysics of the brain could achieve this goal.Keywords
This publication has 151 references indexed in Scilit:
- The rise of the Enterococcus: beyond vancomycin resistanceNature Reviews Microbiology, 2012
- Self-Organized Complexity and Coherent Infomax from the Viewpoint of Jaynes’s Probability TheoryInformation, 2012
- Associations Between Multidrug Resistance, Plasmid Content, and Virulence Potential Among Extraintestinal Pathogenic and CommensalEscherichia colifrom Humans and PoultryFoodborne Pathogens & Disease, 2012
- Understanding vision in wholly empirical termsProceedings of the National Academy of Sciences, 2011
- Aminoglycoside modifying enzymesDrug Resistance Updates, 2010
- Sensitivity to perturbations in vivo implies high noise and suggests rate coding in cortexNature, 2010
- Human Health Hazards from Antimicrobial‐ResistantEscherichia coliof Animal OriginClinical Infectious Diseases, 2009
- Probabilistic Population Codes for Bayesian Decision MakingNeuron, 2008
- The Bayesian brain: the role of uncertainty in neural coding and computationTrends in Neurosciences, 2004
- Humans integrate visual and haptic information in a statistically optimal fashionNature, 2002