Bayesian approaches to brain function

From Wikipedia, the free encyclopedia
Jump to: navigation, search

Bayesian approaches to brain function investigate the capacity of the nervous system to operate in situations of uncertainty in a fashion that is close to the optimal prescribed by Bayesian statistics. This term is used in behavioural sciences and neuroscience and studies associated with this term often strive to explain the brain's cognitive abilities based on statistical principles. It is frequently assumed that the nervous system maintains internal probabilistic models that are updated by neural processing of sensory information using methods approximating those of Bayesian probability.[1][2]

Origins[edit]

This field of study has its historical roots in numerous disciplines including machine learning, experimental psychology and Bayesian statistics. As early as the 1860s, with the work of Hermann Helmholtz in experimental psychology the brain's ability to extract perceptual information from sensory data was modeled in terms of probabilistic estimation.[3][4] The basic idea is that the nervous system needs to organize sensory data into an accurate internal model of the outside world.

Bayesian probability has been developed by many important contributors. Pierre-Simon Laplace, Thomas Bayes, Harold Jeffreys, Richard Cox and Edwin Jaynes developed mathematical techniques and procedures for treating probability as the degree of plausibility that could be assigned to a given supposition or hypothesis based on the available evidence.[5] In 1988 E.T. Jaynes presented a framework for using Bayesian Probability to model mental processes.[6] It was thus realized early on that the Bayesian statistical framework holds the potential to lead to insights into the function of the nervous system.

This idea was taken up in research on unsupervised learning, in particular the Analysis by Synthesis approach, branches of machine learning.[7][8] In 1983 Geoffrey Hinton and colleagues proposed the brain could be seen as a machine making decisions based on the uncertainties of the outside world.[9] During the 1990s researchers including Peter Dayan, Geoffrey Hinton and Richard Zemel proposed that the brain represents knowledge of the world in terms of probabilities and made specific proposals for tractable neural processes that could manifest such a Helmholtz Machine.[10][11][12]

Psychophysics[edit]

A wide range of studies interpret the results of psychophysical experiments in light of Bayesian perceptual models. Many aspects of human perceptual and motor behavior can be modeled with Bayesian statistics. This approach, with its emphasis on behavioral outcomes as the ultimate expressions of neural information processing, is also known for modeling sensory and motor decisions using Bayesian decision theory. Examples are the work of Landy,[13][14] Jacobs,[15][16] Jordan, Knill,[17][18] Kording and Wolpert,[19][20] and Goldreich.[21][22][23]

Also see Bayesian inference in motor learning

Neural coding[edit]

Many theoretical studies ask how the nervous system could implement Bayesian algorithms. Examples are the work of Pouget, Zemel, Deneve, Latham, Hinton and Dayan. George and Hawkins published a paper that establishes a model of cortical information processing called hierarchical temporal memory that is based on Bayesian network of Markov chains. They further map this mathematical model to the existing knowledge about the architecture of cortex and show how neurons could recognize patterns by hierarchical Bayesian inference.[24]

Electrophysiology[edit]

A number of recent electrophysiological studies focus on the representation of probabilities in the nervous system. Examples are the work of Shadlen and Schultz.

Predictive coding[edit]

Predictive coding is a neurobiologically plausible scheme for inferring the causes of sensory input based on minimizing prediction error.[25] These schemes are related formally to Kalman filtering and other Bayesian update schemes.

Free energy[edit]

During the 1990s some researchers such as Geoffrey Hinton and Karl Friston began examining the concept of free energy as a calculably tractable measure of the discrepancy between actual features of the world and representations of those features captured by neural network models.[26] A synthesis has been attempted recently[27] by Karl Friston, in which the Bayesian brain emerges from a general principle of free energy minimisation.[28] In this framework, both action and perception are seen as a consequence of suppressing free-energy, leading to perceptual[29] and active inference[30] and a more embodied (enactive) view of the Bayesian brain. Using variational Bayesian methods, it can be shown how internal models of the world are updated by sensory information to minimize free energy or the discrepancy between sensory input and predictions of that input. This can be cast (in neurobiologically plausible terms) as predictive coding or, more generally, Bayesian filtering.

According to Friston:[31]

"The free-energy considered here represents a bound on the surprise inherent in any exchange with the environment, under expectations encoded by its state or configuration. A system can minimise free energy by changing its configuration to change the way it samples the environment, or to change its expectations. These changes correspond to action and perception, respectively, and lead to an adaptive exchange with the environment that is characteristic of biological systems. This treatment implies that the system’s state and structure encode an implicit and probabilistic model of the environment."[31]

This area of research was summarized in terms understandable by the layperson in a 2008 article in New Scientist that offered a unifying theory of brain function.[32] Friston makes the following claims about the explanatory power of the theory:

"This model of brain function can explain a wide range of anatomical and physiological aspects of brain systems; for example, the hierarchical deployment of cortical areas, recurrent architectures using forward and backward connections and functional asymmetries in these connections. In terms of synaptic physiology, it predicts associative plasticity and, for dynamic models, spike-timing-dependent plasticity. In terms of electrophysiology it accounts for classical and extra-classical receptive field effects and long-latency or endogenous components of evoked cortical responses. It predicts the attenuation of responses encoding prediction error with perceptual learning and explains many phenomena like repetition suppression, mismatch negativity and the P300 in electroencephalography. In psychophysical terms, it accounts for the behavioural correlates of these physiological phenomena, e.g., priming, and global precedence."[31]

"It is fairly easy to show that both perceptual inference and learning rest on a minimisation of free energy or suppression of prediction error."[31]

See also[edit]

References[edit]

  1. ^ Kenji Doya (Editor), Shin Ishii (Editor), Alexandre Pouget (Editor), Rajesh P. N. Rao (Editor) (2007), Bayesian Brain: Probabilistic Approaches to Neural Coding, The MIT Press; 1 edition (Jan 1 2007)
  2. ^ Knill David,Pouget Alexandre (2004), The Bayesian brain: the role of uncertainty in neural coding and computation,TRENDS in Neurosciences Vol.27 No.12 December 2004
  3. ^ Helmholtz, H. (1860/1962). Handbuch der physiologischen optik (Southall, J. P. C. (Ed.), English trans.),Vol. 3. New York: Dover.
  4. ^ Westheimer, G. (2008) Was Helmholtz a Bayesian?" Perception 39, 642–50
  5. ^ Jaynes, E. T., 1986, `Bayesian Methods: General Background,' in Maximum-Entropy and Bayesian Methods in Applied Statistics, J. H. Justice (ed.), Cambridge Univ. Press, Cambridge
  6. ^ Jaynes, E. T., 1988, `How Does the Brain Do Plausible Reasoning?', in Maximum-Entropy and Bayesian Methods in Science and Engineering, 1, G. J. Erickson and C. R. Smith (eds.)
  7. ^ Ghahramani, Z. (2004). Unsupervised learning. In O. Bousquet, G. Raetsch, & U. von Luxburg (Eds.), Advanced lectures on machine learning. Berlin: Springer-Verlag.
  8. ^ Neisser, U., 1967. Cognitive Psychology. Appleton-Century-Crofts, New York.
  9. ^ Fahlman, S.E., Hinton, G.E. and Sejnowski, T.J.(1983). Massively parallel architectures for A.I.: Netl, Thistle, and Boltzmann machines. Proceedings of the National Conference on Artificial Intelligence, Washington DC.
  10. ^ Dayan, P., Hinton, G. E., & Neal, R. M. (1995). The Helmholtz machine. Neural Computation, 7, 889–904.
  11. ^ Dayan, P. and Hinton, G. E. (1996), Varieties of Helmholtz machines. , Neural Networks, 9 1385–1403.
  12. ^ Hinton, G. E., Dayan, P., To, A. and Neal R. M. (1995), The Helmholtz machine through time., Fogelman-Soulie and R. Gallinari (editors) ICANN-95, 483–490
  13. ^ Tassinari H, Hudson TE & Landy MS. (2006). Combining priors and noisy visual cues in a rapid pointing task" Journal of Neuroscience 26(40), 10154–10163.
  14. ^ Hudson TE, Maloney LT & Landy MS. (2008). Optimal compensation for temporal uncertainty in movement planning. PLoS Computational Biology, 4(7).
  15. ^ Jacobs RA (1999). Optimal integration of texture and motion cues to depth" Vision Research 39(21), 3621–9.
  16. ^ Battaglia PW, Jacobs RA & Aslin RN (2003). Bayesian integration of visual and auditory signals for spatial localization. Journal of the Optical Society of America, 20(7), 1391–7.
  17. ^ Knill DC (2005). Reaching for visual cues to depth: The brain combines depth cues differently for motor control and perception. Journal of Vision, 5(2), 103:15.
  18. ^ Knill DC (2007). Learning Bayesian priors for depth perception. Journal of Vision, 7(8), 1–20.
  19. ^ Koerding KP & Wolpert DM (2004). Bayesian integration in sensorimotor learning. Nature, 427, 244–7.
  20. ^ Koerding KP, Ku S & Wolpert DM (2004). Bayesian integration in force estimation" Journal of Neurophysiology 92, 3161–5.
  21. ^ Goldreich, D (Mar 28, 2007). "A Bayesian perceptual model replicates the cutaneous rabbit and other tactile spatiotemporal illusions.". PLoS ONE 2 (3): e333. doi:10.1371/journal.pone.0000333. PMC 1828626. PMID 17389923. 
  22. ^ Goldreich, Daniel; Tong, Jonathan (10 May 2013). "Prediction, Postdiction, and Perceptual Length Contraction: A Bayesian Low-Speed Prior Captures the Cutaneous Rabbit and Related Illusions". Frontiers in Psychology 4 (221). doi:10.3389/fpsyg.2013.00221. 
  23. ^ Goldreich, D; Peterson, MA (2012). "A Bayesian observer replicates convexity context effects in figure-ground perception.". Seeing and perceiving 25 (3-4): 365–95. doi:10.1163/187847612X634445. PMID 22564398. 
  24. ^ George D, Hawkins J, 2009 Towards a Mathematical Theory of Cortical Micro-circuits" PLoS Comput Biol 5(10) e1000532. doi:10.1371/journal.pcbi.1000532
  25. ^ Rao RPN, Ballard DH. Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects. Nature Neuroscience. 1999. 2:79–87
  26. ^ Hinton, G. E. and Zemel, R. S.(1994), Autoencoders, minimum description length, and Helmholtz free energy. Advances in Neural Information Processing Systems 6. J. D. Cowan, G. Tesauro and J. Alspector (Eds.), Morgan Kaufmann: San Mateo, CA.
  27. ^ Friston K, The free-energy principle: A unified brain theory?, Nat Rev Neurosci. 2010. 11:127–38
  28. ^ Friston K, Kilner J, Harrison L. A free energy principle for the brain, J Physiol Paris. 2006. 100:70–87
  29. ^ Friston K, A theory of cortical responses, Philos Trans R Soc Lond B Biol Sci. 2005. 360:815–36.
  30. ^ Friston KJ, Daunizeau J, Kilner J, Kiebel SJ. Action and behavior: A free-energy formulation, Biol Cybern. 2010. 102:227–60
  31. ^ a b c d Friston K, Stephan KE., Free energy and the brain, Synthese. 2007. 159:417–458
  32. ^ Huang Gregory (2008), "Is This a Unified Theory of the Brain?", New Scientist. May 23, 2008.

External links[edit]