Jump to content

Computational neuroscience: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
No edit summary
Citation bot (talk | contribs)
Altered doi-broken-date. | Use this bot. Report bugs. | #UCB_CommandLine
 
(517 intermediate revisions by more than 100 users not shown)
Line 1: Line 1:
{{short description|Branch of neuroscience}}
'''Computational neuroscience''' is an interdisciplinary science that links the diverse fields of [[neuroscience]], [[cognitive science]], [[electrical engineering]], [[computer science]], [[physics]] and [[mathematics]]. Historically, the term was introduced by [[Eric L. Schwartz]], who organized a conference, held in 1985 in Carmel, California at the request of the Systems Development Foundation, to provide a summary of the current status of a field which until that point was referred to by a variety of names, such as neural modeling, brain theory and neural networks. The proceedings of this definitional meeting were later published as the book "Computational Neuroscience", MIT Press(1990). The early historical roots of the field can be traced to the work of people such as [[Alan Hodgkin|Hodgkin]] & [[Andrew Huxley|Huxley]], [[David H. Hubel|Hubel]] & [[Torsten Wiesel|Wiesel]], and [[David Marr (psychologist)|David Marr]], to name but a few. Hodgkin & Huxley developed the voltage clamp and created the first mathematical model of the [[action potential]]. Hubel & Wiesel discovered that neurons in [[primary visual cortex]], the first cortical area to process information coming from the [[retina]], have oriented receptive fields and are organized in columns (Hubel & Wiesel, 1962). David Marr's work focused on the interactions between neurons, suggesting computational approaches to the study of how functional groups of neurons within the [[hippocampus]] and neocortex interact, store, process, and transmit information. Computational modeling of biophysically realistic neurons and dendrites began with the work of Wilfrid Rall, with the first multicompartmental model using [[cable theory]].
'''Computational neuroscience''' (also known as '''theoretical neuroscience''' or '''mathematical neuroscience''') is a branch of&nbsp;[[neuroscience]]&nbsp;which employs [[mathematics]], [[computer science]], theoretical analysis and abstractions of the brain to understand the principles that govern the [[Developmental neuroscience|development]], [[Neuroanatomy|structure]], [[Neurophysiology|physiology]] and [[Cognitive neuroscience|cognitive abilities]] of the [[nervous system]].<ref name=":0">{{Cite book|title=Fundamentals of Computational Neuroscience|url=https://archive.org/details/fundamentalscomp00ttra|url-access=limited|last=Trappenberg|first=Thomas P.|publisher=Oxford University Press Inc.|year=2010|isbn=978-0-19-851582-1|location=United States|pages=[https://archive.org/details/fundamentalscomp00ttra/page/n17 2]}}</ref><ref>{{cite book |chapter=What is computational neuroscience? |author1=Patricia S. Churchland |author2=Christof Koch |author3=Terrence J. Sejnowski |title=Computational Neuroscience |pages=46–55 |editor1=Eric L. Schwartz |year=1993 |publisher=MIT Press |url=http://mitpress.mit.edu/catalog/item/default.asp?ttype=2&tid=7195 |access-date=2009-06-11 |url-status=dead |archive-url=https://web.archive.org/web/20110604124206/http://mitpress.mit.edu/catalog/item/default.asp?ttype=2&tid=7195 |archive-date=2011-06-04 }}</ref><ref>{{cite book |author1=Dayan P. |author-link=Peter Dayan |author2=Abbott, L. F. |author2-link=Larry Abbott |title=Theoretical neuroscience: computational and mathematical modeling of neural systems |publisher=MIT Press |location=Cambridge, Mass |year=2001 |isbn=978-0-262-04199-7 }}</ref><ref>{{ cite book | author1= Gerstner, W. | author2 = Kistler, W. | author3 = Naud, R. | author4 = Paninski, L.| title = Neuronal Dynamics | publisher = Cambridge University Press | location = Cambridge, UK | year = 2014 | isbn = 9781107447615}}</ref>


Computational neuroscience employs computational simulations<ref>{{Cite journal |last1=Fan |first1=Xue |last2=Markram |first2=Henry |date=2019 |title=A Brief History of Simulation Neuroscience |journal=Frontiers in Neuroinformatics |volume=13 |page=32 |doi=10.3389/fninf.2019.00032 |doi-access=free |issn=1662-5196 |pmc=6513977 |pmid=31133838}}</ref> to validate and solve mathematical models, and so can be seen as a sub-field of theoretical neuroscience; however, the two fields are often synonymous.<ref>{{Cite book|last=Thomas|first=Trappenberg|url=https://books.google.com/books?id=4PDsA1EVCx0C|title=Fundamentals of Computational Neuroscience|publisher=OUP Oxford|year=2010|isbn=978-0199568413|location=|pages=2|access-date=17 January 2017}}</ref> The term mathematical neuroscience is also used sometimes, to stress the quantitative nature of the field.<ref>{{Cite journal|last1=Gutkin|first1=Boris|last2=Pinto|first2=David|last3=Ermentrout|first3=Bard|date=2003-03-01|title=Mathematical neuroscience: from neurons to circuits to systems|journal=Journal of Physiology-Paris|series=Neurogeometry and visual perception|volume=97|issue=2|pages=209–219|doi=10.1016/j.jphysparis.2003.09.005|pmid=14766142|s2cid=10040483|issn=0928-4257}}</ref>
Computational neuroscience is distinct from psychological [[connectionism]] and theories of learning from disciplines such as [[machine learning]], [[neural networks]] and [[statistical learning theory]] in that it emphasizes descriptions of functional and biologically realistic neurons (and neural systems) and their physiology and dynamics. These models capture the essential features of the biological system at multiple spatial-temporal scales, from membrane currents, protein and chemical coupling to network oscillations, columnar and topographic architecture and learning and memory. These computational models are used to test hypotheses that can be directly verified by current or future biological experiments.


Computational neuroscience focuses on the description of [[Biology|biologically]] plausible [[neuron]]s (and [[Nervous system|neural systems]]) and their physiology and dynamics, and it is therefore not directly concerned with biologically unrealistic models used in [[connectionism]], [[control theory]], [[cybernetics]], [[quantitative psychology]], [[machine learning]], [[artificial neural network]]s, [[artificial intelligence]] and [[computational learning theory]];<ref>{{Cite journal|last1=Kriegeskorte|first1=Nikolaus|last2=Douglas|first2=Pamela K.|date=September 2018|title=Cognitive computational neuroscience|journal=Nature Neuroscience|language=en|volume=21|issue=9|pages=1148–1160|doi=10.1038/s41593-018-0210-5|pmid=30127428|pmc=6706072|arxiv=1807.11819|bibcode=2018arXiv180711819K|issn=1546-1726}}</ref><ref>{{Citation |last=Paolo |first=E. D. |title=Organismically-inspired robotics: homeostatic adaptation and teleology beyond the closed sensorimotor loop | journal=Dynamical Systems Approach to Embodiment and Sociality |s2cid=15349751 }}</ref>
Currently, the field is undergoing a rapid expansion. There are many software packages, such as [[GENESIS_%28software%29|GENESIS]] and [[Neuron_%28software%29|NEURON]], that allow rapid and systematic in silico modeling of realistic neurons. [[Blue Brain]], a collaboration between [[IBM]] and [[École Polytechnique Fédérale de Lausanne]], aims to construct a biophysically detailed simulation of a cortical column on the [[Blue Gene]] [[supercomputer]].
<ref>{{Cite journal|last1=Brooks|first1=R.|last2=Hassabis|first2=D.|last3=Bray|first3=D.|last4=Shashua|first4=A.|date=2012-02-22|title=Turing centenary: Is the brain a good model for machine intelligence?|journal=Nature|language=En|volume=482|issue=7386|pages=462–463|bibcode=2012Natur.482..462.|doi=10.1038/482462a|issn=0028-0836|pmid=22358812|s2cid=205070106|doi-access=free}}</ref> although mutual inspiration exists and sometimes there is no strict limit between fields,<ref>{{Cite book|url=https://books.google.com/books?id=uV9TZzOITMwC&q=%22biological%20plausibility%22&pg=PA17|title=Neural Network Perspectives on Cognition and Adaptive Robotics|last=Browne|first=A.|date=1997-01-01|publisher=CRC Press|isbn=9780750304559|language=en}}</ref><ref>{{Cite journal|last1=Zorzi|first1=Marco|last2=Testolin|first2=Alberto|last3=Stoianov|first3=Ivilin P.|date=2013-08-20|title=Modeling language and cognition with deep unsupervised learning: a tutorial overview|journal=Frontiers in Psychology|volume=4|pages=515|doi=10.3389/fpsyg.2013.00515|issn=1664-1078|pmc=3747356|pmid=23970869|doi-access=free}}</ref><ref>{{Cite journal|last1=Shai|first1=Adam|last2=Larkum|first2=Matthew Evan|date=2017-12-05|title=Branching into brains|journal=eLife|language=en|volume=6|doi=10.7554/eLife.33066|issn=2050-084X|pmc=5716658|pmid=29205152 |doi-access=free }}</ref> with model abstraction in computational neuroscience depending on research scope and the granularity at which biological entities are analyzed.


Models in theoretical neuroscience are aimed at capturing the essential features of the biological system at multiple spatial-temporal scales, from membrane currents, and chemical coupling via [[neural oscillation|network oscillations]], columnar and topographic architecture, nuclei, all the way up to psychological faculties like memory, learning and behavior. These computational models frame hypotheses that can be directly tested by biological or psychological experiments.
==Major Topics==
Research in computational neuroscience can be roughly categorized into several lines of inquiries. Most computational neuroscientists collaborate closely with experimentalists in analyzing novel data and synthesizing new models of biological phenomena.


==History==
===Single Neuron Modeling===
The term 'computational neuroscience' was introduced by [[Eric L. Schwartz]], who organized a conference, held in 1985 in [[Carmel, California]], at the request of the Systems Development Foundation to provide a summary of the current status of a field which until that point was referred to by a variety of names, such as neural modeling, brain theory and neural networks. The proceedings of this definitional meeting were published in 1990 as the book ''Computational Neuroscience''.<ref>{{cite book |author=Schwartz, Eric |title=Computational neuroscience |publisher=MIT Press |location=Cambridge, Mass |year=1990 |isbn=978-0-262-19291-0 }}</ref> The first of the annual open international meetings focused on Computational Neuroscience was organized by [[James M. Bower]] and John Miller in [[San Francisco, California]] in 1989.<ref>{{cite book |author=Bower, James M. |title=20 years of Computational neuroscience |publisher=Springer |location=Berlin, Germany |year=2013 |isbn=978-1461414230}}</ref> The first graduate educational program in computational neuroscience was organized as the Computational and Neural Systems Ph.D. program at the [[California Institute of Technology]] in 1985.

The early historical roots of the field<ref>{{Cite journal |last1=Fan |first1=Xue |last2=Markram |first2=Henry |date=2019 |title=A Brief History of Simulation Neuroscience |journal=Frontiers in Neuroinformatics |volume=13 |page=32 |doi=10.3389/fninf.2019.00032 |doi-access=free |issn=1662-5196 |pmc=6513977 |pmid=31133838}}</ref> can be traced to the work of people including [[Louis Lapicque]], [[Alan Hodgkin|Hodgkin]] & [[Andrew Huxley|Huxley]], [[David H. Hubel|Hubel]] and [[Torsten Wiesel|Wiesel]], and [[David Marr (psychologist)|David Marr]]. Lapicque introduced the [[integrate and fire]] model of the neuron in a seminal article published in 1907,<ref>{{cite journal |author=Lapicque L |title= Recherches quantitatives sur l'excitation électrique des nerfs traitée comme une polarisation |journal=J. Physiol. Pathol. Gen. |volume=9 |pages=620–635 |year=1907}}</ref> a model still popular for [[artificial neural network]]s studies because of its simplicity (see a recent review<ref>{{cite journal |vauthors=Brunel N, Van Rossum MC |title= Lapicque's 1907 paper: from frogs to integrate-and-fire |journal=Biol. Cybern. |volume=97 |pages=337–339 |year=2007 |pmid=17968583 |doi=10.1007/s00422-007-0190-0 |issue=5–6|s2cid= 17816096 }}</ref>).

About 40 years later, [[Alan Hodgkin|Hodgkin]] and [[Andrew Huxley|Huxley]] developed the [[voltage clamp]] and created the first biophysical model of the [[action potential]]. [[David H. Hubel|Hubel]] and [[Torsten Wiesel|Wiesel]] discovered that neurons in the [[primary visual cortex]], the first cortical area to process information coming from the [[retina]], have oriented receptive fields and are organized in columns.<ref>{{cite journal |vauthors=Hubel DH, Wiesel TN |title=Receptive fields, binocular interaction and functional architecture in the cat's visual cortex |journal=J. Physiol. |volume=160 |pages=106–54 |year=1962 |pmid=14449617 |pmc=1359523 |doi= 10.1113/jphysiol.1962.sp006837|url=http://www.jphysiol.org/cgi/pmidlookup?view=long&pmid=14449617 |issue=1}}</ref> David Marr's work focused on the interactions between neurons, suggesting computational approaches to the study of how functional groups of neurons within the [[hippocampus]] and [[neocortex]] interact, store, process, and transmit information. Computational modeling of biophysically realistic neurons and dendrites began with the work of [[Wilfrid Rall]], with the first multicompartmental model using [[cable theory]].

==Major topics==
Research in computational neuroscience can be roughly categorized into several lines of inquiry. Most computational neuroscientists collaborate closely with experimentalists in analyzing novel data and synthesizing new models of biological phenomena.

===Single-neuron modeling===
{{main|Biological neuron models}}
{{main|Biological neuron models}}
Even single neurons have complex biophysical characteristics. Hodgkin and Huxley's [[Hodgkin-Huxley model|original model]] only employed two voltage-sensitive currents, the fast-acting sodium and the inward-rectifying potassium. Though successful in predicting the timing and qualitative features of the action potential, it nevertheless failed to predict a number of important features such as adaptation and shunting. Scientists now believe that there are a wide variety of voltage-sensitive currents, and the implications of the differing dynamics, modulations and sensitivity of these currents is an important topic of computational neuroscience (for reference, see Johnston and Wu, 1994).
Even a single neuron has complex biophysical characteristics and can perform computations (e.g.<ref>{{cite journal |author=Forrest MD |title=Intracellular Calcium Dynamics Permit a Purkinje Neuron Model to Perform Toggle and Gain Computations Upon its Inputs. |journal=Frontiers in Computational Neuroscience |volume=8 |pages=86 |year=2014 | doi=10.3389/fncom.2014.00086 |pmid=25191262 |pmc=4138505|doi-access=free }}</ref>). Hodgkin and Huxley's [[Hodgkin–Huxley model|original model]] only employed two voltage-sensitive currents (Voltage sensitive ion channels are glycoprotein molecules which extend through the lipid bilayer, allowing ions to traverse under certain conditions through the axolemma), the fast-acting sodium and the inward-rectifying potassium. Though successful in predicting the timing and qualitative features of the action potential, it nevertheless failed to predict a number of important features such as adaptation and [[Shunting (neurophysiology)|shunting]]. Scientists now believe that there are a wide variety of voltage-sensitive currents, and the implications of the differing dynamics, modulations, and sensitivity of these currents is an important topic of computational neuroscience.<ref>{{cite book |author1=Wu, Samuel Miao-sin |author2=Johnston, Daniel |title=Foundations of cellular neurophysiology |publisher=MIT Press |location=Cambridge, Mass |year=1995 |isbn=978-0-262-10053-3 }}</ref>


The computational functions of complex [[dendrites]] are also under intense investigation. There is a large body of literature regarding how different currents interact with geometric properties of neurons (for reference, see Koch, 1998).
The computational functions of complex [[dendrites]] are also under intense investigation. There is a large body of literature regarding how different currents interact with geometric properties of neurons.<ref>{{cite book |author=Koch, Christof |title=Biophysics of computation: information processing in single neurons |publisher=Oxford University Press |location=Oxford [Oxfordshire] |year=1999 |isbn=978-0-19-510491-2 }}</ref>


There are many software packages, such as [[GENESIS (software)|GENESIS]] and [[Neuron (software)|NEURON]], that allow rapid and systematic ''in silico'' modeling of realistic neurons. [[Blue Brain]], a project founded by [[Henry Markram]] from the [[École Polytechnique Fédérale de Lausanne]], aims to construct a biophysically detailed simulation of a [[cortical column]] on the [[Blue Gene]] [[supercomputer]].
Some models are also tracking biochemical pathways at very small scales such as spines or synaptic clefts.


Modeling the richness of biophysical properties on the single-neuron scale can supply mechanisms that serve as the building blocks for network dynamics.<ref>{{cite journal|author=Forrest MD|year=2014|title=Intracellular Calcium Dynamics Permit a Purkinje Neuron Model to Perform Toggle and Gain Computations Upon its Inputs.|journal=Frontiers in Computational Neuroscience|volume=8|pages=86|doi=10.3389/fncom.2014.00086|pmc=4138505|pmid=25191262|doi-access=free}}</ref> However, detailed neuron descriptions are computationally expensive and this computing cost can limit the pursuit of realistic network investigations, where many neurons need to be simulated. As a result, researchers that study large neural circuits typically represent each neuron and synapse with an artificially simple model, ignoring much of the biological detail. Hence there is a drive to produce simplified neuron models that can retain significant biological fidelity at a low computational overhead. Algorithms have been developed to produce faithful, faster running, simplified surrogate neuron models from computationally expensive, detailed neuron models.<ref>{{cite journal |author=Forrest MD |title=Simulation of alcohol action upon a detailed Purkinje neuron model and a simpler surrogate model that runs >400 times faster |journal= BMC Neuroscience | volume=16 |issue=27 |pages=27 | date=April 2015 |doi=10.1186/s12868-015-0162-6 |pmid=25928094 |pmc=4417229 |doi-access=free }}</ref>
===Development, Axonal Patterning and Guidance===
How do [[axons]] and [[dendrites]] form during development? How do axons know where to target and how to reach these targets? How do neurons migrate to the proper position in the central and peripheral systems? How do synapses form? We know from molecular biology that distinct parts of the nervous system release distinct chemical cues, from [[growth factors]] to [[hormones]] that modulate and influence the growth and development of functional connections between neurons.


=== Modeling Neuron-glia interactions ===
Theoretical investigations into the formation and patterning of synaptic connection and morphology is still nascent. One hypothesis that has recently garnered some attention is the ''minimal wiring hypothesis'', which postulates that the formation of axons and dendrites effectively minimizes resource allocation while maintaining maximal information storage. (for a review, see [http://www.ncbi.nlm.nih.gov/sites/entrez?Db=pubmed&Cmd=ShowDetailView&TermToSearch=15483599&ordinalpos=1&itool=EntrezSystem2.PEntrez.Pubmed.Pubmed_ResultsPanel.Pubmed_RVDocSum Chklovskii, 2004])
Glial cells participate significantly in the regulation of neuronal activity at both the cellular and the network level. Modeling this interaction allows to clarify the [[potassium cycle]],<ref>{{Cite web |title=Dynamics of Ion Fluxes between Neurons, Astrocytes and the Extracellular Space during Neurotransmission |url=https://cyberleninka.ru/article/n/dynamics-of-ion-fluxes-between-neurons-astrocytes-and-the-extracellular-space-during-neurotransmission/viewer |access-date=2023-03-14 |website=cyberleninka.ru}}</ref><ref>{{Cite journal |last1=Sibille |first1=Jérémie |last2=Duc |first2=Khanh Dao |last3=Holcman |first3=David |last4=Rouach |first4=Nathalie |date=2015-03-31 |title=The Neuroglial Potassium Cycle during Neurotransmission: Role of Kir4.1 Channels |journal=PLOS Computational Biology |language=en |volume=11 |issue=3 |pages=e1004137 |doi=10.1371/journal.pcbi.1004137 |issn=1553-7358 |pmc=4380507 |pmid=25826753|bibcode=2015PLSCB..11E4137S |doi-access=free }}</ref> so important for maintaining homeostasis and to prevent epileptic seizures. Modeling reveals the role of glial protrusions that can penetrate in some cases the synaptic cleft to interfere with the synaptic transmission and thus control synaptic communication.<ref>{{Cite journal |last1=Pannasch |first1=Ulrike |last2=Freche |first2=Dominik |last3=Dallérac |first3=Glenn |last4=Ghézali |first4=Grégory |last5=Escartin |first5=Carole |last6=Ezan |first6=Pascal |last7=Cohen-Salmon |first7=Martine |last8=Benchenane |first8=Karim |last9=Abudara |first9=Veronica |last10=Dufour |first10=Amandine |last11=Lübke |first11=Joachim H. R. |last12=Déglon |first12=Nicole |last13=Knott |first13=Graham |last14=Holcman |first14=David |last15=Rouach |first15=Nathalie |date=April 2014 |title=Connexin 30 sets synaptic strength by controlling astroglial synapse invasion |url=https://www.nature.com/articles/nn.3662 |journal=Nature Neuroscience |language=en |volume=17 |issue=4 |pages=549–558 |doi=10.1038/nn.3662 |pmid=24584052 |s2cid=554918 |issn=1546-1726}}</ref>

===Development, axonal patterning, and guidance===
Computational neuroscience aims to address a wide array of questions, including: How do [[axons]] and [[dendrites]] form during development? How do axons know where to target and how to reach these targets? How do neurons migrate to the proper position in the central and peripheral systems? How do synapses form? We know from [[molecular biology]] that distinct parts of the nervous system release distinct chemical cues, from [[growth factors]] to [[hormones]] that modulate and influence the growth and development of functional connections between neurons.

Theoretical investigations into the formation and patterning of synaptic connection and morphology are still nascent. One hypothesis that has recently garnered some attention is the ''minimal wiring hypothesis'', which postulates that the formation of axons and dendrites effectively minimizes resource allocation while maintaining maximal information storage.<ref>{{cite journal|author3-link=Karel Svoboda (scientist) |vauthors=Chklovskii DB, Mel BW, Svoboda K |title=Cortical rewiring and information storage |journal=Nature |volume=431 |issue=7010 |pages=782–8 |date=October 2004|pmid=15483599 |doi=10.1038/nature03012 |bibcode = 2004Natur.431..782C |s2cid=4430167 }}<br/>Review article</ref>


===Sensory processing===
===Sensory processing===
Early models of sensory processing understood within a theoretical framework is credited to [[Horace Barlow]]. Somewhat similar to the minimal wiring hypothesis described in the preceding section, Barlow understood the processing of the early sensory systems to be a form of [[efficient coding hypothesis|efficient coding]], where the neurons encoded information which minimized the number of spikes. Experimental and computational work have since supported this hypothesis in one form or another.
Early models on sensory processing understood within a theoretical framework are credited to [[Horace Barlow]]. Somewhat similar to the minimal wiring hypothesis described in the preceding section, Barlow understood the processing of the early sensory systems to be a form of [[efficient coding hypothesis|efficient coding]], where the neurons encoded information which minimized the number of spikes. Experimental and computational work have since supported this hypothesis in one form or another. For the example of visual processing, efficient coding is manifested in the
forms of efficient spatial coding, color coding, temporal/motion coding, stereo coding, and combinations of them.<ref>Zhaoping L. 2014, [https://www.oxfordscholarship.com/view/10.1093/acprof:oso/9780199564668.001.0001/acprof-9780199564668-chapter-3 The efficient coding principle ], chapter 3, of the textbook [https://global.oup.com/academic/product/understanding-vision-9780198829362?cc=de&lang=en& Understanding vision: theory, models, and data ]</ref>


Further along the visual pathway, even the efficiently coded visual information is too much for the capacity of the information bottleneck, the visual attentional bottleneck.<ref>see visual spational attention https://en.wikipedia.org/wiki/Visual_spatial_attention</ref> A subsequent theory, [[V1 Saliency Hypothesis|V1 Saliency Hypothesis (V1SH)]], has been developed on exogenous attentional selection of a fraction of visual input for further processing, guided by a bottom-up saliency map in the primary visual cortex.<ref name=Li2002>Li. Z. 2002 [https://www.sciencedirect.com/science/article/abs/pii/S1364661300018179 A saliency map in primary visual cortex] Trends in Cognitive Sciences
Current research in sensory processing is divided among biophysical modelling of different subsystems and more theoretical modelling function of perception. Current models of perception have suggested that the brain performs some form of [[Bayesian inference]] and integration of different sensory information in generating our perception of the physical world.
vol. 6, Pages 9-16, and Zhaoping, L. 2014, [https://www.oxfordscholarship.com/view/10.1093/acprof:oso/9780199564668.001.0001/acprof-9780199564668-chapter-5 The V1 hypothesis—creating a bottom-up saliency map for preattentive selection and segmentation] in the book [https://www.oxfordscholarship.com/view/10.1093/acprof:oso/9780199564668.001.0001/acprof-9780199564668 Understanding Vision: Theory, Models, and Data]</ref>

Current research in sensory processing is divided among a biophysical modeling of different subsystems and a more theoretical modeling of perception. Current models of perception have suggested that the brain performs some form of [[Bayesian approaches to brain function|Bayesian inference]] and integration of different sensory information in generating our perception of the physical world.<ref>{{cite journal|last1=Weiss|first1=Yair|last2=Simoncelli|first2=Eero P.|last3=Adelson|first3=Edward H.|title=Motion illusions as optimal percepts|journal=Nature Neuroscience|date=20 May 2002|volume=5|issue=6|pages=598–604|doi=10.1038/nn0602-858|pmid=12021763|s2cid=2777968}}</ref><ref>{{cite journal|last1=Ernst|first1=Marc O.|last2=Bülthoff|first2=Heinrich H.|title=Merging the senses into a robust percept|journal=Trends in Cognitive Sciences|date=April 2004|volume=8|issue=4|pages=162–169|doi=10.1016/j.tics.2004.02.002|pmid=15050512|citeseerx=10.1.1.299.4638|s2cid=7837073}}</ref>

===Motor control===
Many models of the way the brain controls movement have been developed. This includes models of processing in the brain such as the cerebellum's role for error correction, skill learning in motor cortex and the basal ganglia, or the control of the vestibulo ocular reflex. This also includes many normative models, such as those of the Bayesian or optimal control flavor which are built on the idea that the brain efficiently solves its problems.


===Memory and synaptic plasticity===
===Memory and synaptic plasticity===
{{main|Synaptic plasticity}}
{{main|Synaptic plasticity}}
Earlier models of memory are primarily based on the postulates of [[Hebbian learning]]. Biologically relevant models such as [[Hopfield net]] have been developed to address the properties of associative, rather than content-addressable style of memory that occur in biological systems. These attempts are primarily focusing on the formation of medium-term and long-term memory, localizing in the hippocampus. Models of working memory, relying on theories of network oscillations and persistent activity, have been built to capture some features of the prefrontal cortex in context-related memory. (For review, see Durstewitz et al, 2000)
Earlier models of [[memory]] are primarily based on the postulates of [[Hebbian learning]]. Biologically relevant models such as [[Hopfield net]] have been developed to address the properties of associative (also known as "content-addressable") style of memory that occur in biological systems. These attempts are primarily focusing on the formation of medium- and [[long-term memory]], localizing in the [[hippocampus]].


One of the major problems in biological memory is how it is maintained and changed through multiple time scales. Unstable [[synapses]] are easy to train but also prone to stochastic disruption. Stable [[synapses]] forget less easily, but they are also harder to consolidate. One recent computational hypothesis involves cascades of plasticity (Fusi et al, 2005) that allow synapses to function at multiple time scales. Stereochemically detailed models of the [[acetylcholine receptor]]-based synapse with [[Monte Carlo method]], working at the time scale of microseconds, have been built (Coggan et al, 2005). It is likely that computational tools will contribute greatly to our understanding of how synapses function and change in relation to external stimulus in the coming decades.
One of the major problems in neurophysiological memory is how it is maintained and changed through multiple time scales. Unstable [[synapses]] are easy to train but also prone to stochastic disruption. Stable [[synapses]] forget less easily, but they are also harder to consolidate. It is likely that computational tools will contribute greatly to our understanding of how synapses function and change in relation to external stimulus in the coming decades.


===Behaviors of Networks===
===Behaviors of networks===
Biological neurons are connected to each other in a complex, recurrent fashion. These connections are, unlike most [[artificial neural networks]], sparse and most likely, specific. It is not known how information is transmitted through such sparsely connected networks. It is also unknown what the computational functions, if any, of these specific connectivity patterns are.
Biological neurons are connected to each other in a complex, recurrent fashion. These connections are, unlike most [[artificial neural networks]], sparse and usually specific. It is not known how information is transmitted through such sparsely connected networks, although specific areas of the brain, such as the [[visual cortex]], are understood in some detail.<ref>{{Cite journal|last1=Olshausen|first1=Bruno A.|last2=Field|first2=David J.|date=1997-12-01|title=Sparse coding with an overcomplete basis set: A strategy employed by V1?|journal=Vision Research|volume=37|issue=23|pages=3311–3325|doi=10.1016/S0042-6989(97)00169-7|pmid=9425546|s2cid=14208692|doi-access=free}}</ref> It is also unknown what the computational functions of these specific connectivity patterns are, if any.


The interactions of neurons in a small network can be often reduced to simple models such as the [[Ising model]]. The [[statistical mechanics]] of such simple systems are well-characterized theoretically. There have been some recent evidence that suggests that dynamics of arbitrary neuronal networks can be reduced to pairwise interactions (Schneidman et al, 2006; Shlens et al, 2006.) It's unknown, however, whether such descriptive dynamics impart any important computational function. With the emergence of [[two-photon microscopy]] and [[calcium imaging]], we now have powerful experimental methods with which to test the new theories regarding neuronal networks.
The interactions of neurons in a small network can be often reduced to simple models such as the [[Ising model]]. The [[statistical mechanics]] of such simple systems are well-characterized theoretically. Some recent evidence suggests that dynamics of arbitrary neuronal networks can be reduced to pairwise interactions.<ref>{{cite journal |vauthors=Schneidman E, Berry MJ, Segev R, Bialek W |title=Weak pairwise correlations imply strongly correlated network states in a neural population |journal=Nature |volume=440 |issue=7087 |pages=1007–12 |year=2006 |pmid=16625187 |pmc=1785327 |doi=10.1038/nature04701 |bibcode=2006Natur.440.1007S|arxiv = q-bio/0512013 }}</ref> It is not known, however, whether such descriptive dynamics impart any important computational function. With the emergence of [[two-photon microscopy]] and [[calcium imaging]], we now have powerful experimental methods with which to test the new theories regarding neuronal networks.


In some cases the complex interactions between ''inhibitory'' and ''excitatory'' neurons can be simplified using [[mean-field theory]], which gives rise to the [[Wilson–Cowan model|population model]] of neural networks.<ref>{{cite journal |author1=Wilson, H. R. |author2=Cowan, J.D. |title=A mathematical theory of the functional dynamics of cortical and thalamic nervous tissue
While many neuro-theorists prefer models with reduced complexity, others argue that uncovering structure function relations depends on including as much neuronal and network structure as possible. Models of this type are typically built in large simulations platforms like [[GENESIS_%28software%29]] or [[Neuron_%28software%29]]. There have been some attempts to provide unified methods that bridge, and integrate, these levels of complexity (Eliasmith & Anderson, 2003).
|journal=Kybernetik |volume=13 |issue=2 |pages=55–80 |year=1973 |doi= 10.1007/BF00288786|pmid=4767470 |s2cid=292546 }}</ref> While many neurotheorists prefer such models with reduced complexity, others argue that uncovering structural-functional relations depends on including as much neuronal and network structure as possible. Models of this type are typically built in large simulation platforms like GENESIS or NEURON. There have been some attempts to provide unified methods that bridge and integrate these levels of complexity.<ref>{{cite book |author1=Anderson, Charles H. |author2=Eliasmith, Chris |title=Neural Engineering: Computation, Representation, and Dynamics in Neurobiological Systems (Computational Neuroscience) |publisher=The MIT Press |location=Cambridge, Mass |year=2004 |isbn=978-0-262-55060-4 }}</ref>


===Cognition, Discrimination and Learning===
===Visual attention, identification, and categorization===
Visual attention can be described as a set of mechanisms that limit some processing to a subset of incoming stimuli.<ref>{{cite book|author1=Marvin M. Chun |author2=Jeremy M. Wolfe |author3=E. B. Goldstein | title=Blackwell Handbook of Sensation and Perception |url=https://archive.org/details/blackwellhandboo00gold |url-access=limited | publisher=Blackwell Publishing Ltd |year=2001 | pages=[https://archive.org/details/blackwellhandboo00gold/page/n284 272]–310 |isbn=978-0-631-20684-2}}</ref> Attentional mechanisms shape what we see and what we can act upon. They allow for concurrent selection of some (preferably, relevant) information and inhibition of other information. In order to have a more concrete specification of the mechanism underlying visual attention and the binding of features, a number of computational models have been proposed aiming to explain psychophysical findings. In general, all models postulate the existence of a saliency or priority map for registering the potentially interesting areas of the retinal input, and a gating mechanism for reducing the amount of incoming visual information, so that the limited computational resources of the brain can handle it.<ref>{{cite book|author1=Edmund Rolls |author2=Gustavo Deco | title=Computational Neuroscience of Vision | publisher=Oxford Scholarship Online | year=2012 | isbn=978-0-198-52488-5}}</ref>
Computational modeling of higher cognitive functions has only begun recently. Experimental data comes primarily from [[single unit recording]] in [[primates]]. The [[frontal lobe]] and [[parietal lobe]] function as integrators of information from multiple sensory modalities. There are some tentative ideas regarding how simple mutually inhibitory functional circuits in these areas may carry out biologically relevant computation (Machens et al, 2005).
An example theory that is being extensively tested behaviorally and physiologically is the [[V1 Saliency Hypothesis]] that a bottom-up saliency map is created in the primary visual cortex to guide attention exogenously.<ref name=Li2002 /> Computational neuroscience provides a mathematical framework for studying the mechanisms involved in brain function and allows complete simulation and prediction of neuropsychological syndromes.


===Cognition, discrimination, and learning===
The [[brain]] seems to be able to discriminate and adapt particularly well in certain contexts. For instance, human beings seem to have an enormous capacity for memorizing and recognizing faces. One of the key goals of computational neuroscience is to dissect how biological systems carry out these complex computations efficiently and potentially replicate these processes in building intelligent machines.
Computational modeling of higher cognitive functions has only recently{{When|date=February 2016}} begun. Experimental data comes primarily from [[single-unit recording]] in [[primates]]. The [[frontal lobe]] and [[parietal lobe]] function as integrators of information from multiple sensory modalities. There are some tentative ideas regarding how simple mutually inhibitory functional circuits in these areas may carry out biologically relevant computation.<ref>{{cite journal |vauthors=Machens CK, Romo R, Brody CD |title=Flexible control of mutual inhibition: a neural model of two-interval discrimination |journal=Science |volume=307 |issue=5712 |pages=1121–4 |year=2005 |pmid=15718474 |doi=10.1126/science.1104171 |bibcode = 2005Sci...307.1121M |citeseerx=10.1.1.523.4396 |s2cid=45378154 }}</ref>


The [[brain]] seems to be able to discriminate and adapt particularly well in certain contexts. For instance, human beings seem to have an enormous capacity for memorizing and [[face perception|recognizing faces]]. One of the key goals of computational neuroscience is to dissect how biological systems carry out these complex computations efficiently and potentially replicate these processes in building intelligent machines.
===Consciousness===
The ultimate goal of neuroscience is to be able to explain the every day experience of conscious life. [[Francis Crick]] and [[Christof Koch]] made some attempts in formulating a consistent framework for future work in [[neural correlates of consciousness]] (NCC), though much of the work in this field remains speculative. (for a review, see Koch and Crick, 2003). Another attempt is done by [[Andrew & Alexander Fingelkurts]]: they are developing the [http://www.bm-science.com/team/chapt3.pdf Operational Architectonics] theory of brain-mind functioning. This theory treats consciousness as a biological phenomenon in the brain which is realized by the highly organized macro-level electrophysiological (EEG) phenomena (metastable operational modules), which are brought to existence by the coordinated electrical activity (operational synchrony) of many neuronal populations dispersed throughout the brain (for a review see Fingelkurts An.A. and Fingelkurts Al.A., 2001; 2004; 2006).


The brain's large-scale organizational principles are illuminated by many fields, including biology, psychology, and clinical practice. [[Integrative neuroscience]] attempts to consolidate these observations through unified descriptive models and databases of behavioral measures and recordings. These are the bases for some quantitative modeling of large-scale brain activity.<ref>{{cite journal |vauthors=Robinson PA, Rennie CJ, Rowe DL, O'Connor SC, Gordon E | title=Multiscale brain modelling | journal=Philosophical Transactions of the Royal Society B | volume=360 | issue=1457|pages=1043–1050|year=2005|doi=10.1098/rstb.2005.1638 |pmid=16087447 |pmc=1854922 }}</ref>
==See also==
* [[Connectionism]]
* [[Neural network]]
* [[Biological neuron models]]
* [[Electrophysiology]]
* [[List of publications in biology#neuroscience|Important publications in neuroscience]]
* [[Brain-computer interface]]
* [[Neural engineering]]
* [[Neurotechnology]]
* [[Neuroinformatics]]
* [[Computational Neurogenetic Modeling]]


The Computational Representational Understanding of Mind ([[CRUM]]) is another attempt at modeling human cognition through simulated processes like acquired rule-based systems in decision making and the manipulation of visual representations in decision making.
==References==
* Chklovskii, DB (2004) "Synaptic connectivity and neuronal morphology: two sides of the same coin", Neuron. 43(5):609-17
* Churchland, P. S. & T. J. Sejnowski (1992) ''The Computational Brain'', [[MIT Press]], ISBN 0-262-03188-4.
* Eliasmith, C & C.H. Anderson (2003). ''Neural Engineering: Computation, Representation, and Dynamics in Neurobiological Systems'' [[MIT Press]], ISBN 0-262-55060-1.
* Coggan JS, Bartol TM, Esquenazi E et al, "Evidence for ectopic neurotransmission at a neuronal synapse.", Science.,2005 Jul 15;309(5733):446-51
* Peter Dayan, L.F. Abbott: ''Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems'', [[MIT Press]], 2001, ISBN 0-262-04199-5.
* Durstewitz D, Seamans JK, Sejnowski TJ., (2000) "Neurocomputational models of working memory.", Nat Neurosci. 2000 Nov; Suppl:1184-91.
* Fingelkurts An.A. and Fingelkurts Al.A., (2001) "Operational architectonics of the human brain biopotential field: towards solving the mind-brain problem." Brain and Mind 2:261–296
* Fingelkurts An.A. and Fingelkurts Al.A., (2004) "Making complexity simpler: multivariability and metastability in the Brain." Int. J. Neurosci., vol. 114, pp. 843-862
* Fingelkurts An.A. and Fingelkurts Al.A., (2006) "Timing in cognition and EEG brain dynamics: discreteness versus continuity." Cognitive Proces., vol. 7, pp. 135-162
* Fusi S, Drew PJ, Abbott LF., "Cascade models of synaptically stored memories", Neuron. 2005 Feb 17;45(4):599-611
* Johnston D and Wu SM, ''Foundations of Cellular Neurophysiology'', [[MIT Press]], 1994, ISBN 0-262-10053-3.
* Hodgkin, A. L. and Huxley, A. F. (1952) "A Quantitative Description of Membrane Current and its Application to Conduction and Excitation in Nerve" ''Journal of Physiology'' 117:500-544
* Hubel DH, Wiesel TN (1962) "Receptive fields, binocular interaction and functional architecture in the cat's visual cortex" ''Journal of Physiology'' 160:106-154.
* Koch C ''Biophysics of Computation: Information Processing in Single Neurons'', [[Oxford University Press]], 1998, ISBN 0-19-510491-9.
* Koch, C and Crick, F (2003) "A framework for consciousness", Nat Neurosci. 2003 Feb;6(2):119-26.
* Machens CK, Romo R, Brody CD. (2005) "Flexible control of mutual inhibition: a neural model of two-interval discrimination.", Science. 2005 Feb 18;307(5712):1121-4.
* F. Rieke, D. Warland, W. Bialek and R. de Ruyter van Steveninck: ''Spikes: Exploring the Neural Code'', [[MIT Press]], 1999, ISBN 0-262-68108-0.
* Schneidman E, Berry MJ 2nd, Segev R, Bialek W. (2006) "Weak pairwise correlations imply strongly correlated network states in a neural population.", Nature. 2006 Apr 20;440(7087):1007-12.
* Eric L. Schwartz, ed.: ''Computational Neuroscience'', [[MIT Press]], 1990, ISBN 0-262-19291-8.
* Erik de Schutter, ed.: ''Computational Neuroscience - Realistic Modeling for Experimentalists'', [[CRC Press]], 2000, ISBN 0-8493-2068-2.
* J. L.. van Hemmen, T. J. Sejnowski, eds.: ''23 Problems in Systems Neuroscience'' [[Oxford University Press]], 2005 ISBN 0-19-514822-3.


===[[Consciousness]]===
==External links==
One of the ultimate goals of psychology/neuroscience is to be able to explain the everyday experience of conscious life. [[Francis Crick]], [[Giulio Tononi]] and [[Christof Koch]] made some attempts to formulate consistent frameworks for future work in [[neural correlates of consciousness]] (NCC), though much of the work in this field remains speculative.<ref>{{cite journal |vauthors=Crick F, Koch C |title=A framework for consciousness |journal=Nat. Neurosci. |volume=6 |issue=2 |pages=119–26 |year=2003 |pmid=12555104 |doi=10.1038/nn0203-119|s2cid=13960489 |url= https://zenodo.org/record/852680 }}</ref>
===Books===
* Eric L. Schwartz, ed.: ''Computational Neuroscience'', [[MIT Press]], 1990, ISBN 0-262-19291-8.
===Journals===
* [http://www.informaworld.com/network Network: Computation in Neural Systems]
* [http://www.springerlink.com/openurl.asp?genre=journal&issn=0340-1200 Biological Cybernetics]
* [http://www.kluweronline.com/issn/0929-5313 Journal of Computational Neuroscience]
* [http://neco.mitpress.org/ Neural Computation]
* [http://www.sciencedirect.com/science/journal/08936080 Neural Networks]
* [http://www.elsevier.com/locate/neucom Neurocomputing]
* [http://www.springerlink.com/content/1871-4099/ Cognitive Neurodynamics]


===Computational clinical neuroscience===
===Software===
[[Computational clinical neuroscience]] is a field that brings together experts in neuroscience, [[neurology]], [[psychiatry]], [[decision sciences]] and computational modeling to quantitatively define and investigate problems in [[neurological disorders|neurological]] and [[mental disorders|psychiatric diseases]], and to train scientists and clinicians that wish to apply these models to diagnosis and treatment.<ref>{{cite journal | last1=Adaszewski | first1=Stanisław | last2=Dukart | first2=Juergen | last3=Kherif | first3=Ferath | last4=Frackowiak | first4=Richard | last5=Draganski | first5=Bogdan | author6=Alzheimer's Disease Neuroimaging Initiative |title=How early can we predict Alzheimer's disease using computational anatomy? |journal=Neurobiol Aging |volume=34 |issue=12 |pages=2815–26 |year=2013 |doi=10.1016/j.neurobiolaging.2013.06.015 |pmid=23890839|s2cid=1025210}}</ref><ref>{{cite journal |vauthors=Friston KJ, Stephan KE, Montague R, Dolan RJ |title=Computational psychiatry: the brain as a phantastic organ |journal=Lancet Psychiatry |volume=1 |issue=2 |pages=148–58 |year=2014 |doi=10.1016/S2215-0366(14)70275-5 |pmid=26360579 |s2cid=15504512 }}</ref>
* [http://www.genesis-sim.org/GENESIS/ Genesis], a general neural simulation system
* [http://www.neuron.yale.edu/ Neuron], a neural network simulator
* [http://www.nest-initiative.org NEST], a simulation tool for large neuronal systems.
* [http://www.neuroconstruct.org Neuroconstruct], software for developing biologically realistic 3D neural networks.
* [http://neurofitter.sourceforge.net Neurofitter], a parameter tuning package for electrophysiological neuron models.
* [http://www.neurojet.net Neurojet], a neural network simulator specialized for the hippocampus
* [http://www-2.cs.cmu.edu/~dst/HHsim/ HHsim], a neuronal membrane simulator
* [http://www.mcell.cnl.salk.edu/ MCell], A Monte Carlo Simulator of Cellular Microphysiology
* [[Emergent (software)|Emergent]], neural simulation software
* Python tools : http://neuralensemble.org/
** [http://neuralensemble.org/trac/PyNN pyNN]


=== Predictive computational neuroscience ===
===Conferences===
Predictive computational neuroscience is a recent field that combines signal processing, neuroscience, clinical data and machine learning to predict the brain during coma <ref>{{Cite journal |last1=Floyrac |first1=Aymeric |last2=Doumergue |first2=Adrien |last3=Legriel |first3=Stéphane |last4=Deye |first4=Nicolas |last5=Megarbane |first5=Bruno |last6=Richard |first6=Alexandra |last7=Meppiel |first7=Elodie |last8=Masmoudi |first8=Sana |last9=Lozeron |first9=Pierre |last10=Vicaut |first10=Eric |last11=Kubis |first11=Nathalie |last12=Holcman |first12=David |date=2023 |title=Predicting neurological outcome after cardiac arrest by combining computational parameters extracted from standard and deviant responses from auditory evoked potentials |journal=Frontiers in Neuroscience |volume=17 |page=988394 |doi=10.3389/fnins.2023.988394 |pmid=36875664 |pmc=9975713 |issn=1662-453X|doi-access=free }}</ref> or anesthesia.<ref>{{Cite journal |last1=Sun |first1=Christophe |last2=Holcman |first2=David |date=2022-08-01 |title=Combining transient statistical markers from the EEG signal to predict brain sensitivity to general anesthesia |url=https://www.sciencedirect.com/science/article/pii/S174680942200235X |journal=Biomedical Signal Processing and Control |language=en |volume=77 |pages=103713 |doi=10.1016/j.bspc.2022.103713 |s2cid=248488365 |issn=1746-8094}}</ref> For example, it is possible to anticipate deep brain states using the EEG signal. These states can be used to anticipate hypnotic concentration to administrate to the patient.
* [http://www.cosyne.org Computational and Systems Neuroscience (COSYNE)] - a computational neuroscience meeting with a systems neuroscience focus.
* [http://www.cnsorg.org Annual Computational Neuroscience Meeting (CNS)] - a yearly computational neuroscience meeting.
* [http://www.nips.cc Neural Information Processing Systems (NIPS)] - a leading annual conference covering other machine learning topics as well.
* [http://www.ccnconference.org Computational Cognitive Neuroscience Conference (CCNC)] - a yearly conference.
* [http://www.iccn2007.org/ International Conference on Cognitive Neurodynamics (ICCN)] - a yearly conference.


===Computational Psychiatry===
===Websites===
[[Computational psychiatry]] is a new emerging field that brings together experts in [[machine learning]], [[neuroscience]], [[neurology]], [[psychiatry]], [[psychology]] to provide an understanding of psychiatric disorders.<ref>{{cite journal|last1=Montague|first1=P. Read | last2=Dolan| first2=Raymond J. | last3=Friston|first3=Karl J.|author3-link=Karl Friston|last4=Dayan|first4=Peter|author4-link=Peter Dayan|title=Computational psychiatry|journal=[[Trends in Cognitive Sciences]]|date=14 Dec 2011|volume=16|issue=1|pages=72–80|doi=10.1016/j.tics.2011.11.018|pmid=22177032 |pmc=3556822 }}</ref><ref>{{cite journal | last1=Kato | first1=Ayaka | last2=Kunisato | first2=Yoshihiko | last3=Katahira | first3=Kentaro | last4=Okimura | first4=Tsukasa | last5=Yamashita | first5=Yuichi |title=Computational Psychiatry Research Map (CPSYMAP): a new database for visualizing research papers |journal=Frontiers in Psychiatry|volume=11|issue=1360|year=2020 | page=578706 |doi=10.3389/fpsyt.2020.578706| pmid=33343418 | pmc=7746554 | doi-access=free }}</ref><ref>{{cite journal | last1=Huys | first1=Quentin J M | last2=Maia | first2=Tiago V | last3=Frank | first3=Michael J |title=Computational psychiatry as a bridge from neuroscience to clinical applications |journal=Nature Neuroscience |volume=19 |issue=3 |pages=404–413 |year=2016 |doi=10.1038/nn.4238 | pmid=26906507 | pmc=5443409 }}</ref>
* [http://www.neurosecurity.com Neurosecurity], Articles and lectures on Computational neuroscience.
* [http://home.earthlink.net/~perlewitz/ Perlewitz's computational neuroscience on the web]
* [http://www.compneuro.org compneuro.org], books and programs for neural modeling
* [http://www.scholarpedia.org/article/Encyclopedia_of_Computational_Neuroscience Encyclopedia of Computational Neuroscience], part of [[Scholarpedia]], an online expert curated encyclopedia on computational neuroscience, dynamical systems and machine intelligence
* [http://purl.net/net/neurowiki NeuroWiki], a wiki discussion forum about neuroscience research, especially systems, theoretical/computational, and cognitive neuroscience


==Courses==
==Technology==
* [http://purl.net/net/neurowiki/CompNeuroCourses NeuroWiki:CompNeuroCourses], a list of comp neuro courses with material available online
* http://www.mbl.edu/education/courses/special_topics/mcn.html Summer course at the MBL, which features major figures in the field (Abbott, Bialek, Sejnowski, et.al.) as guest faculty.


===Neuromorphic computing===
{{main|Neuromorphic engineering}}
A neuromorphic computer/chip is any device that uses physical artificial neurons (made from silicon) to do computations (See: [[neuromorphic computing]], [[physical neural network]]). One of the advantages of using a [[physical model]] computer such as this is that it takes the computational load of the processor (in the sense that the structural and some of the functional elements don't have to be programmed since they are in hardware). In recent times,<ref>{{cite web |last1=Russell |first1=John |title=Beyond von Neumann, Neuromorphic Computing Steadily Advances |date=21 March 2016 |url=https://www.hpcwire.com/2016/03/21/lacking-breakthrough-neuromorphic-computing-steadily-advance/}}</ref> neuromorphic technology has been used to build supercomputers which are used in international neuroscience collaborations. Examples include the [[Human Brain Project]] [[SpiNNaker]] supercomputer and the BrainScaleS computer.<ref>{{Cite journal|last1=Calimera|first1=Andrea|last2=Macii|first2=Enrico |last3=Poncino|first3=Massimo |date=2013-08-20|title=The human brain project and neuromorphic computing|journal=Functional Neurology|volume=28|issue=3 |pages=191–196 | doi=10.11138/FNeur/2013.28.3.191|doi-broken-date=1 November 2024 |pmid=24139655 |pmc=3812737 }}</ref>


===Research Groups===
==See also==
{{Div col|colwidth=20em}}
* [http://www.bernstein-centers.de/ Bernstein Centers for Computational Neuroscience Germany]
*[[Action potential]]
* [http://www.cns-berlin.org/ Bernstein Center for Computational Neuroscience Berlin]
*[[Biological neuron models]]
* [http://www.bccn-freiburg.de/ Bernstein Center for Computational Neuroscience Freiburg]
*[[Bayesian approaches to brain function|Bayesian brain]]
* [http://www.bccn-goettingen.de/ Bernstein Center for Computational Neuroscience Goettingen]
*[[Brain simulation]]
* [http://www.bccn-munich.de/start.shtml Bernstein Center for Computational Neuroscience Munich]
*[[Computational anatomy]]
* [http://www.bm-science.com BM-Science - Brain & Mind Technologies Research Centre, Finland]
*[[Connectomics]]
* [http://www.neuroengineering.upenn.edu Neuroengineering Laboratory at the University of Pennsylvania]
*[[Differentiable programming]]
* [http://cneuro.rmki.kfki.hu Computational Neuroscience Group at the KFKI RIPNP of the Hungarian Academy of Sciences]
*[[Electrophysiology]]
* [http://www.cnl.salk.edu Computational Neurobiology Laboratory at the Salk Institute (CNL)]
*[[FitzHugh–Nagumo model]]
* [http://ctn.uwaterloo.ca Centre for Theoretical Neuroscience (CTN) at the University of Waterloo]
*[[Goldman equation]]
* [http://neuro.media.mit.edu/ MIT Media Lab, Neuroengineering and Neuromedia Group]
*[[Hodgkin–Huxley model]]
* [http://itb.biologie.hu-berlin.de/ Institute for Theoretical Biology, Humboldt-Universitaet zu Berlin]
*[[Information theory]]
* [http://www.mth.kcl.ac.uk/research/cns/cns Computational Neuroscience Group at King's College London]
*[[Mathematical model]]
* [http://cbcl.mit.edu/cbcl/index.html MIT Center for Biological & Computational Learning (CBCL)]
*[[Nonlinear dynamics]]
* [http://neurotheory.columbia.edu Center for Theoretical Neuroscience at Columbia University]
*[[Neural coding]]
* [http://icnc.huji.ac.il/ Interdisciplinary Center for Neural Computation at Hebrew University]
*[[Neural decoding]]
* [http://www.gatsby.ucl.ac.uk/ Gatsby Computational Neuroscience Unit at University College London]
*[[Neural oscillation]]
* [http://www.martinos.org/compneuro Martinos Computational Neuroscience Center] for integrating neuroimaging and computational neuroscience
*[[Neuroinformatics]]
* [http://riesenhuberlab.neuro.georgetown.edu Georgetown Laboratory for Computational Cognitive Neuroscience]
*[[Neuromimetic intelligence]]
* [http://www.uni-tuebingen.de/uni/knv/arl Hertie Center for Clinical Brain Research, Laboratory for Action Representation and Learning]
*[[Neuroplasticity]]
* [http://cns.qbi.uq.edu.au Computational Neuroscience Lab, University of Queensland]
*[[Neurophysiology]]
* [http://ccnlab.colorado.edu Computational Cognitive Neuroscience Lab, University of Colorado at Boulder]
*[[Systems neuroscience]]
* [http://tng.ccs.fau.edu Theoretical Neuroscience Group, Florida Atlantic University]
*[[Mathematical and theoretical biology|Theoretical biology]]
* [http://www.cs.kent.ac.uk/projects/cncs/index.html Centre for Cognitive Neuroscience and Cognitive Systems at the University of Kent]
*[[Theta model]]
*[http://www.cnel.ufl.edu Computational Neuroscience Engineering Lab, University of Florida]
{{Div col end}}
*[http://www.anc.ed.ac.uk/people/ Institute for Adaptive and Neural Computation, University of Edinburgh]
*[http://www.plymneuro.org.uk/ Centre for Theoretical and Computational Neuroscience, University of Plymouth]
*[http://www.tnb.ua.ac.be Theoretical Neurobiology Lab, University of Antwerp]
* [http://www.irp.oist.jp/cns/ Computational Neuroscience Unit, Okinawa Institute of Science and Technology]
* [http://www.gnt.ens.fr/ Group for Neural Theory, Ecole normale superieure, Paris]


===Papers===
== References ==
{{Reflist|2}}
* [http://papers.cnl.salk.edu/PDFs/Computational%20Neuroscience%201988-3883.pdf Review] - Sejnowski, T. J.; Koch, C.; Churchland, P. S.; Computational Neuroscience, Science, 241, 1299-1306, 1988

* [http://cbcl.mit.edu/projects/cbcl/publications/ai-publications/2005/AIM-2005-036.pdf A Theory of Object Recognition: Computations and Circuits in the Feedforward Path of the Ventral Stream in Primate Visual Cortex] - Biologically-based vision algorithm
== Bibliography ==
* {{cite journal |author=Chklovskii DB |title=Synaptic connectivity and neuronal morphology: two sides of the same coin |journal=Neuron |volume=43 |issue=5 |pages=609–17 |year=2004 |pmid=15339643 |doi=10.1016/j.neuron.2004.08.012 |s2cid=16217065 |doi-access=free }}
* {{cite book |author1=Sejnowski, Terrence J. |author-link=Terrence Sejnowski |author2=Churchland, Patricia Smith |title=The computational brain |publisher=[[MIT Press]] |location=Cambridge, Mass |year=1992 |isbn=978-0-262-03188-2 }}
* {{ cite book | author1= Gerstner, W. | author2 = Kistler, W. | author3 = Naud, R. | author4 = Paninski, L.| title = Neuronal Dynamics | publisher = Cambridge University Press | location = Cambridge, UK | year = 2014 | isbn = 9781107447615}}
* {{cite book |author1=Dayan P. |author-link=Peter Dayan |author2=Abbott, L. F. |author2-link=Larry Abbott |title=Theoretical neuroscience: computational and mathematical modeling of neural systems |publisher=MIT Press |location=Cambridge, Mass |year=2001 |isbn=978-0-262-04199-7 }}
* {{cite book |author1=Eliasmith, Chris |author2=Anderson, Charles H. |title=Neural engineering: Representation, computation, and dynamics in neurobiological systems |publisher=[[MIT Press]] |location=Cambridge, Mass |year=2003 |isbn=978-0-262-05071-5 }}
* {{cite journal |vauthors=[[Alan Hodgkin|Hodgkin AL]], [[Andrew Huxley|Huxley AF]] |title=A quantitative description of membrane current and its application to conduction and excitation in nerve |journal=J. Physiol. |volume=117 |issue=4 |pages=500–44 |date=28 August 1952|pmid=12991237 |pmc=1392413 |url=http://www.jphysiol.org/cgi/pmidlookup?view=long&pmid=12991237 |doi=10.1113/jphysiol.1952.sp004764}}
* {{cite book |author1=William Bialek |author2=Rieke, Fred |author3=David Warland |author4=Rob de Ruyter van Steveninck |title=Spikes: exploring the neural code |publisher=MIT |location=Cambridge, Mass |year=1999 |isbn=978-0-262-68108-7 |author1-link=William Bialek }}
* {{cite book |author=Schutter, Erik de |title=Computational neuroscience: realistic modeling for experimentalists |publisher=CRC |location=Boca Raton |year=2001 |isbn=978-0-8493-2068-2 }}
* {{cite book |author1=Sejnowski, Terrence J. |author2=Hemmen, J. L. van |title=23 problems in systems neuroscience |publisher=Oxford University Press |location=Oxford [Oxfordshire] |year=2006 |isbn=978-0-19-514822-0 }}
* {{cite book |author1=Michael A. Arbib |author2=Shun-ichi Amari |author3=Prudence H. Arbib | title=The Handbook of Brain Theory and Neural Networks|publisher=The MIT Press|location=Cambridge, Massachusetts |year=2002 |isbn=978-0-262-01197-6}}
* {{cite book|author1=Zhaoping|author-link=Li Zhaoping|first=Li |title=Understanding vision: theory, models, and data|publisher=Oxford University Press|location=Oxford, UK|year=2014|isbn=978-0199564668}}

==See also==

===Software===
* [[Brian (software)|BRIAN]], a [[Python (programming language)|Python]] based simulator
* [[Budapest Reference Connectome]], web based 3D visualization tool to browse connections in the human brain
* [[Emergent (software)|Emergent]], neural simulation software.
* [[GENESIS (software)|GENESIS]], a general neural simulation system.
* [[NEST (software)|NEST]] is a simulator for spiking neural network models that focuses on the dynamics, size and structure of neural systems rather than on the exact morphology of individual neurons.

==External links==
{{Commons category}}

===Journals===
*[https://www.springer.com/mathematics/journal/13408 Journal of Mathematical Neuroscience]
*[https://www.springer.com/10827 Journal of Computational Neuroscience]
* [http://www.mitpressjournals.org/loi/neco Neural Computation]
* [[Cognitive Neurodynamics]]
* [http://frontiersin.org/neuroscience/computationalneuroscience/ Frontiers in Computational Neuroscience]
* [http://www.ploscompbiol.org/home.action PLoS Computational Biology]
* [http://www.frontiersin.org/Journal/specialty.aspx?s=752&name=neuroinformatics&x=y Frontiers in Neuroinformatics]

===Conferences===
* [[Computational and Systems Neuroscience]] (COSYNE) – a computational neuroscience meeting with a systems neuroscience focus.
* [http://www.cnsorg.org Annual Computational Neuroscience Meeting (CNS)] – a yearly computational neuroscience meeting.
* [http://www.nips.cc Neural Information Processing Systems (NIPS)]– a leading annual conference covering mostly machine learning.
* [https://ccneuro.org/ Cognitive Computational Neuroscience (CCN)] – a computational neuroscience meeting focusing on computational models capable of cognitive tasks.
* [https://web.archive.org/web/20070309063503/http://www.iccn2007.org/ International Conference on Cognitive Neurodynamics (ICCN)] – a yearly conference.
* [https://web.archive.org/web/20080705131550/http://www.icms.org.uk/workshops/mathneuro UK Mathematical Neurosciences Meeting]– a yearly conference, focused on mathematical aspects.
* [https://web.archive.org/web/20110429094455/http://www.nncn.de/Aktuelles-en/bernsteinsymposium/Symposium/view?set_language=en Bernstein Conference on Computational Neuroscience (BCCN)]– a yearly computational neuroscience conference ].
* [http://www.areadne.org/index.html AREADNE Conferences]– a biennial meeting that includes theoretical and experimental results.

===Websites===


* [http://www.scholarpedia.org/article/Encyclopedia_of_Computational_Neuroscience Encyclopedia of Computational Neuroscience], part of [[Scholarpedia]], an online expert curated encyclopedia on computational neuroscience and dynamical systems
{{Neuroscience-footer}}
{{Cybernetics}}
{{Neuroscience}}
[[Category:Computational neuroscience|*]]
{{Computational science}}
[[Category:Cybernetics]]


{{DEFAULTSORT:Computational Neuroscience}}
[[es:Neurociencia computacional]]
[[Category:Computational fields of study]]
[[fa:عصب‌شناسی محاسباتی]]
[[Category:Computational neuroscience]]
[[fr:Neurosciences computationnelles]]
[[Category:Mathematical and theoretical biology]]
[[ja:計算論的神経科学]]
[[lt:Neuroinformatika]]

Latest revision as of 05:29, 2 November 2024

Computational neuroscience (also known as theoretical neuroscience or mathematical neuroscience) is a branch of neuroscience which employs mathematics, computer science, theoretical analysis and abstractions of the brain to understand the principles that govern the development, structure, physiology and cognitive abilities of the nervous system.[1][2][3][4]

Computational neuroscience employs computational simulations[5] to validate and solve mathematical models, and so can be seen as a sub-field of theoretical neuroscience; however, the two fields are often synonymous.[6] The term mathematical neuroscience is also used sometimes, to stress the quantitative nature of the field.[7]

Computational neuroscience focuses on the description of biologically plausible neurons (and neural systems) and their physiology and dynamics, and it is therefore not directly concerned with biologically unrealistic models used in connectionism, control theory, cybernetics, quantitative psychology, machine learning, artificial neural networks, artificial intelligence and computational learning theory;[8][9] [10] although mutual inspiration exists and sometimes there is no strict limit between fields,[11][12][13] with model abstraction in computational neuroscience depending on research scope and the granularity at which biological entities are analyzed.

Models in theoretical neuroscience are aimed at capturing the essential features of the biological system at multiple spatial-temporal scales, from membrane currents, and chemical coupling via network oscillations, columnar and topographic architecture, nuclei, all the way up to psychological faculties like memory, learning and behavior. These computational models frame hypotheses that can be directly tested by biological or psychological experiments.

History

[edit]

The term 'computational neuroscience' was introduced by Eric L. Schwartz, who organized a conference, held in 1985 in Carmel, California, at the request of the Systems Development Foundation to provide a summary of the current status of a field which until that point was referred to by a variety of names, such as neural modeling, brain theory and neural networks. The proceedings of this definitional meeting were published in 1990 as the book Computational Neuroscience.[14] The first of the annual open international meetings focused on Computational Neuroscience was organized by James M. Bower and John Miller in San Francisco, California in 1989.[15] The first graduate educational program in computational neuroscience was organized as the Computational and Neural Systems Ph.D. program at the California Institute of Technology in 1985.

The early historical roots of the field[16] can be traced to the work of people including Louis Lapicque, Hodgkin & Huxley, Hubel and Wiesel, and David Marr. Lapicque introduced the integrate and fire model of the neuron in a seminal article published in 1907,[17] a model still popular for artificial neural networks studies because of its simplicity (see a recent review[18]).

About 40 years later, Hodgkin and Huxley developed the voltage clamp and created the first biophysical model of the action potential. Hubel and Wiesel discovered that neurons in the primary visual cortex, the first cortical area to process information coming from the retina, have oriented receptive fields and are organized in columns.[19] David Marr's work focused on the interactions between neurons, suggesting computational approaches to the study of how functional groups of neurons within the hippocampus and neocortex interact, store, process, and transmit information. Computational modeling of biophysically realistic neurons and dendrites began with the work of Wilfrid Rall, with the first multicompartmental model using cable theory.

Major topics

[edit]

Research in computational neuroscience can be roughly categorized into several lines of inquiry. Most computational neuroscientists collaborate closely with experimentalists in analyzing novel data and synthesizing new models of biological phenomena.

Single-neuron modeling

[edit]

Even a single neuron has complex biophysical characteristics and can perform computations (e.g.[20]). Hodgkin and Huxley's original model only employed two voltage-sensitive currents (Voltage sensitive ion channels are glycoprotein molecules which extend through the lipid bilayer, allowing ions to traverse under certain conditions through the axolemma), the fast-acting sodium and the inward-rectifying potassium. Though successful in predicting the timing and qualitative features of the action potential, it nevertheless failed to predict a number of important features such as adaptation and shunting. Scientists now believe that there are a wide variety of voltage-sensitive currents, and the implications of the differing dynamics, modulations, and sensitivity of these currents is an important topic of computational neuroscience.[21]

The computational functions of complex dendrites are also under intense investigation. There is a large body of literature regarding how different currents interact with geometric properties of neurons.[22]

There are many software packages, such as GENESIS and NEURON, that allow rapid and systematic in silico modeling of realistic neurons. Blue Brain, a project founded by Henry Markram from the École Polytechnique Fédérale de Lausanne, aims to construct a biophysically detailed simulation of a cortical column on the Blue Gene supercomputer.

Modeling the richness of biophysical properties on the single-neuron scale can supply mechanisms that serve as the building blocks for network dynamics.[23] However, detailed neuron descriptions are computationally expensive and this computing cost can limit the pursuit of realistic network investigations, where many neurons need to be simulated. As a result, researchers that study large neural circuits typically represent each neuron and synapse with an artificially simple model, ignoring much of the biological detail. Hence there is a drive to produce simplified neuron models that can retain significant biological fidelity at a low computational overhead. Algorithms have been developed to produce faithful, faster running, simplified surrogate neuron models from computationally expensive, detailed neuron models.[24]

Modeling Neuron-glia interactions

[edit]

Glial cells participate significantly in the regulation of neuronal activity at both the cellular and the network level. Modeling this interaction allows to clarify the potassium cycle,[25][26] so important for maintaining homeostasis and to prevent epileptic seizures. Modeling reveals the role of glial protrusions that can penetrate in some cases the synaptic cleft to interfere with the synaptic transmission and thus control synaptic communication.[27]

Development, axonal patterning, and guidance

[edit]

Computational neuroscience aims to address a wide array of questions, including: How do axons and dendrites form during development? How do axons know where to target and how to reach these targets? How do neurons migrate to the proper position in the central and peripheral systems? How do synapses form? We know from molecular biology that distinct parts of the nervous system release distinct chemical cues, from growth factors to hormones that modulate and influence the growth and development of functional connections between neurons.

Theoretical investigations into the formation and patterning of synaptic connection and morphology are still nascent. One hypothesis that has recently garnered some attention is the minimal wiring hypothesis, which postulates that the formation of axons and dendrites effectively minimizes resource allocation while maintaining maximal information storage.[28]

Sensory processing

[edit]

Early models on sensory processing understood within a theoretical framework are credited to Horace Barlow. Somewhat similar to the minimal wiring hypothesis described in the preceding section, Barlow understood the processing of the early sensory systems to be a form of efficient coding, where the neurons encoded information which minimized the number of spikes. Experimental and computational work have since supported this hypothesis in one form or another. For the example of visual processing, efficient coding is manifested in the forms of efficient spatial coding, color coding, temporal/motion coding, stereo coding, and combinations of them.[29]

Further along the visual pathway, even the efficiently coded visual information is too much for the capacity of the information bottleneck, the visual attentional bottleneck.[30] A subsequent theory, V1 Saliency Hypothesis (V1SH), has been developed on exogenous attentional selection of a fraction of visual input for further processing, guided by a bottom-up saliency map in the primary visual cortex.[31]

Current research in sensory processing is divided among a biophysical modeling of different subsystems and a more theoretical modeling of perception. Current models of perception have suggested that the brain performs some form of Bayesian inference and integration of different sensory information in generating our perception of the physical world.[32][33]

Motor control

[edit]

Many models of the way the brain controls movement have been developed. This includes models of processing in the brain such as the cerebellum's role for error correction, skill learning in motor cortex and the basal ganglia, or the control of the vestibulo ocular reflex. This also includes many normative models, such as those of the Bayesian or optimal control flavor which are built on the idea that the brain efficiently solves its problems.

Memory and synaptic plasticity

[edit]

Earlier models of memory are primarily based on the postulates of Hebbian learning. Biologically relevant models such as Hopfield net have been developed to address the properties of associative (also known as "content-addressable") style of memory that occur in biological systems. These attempts are primarily focusing on the formation of medium- and long-term memory, localizing in the hippocampus.

One of the major problems in neurophysiological memory is how it is maintained and changed through multiple time scales. Unstable synapses are easy to train but also prone to stochastic disruption. Stable synapses forget less easily, but they are also harder to consolidate. It is likely that computational tools will contribute greatly to our understanding of how synapses function and change in relation to external stimulus in the coming decades.

Behaviors of networks

[edit]

Biological neurons are connected to each other in a complex, recurrent fashion. These connections are, unlike most artificial neural networks, sparse and usually specific. It is not known how information is transmitted through such sparsely connected networks, although specific areas of the brain, such as the visual cortex, are understood in some detail.[34] It is also unknown what the computational functions of these specific connectivity patterns are, if any.

The interactions of neurons in a small network can be often reduced to simple models such as the Ising model. The statistical mechanics of such simple systems are well-characterized theoretically. Some recent evidence suggests that dynamics of arbitrary neuronal networks can be reduced to pairwise interactions.[35] It is not known, however, whether such descriptive dynamics impart any important computational function. With the emergence of two-photon microscopy and calcium imaging, we now have powerful experimental methods with which to test the new theories regarding neuronal networks.

In some cases the complex interactions between inhibitory and excitatory neurons can be simplified using mean-field theory, which gives rise to the population model of neural networks.[36] While many neurotheorists prefer such models with reduced complexity, others argue that uncovering structural-functional relations depends on including as much neuronal and network structure as possible. Models of this type are typically built in large simulation platforms like GENESIS or NEURON. There have been some attempts to provide unified methods that bridge and integrate these levels of complexity.[37]

Visual attention, identification, and categorization

[edit]

Visual attention can be described as a set of mechanisms that limit some processing to a subset of incoming stimuli.[38] Attentional mechanisms shape what we see and what we can act upon. They allow for concurrent selection of some (preferably, relevant) information and inhibition of other information. In order to have a more concrete specification of the mechanism underlying visual attention and the binding of features, a number of computational models have been proposed aiming to explain psychophysical findings. In general, all models postulate the existence of a saliency or priority map for registering the potentially interesting areas of the retinal input, and a gating mechanism for reducing the amount of incoming visual information, so that the limited computational resources of the brain can handle it.[39] An example theory that is being extensively tested behaviorally and physiologically is the V1 Saliency Hypothesis that a bottom-up saliency map is created in the primary visual cortex to guide attention exogenously.[31] Computational neuroscience provides a mathematical framework for studying the mechanisms involved in brain function and allows complete simulation and prediction of neuropsychological syndromes.

Cognition, discrimination, and learning

[edit]

Computational modeling of higher cognitive functions has only recently[when?] begun. Experimental data comes primarily from single-unit recording in primates. The frontal lobe and parietal lobe function as integrators of information from multiple sensory modalities. There are some tentative ideas regarding how simple mutually inhibitory functional circuits in these areas may carry out biologically relevant computation.[40]

The brain seems to be able to discriminate and adapt particularly well in certain contexts. For instance, human beings seem to have an enormous capacity for memorizing and recognizing faces. One of the key goals of computational neuroscience is to dissect how biological systems carry out these complex computations efficiently and potentially replicate these processes in building intelligent machines.

The brain's large-scale organizational principles are illuminated by many fields, including biology, psychology, and clinical practice. Integrative neuroscience attempts to consolidate these observations through unified descriptive models and databases of behavioral measures and recordings. These are the bases for some quantitative modeling of large-scale brain activity.[41]

The Computational Representational Understanding of Mind (CRUM) is another attempt at modeling human cognition through simulated processes like acquired rule-based systems in decision making and the manipulation of visual representations in decision making.

One of the ultimate goals of psychology/neuroscience is to be able to explain the everyday experience of conscious life. Francis Crick, Giulio Tononi and Christof Koch made some attempts to formulate consistent frameworks for future work in neural correlates of consciousness (NCC), though much of the work in this field remains speculative.[42]

Computational clinical neuroscience

[edit]

Computational clinical neuroscience is a field that brings together experts in neuroscience, neurology, psychiatry, decision sciences and computational modeling to quantitatively define and investigate problems in neurological and psychiatric diseases, and to train scientists and clinicians that wish to apply these models to diagnosis and treatment.[43][44]

Predictive computational neuroscience

[edit]

Predictive computational neuroscience is a recent field that combines signal processing, neuroscience, clinical data and machine learning to predict the brain during coma [45] or anesthesia.[46] For example, it is possible to anticipate deep brain states using the EEG signal. These states can be used to anticipate hypnotic concentration to administrate to the patient.

Computational Psychiatry

[edit]

Computational psychiatry is a new emerging field that brings together experts in machine learning, neuroscience, neurology, psychiatry, psychology to provide an understanding of psychiatric disorders.[47][48][49]

Technology

[edit]

Neuromorphic computing

[edit]

A neuromorphic computer/chip is any device that uses physical artificial neurons (made from silicon) to do computations (See: neuromorphic computing, physical neural network). One of the advantages of using a physical model computer such as this is that it takes the computational load of the processor (in the sense that the structural and some of the functional elements don't have to be programmed since they are in hardware). In recent times,[50] neuromorphic technology has been used to build supercomputers which are used in international neuroscience collaborations. Examples include the Human Brain Project SpiNNaker supercomputer and the BrainScaleS computer.[51]

See also

[edit]

References

[edit]
  1. ^ Trappenberg, Thomas P. (2010). Fundamentals of Computational Neuroscience. United States: Oxford University Press Inc. pp. 2. ISBN 978-0-19-851582-1.
  2. ^ Patricia S. Churchland; Christof Koch; Terrence J. Sejnowski (1993). "What is computational neuroscience?". In Eric L. Schwartz (ed.). Computational Neuroscience. MIT Press. pp. 46–55. Archived from the original on 2011-06-04. Retrieved 2009-06-11.
  3. ^ Dayan P.; Abbott, L. F. (2001). Theoretical neuroscience: computational and mathematical modeling of neural systems. Cambridge, Mass: MIT Press. ISBN 978-0-262-04199-7.
  4. ^ Gerstner, W.; Kistler, W.; Naud, R.; Paninski, L. (2014). Neuronal Dynamics. Cambridge, UK: Cambridge University Press. ISBN 9781107447615.
  5. ^ Fan, Xue; Markram, Henry (2019). "A Brief History of Simulation Neuroscience". Frontiers in Neuroinformatics. 13: 32. doi:10.3389/fninf.2019.00032. ISSN 1662-5196. PMC 6513977. PMID 31133838.
  6. ^ Thomas, Trappenberg (2010). Fundamentals of Computational Neuroscience. OUP Oxford. p. 2. ISBN 978-0199568413. Retrieved 17 January 2017.
  7. ^ Gutkin, Boris; Pinto, David; Ermentrout, Bard (2003-03-01). "Mathematical neuroscience: from neurons to circuits to systems". Journal of Physiology-Paris. Neurogeometry and visual perception. 97 (2): 209–219. doi:10.1016/j.jphysparis.2003.09.005. ISSN 0928-4257. PMID 14766142. S2CID 10040483.
  8. ^ Kriegeskorte, Nikolaus; Douglas, Pamela K. (September 2018). "Cognitive computational neuroscience". Nature Neuroscience. 21 (9): 1148–1160. arXiv:1807.11819. Bibcode:2018arXiv180711819K. doi:10.1038/s41593-018-0210-5. ISSN 1546-1726. PMC 6706072. PMID 30127428.
  9. ^ Paolo, E. D., "Organismically-inspired robotics: homeostatic adaptation and teleology beyond the closed sensorimotor loop", Dynamical Systems Approach to Embodiment and Sociality, S2CID 15349751
  10. ^ Brooks, R.; Hassabis, D.; Bray, D.; Shashua, A. (2012-02-22). "Turing centenary: Is the brain a good model for machine intelligence?". Nature. 482 (7386): 462–463. Bibcode:2012Natur.482..462.. doi:10.1038/482462a. ISSN 0028-0836. PMID 22358812. S2CID 205070106.
  11. ^ Browne, A. (1997-01-01). Neural Network Perspectives on Cognition and Adaptive Robotics. CRC Press. ISBN 9780750304559.
  12. ^ Zorzi, Marco; Testolin, Alberto; Stoianov, Ivilin P. (2013-08-20). "Modeling language and cognition with deep unsupervised learning: a tutorial overview". Frontiers in Psychology. 4: 515. doi:10.3389/fpsyg.2013.00515. ISSN 1664-1078. PMC 3747356. PMID 23970869.
  13. ^ Shai, Adam; Larkum, Matthew Evan (2017-12-05). "Branching into brains". eLife. 6. doi:10.7554/eLife.33066. ISSN 2050-084X. PMC 5716658. PMID 29205152.
  14. ^ Schwartz, Eric (1990). Computational neuroscience. Cambridge, Mass: MIT Press. ISBN 978-0-262-19291-0.
  15. ^ Bower, James M. (2013). 20 years of Computational neuroscience. Berlin, Germany: Springer. ISBN 978-1461414230.
  16. ^ Fan, Xue; Markram, Henry (2019). "A Brief History of Simulation Neuroscience". Frontiers in Neuroinformatics. 13: 32. doi:10.3389/fninf.2019.00032. ISSN 1662-5196. PMC 6513977. PMID 31133838.
  17. ^ Lapicque L (1907). "Recherches quantitatives sur l'excitation électrique des nerfs traitée comme une polarisation". J. Physiol. Pathol. Gen. 9: 620–635.
  18. ^ Brunel N, Van Rossum MC (2007). "Lapicque's 1907 paper: from frogs to integrate-and-fire". Biol. Cybern. 97 (5–6): 337–339. doi:10.1007/s00422-007-0190-0. PMID 17968583. S2CID 17816096.
  19. ^ Hubel DH, Wiesel TN (1962). "Receptive fields, binocular interaction and functional architecture in the cat's visual cortex". J. Physiol. 160 (1): 106–54. doi:10.1113/jphysiol.1962.sp006837. PMC 1359523. PMID 14449617.
  20. ^ Forrest MD (2014). "Intracellular Calcium Dynamics Permit a Purkinje Neuron Model to Perform Toggle and Gain Computations Upon its Inputs". Frontiers in Computational Neuroscience. 8: 86. doi:10.3389/fncom.2014.00086. PMC 4138505. PMID 25191262.
  21. ^ Wu, Samuel Miao-sin; Johnston, Daniel (1995). Foundations of cellular neurophysiology. Cambridge, Mass: MIT Press. ISBN 978-0-262-10053-3.
  22. ^ Koch, Christof (1999). Biophysics of computation: information processing in single neurons. Oxford [Oxfordshire]: Oxford University Press. ISBN 978-0-19-510491-2.
  23. ^ Forrest MD (2014). "Intracellular Calcium Dynamics Permit a Purkinje Neuron Model to Perform Toggle and Gain Computations Upon its Inputs". Frontiers in Computational Neuroscience. 8: 86. doi:10.3389/fncom.2014.00086. PMC 4138505. PMID 25191262.
  24. ^ Forrest MD (April 2015). "Simulation of alcohol action upon a detailed Purkinje neuron model and a simpler surrogate model that runs >400 times faster". BMC Neuroscience. 16 (27): 27. doi:10.1186/s12868-015-0162-6. PMC 4417229. PMID 25928094.
  25. ^ "Dynamics of Ion Fluxes between Neurons, Astrocytes and the Extracellular Space during Neurotransmission". cyberleninka.ru. Retrieved 2023-03-14.
  26. ^ Sibille, Jérémie; Duc, Khanh Dao; Holcman, David; Rouach, Nathalie (2015-03-31). "The Neuroglial Potassium Cycle during Neurotransmission: Role of Kir4.1 Channels". PLOS Computational Biology. 11 (3): e1004137. Bibcode:2015PLSCB..11E4137S. doi:10.1371/journal.pcbi.1004137. ISSN 1553-7358. PMC 4380507. PMID 25826753.
  27. ^ Pannasch, Ulrike; Freche, Dominik; Dallérac, Glenn; Ghézali, Grégory; Escartin, Carole; Ezan, Pascal; Cohen-Salmon, Martine; Benchenane, Karim; Abudara, Veronica; Dufour, Amandine; Lübke, Joachim H. R.; Déglon, Nicole; Knott, Graham; Holcman, David; Rouach, Nathalie (April 2014). "Connexin 30 sets synaptic strength by controlling astroglial synapse invasion". Nature Neuroscience. 17 (4): 549–558. doi:10.1038/nn.3662. ISSN 1546-1726. PMID 24584052. S2CID 554918.
  28. ^ Chklovskii DB, Mel BW, Svoboda K (October 2004). "Cortical rewiring and information storage". Nature. 431 (7010): 782–8. Bibcode:2004Natur.431..782C. doi:10.1038/nature03012. PMID 15483599. S2CID 4430167.
    Review article
  29. ^ Zhaoping L. 2014, The efficient coding principle , chapter 3, of the textbook Understanding vision: theory, models, and data
  30. ^ see visual spational attention https://en.wikipedia.org/wiki/Visual_spatial_attention
  31. ^ a b Li. Z. 2002 A saliency map in primary visual cortex Trends in Cognitive Sciences vol. 6, Pages 9-16, and Zhaoping, L. 2014, The V1 hypothesis—creating a bottom-up saliency map for preattentive selection and segmentation in the book Understanding Vision: Theory, Models, and Data
  32. ^ Weiss, Yair; Simoncelli, Eero P.; Adelson, Edward H. (20 May 2002). "Motion illusions as optimal percepts". Nature Neuroscience. 5 (6): 598–604. doi:10.1038/nn0602-858. PMID 12021763. S2CID 2777968.
  33. ^ Ernst, Marc O.; Bülthoff, Heinrich H. (April 2004). "Merging the senses into a robust percept". Trends in Cognitive Sciences. 8 (4): 162–169. CiteSeerX 10.1.1.299.4638. doi:10.1016/j.tics.2004.02.002. PMID 15050512. S2CID 7837073.
  34. ^ Olshausen, Bruno A.; Field, David J. (1997-12-01). "Sparse coding with an overcomplete basis set: A strategy employed by V1?". Vision Research. 37 (23): 3311–3325. doi:10.1016/S0042-6989(97)00169-7. PMID 9425546. S2CID 14208692.
  35. ^ Schneidman E, Berry MJ, Segev R, Bialek W (2006). "Weak pairwise correlations imply strongly correlated network states in a neural population". Nature. 440 (7087): 1007–12. arXiv:q-bio/0512013. Bibcode:2006Natur.440.1007S. doi:10.1038/nature04701. PMC 1785327. PMID 16625187.
  36. ^ Wilson, H. R.; Cowan, J.D. (1973). "A mathematical theory of the functional dynamics of cortical and thalamic nervous tissue". Kybernetik. 13 (2): 55–80. doi:10.1007/BF00288786. PMID 4767470. S2CID 292546.
  37. ^ Anderson, Charles H.; Eliasmith, Chris (2004). Neural Engineering: Computation, Representation, and Dynamics in Neurobiological Systems (Computational Neuroscience). Cambridge, Mass: The MIT Press. ISBN 978-0-262-55060-4.
  38. ^ Marvin M. Chun; Jeremy M. Wolfe; E. B. Goldstein (2001). Blackwell Handbook of Sensation and Perception. Blackwell Publishing Ltd. pp. 272–310. ISBN 978-0-631-20684-2.
  39. ^ Edmund Rolls; Gustavo Deco (2012). Computational Neuroscience of Vision. Oxford Scholarship Online. ISBN 978-0-198-52488-5.
  40. ^ Machens CK, Romo R, Brody CD (2005). "Flexible control of mutual inhibition: a neural model of two-interval discrimination". Science. 307 (5712): 1121–4. Bibcode:2005Sci...307.1121M. CiteSeerX 10.1.1.523.4396. doi:10.1126/science.1104171. PMID 15718474. S2CID 45378154.
  41. ^ Robinson PA, Rennie CJ, Rowe DL, O'Connor SC, Gordon E (2005). "Multiscale brain modelling". Philosophical Transactions of the Royal Society B. 360 (1457): 1043–1050. doi:10.1098/rstb.2005.1638. PMC 1854922. PMID 16087447.
  42. ^ Crick F, Koch C (2003). "A framework for consciousness". Nat. Neurosci. 6 (2): 119–26. doi:10.1038/nn0203-119. PMID 12555104. S2CID 13960489.
  43. ^ Adaszewski, Stanisław; Dukart, Juergen; Kherif, Ferath; Frackowiak, Richard; Draganski, Bogdan; Alzheimer's Disease Neuroimaging Initiative (2013). "How early can we predict Alzheimer's disease using computational anatomy?". Neurobiol Aging. 34 (12): 2815–26. doi:10.1016/j.neurobiolaging.2013.06.015. PMID 23890839. S2CID 1025210.
  44. ^ Friston KJ, Stephan KE, Montague R, Dolan RJ (2014). "Computational psychiatry: the brain as a phantastic organ". Lancet Psychiatry. 1 (2): 148–58. doi:10.1016/S2215-0366(14)70275-5. PMID 26360579. S2CID 15504512.
  45. ^ Floyrac, Aymeric; Doumergue, Adrien; Legriel, Stéphane; Deye, Nicolas; Megarbane, Bruno; Richard, Alexandra; Meppiel, Elodie; Masmoudi, Sana; Lozeron, Pierre; Vicaut, Eric; Kubis, Nathalie; Holcman, David (2023). "Predicting neurological outcome after cardiac arrest by combining computational parameters extracted from standard and deviant responses from auditory evoked potentials". Frontiers in Neuroscience. 17: 988394. doi:10.3389/fnins.2023.988394. ISSN 1662-453X. PMC 9975713. PMID 36875664.
  46. ^ Sun, Christophe; Holcman, David (2022-08-01). "Combining transient statistical markers from the EEG signal to predict brain sensitivity to general anesthesia". Biomedical Signal Processing and Control. 77: 103713. doi:10.1016/j.bspc.2022.103713. ISSN 1746-8094. S2CID 248488365.
  47. ^ Montague, P. Read; Dolan, Raymond J.; Friston, Karl J.; Dayan, Peter (14 Dec 2011). "Computational psychiatry". Trends in Cognitive Sciences. 16 (1): 72–80. doi:10.1016/j.tics.2011.11.018. PMC 3556822. PMID 22177032.
  48. ^ Kato, Ayaka; Kunisato, Yoshihiko; Katahira, Kentaro; Okimura, Tsukasa; Yamashita, Yuichi (2020). "Computational Psychiatry Research Map (CPSYMAP): a new database for visualizing research papers". Frontiers in Psychiatry. 11 (1360): 578706. doi:10.3389/fpsyt.2020.578706. PMC 7746554. PMID 33343418.
  49. ^ Huys, Quentin J M; Maia, Tiago V; Frank, Michael J (2016). "Computational psychiatry as a bridge from neuroscience to clinical applications". Nature Neuroscience. 19 (3): 404–413. doi:10.1038/nn.4238. PMC 5443409. PMID 26906507.
  50. ^ Russell, John (21 March 2016). "Beyond von Neumann, Neuromorphic Computing Steadily Advances".
  51. ^ Calimera, Andrea; Macii, Enrico; Poncino, Massimo (2013-08-20). "The human brain project and neuromorphic computing". Functional Neurology. 28 (3): 191–196. doi:10.11138/FNeur/2013.28.3.191 (inactive 1 November 2024). PMC 3812737. PMID 24139655.{{cite journal}}: CS1 maint: DOI inactive as of November 2024 (link)

Bibliography

[edit]

See also

[edit]

Software

[edit]
  • BRIAN, a Python based simulator
  • Budapest Reference Connectome, web based 3D visualization tool to browse connections in the human brain
  • Emergent, neural simulation software.
  • GENESIS, a general neural simulation system.
  • NEST is a simulator for spiking neural network models that focuses on the dynamics, size and structure of neural systems rather than on the exact morphology of individual neurons.
[edit]

Journals

[edit]

Conferences

[edit]

Websites

[edit]