1
|
Zhao D, Si B. Formation of cognitive maps in large-scale environments by sensorimotor integration. Cogn Neurodyn 2025; 19:19. [PMID: 39801918 PMCID: PMC11717777 DOI: 10.1007/s11571-024-10200-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2024] [Revised: 10/08/2024] [Accepted: 10/26/2024] [Indexed: 01/16/2025] Open
Abstract
Hippocampus in the mammalian brain supports navigation by building a cognitive map of the environment. However, only a few studies have investigated cognitive maps in large-scale arenas. To reveal the computational mechanisms underlying the formation of cognitive maps in large-scale environments, we propose a neural network model of the entorhinal-hippocampal neural circuit that integrates both spatial and non-spatial information. Spatial information is relayed from the grid units in medial entorhinal cortex (MEC) by integrating multimodal sensory-motor signals. Non-spatial, such as object, information is imparted from the visual units in lateral entorhinal cortex (LEC) by encoding visual scenes through a deep neural network. The synaptic weights from the grid units and the visual units to the place units in the hippocampus are learned by a competitive learning rule. We simulated the model in a large box maze. The place units in the model form irregularly-spaced multiple fields across the environment. When the strength of visual inputs is dominant, the responses of place units become conjunctive and egocentric. These results point to the key role of the hippocampus in balancing spatial and non-spatial information relayed via LEC and MEC.
Collapse
Affiliation(s)
- Dongye Zhao
- Information Science Academy, China Electronics Technology Group Corporation, Beijing, 100086 China
| | - Bailu Si
- School of Systems Science, Beijing Normal University, Beijing, 100875 China
- Chinese Institute for Brain Research, Beijing, 102206 China
| |
Collapse
|
2
|
Rolls ET. Hippocampal Discoveries: Spatial View Cells, Connectivity, and Computations for Memory and Navigation, in Primates Including Humans. Hippocampus 2025; 35:e23666. [PMID: 39690918 DOI: 10.1002/hipo.23666] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/01/2024] [Revised: 10/19/2024] [Accepted: 11/26/2024] [Indexed: 12/19/2024]
Abstract
Two key series of discoveries about the hippocampus are described. One is the discovery of hippocampal spatial view cells in primates. This discovery opens the way to a much better understanding of human episodic memory, for episodic memory prototypically involves a memory of where people or objects or rewards have been seen in locations "out there" which could never be implemented by the place cells that encode the location of a rat or mouse. Further, spatial view cells are valuable for navigation using vision and viewed landmarks, and provide for much richer, vision-based, navigation than the place to place self-motion update performed by rats and mice who live in dark underground tunnels. Spatial view cells thus offer a revolution in our understanding of the functions of the hippocampus in memory and navigation in humans and other primates with well-developed foveate vision. The second discovery describes a computational theory of the hippocampal-neocortical memory system that includes the only quantitative theory of how information is recalled from the hippocampus to the neocortex. It is shown how foundations for this research were the discovery of reward neurons for food reward, and non-reward, in the primate orbitofrontal cortex, and representations of value including of monetary value in the human orbitofrontal cortex; and the discovery of face identity and face expression cells in the primate inferior temporal visual cortex and how they represent transform-invariant information. This research illustrates how in order to understand a brain computation, a whole series of integrated interdisciplinary discoveries is needed to build a theory of the operation of each neural system.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK
- Department of Computer Science, University of Warwick, Coventry, UK
| |
Collapse
|
3
|
Rolls ET, Treves A. A theory of hippocampal function: New developments. Prog Neurobiol 2024; 238:102636. [PMID: 38834132 DOI: 10.1016/j.pneurobio.2024.102636] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2024] [Revised: 04/15/2024] [Accepted: 05/30/2024] [Indexed: 06/06/2024]
Abstract
We develop further here the only quantitative theory of the storage of information in the hippocampal episodic memory system and its recall back to the neocortex. The theory is upgraded to account for a revolution in understanding of spatial representations in the primate, including human, hippocampus, that go beyond the place where the individual is located, to the location being viewed in a scene. This is fundamental to much primate episodic memory and navigation: functions supported in humans by pathways that build 'where' spatial view representations by feature combinations in a ventromedial visual cortical stream, separate from those for 'what' object and face information to the inferior temporal visual cortex, and for reward information from the orbitofrontal cortex. Key new computational developments include the capacity of the CA3 attractor network for storing whole charts of space; how the correlations inherent in self-organizing continuous spatial representations impact the storage capacity; how the CA3 network can combine continuous spatial and discrete object and reward representations; the roles of the rewards that reach the hippocampus in the later consolidation into long-term memory in part via cholinergic pathways from the orbitofrontal cortex; and new ways of analysing neocortical information storage using Potts networks.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK; Department of Computer Science, University of Warwick, Coventry CV4 7AL, UK.
| | | |
Collapse
|
4
|
Rolls ET. Hippocampal spatial view cells for memory and navigation, and their underlying connectivity in humans. Hippocampus 2023; 33:533-572. [PMID: 36070199 PMCID: PMC10946493 DOI: 10.1002/hipo.23467] [Citation(s) in RCA: 33] [Impact Index Per Article: 16.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2022] [Revised: 08/16/2022] [Accepted: 08/16/2022] [Indexed: 01/08/2023]
Abstract
Hippocampal and parahippocampal gyrus spatial view neurons in primates respond to the spatial location being looked at. The representation is allocentric, in that the responses are to locations "out there" in the world, and are relatively invariant with respect to retinal position, eye position, head direction, and the place where the individual is located. The underlying connectivity in humans is from ventromedial visual cortical regions to the parahippocampal scene area, leading to the theory that spatial view cells are formed by combinations of overlapping feature inputs self-organized based on their closeness in space. Thus, although spatial view cells represent "where" for episodic memory and navigation, they are formed by ventral visual stream feature inputs in the parahippocampal gyrus in what is the parahippocampal scene area. A second "where" driver of spatial view cells are parietal inputs, which it is proposed provide the idiothetic update for spatial view cells, used for memory recall and navigation when the spatial view details are obscured. Inferior temporal object "what" inputs and orbitofrontal cortex reward inputs connect to the human hippocampal system, and in macaques can be associated in the hippocampus with spatial view cell "where" representations to implement episodic memory. Hippocampal spatial view cells also provide a basis for navigation to a series of viewed landmarks, with the orbitofrontal cortex reward inputs to the hippocampus providing the goals for navigation, which can then be implemented by hippocampal connectivity in humans to parietal cortex regions involved in visuomotor actions in space. The presence of foveate vision and the highly developed temporal lobe for object and scene processing in primates including humans provide a basis for hippocampal spatial view cells to be key to understanding episodic memory in the primate and human hippocampus, and the roles of this system in primate including human navigation.
Collapse
Affiliation(s)
- Edmund T. Rolls
- Oxford Centre for Computational NeuroscienceOxfordUK
- Department of Computer ScienceUniversity of WarwickCoventryUK
| |
Collapse
|
5
|
Verhagen JV, Baker KL, Vasan G, Pieribone VA, Rolls ET. Odor encoding by signals in the olfactory bulb. J Neurophysiol 2023; 129:431-444. [PMID: 36598147 PMCID: PMC9925169 DOI: 10.1152/jn.00449.2022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2022] [Revised: 12/21/2022] [Accepted: 12/22/2022] [Indexed: 01/05/2023] Open
Abstract
To understand the operation of the olfactory system, it is essential to know how information is encoded in the olfactory bulb. We applied Shannon information theoretic methods to address this, with signals from up to 57 glomeruli simultaneously optically imaged from presynaptic inputs in glomeruli in the mouse dorsal (dOB) and lateral (lOB) olfactory bulb, in response to six exemplar pure chemical odors. We discovered that, first, the tuning of these signals from glomeruli to a set of odors is remarkably broad, with a mean sparseness of 0.83 and a mean signal correlation of 0.64. Second, both of these factors contribute to the low information that is available from the responses of even populations of many tens of glomeruli, which was only 1.35 bits across 33 glomeruli on average, compared with the 2.58 bits required to perfectly encode these six odors. Third, although there is considerable interest in the possibility of temporal encoding of stimulus including odor identity, the amount of information in the temporal aspects of the presynaptic glomerular responses was low (mean 0.11 bits) and, importantly, was redundant with respect to the information available from the rates. Fourth, the information from simultaneously recorded glomeruli asymptotes very gradually and nonlinearly, showing that glomeruli do not have independent responses. Fifth, the information from a population became available quite rapidly, within 100 ms of sniff onset, and the peak of the glomerular response was at 200 ms. Sixth, the information from the lOB was not additive with that of the dOB.NEW & NOTEWORTHY We report broad tuning and low odor information available across the lateral and dorsal bulb populations of glomeruli. Even though response latencies can be significantly predictive of stimulus identity, such contained very little information and none that was not redundant with information based on rate coding alone. Last, in line with the emerging notion of the important role of earliest stages of responses ("primacy"), we report a very rapid rise in information after each inhalation.
Collapse
Affiliation(s)
- Justus V Verhagen
- The John B. Pierce Laboratory, New Haven, Connecticut
- Department of Neuroscience, Yale University, New Haven, Connecticut
| | - Keeley L Baker
- The John B. Pierce Laboratory, New Haven, Connecticut
- Department of Neuroscience, Yale University, New Haven, Connecticut
| | - Ganesh Vasan
- The John B. Pierce Laboratory, New Haven, Connecticut
- Department of Neuroscience, Yale University, New Haven, Connecticut
| | - Vincent A Pieribone
- The John B. Pierce Laboratory, New Haven, Connecticut
- Department of Neuroscience, Yale University, New Haven, Connecticut
- Department of Cellular and Molecular Physiology, Yale School of Medicine, New Haven, Connecticut
| | - Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, United Kingdom
- University of Warwick, Coventry, United Kingdom
| |
Collapse
|
6
|
|
7
|
Rolls ET. Learning Invariant Object and Spatial View Representations in the Brain Using Slow Unsupervised Learning. Front Comput Neurosci 2021; 15:686239. [PMID: 34366818 PMCID: PMC8335547 DOI: 10.3389/fncom.2021.686239] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2021] [Accepted: 06/29/2021] [Indexed: 11/13/2022] Open
Abstract
First, neurophysiological evidence for the learning of invariant representations in the inferior temporal visual cortex is described. This includes object and face representations with invariance for position, size, lighting, view and morphological transforms in the temporal lobe visual cortex; global object motion in the cortex in the superior temporal sulcus; and spatial view representations in the hippocampus that are invariant with respect to eye position, head direction, and place. Second, computational mechanisms that enable the brain to learn these invariant representations are proposed. For the ventral visual system, one key adaptation is the use of information available in the statistics of the environment in slow unsupervised learning to learn transform-invariant representations of objects. This contrasts with deep supervised learning in artificial neural networks, which uses training with thousands of exemplars forced into different categories by neuronal teachers. Similar slow learning principles apply to the learning of global object motion in the dorsal visual system leading to the cortex in the superior temporal sulcus. The learning rule that has been explored in VisNet is an associative rule with a short-term memory trace. The feed-forward architecture has four stages, with convergence from stage to stage. This type of slow learning is implemented in the brain in hierarchically organized competitive neuronal networks with convergence from stage to stage, with only 4-5 stages in the hierarchy. Slow learning is also shown to help the learning of coordinate transforms using gain modulation in the dorsal visual system extending into the parietal cortex and retrosplenial cortex. Representations are learned that are in allocentric spatial view coordinates of locations in the world and that are independent of eye position, head direction, and the place where the individual is located. This enables hippocampal spatial view cells to use idiothetic, self-motion, signals for navigation when the view details are obscured for short periods.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, United Kingdom.,Department of Computer Science, University of Warwick, Coventry, United Kingdom
| |
Collapse
|
8
|
Rolls ET. Attractor cortical neurodynamics, schizophrenia, and depression. Transl Psychiatry 2021; 11:215. [PMID: 33846293 PMCID: PMC8041760 DOI: 10.1038/s41398-021-01333-7] [Citation(s) in RCA: 21] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/01/2021] [Revised: 03/09/2021] [Accepted: 03/24/2021] [Indexed: 12/17/2022] Open
Abstract
The local recurrent collateral connections between cortical neurons provide a basis for attractor neural networks for memory, attention, decision-making, and thereby for many aspects of human behavior. In schizophrenia, a reduction of the firing rates of cortical neurons, caused for example by reduced NMDA receptor function or reduced spines on neurons, can lead to instability of the high firing rate attractor states that normally implement short-term memory and attention in the prefrontal cortex, contributing to the cognitive symptoms. Reduced NMDA receptor function in the orbitofrontal cortex by reducing firing rates may produce negative symptoms, by reducing reward, motivation, and emotion. Reduced functional connectivity between some brain regions increases the temporal variability of the functional connectivity, contributing to the reduced stability and more loosely associative thoughts. Further, the forward projections have decreased functional connectivity relative to the back projections in schizophrenia, and this may reduce the effects of external bottom-up inputs from the world relative to internal top-down thought processes. Reduced cortical inhibition, caused by a reduction of GABA neurotransmission, can lead to instability of the spontaneous firing states of cortical networks, leading to a noise-induced jump to a high firing rate attractor state even in the absence of external inputs, contributing to the positive symptoms of schizophrenia. In depression, the lateral orbitofrontal cortex non-reward attractor network system is over-connected and has increased sensitivity to non-reward, providing a new approach to understanding depression. This is complemented by under-sensitivity and under-connectedness of the medial orbitofrontal cortex reward system in depression.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK.
- Department of Computer Science, University of Warwick, Coventry, UK.
| |
Collapse
|
9
|
Rolls ET. Neurons including hippocampal spatial view cells, and navigation in primates including humans. Hippocampus 2021; 31:593-611. [PMID: 33760309 DOI: 10.1002/hipo.23324] [Citation(s) in RCA: 22] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2020] [Revised: 03/01/2021] [Accepted: 03/13/2021] [Indexed: 01/11/2023]
Abstract
A new theory is proposed of mechanisms of navigation in primates including humans in which spatial view cells found in the primate hippocampus and parahippocampal gyrus are used to guide the individual from landmark to landmark. The navigation involves approach to each landmark in turn (taxis), using spatial view cells to identify the next landmark in the sequence, and does not require a topological map. Two other cell types found in primates, whole body motion cells, and head direction cells, can be utilized in the spatial view cell navigational mechanism, but are not essential. If the landmarks become obscured, then the spatial view representations can be updated by self-motion (idiothetic) path integration using spatial coordinate transform mechanisms in the primate dorsal visual system to transform from egocentric to allocentric spatial view coordinates. A continuous attractor network or time cells or working memory is used in this approach to navigation to encode and recall the spatial view sequences involved. I also propose how navigation can be performed using a further type of neuron found in primates, allocentric-bearing-to-a-landmark neurons, in which changes of direction are made when a landmark reaches a particular allocentric bearing. This is useful if a landmark cannot be approached. The theories are made explicit in models of navigation, which are then illustrated by computer simulations. These types of navigation are contrasted with triangulation, which requires a topological map. It is proposed that the first strategy utilizing spatial view cells is used frequently in humans, and is relatively simple because primates have spatial view neurons that respond allocentrically to locations in spatial scenes. An advantage of this approach to navigation is that hippocampal spatial view neurons are also useful for episodic memory, and for imagery.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK.,Department of Computer Science, University of Warwick, Coventry, UK
| |
Collapse
|
10
|
Lehky SR, Tanaka K, Sereno AB. Pseudosparse neural coding in the visual system of primates. Commun Biol 2021; 4:50. [PMID: 33420410 PMCID: PMC7794537 DOI: 10.1038/s42003-020-01572-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2020] [Accepted: 12/04/2020] [Indexed: 11/09/2022] Open
Abstract
When measuring sparseness in neural populations as an indicator of efficient coding, an implicit assumption is that each stimulus activates a different random set of neurons. In other words, population responses to different stimuli are, on average, uncorrelated. Here we examine neurophysiological data from four lobes of macaque monkey cortex, including V1, V2, MT, anterior inferotemporal cortex, lateral intraparietal cortex, the frontal eye fields, and perirhinal cortex, to determine how correlated population responses are. We call the mean correlation the pseudosparseness index, because high pseudosparseness can mimic statistical properties of sparseness without being authentically sparse. In every data set we find high levels of pseudosparseness ranging from 0.59-0.98, substantially greater than the value of 0.00 for authentic sparseness. This was true for synthetic and natural stimuli, as well as for single-electrode and multielectrode data. A model indicates that a key variable producing high pseudosparseness is the standard deviation of spontaneous activity across the population. Consistently high values of pseudosparseness in the data demand reconsideration of the sparse coding literature as well as consideration of the degree to which authentic sparseness provides a useful framework for understanding neural coding in the cortex.
Collapse
Affiliation(s)
- Sidney R Lehky
- Cognitive Brain Mapping Laboratory, RIKEN Center for Brain Science, Wako-shi, Saitama, 351-0198, Japan. .,Computational Neurobiology Laboratory, The Salk Institute, La Jolla, CA, 92037, USA.
| | - Keiji Tanaka
- Cognitive Brain Mapping Laboratory, RIKEN Center for Brain Science, Wako-shi, Saitama, 351-0198, Japan
| | - Anne B Sereno
- Department of Psychological Sciences, Purdue University, West Lafayette, IN, 47907, USA.,Weldon School of Biomedical Engineering, Purdue University, West Lafayette, IN, 47907, USA
| |
Collapse
|
11
|
The Generation of Time in the Hippocampal Memory System. Cell Rep 2020; 28:1649-1658.e6. [PMID: 31412236 DOI: 10.1016/j.celrep.2019.07.042] [Citation(s) in RCA: 40] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2019] [Revised: 05/30/2019] [Accepted: 07/14/2019] [Indexed: 11/21/2022] Open
Abstract
We propose that ramping time cells in the lateral entorhinal cortex can be produced by synaptic adaptation and demonstrate this in an integrate-and-fire attractor network model. We propose that competitive networks in the hippocampal system can convert these entorhinal ramping cells into hippocampal time cells and demonstrate this in a competitive network. We propose that this conversion is necessary to provide orthogonal hippocampal time representations to encode the temporal sequence of events in hippocampal episodic memory, and we support that with analytic arguments. We demonstrate that this processing can produce hippocampal neuronal ensembles that not only show replay of the sequence later on, but can also do this in reverse order in reverse replay. This research addresses a major issue in neuroscience: the mechanisms by which time is encoded in the brain and how the time representations are then useful in the hippocampal memory of events and their order.
Collapse
|
12
|
Bowers JS, Martin ND, Gale EM. Researchers Keep Rejecting Grandmother Cells after Running the Wrong Experiments: The Issue Is How Familiar Stimuli Are Identified. Bioessays 2020; 41:e1800248. [PMID: 31322760 DOI: 10.1002/bies.201800248] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2018] [Revised: 05/01/2019] [Indexed: 02/01/2023]
Abstract
There is widespread agreement in neuroscience and psychology that the visual system identifies objects and faces based on a pattern of activation over many neurons, each neuron being involved in representing many different categories. The hypothesis that the visual system includes finely tuned neurons for specific objects or faces for the sake of identification, so-called "grandmother cells", is widely rejected. Here it is argued that the rejection of grandmother cells is premature. Grandmother cells constitute a hypothesis of how familiar visual categories are identified, but the primary evidence against this hypothesis comes from studies that have failed to observe neurons that selectively respond to unfamiliar stimuli. These findings are reviewed and it is shown that they are irrelevant. Neuroscientists need to better understand existing models of face and object identification that include grandmother cells and then compare the selectivity of these units with single neurons responding to stimuli that can be identified.
Collapse
Affiliation(s)
- Jeffrey S Bowers
- School of Psychological Science, University of Bristol, Bristol, BS8 1TU, UK
| | - Nicolas D Martin
- School of Psychological Science, University of Bristol, Bristol, BS8 1TU, UK
| | - Ella M Gale
- School of Psychological Science, University of Bristol, Bristol, BS8 1TU, UK
| |
Collapse
|
13
|
Rolls ET. Spatial coordinate transforms linking the allocentric hippocampal and egocentric parietal primate brain systems for memory, action in space, and navigation. Hippocampus 2019; 30:332-353. [PMID: 31697002 DOI: 10.1002/hipo.23171] [Citation(s) in RCA: 24] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2019] [Revised: 10/05/2019] [Accepted: 10/09/2019] [Indexed: 01/03/2023]
Abstract
A theory and model of spatial coordinate transforms in the dorsal visual system through the parietal cortex that enable an interface via posterior cingulate and related retrosplenial cortex to allocentric spatial representations in the primate hippocampus is described. First, a new approach to coordinate transform learning in the brain is proposed, in which the traditional gain modulation is complemented by temporal trace rule competitive network learning. It is shown in a computational model that the new approach works much more precisely than gain modulation alone, by enabling neurons to represent the different combinations of signal and gain modulator more accurately. This understanding may have application to many brain areas where coordinate transforms are learned. Second, a set of coordinate transforms is proposed for the dorsal visual system/parietal areas that enables a representation to be formed in allocentric spatial view coordinates. The input stimulus is merely a stimulus at a given position in retinal space, and the gain modulation signals needed are eye position, head direction, and place, all of which are present in the primate brain. Neurons that encode the bearing to a landmark are involved in the coordinate transforms. Part of the importance here is that the coordinates of the allocentric view produced in this model are the same as those of spatial view cells that respond to allocentric view recorded in the primate hippocampus and parahippocampal cortex. The result is that information from the dorsal visual system can be used to update the spatial input to the hippocampus in the appropriate allocentric coordinate frame, including providing for idiothetic update to allow for self-motion. It is further shown how hippocampal spatial view cells could be useful for the transform from hippocampal allocentric coordinates to egocentric coordinates useful for actions in space and for navigation.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK.,Department of Computer Science, University of Warwick, Coventry, UK
| |
Collapse
|
14
|
Rolls ET. The storage and recall of memories in the hippocampo-cortical system. Cell Tissue Res 2018; 373:577-604. [PMID: 29218403 PMCID: PMC6132650 DOI: 10.1007/s00441-017-2744-3] [Citation(s) in RCA: 112] [Impact Index Per Article: 16.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2017] [Accepted: 11/12/2017] [Indexed: 02/07/2023]
Abstract
A quantitative computational theory of the operation of the hippocampus as an episodic memory system is described. The CA3 system operates as a single attractor or autoassociation network (1) to enable rapid one-trial associations between any spatial location (place in rodents or spatial view in primates) and an object or reward and (2) to provide for completion of the whole memory during recall from any part. The theory is extended to associations between time and object or reward to implement temporal order memory, which is also important in episodic memory. The dentate gyrus performs pattern separation by competitive learning to create sparse representations producing, for example, neurons with place-like fields from entorhinal cortex grid cells. The dentate granule cells generate, by the very small number of mossy fibre connections to CA3, a randomizing pattern separation effect that is important during learning but not recall and that separates out the patterns represented by CA3 firing as being very different from each other. This is optimal for an unstructured episodic memory system in which each memory must be kept distinct from other memories. The direct perforant path input to CA3 is quantitatively appropriate for providing the cue for recall in CA3 but not for learning. The CA1 recodes information from CA3 to set up associatively learned backprojections to the neocortex to allow the subsequent retrieval of information to the neocortex, giving a quantitative account of the large number of hippocampo-neocortical and neocortical-neocortical backprojections. Tests of the theory including hippocampal subregion analyses and hippocampal NMDA receptor knockouts are described and support the theory.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, England.
- Department of Computer Science, University of Warwick, Coventry, England.
| |
Collapse
|
15
|
Rolls ET, Mills WPC. Non-accidental properties, metric invariance, and encoding by neurons in a model of ventral stream visual object recognition, VisNet. Neurobiol Learn Mem 2018; 152:20-31. [PMID: 29723671 DOI: 10.1016/j.nlm.2018.04.017] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2018] [Revised: 04/02/2018] [Accepted: 04/27/2018] [Indexed: 11/18/2022]
Abstract
When objects transform into different views, some properties are maintained, such as whether the edges are convex or concave, and these non-accidental properties are likely to be important in view-invariant object recognition. The metric properties, such as the degree of curvature, may change with different views, and are less likely to be useful in object recognition. It is shown that in a model of invariant visual object recognition in the ventral visual stream, VisNet, non-accidental properties are encoded much more than metric properties by neurons. Moreover, it is shown how with the temporal trace rule training in VisNet, non-accidental properties of objects become encoded by neurons, and how metric properties are treated invariantly. We also show how VisNet can generalize between different objects if they have the same non-accidental property, because the metric properties are likely to overlap. VisNet is a 4-layer unsupervised model of visual object recognition trained by competitive learning that utilizes a temporal trace learning rule to implement the learning of invariance using views that occur close together in time. A second crucial property of this model of object recognition is, when neurons in the level corresponding to the inferior temporal visual cortex respond selectively to objects, whether neurons in the intermediate layers can respond to combinations of features that may be parts of two or more objects. In an investigation using the four sides of a square presented in every possible combination, it was shown that even though different layer 4 neurons are tuned to encode each feature or feature combination orthogonally, neurons in the intermediate layers can respond to features or feature combinations present is several objects. This property is an important part of the way in which high capacity can be achieved in the four-layer ventral visual cortical pathway. These findings concerning non-accidental properties and the use of neurons in intermediate layers of the hierarchy help to emphasise fundamental underlying principles of the computations that may be implemented in the ventral cortical visual stream used in object recognition.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK.
| | - W Patrick C Mills
- University of Warwick, Department of Computer Science, Coventry, UK. http://www.oxcns.org
| |
Collapse
|
16
|
Dong Q, Wang H, Hu Z. Statistics of Visual Responses to Image Object Stimuli from Primate AIT Neurons to DNN Neurons. Neural Comput 2017; 30:447-476. [PMID: 29162010 DOI: 10.1162/neco_a_01039] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Under the goal-driven paradigm, Yamins et al. ( 2014 ; Yamins & DiCarlo, 2016 ) have shown that by optimizing only the final eight-way categorization performance of a four-layer hierarchical network, not only can its top output layer quantitatively predict IT neuron responses but its penultimate layer can also automatically predict V4 neuron responses. Currently, deep neural networks (DNNs) in the field of computer vision have reached image object categorization performance comparable to that of human beings on ImageNet, a data set that contains 1.3 million training images of 1000 categories. We explore whether the DNN neurons (units in DNNs) possess image object representational statistics similar to monkey IT neurons, particularly when the network becomes deeper and the number of image categories becomes larger, using VGG19, a typical and widely used deep network of 19 layers in the computer vision field. Following Lehky, Kiani, Esteky, and Tanaka ( 2011 , 2014 ), where the response statistics of 674 IT neurons to 806 image stimuli are analyzed using three measures (kurtosis, Pareto tail index, and intrinsic dimensionality), we investigate the three issues in this letter using the same three measures: (1) the similarities and differences of the neural response statistics between VGG19 and primate IT cortex, (2) the variation trends of the response statistics of VGG19 neurons at different layers from low to high, and (3) the variation trends of the response statistics of VGG19 neurons when the numbers of stimuli and neurons increase. We find that the response statistics on both single-neuron selectivity and population sparseness of VGG19 neurons are fundamentally different from those of IT neurons in most cases; by increasing the number of neurons in different layers and the number of stimuli, the response statistics of neurons at different layers from low to high do not substantially change; and the estimated intrinsic dimensionality values at the low convolutional layers of VGG19 are considerably larger than the value of approximately 100 reported for IT neurons in Lehky et al. ( 2014 ), whereas those at the high fully connected layers are close to or lower than 100. To the best of our knowledge, this work is the first attempt to analyze the response statistics of DNN neurons with respect to primate IT neurons in image object representation.
Collapse
Affiliation(s)
- Qiulei Dong
- National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China; University of Chinese Academy of Sciences, Beijing 100049, China; and CAS Center for Excellence in Brain Science and Intelligence Technology, Beijing 100190, China
| | - Hong Wang
- University of Chinese Academy of Sciences, Beijing 100049, China
| | - Zhanyi Hu
- National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China; University of Chinese Academy of Sciences, Beijing 100049, China; and CAS Center for Excellence in Brain Science and Intelligence Technology, Beijing 100190, China
| |
Collapse
|
17
|
Dong Q, Liu B, Hu Z. Comparison of IT Neural Response Statistics with Simulations. Front Comput Neurosci 2017; 11:60. [PMID: 28747882 PMCID: PMC5506183 DOI: 10.3389/fncom.2017.00060] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2017] [Accepted: 06/23/2017] [Indexed: 11/13/2022] Open
Abstract
Lehky et al. (2011) provided a statistical analysis on the responses of the recorded 674 neurons to 806 image stimuli in anterior inferotemporalm (AIT) cortex of two monkeys. In terms of kurtosis and Pareto tail index, they observed that the population sparseness of both unnormalized and normalized responses is always larger than their single-neuron selectivity, hence concluded that the critical features for individual neurons in primate AIT cortex are not very complex, but there is an indefinitely large number of them. In this work, we explore an "inverse problem" by simulation, that is, by simulating each neuron indeed only responds to a very limited number of stimuli among a very large number of neurons and stimuli, to assess whether the population sparseness is always larger than the single-neuron selectivity. Our simulation results show that the population sparseness exceeds the single-neuron selectivity in most cases even if the number of neurons and stimuli are much larger than several hundreds, which confirms the observations in Lehky et al. (2011). In addition, we found that the variances of the computed kurtosis and Pareto tail index are quite large in some cases, which reveals some limitations of these two criteria when used for neuron response evaluation.
Collapse
Affiliation(s)
- Qiulei Dong
- National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of SciencesBeijing, China.,Department of Artificial Intelligence, University of Chinese Academy of SciencesBeijing, China.,Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of SciencesBeijing, China
| | - Bo Liu
- National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of SciencesBeijing, China.,Department of Artificial Intelligence, University of Chinese Academy of SciencesBeijing, China
| | - Zhanyi Hu
- National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of SciencesBeijing, China.,Department of Artificial Intelligence, University of Chinese Academy of SciencesBeijing, China.,Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of SciencesBeijing, China
| |
Collapse
|
18
|
Rolls ET, Deco G. Non-reward neural mechanisms in the orbitofrontal cortex. Cortex 2016; 83:27-38. [PMID: 27474915 DOI: 10.1080/23273798.2016.1203443] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2016] [Revised: 05/03/2016] [Accepted: 06/24/2016] [Indexed: 05/27/2023]
Abstract
Single neurons in the primate orbitofrontal cortex respond when an expected reward is not obtained, and behaviour must change. The human lateral orbitofrontal cortex is activated when non-reward, or loss occurs. The neuronal computation of this negative reward prediction error is fundamental for the emotional changes associated with non-reward, and with changing behaviour. Little is known about the neuronal mechanism. Here we propose a mechanism, which we formalize into a neuronal network model, which is simulated to enable the operation of the mechanism to be investigated. A single attractor network has a reward population (or pool) of neurons that is activated by expected reward, and maintain their firing until, after a time, synaptic depression reduces the firing rate in this neuronal population. If a reward outcome is not received, the decreasing firing in the reward neurons releases the inhibition implemented by inhibitory neurons, and this results in a second population of non-reward neurons to start and continue firing encouraged by the spiking-related noise in the network. If a reward outcome is received, this keeps the reward attractor active, and this through the inhibitory neurons prevents the non-reward attractor neurons from being activated. If an expected reward has been signalled, and the reward attractor neurons are active, their firing can be directly inhibited by a non-reward outcome, and the non-reward neurons become activated because the inhibition on them is released. The neuronal mechanisms in the orbitofrontal cortex for computing negative reward prediction error are important, for this system may be over-reactive in depression, under-reactive in impulsive behaviour, and may influence the dopaminergic 'prediction error' neurons.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK; University of Warwick, Department of Computer Science, Coventry, UK. http://www.oxcns.org
| | - Gustavo Deco
- Universitat Pompeu Fabra, Theoretical and Computational Neuroscience, Barcelona, Spain; Institucio Catalana de Recerca i Estudis Avancats (ICREA), Spain
| |
Collapse
|
19
|
Zhou S, Migliore M, Yu Y. Odor Experience Facilitates Sparse Representations of New Odors in a Large-Scale Olfactory Bulb Model. Front Neuroanat 2016; 10:10. [PMID: 26903819 PMCID: PMC4749983 DOI: 10.3389/fnana.2016.00010] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2015] [Accepted: 01/27/2016] [Indexed: 01/11/2023] Open
Abstract
Prior odor experience has a profound effect on the coding of new odor inputs by animals. The olfactory bulb, the first relay of the olfactory pathway, can substantially shape the representations of odor inputs. How prior odor experience affects the representation of new odor inputs in olfactory bulb and its underlying network mechanism are still unclear. Here we carried out a series of simulations based on a large-scale realistic mitral-granule network model and found that prior odor experience not only accelerated formation of the network, but it also significantly strengthened sparse responses in the mitral cell network while decreasing sparse responses in the granule cell network. This modulation of sparse representations may be due to the increase of inhibitory synaptic weights. Correlations among mitral cells within the network and correlations between mitral network responses to different odors decreased gradually when the number of prior training odors was increased, resulting in a greater decorrelation of the bulb representations of input odors. Based on these findings, we conclude that the degree of prior odor experience facilitates degrees of sparse representations of new odors by the mitral cell network through experience-enhanced inhibition mechanism.
Collapse
Affiliation(s)
- Shanglin Zhou
- School of Life Science and The Collaborative Innovation Center for Brain Science, The Center for Computational Systems Biology, Fudan University Shanghai, China
| | - Michele Migliore
- Division of Palermo, Institute of Biophysics, National Research CouncilPalermo, Italy; Department of Neurobiology, Yale University School of MedicineNew Haven, CT, USA
| | - Yuguo Yu
- School of Life Science and The Collaborative Innovation Center for Brain Science, The Center for Computational Systems Biology, Fudan University Shanghai, China
| |
Collapse
|
20
|
Robinson L, Rolls ET. Invariant visual object recognition: biologically plausible approaches. BIOLOGICAL CYBERNETICS 2015; 109:505-35. [PMID: 26335743 PMCID: PMC4572081 DOI: 10.1007/s00422-015-0658-2] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/03/2015] [Accepted: 08/14/2015] [Indexed: 06/01/2023]
Abstract
Key properties of inferior temporal cortex neurons are described, and then, the biological plausibility of two leading approaches to invariant visual object recognition in the ventral visual system is assessed to investigate whether they account for these properties. Experiment 1 shows that VisNet performs object classification with random exemplars comparably to HMAX, except that the final layer C neurons of HMAX have a very non-sparse representation (unlike that in the brain) that provides little information in the single-neuron responses about the object class. Experiment 2 shows that VisNet forms invariant representations when trained with different views of each object, whereas HMAX performs poorly when assessed with a biologically plausible pattern association network, as HMAX has no mechanism to learn view invariance. Experiment 3 shows that VisNet neurons do not respond to scrambled images of faces, and thus encode shape information. HMAX neurons responded with similarly high rates to the unscrambled and scrambled faces, indicating that low-level features including texture may be relevant to HMAX performance. Experiment 4 shows that VisNet can learn to recognize objects even when the view provided by the object changes catastrophically as it transforms, whereas HMAX has no learning mechanism in its S-C hierarchy that provides for view-invariant learning. This highlights some requirements for the neurobiological mechanisms of high-level vision, and how some different approaches perform, in order to help understand the fundamental underlying principles of invariant visual object recognition in the ventral visual stream.
Collapse
Affiliation(s)
- Leigh Robinson
- Department of Computer Science, University of Warwick, Coventry, UK
| | - Edmund T Rolls
- Department of Computer Science, University of Warwick, Coventry, UK.
- Oxford Centre for Computational Neuroscience, Oxford, UK.
| |
Collapse
|
21
|
Rolls ET. Pattern separation, completion, and categorisation in the hippocampus and neocortex. Neurobiol Learn Mem 2015; 129:4-28. [PMID: 26190832 DOI: 10.1016/j.nlm.2015.07.008] [Citation(s) in RCA: 123] [Impact Index Per Article: 12.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2015] [Revised: 07/02/2015] [Accepted: 07/11/2015] [Indexed: 12/22/2022]
Abstract
The mechanisms for pattern completion and pattern separation are described in the context of a theory of hippocampal function in which the hippocampal CA3 system operates as a single attractor or autoassociation network to enable rapid, one-trial, associations between any spatial location (place in rodents, or spatial view in primates) and an object or reward, and to provide for completion of the whole memory during recall from any part. The factors important in the pattern completion in CA3 and also a large number of independent memories stored in CA3 include: a sparse distributed representation, representations that are independent due to the randomizing effect of the mossy fibres, heterosynaptic long-term depression as well as long-term potentiation in the recurrent collateral synapses, and diluted connectivity to minimize the number of multiple synapses between any pair of CA3 neurons which otherwise distort the basins of attraction. Recall of information from CA3 is implemented by the entorhinal cortex perforant path synapses to CA3 cells, which in acting as a pattern associator allow some pattern generalization. Pattern separation is performed in the dentate granule cells using competitive learning to convert grid-like entorhinal cortex firing to place-like fields, and in the dentate to CA3 connections that have diluted connectivity. Recall to the neocortex is achieved by a reverse hierarchical series of pattern association networks implemented by the hippocampo-cortical backprojections, each one of which performs some pattern generalization, to retrieve a complete pattern of cortical firing in higher-order cortical areas. New results on competitive networks show which factors contribute to their ability to perform pattern separation, pattern clustering, and pattern categorisation, and how these apply in different hippocampal and neocortical systems.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, England, United Kingdom; University of Warwick, Department of Computer Science, Coventry CV4 7AL, England, United Kingdom.
| |
Collapse
|
22
|
Sparse coding and lateral inhibition arising from balanced and unbalanced dendrodendritic excitation and inhibition. J Neurosci 2015; 34:13701-13. [PMID: 25297097 DOI: 10.1523/jneurosci.1834-14.2014] [Citation(s) in RCA: 38] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/27/2022] Open
Abstract
The precise mechanism by which synaptic excitation and inhibition interact with each other in odor coding through the unique dendrodendritic synaptic microcircuits present in olfactory bulb is unknown. Here a scaled-up model of the mitral-granule cell network in the rodent olfactory bulb is used to analyze dendrodendritic processing of experimentally determined odor patterns. We found that the interaction between excitation and inhibition is responsible for two fundamental computational mechanisms: (1) a balanced excitation/inhibition in strongly activated mitral cells, leading to a sparse representation of odorant input, and (2) an unbalanced excitation/inhibition (inhibition dominated) in surrounding weakly activated mitral cells, leading to lateral inhibition. These results suggest how both mechanisms can carry information about the input patterns, with optimal level of synaptic excitation and inhibition producing the highest level of sparseness and decorrelation in the network response. The results suggest how the learning process, through the emergent development of these mechanisms, can enhance odor representation of olfactory bulb.
Collapse
|
23
|
Abstract
It remains unclear how single neurons in the human brain represent whole-object visual stimuli. While recordings in both human and nonhuman primates have shown distributed representations of objects (many neurons encoding multiple objects), recordings of single neurons in the human medial temporal lobe, taken as subjects' discriminated objects during multiple presentations, have shown gnostic representations (single neurons encoding one object). Because some studies suggest that repeated viewing may enhance neural selectivity for objects, we had human subjects discriminate objects in a single, more naturalistic viewing session. We found that, across 432 well isolated neurons recorded in the hippocampus and amygdala, the average fraction of objects encoded was 26%. We also found that more neurons encoded several objects versus only one object in the hippocampus (28 vs 18%, p < 0.001) and in the amygdala (30 vs 19%, p < 0.001). Thus, during realistic viewing experiences, typical neurons in the human medial temporal lobe code for a considerable range of objects, across multiple semantic categories.
Collapse
|
24
|
A computational theory of hippocampal function, and tests of the theory: New developments. Neurosci Biobehav Rev 2015; 48:92-147. [DOI: 10.1016/j.neubiorev.2014.11.009] [Citation(s) in RCA: 226] [Impact Index Per Article: 22.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2014] [Revised: 10/24/2014] [Accepted: 11/12/2014] [Indexed: 01/01/2023]
|
25
|
Networks for memory, perception, and decision-making, and beyond to how the syntax for language might be implemented in the brain. Brain Res 2014; 1621:316-34. [PMID: 25239476 DOI: 10.1016/j.brainres.2014.09.021] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2014] [Revised: 09/05/2014] [Accepted: 09/08/2014] [Indexed: 12/24/2022]
Abstract
Neural principles that provide a foundation for memory, perception, and decision-making include place coding with sparse distributed representations, associative synaptic modification, and attractor networks in which the storage capacity is in the order of the number of associatively modifiable recurrent synapses on any one neuron. Based on those and further principles of cortical computation, hypotheses are explored in which syntax is encoded in the cortex using sparse distributed place coding. Each cortical module 2-3 mm in diameter is proposed to be formed of a local attractor neuronal network with a capacity in the order of 10,000 words (e.g. subjects, verbs or objects depending on the module). Such a system may form a deep language-of-thought layer. For the information to be communicated to other people, the modules in which the neurons are firing which encode the syntactic role, as well as which neurons are firing to specify the words, must be communicated. It is proposed that one solution to this (used in English) is temporal order encoding, for example subject-verb-object. It is shown with integrate-and-fire simulations that this order encoding could be implemented by weakly forward-coupled subject-verb-object modules. A related system can decode a temporal sequence. This approach based on known principles of cortical computation needs to be extended to investigate further whether it could form a biological foundation for the implementation of language in the brain. This article is part of a Special Issue entitled SI: Brain and Memory.
Collapse
|
26
|
Rolls ET, Webb TJ. Finding and recognizing objects in natural scenes: complementary computations in the dorsal and ventral visual systems. Front Comput Neurosci 2014; 8:85. [PMID: 25161619 PMCID: PMC4130325 DOI: 10.3389/fncom.2014.00085] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2014] [Accepted: 07/16/2014] [Indexed: 01/09/2023] Open
Abstract
Searching for and recognizing objects in complex natural scenes is implemented by multiple saccades until the eyes reach within the reduced receptive field sizes of inferior temporal cortex (IT) neurons. We analyze and model how the dorsal and ventral visual streams both contribute to this. Saliency detection in the dorsal visual system including area LIP is modeled by graph-based visual saliency, and allows the eyes to fixate potential objects within several degrees. Visual information at the fixated location subtending approximately 9° corresponding to the receptive fields of IT neurons is then passed through a four layer hierarchical model of the ventral cortical visual system, VisNet. We show that VisNet can be trained using a synaptic modification rule with a short-term memory trace of recent neuronal activity to capture both the required view and translation invariances to allow in the model approximately 90% correct object recognition for 4 objects shown in any view across a range of 135° anywhere in a scene. The model was able to generalize correctly within the four trained views and the 25 trained translations. This approach analyses the principles by which complementary computations in the dorsal and ventral visual cortical streams enable objects to be located and recognized in complex natural scenes.
Collapse
Affiliation(s)
- Edmund T. Rolls
- Department of Computer Science, University of WarwickCoventry, UK
- Oxford Centre for Computational NeuroscienceOxford, UK
| | - Tristan J. Webb
- Department of Computer Science, University of WarwickCoventry, UK
| |
Collapse
|
27
|
Elliott T. Sparseness, antisparseness and anything in between: the operating point of a neuron determines its computational repertoire. Neural Comput 2014; 26:1924-72. [PMID: 24922502 DOI: 10.1162/neco_a_00630] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
A recent model of intrinsic plasticity coupled to Hebbian synaptic plasticity proposes that adaptation of a neuron's threshold and gain in a sigmoidal response function to achieve a sparse, exponential output firing rate distribution facilitates the discovery of heavy-tailed or super- gaussian sources in the neuron's inputs. We show that the exponential output distribution is irrelevant to these dynamics and that, furthermore, while sparseness is sufficient, it is not necessary. The intrinsic plasticity mechanism drives the neuron's threshold large and positive, and we prove that in such a regime, the neuron will find supergaussian sources; equally, however, if the threshold is large and negative (an antisparse regime), it will also find supergaussian sources. Away from such extremes, the neuron can also discover subgaussian sources. By examining a neuron with a fixed sigmoidal nonlinearity and considering the synaptic strength fixed-point structure in the two-dimensional parameter space defined by the neuron's threshold and gain, we show that this space is carved up into sub- and supergaussian-input-finding regimes, possibly with regimes of simultaneous stability of sub- and supergaussian sources or regimes of instability of all sources; a single gaussian source may also be stabilized by the presence of a nongaussian source. A neuron's operating point (essentially its threshold and gain coupled with its input statistics) therefore critically determines its computational repertoire. Intrinsic plasticity mechanisms induce trajectories in this parameter space but do not fundamentally modify it. Unless the trajectories cross critical boundaries in this space, intrinsic plasticity is irrelevant and the neuron's nonlinearity may be frozen with identical receptive field refinement dynamics.
Collapse
Affiliation(s)
- Terry Elliott
- Department of Electronics and Computer Science, University of Southampton, Highfield, Southampton, SO17 1BJ, U.K.
| |
Collapse
|
28
|
Webb TJ, Rolls ET. Deformation-specific and deformation-invariant visual object recognition: pose vs. identity recognition of people and deforming objects. Front Comput Neurosci 2014; 8:37. [PMID: 24744725 PMCID: PMC3978248 DOI: 10.3389/fncom.2014.00037] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2013] [Accepted: 03/12/2014] [Indexed: 11/18/2022] Open
Abstract
When we see a human sitting down, standing up, or walking, we can recognize one of these poses independently of the individual, or we can recognize the individual person, independently of the pose. The same issues arise for deforming objects. For example, if we see a flag deformed by the wind, either blowing out or hanging languidly, we can usually recognize the flag, independently of its deformation; or we can recognize the deformation independently of the identity of the flag. We hypothesize that these types of recognition can be implemented by the primate visual system using temporo-spatial continuity as objects transform as a learning principle. In particular, we hypothesize that pose or deformation can be learned under conditions in which large numbers of different people are successively seen in the same pose, or objects in the same deformation. We also hypothesize that person-specific representations that are independent of pose, and object-specific representations that are independent of deformation and view, could be built, when individual people or objects are observed successively transforming from one pose or deformation and view to another. These hypotheses were tested in a simulation of the ventral visual system, VisNet, that uses temporal continuity, implemented in a synaptic learning rule with a short-term memory trace of previous neuronal activity, to learn invariant representations. It was found that depending on the statistics of the visual input, either pose-specific or deformation-specific representations could be built that were invariant with respect to individual and view; or that identity-specific representations could be built that were invariant with respect to pose or deformation and view. We propose that this is how pose-specific and pose-invariant, and deformation-specific and deformation-invariant, perceptual representations are built in the brain.
Collapse
Affiliation(s)
- Tristan J. Webb
- Department of Computer Science, University of WarwickCoventry, UK
| | - Edmund T. Rolls
- Department of Computer Science, University of WarwickCoventry, UK
- Oxford Centre for Computational NeuroscienceOxford, UK
| |
Collapse
|
29
|
Samonds JM, Potetz BR, Lee TS. Sample skewness as a statistical measurement of neuronal tuning sharpness. Neural Comput 2014; 26:860-906. [PMID: 24555451 DOI: 10.1162/neco_a_00582] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
We propose using the statistical measurement of the sample skewness of the distribution of mean firing rates of a tuning curve to quantify sharpness of tuning. For some features, like binocular disparity, tuning curves are best described by relatively complex and sometimes diverse functions, making it difficult to quantify sharpness with a single function and parameter. Skewness provides a robust nonparametric measure of tuning curve sharpness that is invariant with respect to the mean and variance of the tuning curve and is straightforward to apply to a wide range of tuning, including simple orientation tuning curves and complex object tuning curves that often cannot even be described parametrically. Because skewness does not depend on a specific model or function of tuning, it is especially appealing to cases of sharpening where recurrent interactions among neurons produce sharper tuning curves that deviate in a complex manner from the feedforward function of tuning. Since tuning curves for all neurons are not typically well described by a single parametric function, this model independence additionally allows skewness to be applied to all recorded neurons, maximizing the statistical power of a set of data. We also compare skewness with other nonparametric measures of tuning curve sharpness and selectivity. Compared to these other nonparametric measures tested, skewness is best used for capturing the sharpness of multimodal tuning curves defined by narrow peaks (maximum) and broad valleys (minima). Finally, we provide a more formal definition of sharpness using a shape-based information gain measure and derive and show that skewness is correlated with this definition.
Collapse
Affiliation(s)
- Jason M Samonds
- Center for the Neural Basis of Cognition and Computer Science Department, Carnegie Mellon University, Pittsburgh, PA 15213, U.S.A.
| | | | | |
Collapse
|
30
|
Rolls ET. The mechanisms for pattern completion and pattern separation in the hippocampus. Front Syst Neurosci 2013; 7:74. [PMID: 24198767 PMCID: PMC3812781 DOI: 10.3389/fnsys.2013.00074] [Citation(s) in RCA: 285] [Impact Index Per Article: 23.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2013] [Accepted: 10/14/2013] [Indexed: 12/30/2022] Open
Abstract
The mechanisms for pattern completion and pattern separation are described in the context of a theory of hippocampal function in which the hippocampal CA3 system operates as a single attractor or autoassociation network to enable rapid, one-trial, associations between any spatial location (place in rodents, or spatial view in primates) and an object or reward, and to provide for completion of the whole memory during recall from any part. The factors important in the pattern completion in CA3 together with a large number of independent memories stored in CA3 include a sparse distributed representation which is enhanced by the graded firing rates of CA3 neurons, representations that are independent due to the randomizing effect of the mossy fibers, heterosynaptic long-term depression as well as long-term potentiation in the recurrent collateral synapses, and diluted connectivity to minimize the number of multiple synapses between any pair of CA3 neurons which otherwise distort the basins of attraction. Recall of information from CA3 is implemented by the entorhinal cortex perforant path synapses to CA3 cells, which in acting as a pattern associator allow some pattern generalization. Pattern separation is performed in the dentate granule cells using competitive learning to convert grid-like entorhinal cortex firing to place-like fields. Pattern separation in CA3, which is important for completion of any one of the stored patterns from a fragment, is provided for by the randomizing effect of the mossy fiber synapses to which neurogenesis may contribute, by the large number of dentate granule cells each with a sparse representation, and by the sparse independent representations in CA3. Recall to the neocortex is achieved by a reverse hierarchical series of pattern association networks implemented by the hippocampo-cortical backprojections, each one of which performs some pattern generalization, to retrieve a complete pattern of cortical firing in higher-order cortical areas.
Collapse
Affiliation(s)
- Edmund T. Rolls
- Oxford Centre for Computational NeuroscienceOxford, UK
- Department of Computer Science, University of WarwickCoventry, UK
| |
Collapse
|
31
|
Rolls ET. A quantitative theory of the functions of the hippocampal CA3 network in memory. Front Cell Neurosci 2013; 7:98. [PMID: 23805074 PMCID: PMC3691555 DOI: 10.3389/fncel.2013.00098] [Citation(s) in RCA: 71] [Impact Index Per Article: 5.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2013] [Accepted: 06/05/2013] [Indexed: 12/21/2022] Open
Abstract
A quantitative computational theory of the operation of the hippocampal CA3 system as an autoassociation or attractor network used in episodic memory system is described. In this theory, the CA3 system operates as a single attractor or autoassociation network to enable rapid, one-trial, associations between any spatial location (place in rodents, or spatial view in primates) and an object or reward, and to provide for completion of the whole memory during recall from any part. The theory is extended to associations between time and object or reward to implement temporal order memory, also important in episodic memory. The dentate gyrus (DG) performs pattern separation by competitive learning to produce sparse representations suitable for setting up new representations in CA3 during learning, producing for example neurons with place-like fields from entorhinal cortex grid cells. The dentate granule cells produce by the very small number of mossy fiber (MF) connections to CA3 a randomizing pattern separation effect important during learning but not recall that separates out the patterns represented by CA3 firing to be very different from each other, which is optimal for an unstructured episodic memory system in which each memory must be kept distinct from other memories. The direct perforant path (pp) input to CA3 is quantitatively appropriate to provide the cue for recall in CA3, but not for learning. Tests of the theory including hippocampal subregion analyses and hippocampal NMDA receptor knockouts are described, and support the theory.
Collapse
Affiliation(s)
- Edmund T. Rolls
- Oxford Centre for Computational NeuroscienceOxford, UK
- Department of Computer Science, University of WarwickCoventry, UK
| |
Collapse
|
32
|
Palm G. Neural associative memories and sparse coding. Neural Netw 2013; 37:165-71. [DOI: 10.1016/j.neunet.2012.08.013] [Citation(s) in RCA: 80] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2012] [Revised: 08/17/2012] [Accepted: 08/22/2012] [Indexed: 11/16/2022]
|
33
|
Abstract
Ciliates become highly social, even displaying animal-like qualities, in the joint presence of aroused conspecifics and nonself mating pheromones. Pheromone detection putatively helps trigger instinctual and learned courtship and dominance displays from which social judgments are made about the availability, compatibility, and fitness representativeness or likelihood of prospective mates and rivals. In earlier studies, I demonstrated the heterotrich Spirostomum ambiguum improves mating competence by effecting preconjugal strategies and inferences in mock social trials via behavioral heuristics built from Hebbian-like associative learning. Heuristics embody serial patterns of socially relevant action that evolve into ordered, topologically invariant computational networks supporting intra- and intermate selection. S. ambiguum employs heuristics to acquire, store, plan, compare, modify, select, and execute sets of mating propaganda. One major adaptive constraint over formation and use of heuristics involves a ciliate’s initial subjective bias, responsiveness, or preparedness, as defined by Stevens’ Law of subjective stimulus intensity, for perceiving the meaningfulness of mechanical pressures accompanying cell-cell contacts and additional perimating events. This bias controls durations and valences of nonassociative learning, search rates for appropriate mating strategies, potential net reproductive payoffs, levels of social honesty and deception, successful error diagnosis and correction of mating signals, use of insight or analysis to solve mating dilemmas, bioenergetics expenditures, and governance of mating decisions by classical or quantum statistical mechanics. I now report this same social bias also differentially affects the spatiotemporal sparseness, as measured with metric entropy, of ciliate heuristics. Sparseness plays an important role in neural systems through optimizing the specificity, efficiency, and capacity of memory representations. The present findings indicate sparseness performs a similar function in single aneural cells by tuning the size and density of encoded computational architectures useful for decision making in social contexts.
Collapse
|
34
|
Rolls ET. Invariant Visual Object and Face Recognition: Neural and Computational Bases, and a Model, VisNet. Front Comput Neurosci 2012; 6:35. [PMID: 22723777 PMCID: PMC3378046 DOI: 10.3389/fncom.2012.00035] [Citation(s) in RCA: 59] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2011] [Accepted: 05/23/2012] [Indexed: 11/13/2022] Open
Abstract
Neurophysiological evidence for invariant representations of objects and faces in the primate inferior temporal visual cortex is described. Then a computational approach to how invariant representations are formed in the brain is described that builds on the neurophysiology. A feature hierarchy model in which invariant representations can be built by self-organizing learning based on the temporal and spatial statistics of the visual input produced by objects as they transform in the world is described. VisNet can use temporal continuity in an associative synaptic learning rule with a short-term memory trace, and/or it can use spatial continuity in continuous spatial transformation learning which does not require a temporal trace. The model of visual processing in the ventral cortical stream can build representations of objects that are invariant with respect to translation, view, size, and also lighting. The model has been extended to provide an account of invariant representations in the dorsal visual system of the global motion produced by objects such as looming, rotation, and object-based movement. The model has been extended to incorporate top-down feedback connections to model the control of attention by biased competition in, for example, spatial and object search tasks. The approach has also been extended to account for how the visual system can select single objects in complex visual scenes, and how multiple objects can be represented in a scene. The approach has also been extended to provide, with an additional layer, for the development of representations of spatial scenes of the type found in the hippocampus.
Collapse
Affiliation(s)
- Edmund T. Rolls
- Oxford Centre for Computational NeuroscienceOxford, UK
- Department of Computer Science, University of WarwickCoventry, UK
| |
Collapse
|
35
|
Rolls ET, Treves A. The neuronal encoding of information in the brain. Prog Neurobiol 2011; 95:448-90. [PMID: 21907758 DOI: 10.1016/j.pneurobio.2011.08.002] [Citation(s) in RCA: 163] [Impact Index Per Article: 11.6] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2010] [Revised: 08/03/2011] [Accepted: 08/15/2011] [Indexed: 11/16/2022]
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK
| | | |
Collapse
|
36
|
Ison MJ, Mormann F, Cerf M, Koch C, Fried I, Quiroga RQ. Selectivity of pyramidal cells and interneurons in the human medial temporal lobe. J Neurophysiol 2011; 106:1713-21. [PMID: 21715671 PMCID: PMC3191845 DOI: 10.1152/jn.00576.2010] [Citation(s) in RCA: 42] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2010] [Accepted: 06/27/2011] [Indexed: 11/22/2022] Open
Abstract
Neurons in the medial temporal lobe (MTL) respond selectively to pictures of specific individuals, objects, and places. However, the underlying mechanisms leading to such degree of stimulus selectivity are largely unknown. A necessary step to move forward in this direction involves the identification and characterization of the different neuron types present in MTL circuitry. We show that putative principal cells recorded in vivo from the human MTL are more selective than putative interneurons. Furthermore, we report that putative hippocampal pyramidal cells exhibit the highest degree of selectivity within the MTL, reflecting the hierarchical processing of visual information. We interpret these differences in selectivity as a plausible mechanism for generating sparse responses.
Collapse
Affiliation(s)
- Matias J Ison
- Department of Engineering, University of Leicester, Leicester,UK.
| | | | | | | | | | | |
Collapse
|
37
|
Webb TJ, Rolls ET, Deco G, Feng J. Noise in attractor networks in the brain produced by graded firing rate representations. PLoS One 2011; 6:e23630. [PMID: 21931607 PMCID: PMC3169549 DOI: 10.1371/journal.pone.0023630] [Citation(s) in RCA: 13] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2011] [Accepted: 07/20/2011] [Indexed: 11/19/2022] Open
Abstract
Representations in the cortex are often distributed with graded firing rates in the neuronal populations. The firing rate probability distribution of each neuron to a set of stimuli is often exponential or gamma. In processes in the brain, such as decision-making, that are influenced by the noise produced by the close to random spike timings of each neuron for a given mean rate, the noise with this graded type of representation may be larger than with the binary firing rate distribution that is usually investigated. In integrate-and-fire simulations of an attractor decision-making network, we show that the noise is indeed greater for a given sparseness of the representation for graded, exponential, than for binary firing rate distributions. The greater noise was measured by faster escaping times from the spontaneous firing rate state when the decision cues are applied, and this corresponds to faster decision or reaction times. The greater noise was also evident as less stability of the spontaneous firing state before the decision cues are applied. The implication is that spiking-related noise will continue to be a factor that influences processes such as decision-making, signal detection, short-term memory, and memory recall even with the quite large networks found in the cerebral cortex. In these networks there are several thousand recurrent collateral synapses onto each neuron. The greater noise with graded firing rate distributions has the advantage that it can increase the speed of operation of cortical circuitry.
Collapse
Affiliation(s)
- Tristan J. Webb
- Department of Computer Science and Complexity Science Centre, University of Warwick, Coventry, United Kingdom
| | - Edmund T. Rolls
- Oxford Centre for Computational Neuroscience, Oxford, United Kingdom
- Department of Computer Science, University of Warwick, Coventry, United Kingdom
| | - Gustavo Deco
- Theoretical and Computational Neuroscience, Universitat Pompeu Fabra, Barcelona, Spain
| | - Jianfeng Feng
- Department of Computer Science and Complexity Science Centre, University of Warwick, Coventry, United Kingdom
| |
Collapse
|
38
|
Cortical attractor network dynamics with diluted connectivity. Brain Res 2011; 1434:212-25. [PMID: 21875702 DOI: 10.1016/j.brainres.2011.08.002] [Citation(s) in RCA: 20] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2011] [Revised: 07/29/2011] [Accepted: 08/02/2011] [Indexed: 11/23/2022]
Abstract
The connectivity of the cerebral cortex is diluted, with the probability of excitatory connections between even nearby pyramidal cells rarely more than 0.1, and in the hippocampus 0.04. To investigate the extent to which this diluted connectivity affects the dynamics of attractor networks in the cerebral cortex, we simulated an integrate-and-fire attractor network taking decisions between competing inputs with diluted connectivity of 0.25 or 0.1, and with the same number of synaptic connections per neuron for the recurrent collateral synapses within an attractor population as for full connectivity. The results indicated that there was less spiking-related noise with the diluted connectivity in that the stability of the network when in the spontaneous state of firing increased, and the accuracy of the correct decisions increased. The decision times were a little slower with diluted than with complete connectivity. Given that the capacity of the network is set by the number of recurrent collateral synaptic connections per neuron, on which there is a biological limit, the findings indicate that the stability of cortical networks, and the accuracy of their correct decisions or memory recall operations, can be increased by utilizing diluted connectivity and correspondingly increasing the number of neurons in the network, with little impact on the speed of processing of the cortex. Thus diluted connectivity can decrease cortical spiking-related noise. In addition, we show that the Fano factor for the trial-to-trial variability of the neuronal firing decreases from the spontaneous firing state value when the attractor network makes a decision. This article is part of a Special Issue entitled "Neural Coding".
Collapse
|
39
|
Variability of spike firing during θ-coupled replay of memories in a simulated attractor network. Brain Res 2011; 1434:152-61. [PMID: 21907326 DOI: 10.1016/j.brainres.2011.07.055] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2011] [Revised: 06/16/2011] [Accepted: 07/27/2011] [Indexed: 11/21/2022]
Abstract
Simulation work has recently shown that attractor networks can reproduce Poisson-like variability of single cell spiking, with coefficient of variation (Cv(2)) around unity, consistent with cortical data. However, the use of local variability (Lv) measures has revealed area- and layer-specific deviations from Poisson-like firing. In order to test these findings in silico we used a biophysically detailed attractor network model. We show that Lv well above 1, specifically found in superficial cortical layers and prefrontal areas, can indeed be reproduced in such networks and is consistent with periodic replay rather than persistent firing. The memory replay at the theta time scale provides a framework for a multi-item memory storage in the model. This article is part of a Special Issue entitled Neural Coding.
Collapse
|
40
|
Akrami A, Russo E, Treves A. Lateral thinking, from the Hopfield model to cortical dynamics. Brain Res 2011; 1434:4-16. [PMID: 21839426 DOI: 10.1016/j.brainres.2011.07.030] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2011] [Revised: 07/10/2011] [Accepted: 07/13/2011] [Indexed: 11/28/2022]
Abstract
Self-organizing attractor networks may comprise the building blocks for cortical dynamics, providing the basic operations of categorization, including analog-to-digital conversion, association and auto-association, which are then expressed as components of distinct cognitive functions depending on the contents of the neural codes in each region. To assess the viability of this scenario, we first review how a local cortical patch may be modeled as an attractor network, in which memory representations are not artificially stored as prescribed binary patterns of activity as in the Hopfield model, but self-organize as continuously graded patterns induced by afferent input. Recordings in macaques indicate that such cortical attractor networks may express retrieval dynamics over cognitively plausible rapid time scales, shorter than those dominated by neuronal fatigue. A cortical network encompassing many local attractor networks, and incorporating a realistic description of adaptation dynamics, may be captured by a Potts model. This network model has the capacity to engage long-range associations into sustained iterative attractor dynamics at a cortical scale, in what may be regarded as a mathematical model of spontaneous lateral thought. This article is part of a Special Issue entitled: Neural Coding.
Collapse
Affiliation(s)
- Athena Akrami
- SISSA, Cognitive Neuroscience sector, via Bonomea 265, 34136 Trieste, Italy
| | | | | |
Collapse
|
41
|
Abstract
We consider the mechanisms that enable decisions to be postponed for a period after the evidence has been provided. Using an information theoretic approach, we show that information about the forthcoming action becomes available from the activity of neurons in the medial premotor cortex in a sequential decision-making task after the second stimulus is applied, providing the information for a decision about whether the first or second stimulus is higher in vibrotactile frequency. The information then decays in a 3-s delay period in which the neuronal activity declines before the behavioral response can be made. The information then increases again when the behavioral response is required. We model this neuronal activity using an attractor decision-making network in which information reflecting the decision is maintained at a low level during the delay period, and is then selectively restored by a nonspecific input when the response is required. One mechanism for the short-term memory is synaptic facilitation, which can implement a mechanism for postponed decisions that can be correct even when there is little neuronal firing during the delay period before the postponed decision. Another mechanism is graded firing rates by different neurons in the delay period, with restoration by the nonspecific input of the low-rate activity from the higher-rate neurons still firing in the delay period. These mechanisms can account for the decision making and for the memory of the decision before a response can be made, which are evident in the activity of neurons in the medial premotor cortex.
Collapse
|
42
|
Lehky SR, Kiani R, Esteky H, Tanaka K. Statistics of visual responses in primate inferotemporal cortex to object stimuli. J Neurophysiol 2011; 106:1097-117. [PMID: 21562200 DOI: 10.1152/jn.00990.2010] [Citation(s) in RCA: 33] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
We have characterized selectivity and sparseness in anterior inferotemporal cortex, using a large data set. Responses were collected from 674 monkey inferotemporal cells, each stimulated by 806 object photographs. This 806 × 674 matrix was examined in two ways: columnwise, looking at responses of a single neuron to all images (single-neuron selectivity), and rowwise, looking at the responses of all neurons caused by a single image (population sparseness). Selectivity and sparseness were measured as kurtosis of probability distributions. Population sparseness exceeded single-neuron selectivity, with specific values dependent on the size of the data sample. This difference was principally caused by inclusion, within the population, of neurons with a variety of dynamic ranges (standard deviations of responses over all images). Statistics of large responses were examined by quantifying how quickly the upper tail of the probability distribution decreased (tail heaviness). This analysis demonstrated that population responses had heavier tails than single-neuron responses, consistent with the difference between sparseness and selectivity measurements. Population responses with spontaneous activity subtracted had the heaviest tails, following a power law. The very light tails of single-neuron responses indicate that the critical feature for each neuron is simple enough to have a high probability of occurring within a limited stimulus set. Heavy tails of population responses indicate that there are a large number of different critical features to which different neurons are tuned. These results are inconsistent with some structural models of object recognition that posit that objects are decomposed into a small number of standard features.
Collapse
Affiliation(s)
- Sidney R Lehky
- Cognitive Brain Mapping Laboratory, RIKEN Brain Science Inst., Hirosawa 2-1, Wako-shi, Saitama 351-0198, Japan.
| | | | | | | |
Collapse
|
43
|
Abstract
Many learned behaviors are thought to require the activity of high-level neurons that represent categories of complex signals, such as familiar faces or native speech sounds. How these complex, experience-dependent neural responses emerge within the brain's circuitry is not well understood. The caudomedial mesopallium (CMM), a secondary auditory region in the songbird brain, contains neurons that respond to specific combinations of song components and respond preferentially to the songs that birds have learned to recognize. Here, we examine the transformation of these learned responses across a broader forebrain circuit that includes the caudolateral mesopallium (CLM), an auditory region that provides input to CMM. We recorded extracellular single-unit activity in CLM and CMM in European starlings trained to recognize sets of conspecific songs and compared multiple encoding properties of neurons between these regions. We find that the responses of CMM neurons are more selective between song components, convey more information about song components, and are more variable over repeated components than the responses of CLM neurons. While learning enhances neural encoding of song components in both regions, CMM neurons encode more information about the learned categories associated with songs than do CLM neurons. Collectively, these data suggest that CLM and CMM are part of a functional sensory hierarchy that is modified by learning to yield representations of natural vocal signals that are increasingly informative with respect to behavior.
Collapse
|
44
|
Lehky SR, Sereno AB. Population coding of visual space: modeling. Front Comput Neurosci 2011; 4:155. [PMID: 21344012 PMCID: PMC3034232 DOI: 10.3389/fncom.2010.00155] [Citation(s) in RCA: 22] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2010] [Accepted: 12/09/2010] [Indexed: 11/13/2022] Open
Abstract
We examine how the representation of space is affected by receptive field (RF) characteristics of the encoding population. Spatial responses were defined by overlapping Gaussian RFs. These responses were analyzed using multidimensional scaling to extract the representation of global space implicit in population activity. Spatial representations were based purely on firing rates, which were not labeled with RF characteristics (tuning curve peak location, for example), differentiating this approach from many other population coding models. Because responses were unlabeled, this model represents space using intrinsic coding, extracting relative positions amongst stimuli, rather than extrinsic coding where known RF characteristics provide a reference frame for extracting absolute positions. Two parameters were particularly important: RF diameter and RF dispersion, where dispersion indicates how broadly RF centers are spread out from the fovea. For large RFs, the model was able to form metrically accurate representations of physical space on low-dimensional manifolds embedded within the high-dimensional neural population response space, suggesting that in some cases the neural representation of space may be dimensionally isomorphic with 3D physical space. Smaller RF sizes degraded and distorted the spatial representation, with the smallest RF sizes (present in early visual areas) being unable to recover even a topologically consistent rendition of space on low-dimensional manifolds. Finally, although positional invariance of stimulus responses has long been associated with large RFs in object recognition models, we found RF dispersion rather than RF diameter to be the critical parameter. In fact, at a population level, the modeling suggests that higher ventral stream areas with highly restricted RF dispersion would be unable to achieve positionally-invariant representations beyond this narrow region around fixation.
Collapse
Affiliation(s)
- Sidney R Lehky
- Computational Neuroscience Laboratory, Salk Institute for Biological Studies La Jolla, CA, USA
| | | |
Collapse
|
45
|
|
46
|
Arleo A, Nieus T, Bezzi M, D'Errico A, D'Angelo E, Coenen OJMD. How synaptic release probability shapes neuronal transmission: information-theoretic analysis in a cerebellar granule cell. Neural Comput 2010; 22:2031-58. [PMID: 20438336 DOI: 10.1162/neco_a_00006-arleo] [Citation(s) in RCA: 24] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
A nerve cell receives multiple inputs from upstream neurons by way of its synapses. Neuron processing functions are thus influenced by changes in the biophysical properties of the synapse, such as long-term potentiation (LTP) or depression (LTD). This observation has opened new perspectives on the biophysical basis of learning and memory, but its quantitative impact on the information transmission of a neuron remains partially elucidated. One major obstacle is the high dimensionality of the neuronal input-output space, which makes it unfeasible to perform a thorough computational analysis of a neuron with multiple synaptic inputs. In this work, information theory was employed to characterize the information transmission of a cerebellar granule cell over a region of its excitatory input space following synaptic changes. Granule cells have a small dendritic tree (on average, they receive only four mossy fiber afferents), which greatly bounds the input combinatorial space, reducing the complexity of information-theoretic calculations. Numerical simulations and LTP experiments quantified how changes in neurotransmitter release probability (p) modulated information transmission of a cerebellar granule cell. Numerical simulations showed that p shaped the neurotransmission landscape in unexpected ways. As p increased, the optimality of the information transmission of most stimuli did not increase strictly monotonically; instead it reached a plateau at intermediate p levels. Furthermore, our results showed that the spatiotemporal characteristics of the inputs determine the effect of p on neurotransmission, thus permitting the selection of distinctive preferred stimuli for different p values. These selective mechanisms may have important consequences on the encoding of cerebellar mossy fiber inputs and the plasticity and computation at the next circuit stage, including the parallel fiber-Purkinje cell synapses.
Collapse
Affiliation(s)
- Angelo Arleo
- CNRS, UPMC, UMR 7102 Neurobiology of Adaptive Processes, Paris, France.
| | | | | | | | | | | |
Collapse
|
47
|
Local non-linear interactions in the visual cortex may reflect global decorrelation. J Comput Neurosci 2010; 30:109-24. [PMID: 20422445 DOI: 10.1007/s10827-010-0239-2] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2009] [Revised: 03/12/2010] [Accepted: 04/13/2010] [Indexed: 10/19/2022]
Abstract
The classical receptive field in the primary visual cortex have been successfully explained by sparse activation of relatively independent units, whose tuning properties reflect the statistical dependencies in the natural environment. Robust surround modulation, emerging from stimulation beyond the classical receptive field, has been associated with increase of lifetime sparseness in the V1, but the system-wide modulation of response strength have currently no theoretical explanation. We measured fMRI responses from human visual cortex and quantified the contextual modulation with a decorrelation coefficient (d), derived from a subtractive normalization model. All active cortical areas demonstrated local non-linear summation of responses, which were in line with hypothesis of global decorrelation of voxels responses. In addition, we found sensitivity to surrounding stimulus structure across the ventral stream, and large-scale sensitivity to the number of simultaneous objects. Response sparseness across voxel population increased consistently with larger stimuli. These data suggest that contextual modulation for a stimulus event reflect optimization of the code and perhaps increase in energy efficiency throughout the ventral stream hierarchy. Our model provides a novel prediction that average suppression of response amplitude for simultaneous stimuli across the cortical network is a monotonic function of similarity of response strengths in the network when the stimuli are presented alone.
Collapse
|
48
|
A computational theory of episodic memory formation in the hippocampus. Behav Brain Res 2010; 215:180-96. [PMID: 20307583 DOI: 10.1016/j.bbr.2010.03.027] [Citation(s) in RCA: 168] [Impact Index Per Article: 11.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2010] [Revised: 03/10/2010] [Accepted: 03/13/2010] [Indexed: 11/22/2022]
Abstract
A quantitative computational theory of the operation of the hippocampus as an episodic memory system is described. The CA3 system operates as a single attractor or autoassociation network to enable rapid, one-trial associations between any spatial location (place in rodents or spatial view in primates) and an object or reward and to provide for completion of the whole memory during recall from any part. The theory is extended to associations between time and object or reward to implement temporal order memory, also important in episodic memory. The dentate gyrus performs pattern separation by competitive learning to produce sparse representations, producing for example neurons with place-like fields from entorhinal cortex grid cells. The dentate granule cells produce by the very small number of mossy fibre connections to CA3 a randomizing pattern separation effect important during learning but not recall that separates out the patterns represented by CA3 firing to be very different from each other, which is optimal for an unstructured episodic memory system in which each memory must be kept distinct from other memories. The direct perforant path input to CA3 is quantitatively appropriate to provide the cue for recall in CA3, but not for learning. The CA1 recodes information from CA3 to set up associatively learned backprojections to neocortex to allow subsequent retrieval of information to neocortex, providing a quantitative account of the large number of hippocampo-neocortical and neocortical-neocortical backprojections. Tests of the theory including hippocampal subregion analyses and hippocampal NMDA receptor knockouts are described and support the theory.
Collapse
|
49
|
Tompa T, Sáry G. A review on the inferior temporal cortex of the macaque. ACTA ACUST UNITED AC 2010; 62:165-82. [PMID: 19853626 DOI: 10.1016/j.brainresrev.2009.10.001] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2009] [Revised: 10/14/2009] [Accepted: 10/14/2009] [Indexed: 10/20/2022]
|
50
|
Rolls ET. Attractor networks. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2009; 1:119-134. [PMID: 26272845 DOI: 10.1002/wcs.1] [Citation(s) in RCA: 33] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
Abstract
An attractor network is a network of neurons with excitatory interconnections that can settle into a stable pattern of firing. This article shows how attractor networks in the cerebral cortex are important for long-term memory, short-term memory, attention, and decision making. The article then shows how the random firing of neurons can influence the stability of these networks by introducing stochastic noise, and how these effects are involved in probabilistic decision making, and implicated in some disorders of cortical function such as poor short-term memory and attention, schizophrenia, and obsessive-compulsive disorder. Copyright © 2009 John Wiley & Sons, Ltd. For further resources related to this article, please visit the WIREs website.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK
| |
Collapse
|