1
|
Laurin AS, Ouerfelli-Ethier J, Pisella L, Khan AZ. Reduced spatial attentional distribution in older adults. J Vis 2024; 24:8. [PMID: 38591941 PMCID: PMC11008755 DOI: 10.1167/jov.24.4.8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2023] [Accepted: 02/06/2024] [Indexed: 04/10/2024] Open
Abstract
Older adults show decline in visual search performance, but the underlying cause remains unclear. It has been suggested that older adults' altered performance may be related to reduced spatial attention to peripheral visual information compared with younger adults. In this study, 18 younger (M = 21.6 years) and 16 older (M = 69.1 years) participants performed pop-out and serial visual search tasks with variously sized gaze-contingent artificial central scotomas (3°, 5°, or 7° diameter). By occluding central vision, we measured how attention to the periphery was contributing to the search performance. We also tested the effect of target eccentricity on search times and eye movements. We hypothesized that, if attention is reduced primarily in the periphery in older adults, we would observe longer search times for more eccentric targets and with central occlusion. During the pop-out search, older adults showed a steeper decline in search performance with increasing eccentricity and central scotoma size compared with younger adults. In contrast, during the serial search, older adults had longer search times than younger adults overall, independent of target eccentricity and scotoma size. Longer search times were attributed to higher cost-per-item slopes, indicating increased difficulty in simultaneously processing complex symbols made up of separable features in aging, possibly stemming from challenges in spatially binding individual features. Altogether, our findings point to fewer attentional resources of simultaneous visual processing to distribute over space or separable features of objects, consistent with decreased dorsal visual stream functioning in aging.
Collapse
Affiliation(s)
| | - Julie Ouerfelli-Ethier
- School of Optometry, University of Montreal, Montreal, QC, Canada
- Université Claude Bernard Lyon 1, Centre de Recherche en Neurosciences de Lyon (CRNL), INSERM U1028, Bron, France
| | - Laure Pisella
- Université Claude Bernard Lyon 1, Centre de Recherche en Neurosciences de Lyon (CRNL), INSERM U1028, Bron, France
| | | |
Collapse
|
2
|
Applegate MC, Gutnichenko KS, Aronov D. Topography of inputs into the hippocampal formation of a food-caching bird. J Comp Neurol 2023; 531:1669-1688. [PMID: 37553864 PMCID: PMC10611445 DOI: 10.1002/cne.25533] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2023] [Revised: 07/12/2023] [Accepted: 07/24/2023] [Indexed: 08/10/2023]
Abstract
The mammalian hippocampal formation (HF) is organized into domains associated with different functions. These differences are driven in part by the pattern of input along the hippocampal long axis, such as visual input to the septal hippocampus and amygdalar input to the temporal hippocampus. HF is also organized along the transverse axis, with different patterns of neural activity in the hippocampus and the entorhinal cortex. In some birds, a similar organization has been observed along both of these axes. However, it is not known what role inputs play in this organization. We used retrograde tracing to map inputs into HF of a food-caching bird, the black-capped chickadee. We first compared two locations along the transverse axis: the hippocampus and the dorsolateral hippocampal area (DL), which is analogous to the entorhinal cortex. We found that pallial regions predominantly targeted DL, while some subcortical regions like the lateral hypothalamus (LHy) preferentially targeted the hippocampus. We then examined the hippocampal long axis and found that almost all inputs were topographic along this direction. For example, the anterior hippocampus was preferentially innervated by thalamic regions, while the posterior hippocampus received more amygdalar input. Some of the topographies we found bear a resemblance to those described in the mammalian brain, revealing a remarkable anatomical similarity of phylogenetically distant animals. More generally, our work establishes the pattern of inputs to HF in chickadees. Some of these patterns may be unique to chickadees, laying the groundwork for studying the anatomical basis of these birds' exceptional hippocampal memory.
Collapse
Affiliation(s)
| | | | - Dmitriy Aronov
- Zuckerman Mind Brain Behavior Institute Columbia University
| |
Collapse
|
3
|
Segraves MA. Using Natural Scenes to Enhance our Understanding of the Cerebral Cortex's Role in Visual Search. Annu Rev Vis Sci 2023; 9:435-454. [PMID: 37164028 DOI: 10.1146/annurev-vision-100720-124033] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/12/2023]
Abstract
Using natural scenes is an approach to studying the visual and eye movement systems approximating how these systems function in everyday life. This review examines the results from behavioral and neurophysiological studies using natural scene viewing in humans and monkeys. The use of natural scenes for the study of cerebral cortical activity is relatively new and presents challenges for data analysis. Methods and results from the use of natural scenes for the study of the visual and eye movement cortex are presented, with emphasis on new insights that this method provides enhancing what is known about these cortical regions from the use of conventional methods.
Collapse
Affiliation(s)
- Mark A Segraves
- Department of Neurobiology, Northwestern University, Evanston, Illinois, USA;
| |
Collapse
|
4
|
Rolls ET. Emotion, motivation, decision-making, the orbitofrontal cortex, anterior cingulate cortex, and the amygdala. Brain Struct Funct 2023; 228:1201-1257. [PMID: 37178232 PMCID: PMC10250292 DOI: 10.1007/s00429-023-02644-9] [Citation(s) in RCA: 57] [Impact Index Per Article: 28.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2023] [Accepted: 04/12/2023] [Indexed: 05/15/2023]
Abstract
The orbitofrontal cortex and amygdala are involved in emotion and in motivation, but the relationship between these functions performed by these brain structures is not clear. To address this, a unified theory of emotion and motivation is described in which motivational states are states in which instrumental goal-directed actions are performed to obtain rewards or avoid punishers, and emotional states are states that are elicited when the reward or punisher is or is not received. This greatly simplifies our understanding of emotion and motivation, for the same set of genes and associated brain systems can define the primary or unlearned rewards and punishers such as sweet taste or pain. Recent evidence on the connectivity of human brain systems involved in emotion and motivation indicates that the orbitofrontal cortex is involved in reward value and experienced emotion with outputs to cortical regions including those involved in language, and is a key brain region involved in depression and the associated changes in motivation. The amygdala has weak effective connectivity back to the cortex in humans, and is implicated in brainstem-mediated responses to stimuli such as freezing and autonomic activity, rather than in declarative emotion. The anterior cingulate cortex is involved in learning actions to obtain rewards, and with the orbitofrontal cortex and ventromedial prefrontal cortex in providing the goals for navigation and in reward-related effects on memory consolidation mediated partly via the cholinergic system.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK.
- Department of Computer Science, University of Warwick, Coventry, UK.
| |
Collapse
|
5
|
Rolls ET. Hippocampal spatial view cells for memory and navigation, and their underlying connectivity in humans. Hippocampus 2023; 33:533-572. [PMID: 36070199 PMCID: PMC10946493 DOI: 10.1002/hipo.23467] [Citation(s) in RCA: 33] [Impact Index Per Article: 16.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2022] [Revised: 08/16/2022] [Accepted: 08/16/2022] [Indexed: 01/08/2023]
Abstract
Hippocampal and parahippocampal gyrus spatial view neurons in primates respond to the spatial location being looked at. The representation is allocentric, in that the responses are to locations "out there" in the world, and are relatively invariant with respect to retinal position, eye position, head direction, and the place where the individual is located. The underlying connectivity in humans is from ventromedial visual cortical regions to the parahippocampal scene area, leading to the theory that spatial view cells are formed by combinations of overlapping feature inputs self-organized based on their closeness in space. Thus, although spatial view cells represent "where" for episodic memory and navigation, they are formed by ventral visual stream feature inputs in the parahippocampal gyrus in what is the parahippocampal scene area. A second "where" driver of spatial view cells are parietal inputs, which it is proposed provide the idiothetic update for spatial view cells, used for memory recall and navigation when the spatial view details are obscured. Inferior temporal object "what" inputs and orbitofrontal cortex reward inputs connect to the human hippocampal system, and in macaques can be associated in the hippocampus with spatial view cell "where" representations to implement episodic memory. Hippocampal spatial view cells also provide a basis for navigation to a series of viewed landmarks, with the orbitofrontal cortex reward inputs to the hippocampus providing the goals for navigation, which can then be implemented by hippocampal connectivity in humans to parietal cortex regions involved in visuomotor actions in space. The presence of foveate vision and the highly developed temporal lobe for object and scene processing in primates including humans provide a basis for hippocampal spatial view cells to be key to understanding episodic memory in the primate and human hippocampus, and the roles of this system in primate including human navigation.
Collapse
Affiliation(s)
- Edmund T. Rolls
- Oxford Centre for Computational NeuroscienceOxfordUK
- Department of Computer ScienceUniversity of WarwickCoventryUK
| |
Collapse
|
6
|
Applegate MC, Gutnichenko KS, Aronov D. Topography of inputs into the hippocampal formation of a food-caching bird. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.03.14.532572. [PMID: 36993579 PMCID: PMC10054989 DOI: 10.1101/2023.03.14.532572] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
The mammalian hippocampal formation (HF) is organized into domains associated with different functions. These differences are driven in part by the pattern of input along the hippocampal long axis, such as visual input to the septal hippocampus and amygdalar input to temporal hippocampus. HF is also organized along the transverse axis, with different patterns of neural activity in the hippocampus and the entorhinal cortex. In some birds, a similar organization has been observed along both of these axes. However, it is not known what role inputs play in this organization. We used retrograde tracing to map inputs into HF of a food-caching bird, the black-capped chickadee. We first compared two locations along the transverse axis: the hippocampus and the dorsolateral hippocampal area (DL), which is analogous to the entorhinal cortex. We found that pallial regions predominantly targeted DL, while some subcortical regions like the lateral hypothalamus (LHy) preferentially targeted the hippocampus. We then examined the hippocampal long axis and found that almost all inputs were topographic along this direction. For example, the anterior hippocampus was preferentially innervated by thalamic regions, while posterior hippocampus received more amygdalar input. Some of the topographies we found bear resemblance to those described in the mammalian brain, revealing a remarkable anatomical similarity of phylogenetically distant animals. More generally, our work establishes the pattern of inputs to HF in chickadees. Some of these patterns may be unique to chickadees, laying the groundwork for studying the anatomical basis of these birds ’ exceptional hippocampal memory.
Collapse
Affiliation(s)
| | | | - Dmitriy Aronov
- Zuckerman Mind Brain Behavior Institute, Columbia University
| |
Collapse
|
7
|
Rolls ET. Learning Invariant Object and Spatial View Representations in the Brain Using Slow Unsupervised Learning. Front Comput Neurosci 2021; 15:686239. [PMID: 34366818 PMCID: PMC8335547 DOI: 10.3389/fncom.2021.686239] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2021] [Accepted: 06/29/2021] [Indexed: 11/13/2022] Open
Abstract
First, neurophysiological evidence for the learning of invariant representations in the inferior temporal visual cortex is described. This includes object and face representations with invariance for position, size, lighting, view and morphological transforms in the temporal lobe visual cortex; global object motion in the cortex in the superior temporal sulcus; and spatial view representations in the hippocampus that are invariant with respect to eye position, head direction, and place. Second, computational mechanisms that enable the brain to learn these invariant representations are proposed. For the ventral visual system, one key adaptation is the use of information available in the statistics of the environment in slow unsupervised learning to learn transform-invariant representations of objects. This contrasts with deep supervised learning in artificial neural networks, which uses training with thousands of exemplars forced into different categories by neuronal teachers. Similar slow learning principles apply to the learning of global object motion in the dorsal visual system leading to the cortex in the superior temporal sulcus. The learning rule that has been explored in VisNet is an associative rule with a short-term memory trace. The feed-forward architecture has four stages, with convergence from stage to stage. This type of slow learning is implemented in the brain in hierarchically organized competitive neuronal networks with convergence from stage to stage, with only 4-5 stages in the hierarchy. Slow learning is also shown to help the learning of coordinate transforms using gain modulation in the dorsal visual system extending into the parietal cortex and retrosplenial cortex. Representations are learned that are in allocentric spatial view coordinates of locations in the world and that are independent of eye position, head direction, and the place where the individual is located. This enables hippocampal spatial view cells to use idiothetic, self-motion, signals for navigation when the view details are obscured for short periods.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, United Kingdom.,Department of Computer Science, University of Warwick, Coventry, United Kingdom
| |
Collapse
|
8
|
Rolls ET. Spatial coordinate transforms linking the allocentric hippocampal and egocentric parietal primate brain systems for memory, action in space, and navigation. Hippocampus 2019; 30:332-353. [PMID: 31697002 DOI: 10.1002/hipo.23171] [Citation(s) in RCA: 24] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2019] [Revised: 10/05/2019] [Accepted: 10/09/2019] [Indexed: 01/03/2023]
Abstract
A theory and model of spatial coordinate transforms in the dorsal visual system through the parietal cortex that enable an interface via posterior cingulate and related retrosplenial cortex to allocentric spatial representations in the primate hippocampus is described. First, a new approach to coordinate transform learning in the brain is proposed, in which the traditional gain modulation is complemented by temporal trace rule competitive network learning. It is shown in a computational model that the new approach works much more precisely than gain modulation alone, by enabling neurons to represent the different combinations of signal and gain modulator more accurately. This understanding may have application to many brain areas where coordinate transforms are learned. Second, a set of coordinate transforms is proposed for the dorsal visual system/parietal areas that enables a representation to be formed in allocentric spatial view coordinates. The input stimulus is merely a stimulus at a given position in retinal space, and the gain modulation signals needed are eye position, head direction, and place, all of which are present in the primate brain. Neurons that encode the bearing to a landmark are involved in the coordinate transforms. Part of the importance here is that the coordinates of the allocentric view produced in this model are the same as those of spatial view cells that respond to allocentric view recorded in the primate hippocampus and parahippocampal cortex. The result is that information from the dorsal visual system can be used to update the spatial input to the hippocampus in the appropriate allocentric coordinate frame, including providing for idiothetic update to allow for self-motion. It is further shown how hippocampal spatial view cells could be useful for the transform from hippocampal allocentric coordinates to egocentric coordinates useful for actions in space and for navigation.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK.,Department of Computer Science, University of Warwick, Coventry, UK
| |
Collapse
|
9
|
Yu L, Jin M, Zhou K. Multi-channel biomimetic visual transformation for object feature extraction and recognition of complex scenes. APPL INTELL 2019. [DOI: 10.1007/s10489-019-01550-0] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
10
|
Rolls ET, Wirth S. Spatial representations in the primate hippocampus, and their functions in memory and navigation. Prog Neurobiol 2018; 171:90-113. [DOI: 10.1016/j.pneurobio.2018.09.004] [Citation(s) in RCA: 59] [Impact Index Per Article: 8.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2018] [Revised: 09/10/2018] [Accepted: 09/10/2018] [Indexed: 01/01/2023]
|
11
|
Rolls ET, Mills WPC. Non-accidental properties, metric invariance, and encoding by neurons in a model of ventral stream visual object recognition, VisNet. Neurobiol Learn Mem 2018; 152:20-31. [PMID: 29723671 DOI: 10.1016/j.nlm.2018.04.017] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2018] [Revised: 04/02/2018] [Accepted: 04/27/2018] [Indexed: 11/18/2022]
Abstract
When objects transform into different views, some properties are maintained, such as whether the edges are convex or concave, and these non-accidental properties are likely to be important in view-invariant object recognition. The metric properties, such as the degree of curvature, may change with different views, and are less likely to be useful in object recognition. It is shown that in a model of invariant visual object recognition in the ventral visual stream, VisNet, non-accidental properties are encoded much more than metric properties by neurons. Moreover, it is shown how with the temporal trace rule training in VisNet, non-accidental properties of objects become encoded by neurons, and how metric properties are treated invariantly. We also show how VisNet can generalize between different objects if they have the same non-accidental property, because the metric properties are likely to overlap. VisNet is a 4-layer unsupervised model of visual object recognition trained by competitive learning that utilizes a temporal trace learning rule to implement the learning of invariance using views that occur close together in time. A second crucial property of this model of object recognition is, when neurons in the level corresponding to the inferior temporal visual cortex respond selectively to objects, whether neurons in the intermediate layers can respond to combinations of features that may be parts of two or more objects. In an investigation using the four sides of a square presented in every possible combination, it was shown that even though different layer 4 neurons are tuned to encode each feature or feature combination orthogonally, neurons in the intermediate layers can respond to features or feature combinations present is several objects. This property is an important part of the way in which high capacity can be achieved in the four-layer ventral visual cortical pathway. These findings concerning non-accidental properties and the use of neurons in intermediate layers of the hierarchy help to emphasise fundamental underlying principles of the computations that may be implemented in the ventral cortical visual stream used in object recognition.
Collapse
Affiliation(s)
- Edmund T Rolls
- Oxford Centre for Computational Neuroscience, Oxford, UK.
| | - W Patrick C Mills
- University of Warwick, Department of Computer Science, Coventry, UK. http://www.oxcns.org
| |
Collapse
|
12
|
Galeazzi JM, Navajas J, Mender BMW, Quian Quiroga R, Minini L, Stringer SM. The visual development of hand-centered receptive fields in a neural network model of the primate visual system trained with experimentally recorded human gaze changes. NETWORK (BRISTOL, ENGLAND) 2016; 27:29-51. [PMID: 27253452 PMCID: PMC4926791 DOI: 10.1080/0954898x.2016.1187311] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/07/2016] [Revised: 05/04/2016] [Accepted: 05/04/2016] [Indexed: 06/05/2023]
Abstract
Neurons have been found in the primate brain that respond to objects in specific locations in hand-centered coordinates. A key theoretical challenge is to explain how such hand-centered neuronal responses may develop through visual experience. In this paper we show how hand-centered visual receptive fields can develop using an artificial neural network model, VisNet, of the primate visual system when driven by gaze changes recorded from human test subjects as they completed a jigsaw. A camera mounted on the head captured images of the hand and jigsaw, while eye movements were recorded using an eye-tracking device. This combination of data allowed us to reconstruct the retinal images seen as humans undertook the jigsaw task. These retinal images were then fed into the neural network model during self-organization of its synaptic connectivity using a biologically plausible trace learning rule. A trace learning mechanism encourages neurons in the model to learn to respond to input images that tend to occur in close temporal proximity. In the data recorded from human subjects, we found that the participant's gaze often shifted through a sequence of locations around a fixed spatial configuration of the hand and one of the jigsaw pieces. In this case, trace learning should bind these retinal images together onto the same subset of output neurons. The simulation results consequently confirmed that some cells learned to respond selectively to the hand and a jigsaw piece in a fixed spatial configuration across different retinal views.
Collapse
Affiliation(s)
- Juan M. Galeazzi
- Oxford Centre for Theoretical Neuroscience and Artificial Intelligence, Department of Experimental Psychology, University of Oxford, Oxford, UK
| | - Joaquín Navajas
- Institute of Cognitive Neuroscience, University College London, London, UK
- Centre for Systems Neuroscience, University of Leicester, Leicester, UK
| | - Bedeho M. W. Mender
- Oxford Centre for Theoretical Neuroscience and Artificial Intelligence, Department of Experimental Psychology, University of Oxford, Oxford, UK
| | | | - Loredana Minini
- Oxford Centre for Theoretical Neuroscience and Artificial Intelligence, Department of Experimental Psychology, University of Oxford, Oxford, UK
| | - Simon M. Stringer
- Oxford Centre for Theoretical Neuroscience and Artificial Intelligence, Department of Experimental Psychology, University of Oxford, Oxford, UK
| |
Collapse
|
13
|
Stacho M, Ströckens F, Xiao Q, Güntürkün O. Functional organization of telencephalic visual association fields in pigeons. Behav Brain Res 2016; 303:93-102. [DOI: 10.1016/j.bbr.2016.01.045] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2015] [Revised: 01/15/2016] [Accepted: 01/17/2016] [Indexed: 12/24/2022]
|
14
|
Galeazzi JM, Minini L, Stringer SM. The Development of Hand-Centered Visual Representations in the Primate Brain: A Computer Modeling Study Using Natural Visual Scenes. Front Comput Neurosci 2015; 9:147. [PMID: 26696876 PMCID: PMC4678233 DOI: 10.3389/fncom.2015.00147] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2015] [Accepted: 11/23/2015] [Indexed: 11/23/2022] Open
Abstract
Neurons that respond to visual targets in a hand-centered frame of reference have been found within various areas of the primate brain. We investigate how hand-centered visual representations may develop in a neural network model of the primate visual system called VisNet, when the model is trained on images of the hand seen against natural visual scenes. The simulations show how such neurons may develop through a biologically plausible process of unsupervised competitive learning and self-organization. In an advance on our previous work, the visual scenes consisted of multiple targets presented simultaneously with respect to the hand. Three experiments are presented. First, VisNet was trained with computerized images consisting of a realistic image of a hand and a variety of natural objects, presented in different textured backgrounds during training. The network was then tested with just one textured object near the hand in order to verify if the output cells were capable of building hand-centered representations with a single localized receptive field. We explain the underlying principles of the statistical decoupling that allows the output cells of the network to develop single localized receptive fields even when the network is trained with multiple objects. In a second simulation we examined how some of the cells with hand-centered receptive fields decreased their shape selectivity and started responding to a localized region of hand-centered space as the number of objects presented in overlapping locations during training increases. Lastly, we explored the same learning principles training the network with natural visual scenes collected by volunteers. These results provide an important step in showing how single, localized, hand-centered receptive fields could emerge under more ecologically realistic visual training conditions.
Collapse
Affiliation(s)
- Juan M. Galeazzi
- Department of Experimental Psychology, Oxford Centre for Theoretical Neuroscience and Artificial Intelligence, University of OxfordOxford, UK
| | | | | |
Collapse
|
15
|
Robinson L, Rolls ET. Invariant visual object recognition: biologically plausible approaches. BIOLOGICAL CYBERNETICS 2015; 109:505-35. [PMID: 26335743 PMCID: PMC4572081 DOI: 10.1007/s00422-015-0658-2] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/03/2015] [Accepted: 08/14/2015] [Indexed: 06/01/2023]
Abstract
Key properties of inferior temporal cortex neurons are described, and then, the biological plausibility of two leading approaches to invariant visual object recognition in the ventral visual system is assessed to investigate whether they account for these properties. Experiment 1 shows that VisNet performs object classification with random exemplars comparably to HMAX, except that the final layer C neurons of HMAX have a very non-sparse representation (unlike that in the brain) that provides little information in the single-neuron responses about the object class. Experiment 2 shows that VisNet forms invariant representations when trained with different views of each object, whereas HMAX performs poorly when assessed with a biologically plausible pattern association network, as HMAX has no mechanism to learn view invariance. Experiment 3 shows that VisNet neurons do not respond to scrambled images of faces, and thus encode shape information. HMAX neurons responded with similarly high rates to the unscrambled and scrambled faces, indicating that low-level features including texture may be relevant to HMAX performance. Experiment 4 shows that VisNet can learn to recognize objects even when the view provided by the object changes catastrophically as it transforms, whereas HMAX has no learning mechanism in its S-C hierarchy that provides for view-invariant learning. This highlights some requirements for the neurobiological mechanisms of high-level vision, and how some different approaches perform, in order to help understand the fundamental underlying principles of invariant visual object recognition in the ventral visual stream.
Collapse
Affiliation(s)
- Leigh Robinson
- Department of Computer Science, University of Warwick, Coventry, UK
| | - Edmund T Rolls
- Department of Computer Science, University of Warwick, Coventry, UK.
- Oxford Centre for Computational Neuroscience, Oxford, UK.
| |
Collapse
|