1
|
Homma NY, See JZ, Atencio CA, Hu C, Downer JD, Beitel RE, Cheung SW, Najafabadi MS, Olsen T, Bigelow J, Hasenstaub AR, Malone BJ, Schreiner CE. Receptive-field nonlinearities in primary auditory cortex: a comparative perspective. Cereb Cortex 2024; 34:bhae364. [PMID: 39270676 PMCID: PMC11398879 DOI: 10.1093/cercor/bhae364] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2023] [Revised: 08/14/2024] [Accepted: 08/21/2024] [Indexed: 09/15/2024] Open
Abstract
Cortical processing of auditory information can be affected by interspecies differences as well as brain states. Here we compare multifeature spectro-temporal receptive fields (STRFs) and associated input/output functions or nonlinearities (NLs) of neurons in primary auditory cortex (AC) of four mammalian species. Single-unit recordings were performed in awake animals (female squirrel monkeys, female, and male mice) and anesthetized animals (female squirrel monkeys, rats, and cats). Neuronal responses were modeled as consisting of two STRFs and their associated NLs. The NLs for the STRF with the highest information content show a broad distribution between linear and quadratic forms. In awake animals, we find a higher percentage of quadratic-like NLs as opposed to more linear NLs in anesthetized animals. Moderate sex differences of the shape of NLs were observed between male and female unanesthetized mice. This indicates that the core AC possesses a rich variety of potential computations, particularly in awake animals, suggesting that multiple computational algorithms are at play to enable the auditory system's robust recognition of auditory events.
Collapse
Affiliation(s)
- Natsumi Y Homma
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
- Department of Physiology, Development and Neuroscience, University of Cambridge, Downing Street, Cambridge, UK
| | - Jermyn Z See
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
| | - Craig A Atencio
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
| | - Congcong Hu
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
| | - Joshua D Downer
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
- Center of Neuroscience, University of California Davis, Newton Ct, Davis, CA, USA
| | - Ralph E Beitel
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
| | - Steven W Cheung
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
| | - Mina Sadeghi Najafabadi
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
| | - Timothy Olsen
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
| | - James Bigelow
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
| | - Andrea R Hasenstaub
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
| | - Brian J Malone
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
- Center of Neuroscience, University of California Davis, Newton Ct, Davis, CA, USA
| | - Christoph E Schreiner
- John & Edward Coleman Memorial Laboratory, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, CA, USA
| |
Collapse
|
2
|
Yedjour H, Yedjour D. A spatiotemporal energy model based on spiking neurons for human motion perception. Cogn Neurodyn 2024; 18:2015-2029. [PMID: 39104665 PMCID: PMC11297886 DOI: 10.1007/s11571-024-10068-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2023] [Revised: 11/30/2023] [Accepted: 01/09/2024] [Indexed: 08/07/2024] Open
Abstract
Inspired by the motion processing pathway, this paper proposes a bio-inspired feedforward spiking network model based on Hodgkin-Huxley neurons for human motion perception. The proposed network mimics the mechanisms of direction selectivity found in simple and complex cells of the primary visual cortex. Simple cells' receptive fields are modeled using Gabor energy filters, while complex cells' receptive fields are constructed by integrating the responses of simple cells in an energy model. To generate the motion map, the spiking output of the network integrates motion information encoded by the responses of complex cells with various preferred directions. Simulation results demonstrate that the spiking neuron-based network effectively replicates the directional selectivity operation of the visual cortex when presented with a sequence of time-varying images. We evaluate the proposed model against state-of-the-art spiking neuron-based motion detection models using publicly available datasets. The results highlight the model's capability to extract motion energy from diverse video sequences, akin to human visual motion perception models. Additionally, we showcase the application of the proposed model in motion segmentation tasks and compare its performance with state-of-the-art motion-based segmentation models using challenging video segmentation benchmarks. The results indicate competitive performance. The motion maps generated by the proposed model can be utilized for action recognition in input videos.
Collapse
Affiliation(s)
- Hayat Yedjour
- Faculty of Mathematics and Computer Science, Department of Computer Science, Université des Sciences et de la Technologie d’Oran Mohamed Boudiaf, USTO-MB, EL M’naouer, BP 1505, 31000 Oran, Algeria
| | - Dounia Yedjour
- Faculty of Mathematics and Computer Science, Department of Computer Science, Université des Sciences et de la Technologie d’Oran Mohamed Boudiaf, USTO-MB, EL M’naouer, BP 1505, 31000 Oran, Algeria
| |
Collapse
|
3
|
Quaia C, Krauzlis RJ. Object recognition in primates: what can early visual areas contribute? Front Behav Neurosci 2024; 18:1425496. [PMID: 39070778 PMCID: PMC11272660 DOI: 10.3389/fnbeh.2024.1425496] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2024] [Accepted: 07/01/2024] [Indexed: 07/30/2024] Open
Abstract
Introduction If neuroscientists were asked which brain area is responsible for object recognition in primates, most would probably answer infero-temporal (IT) cortex. While IT is likely responsible for fine discriminations, and it is accordingly dominated by foveal visual inputs, there is more to object recognition than fine discrimination. Importantly, foveation of an object of interest usually requires recognizing, with reasonable confidence, its presence in the periphery. Arguably, IT plays a secondary role in such peripheral recognition, and other visual areas might instead be more critical. Methods To investigate how signals carried by early visual processing areas (such as LGN and V1) could be used for object recognition in the periphery, we focused here on the task of distinguishing faces from non-faces. We tested how sensitive various models were to nuisance parameters, such as changes in scale and orientation of the image, and the type of image background. Results We found that a model of V1 simple or complex cells could provide quite reliable information, resulting in performance better than 80% in realistic scenarios. An LGN model performed considerably worse. Discussion Because peripheral recognition is both crucial to enable fine recognition (by bringing an object of interest on the fovea), and probably sufficient to account for a considerable fraction of our daily recognition-guided behavior, we think that the current focus on area IT and foveal processing is too narrow. We propose that rather than a hierarchical system with IT-like properties as its primary aim, object recognition should be seen as a parallel process, with high-accuracy foveal modules operating in parallel with lower-accuracy and faster modules that can operate across the visual field.
Collapse
Affiliation(s)
- Christian Quaia
- Laboratory of Sensorimotor Research, National Eye Institute, NIH, Bethesda, MD, United States
| | | |
Collapse
|
4
|
Almasi A, Sun SH, Jung YJ, Ibbotson M, Meffin H. Data-driven modelling of visual receptive fields: comparison between the generalized quadratic model and the nonlinear input model. J Neural Eng 2024; 21:046014. [PMID: 38941988 DOI: 10.1088/1741-2552/ad5d15] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2023] [Accepted: 06/28/2024] [Indexed: 06/30/2024]
Abstract
Objective: Neurons in primary visual cortex (V1) display a range of sensitivity in their response to translations of their preferred visual features within their receptive field: from high specificity to a precise position through to complete invariance. This visual feature selectivity and invariance is frequently modeled by applying a selection of linear spatial filters to the input image, that define the feature selectivity, followed by a nonlinear function that combines the filter outputs, that defines the invariance, to predict the neural response. We compare two such classes of model, that are both popular and parsimonious, the generalized quadratic model (GQM) and the nonlinear input model (NIM). These two classes of model differ primarily in that the NIM can accommodate a greater diversity in the form of nonlinearity that is applied to the outputs of the filters.Approach: We compare the two model types by applying them to data from multielectrode recordings from cat primary visual cortex in response to spatially white Gaussian noise After fitting both classes of model to a database of 342 single units (SUs), we analyze the qualitative and quantitative differences in the visual feature processing performed by the two models and their ability to predict neural response.Main results: We find that the NIM predicts response rates on a held-out data at least as well as the GQM for 95% of SUs. Superior performance occurs predominantly for those units with above average spike rates and is largely due to the NIMs ability to capture aspects of the model's nonlinear function cannot be captured with the GQM rather than differences in the visual features being processed by the two different models.Significance: These results can help guide model choice for data-driven receptive field modelling.
Collapse
Affiliation(s)
- Ali Almasi
- National Vision Research Institute, Carlton, VIC 3053, Australia
| | - Shi H Sun
- National Vision Research Institute, Carlton, VIC 3053, Australia
| | - Young Jun Jung
- National Vision Research Institute, Carlton, VIC 3053, Australia
| | - Michael Ibbotson
- National Vision Research Institute, Carlton, VIC 3053, Australia
- Department of Optometry and Vision Sciences, The University of Melbourne, Parkville, VIC 3010, Australia
| | - Hamish Meffin
- National Vision Research Institute, Carlton, VIC 3053, Australia
- Department of Biomedical Engineering, The University of Melbourne, Parkville, VIC 3010, Australia
| |
Collapse
|
5
|
Hassanpour MS, Merlin S, Federer F, Zaidi Q, Angelucci A. Primate V2 Receptive Fields Derived from Anatomically Identified Large-Scale V1 Inputs. RESEARCH SQUARE 2024:rs.3.rs-4139501. [PMID: 38798339 PMCID: PMC11118708 DOI: 10.21203/rs.3.rs-4139501/v1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2024]
Abstract
In the primate visual system, visual object recognition involves a series of cortical areas arranged hierarchically along the ventral visual pathway. As information flows through this hierarchy, neurons become progressively tuned to more complex image features. The circuit mechanisms and computations underlying the increasing complexity of these receptive fields (RFs) remain unidentified. To understand how this complexity emerges in the secondary visual area (V2), we investigated the functional organization of inputs from the primary visual cortex (V1) to V2 by combining retrograde anatomical tracing of these inputs with functional imaging of feature maps in macaque monkey V1 and V2. We found that V1 neurons sending inputs to single V2 orientation columns have a broad range of preferred orientations, but are strongly biased towards the orientation represented at the injected V2 site. For each V2 site, we then constructed a feedforward model based on the linear combination of its anatomically-identified large-scale V1 inputs, and studied the response proprieties of the generated V2 RFs. We found that V2 RFs derived from the linear feedforward model were either elongated versions of V1 filters or had spatially complex structures. These modeled RFs predicted V2 neuron responses to oriented grating stimuli with high accuracy. Remarkably, this simple model also explained the greater selectivity to naturalistic textures of V2 cells compared to their V1 input cells. Our results demonstrate that simple linear combinations of feedforward inputs can account for the orientation selectivity and texture sensitivity of V2 RFs.
Collapse
Affiliation(s)
- Mahlega S Hassanpour
- Dept. of Ophthalmology and Visual Science, Moran Eye Institute, University of Utah
| | - Sam Merlin
- Dept. of Ophthalmology and Visual Science, Moran Eye Institute, University of Utah
- Present address: Dept of Medical Science, School of Science, Western Sydney University
| | - Frederick Federer
- Dept. of Ophthalmology and Visual Science, Moran Eye Institute, University of Utah
| | - Qasim Zaidi
- Graduate Center for Vision Research, State University of New York, College of Optometry
| | - Alessandra Angelucci
- Dept. of Ophthalmology and Visual Science, Moran Eye Institute, University of Utah
| |
Collapse
|
6
|
Hassanpour MS, Merlin S, Federer F, Zaidi Q, Angelucci A. Primate V2 Receptive Fields Derived from Anatomically Identified Large-Scale V1 Inputs. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.03.22.586002. [PMID: 38585792 PMCID: PMC10996519 DOI: 10.1101/2024.03.22.586002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/09/2024]
Abstract
In the primate visual system, visual object recognition involves a series of cortical areas arranged hierarchically along the ventral visual pathway. As information flows through this hierarchy, neurons become progressively tuned to more complex image features. The circuit mechanisms and computations underlying the increasing complexity of these receptive fields (RFs) remain unidentified. To understand how this complexity emerges in the secondary visual area (V2), we investigated the functional organization of inputs from the primary visual cortex (V1) to V2 by combining retrograde anatomical tracing of these inputs with functional imaging of feature maps in macaque monkey V1 and V2. We found that V1 neurons sending inputs to single V2 orientation columns have a broad range of preferred orientations, but are strongly biased towards the orientation represented at the injected V2 site. For each V2 site, we then constructed a feedforward model based on the linear combination of its anatomically-identified large-scale V1 inputs, and studied the response proprieties of the generated V2 RFs. We found that V2 RFs derived from the linear feedforward model were either elongated versions of V1 filters or had spatially complex structures. These modeled RFs predicted V2 neuron responses to oriented grating stimuli with high accuracy. Remarkably, this simple model also explained the greater selectivity to naturalistic textures of V2 cells compared to their V1 input cells. Our results demonstrate that simple linear combinations of feedforward inputs can account for the orientation selectivity and texture sensitivity of V2 RFs.
Collapse
|
7
|
Lian Y, Williams S, Alexander AS, Hasselmo ME, Burkitt AN. Learning the Vector Coding of Egocentric Boundary Cells from Visual Data. J Neurosci 2023; 43:5180-5190. [PMID: 37286350 PMCID: PMC10342228 DOI: 10.1523/jneurosci.1071-22.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/02/2022] [Revised: 05/25/2023] [Accepted: 05/27/2023] [Indexed: 06/09/2023] Open
Abstract
The use of spatial maps to navigate through the world requires a complex ongoing transformation of egocentric views of the environment into position within the allocentric map. Recent research has discovered neurons in retrosplenial cortex and other structures that could mediate the transformation from egocentric views to allocentric views. These egocentric boundary cells respond to the egocentric direction and distance of barriers relative to an animal's point of view. This egocentric coding based on the visual features of barriers would seem to require complex dynamics of cortical interactions. However, computational models presented here show that egocentric boundary cells can be generated with a remarkably simple synaptic learning rule that forms a sparse representation of visual input as an animal explores the environment. Simulation of this simple sparse synaptic modification generates a population of egocentric boundary cells with distributions of direction and distance coding that strikingly resemble those observed within the retrosplenial cortex. Furthermore, some egocentric boundary cells learnt by the model can still function in new environments without retraining. This provides a framework for understanding the properties of neuronal populations in the retrosplenial cortex that may be essential for interfacing egocentric sensory information with allocentric spatial maps of the world formed by neurons in downstream areas, including the grid cells in entorhinal cortex and place cells in the hippocampus.SIGNIFICANCE STATEMENT The computational model presented here demonstrates that the recently discovered egocentric boundary cells in retrosplenial cortex can be generated with a remarkably simple synaptic learning rule that forms a sparse representation of visual input as an animal explores the environment. Additionally, our model generates a population of egocentric boundary cells with distributions of direction and distance coding that strikingly resemble those observed within the retrosplenial cortex. This transformation between sensory input and egocentric representation in the navigational system could have implications for the way in which egocentric and allocentric representations interface in other brain areas.
Collapse
Affiliation(s)
- Yanbo Lian
- Department of Biomedical Engineering, University of Melbourne, Melbourne, Victoria 3010, Australia
| | - Simon Williams
- Department of Electrical and Electronic Engineering, University of Melbourne, Melbourne, Victoria 3010, Australia
| | - Andrew S Alexander
- Center for Systems Neuroscience, Department of Psychological and Brain Sciences, Boston University, Boston, Massachusetts 02215
| | - Michael E Hasselmo
- Center for Systems Neuroscience, Department of Psychological and Brain Sciences, Boston University, Boston, Massachusetts 02215
| | - Anthony N Burkitt
- Department of Biomedical Engineering, University of Melbourne, Melbourne, Victoria 3010, Australia
| |
Collapse
|
8
|
Artificial Visual System for Orientation Detection Based on Hubel–Wiesel Model. Brain Sci 2022; 12:brainsci12040470. [PMID: 35448001 PMCID: PMC9025109 DOI: 10.3390/brainsci12040470] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2022] [Revised: 03/27/2022] [Accepted: 03/30/2022] [Indexed: 01/18/2023] Open
Abstract
The Hubel–Wiesel (HW) model is a classical neurobiological model for explaining the orientation selectivity of cortical cells. However, the HW model still has not been fully proved physiologically, and there are few concise but efficient systems to quantify and simulate the HW model and can be used for object orientation detection applications. To realize a straightforward and efficient quantitive method and validate the HW model’s reasonability and practicality, we use McCulloch-Pitts (MP) neuron model to simulate simple cells and complex cells and implement an artificial visual system (AVS) for two-dimensional object orientation detection. First, we realize four types of simple cells that are only responsible for detecting a specific orientation angle locally. Complex cells are realized with the sum function. Every local orientation information of an object is collected by simple cells and subsequently converged to the corresponding same type complex cells for computing global activation degree. Finally, the global orientation is obtained according to the activation degree of each type of complex cell. Based on this scheme, an AVS for global orientation detection is constructed. We conducted computer simulations to prove the feasibility and effectiveness of our scheme and the AVS. Computer simulations show that the mechanism-based AVS can make accurate orientation discrimination and shows striking biological similarities with the natural visual system, which indirectly proves the rationality of the Hubel–Wiesel model. Furthermore, compared with traditional CNN, we find that our AVS beats CNN on orientation detection tasks in identification accuracy, noise resistance, computation and learning cost, hardware implementation, and reasonability.
Collapse
|