1
|
Dyballa L, Field GD, Stryker MP, Zucker SW. Encoding manifolds constructed from grating responses organize responses to natural scenes in cortical visual areas. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.10.24.620089. [PMID: 39484529 PMCID: PMC11527117 DOI: 10.1101/2024.10.24.620089] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Indexed: 11/03/2024]
Abstract
We have created "encoding manifolds" to reveal the overall responses of a brain area to a variety of stimuli. Encoding manifolds organize response properties globally: each point on an encoding manifold is a neuron, and nearby neurons respond similarly to the stimulus ensemble in time. We previously found, using a large stimulus ensemble including optic flows, that encoding manifolds for the retina were highly clustered, with each cluster corresponding to a different ganglion cell type. In contrast, the topology of the V1 manifold was continuous. Now, using responses of individual neurons from the Allen Institute Visual Coding-Neuropixels dataset in the mouse, we infer encoding manifolds for V1 and for five higher cortical visual areas (VISam, VISal, VISpm, VISlm, and VISrl). We show here that the encoding manifold topology computed only from responses to various grating stimuli is also continuous, not only for V1 but also for the higher visual areas, with smooth coordinates spanning it that include, among others, orientation selectivity and firing-rate magnitude. Surprisingly, the encoding manifold for gratings also provides information about natural scene responses. To investigate whether neurons respond more strongly to gratings or natural scenes, we plot the log ratio of natural scene responses to grating responses (mean firing rates) on the encoding manifold. This reveals a global coordinate axis organizing neurons' preferences between these two stimuli. This coordinate is orthogonal (i.e., uncorrelated) to that organizing firing rate magnitudes in VISp. Analyzing layer responses, a preference for gratings is concentrated in layer 6, whereas preference for natural scenes tends to be higher in layers 2/3 and 4. We also find that preference for natural scenes dominates the responses of neurons that prefer low (0.02 cpd) and high (0.32 cpd) spatial frequencies, rather than intermediate ones (0.04 to 0.16 cpd). Conclusion: while gratings seem limited and natural scenes unconstrained, machine learning algorithms can reveal subtle relationships between them beyond linear techniques.
Collapse
|
2
|
Liao L, Xu K, Wu H, Chen C, Sun W, Yan Q, Jay Kuo CC, Lin W. Blind Video Quality Prediction by Uncovering Human Video Perceptual Representation. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; 33:4998-5013. [PMID: 39236121 DOI: 10.1109/tip.2024.3445738] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/07/2024]
Abstract
Blind video quality assessment (VQA) has become an increasingly demanding problem in automatically assessing the quality of ever-growing in-the-wild videos. Although efforts have been made to measure temporal distortions, the core to distinguish between VQA and image quality assessment (IQA), the lack of modeling of how the human visual system (HVS) relates to the temporal quality of videos hinders the precise mapping of predicted temporal scores to the human perception. Inspired by the recent discovery of the temporal straightness law of natural videos in the HVS, this paper intends to model the complex temporal distortions of in-the-wild videos in a simple and uniform representation by describing the geometric properties of videos in the visual perceptual domain. A novel videolet, with perceptual representation embedding of a few consecutive frames, is designed as the basic quality measurement unit to quantify temporal distortions by measuring the angular and linear displacements from the straightness law. By combining the predicted score on each videolet, a perceptually temporal quality evaluator (PTQE) is formed to measure the temporal quality of the entire video. Experimental results demonstrate that the perceptual representation in the HVS is an efficient way of predicting subjective temporal quality. Moreover, when combined with spatial quality metrics, PTQE achieves top performance over popular in-the-wild video datasets. More importantly, PTQE requires no additional information beyond the video being assessed, making it applicable to any dataset without parameter tuning. Additionally, the generalizability of PTQE is evaluated on video frame interpolation tasks, demonstrating its potential to benefit temporal-related enhancement tasks.
Collapse
|
3
|
Yang R, Zhao P, Wang L, Feng C, Peng C, Wang Z, Zhang Y, Shen M, Shi K, Weng S, Dong C, Zeng F, Zhang T, Chen X, Wang S, Wang Y, Luo Y, Chen Q, Chen Y, Jiang C, Jia S, Yu Z, Liu J, Wang F, Jiang S, Xu W, Li L, Wang G, Mo X, Zheng G, Chen A, Zhou X, Jiang C, Yuan Y, Yan B, Zhang J. Assessment of visual function in blind mice and monkeys with subretinally implanted nanowire arrays as artificial photoreceptors. Nat Biomed Eng 2024; 8:1018-1039. [PMID: 37996614 DOI: 10.1038/s41551-023-01137-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2022] [Accepted: 10/17/2023] [Indexed: 11/25/2023]
Abstract
Retinal prostheses could restore image-forming vision in conditions of photoreceptor degeneration. However, contrast sensitivity and visual acuity are often insufficient. Here we report the performance, in mice and monkeys with induced photoreceptor degeneration, of subretinally implanted gold-nanoparticle-coated titania nanowire arrays providing a spatial resolution of 77.5 μm and a temporal resolution of 3.92 Hz in ex vivo retinas (as determined by patch-clamp recording of retinal ganglion cells). In blind mice, the arrays allowed for the detection of drifting gratings and flashing objects at light-intensity thresholds of 15.70-18.09 μW mm-2, and offered visual acuities of 0.3-0.4 cycles per degree, as determined by recordings of visually evoked potentials and optomotor-response tests. In monkeys, the arrays were stable for 54 weeks, allowed for the detection of a 10-μW mm-2 beam of light (0.5° in beam angle) in visually guided saccade experiments, and induced plastic changes in the primary visual cortex, as indicated by long-term in vivo calcium imaging. Nanomaterials as artificial photoreceptors may ameliorate visual deficits in patients with photoreceptor degeneration.
Collapse
Affiliation(s)
- Ruyi Yang
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Peng Zhao
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Liyang Wang
- Department of Ophthalmology, Zhongshan Hospital, Fudan University, Shanghai, P. R. China
| | - Chenli Feng
- Department of Ophthalmology, Zhongshan Hospital, Fudan University, Shanghai, P. R. China
| | - Chen Peng
- Laboratory of Advanced Materials, Department of Chemistry, Fudan University, Shanghai, P. R. China
| | - Zhexuan Wang
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Yingying Zhang
- Key Laboratory of Brain Functional Genomics (Ministry of Education), East China Normal University, Shanghai, P. R. China
| | - Minqian Shen
- Department of Ophthalmology, Zhongshan Hospital, Fudan University, Shanghai, P. R. China
| | - Kaiwen Shi
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Shijun Weng
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Chunqiong Dong
- Department of Ophthalmology, Zhongshan Hospital, Fudan University, Shanghai, P. R. China
| | - Fu Zeng
- Key Laboratory of Brain Functional Genomics (Ministry of Education), East China Normal University, Shanghai, P. R. China
| | - Tianyun Zhang
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Xingdong Chen
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Shuiyuan Wang
- Shanghai Key Lab for Future Computing Hardware and System, School of Microelectronics, Fudan University, Shanghai, P. R. China
| | - Yiheng Wang
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Yuanyuan Luo
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Qingyuan Chen
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Yuqing Chen
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Chengyong Jiang
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Shanshan Jia
- School of Computer Science, Institute for Artificial Intelligence, Peking University, Beijing, P.R. China
| | - Zhaofei Yu
- School of Computer Science, Institute for Artificial Intelligence, Peking University, Beijing, P.R. China
| | - Jian Liu
- School of Computer Science, University of Birmingham, Birmingham, UK
| | - Fei Wang
- Department of Hand Surgery, the National Clinical Research Center for Aging and Medicine, Huashan Hospital, Fudan University, Shanghai, P. R. China
| | - Su Jiang
- Department of Hand Surgery, the National Clinical Research Center for Aging and Medicine, Huashan Hospital, Fudan University, Shanghai, P. R. China
| | - Wendong Xu
- Department of Hand Surgery, the National Clinical Research Center for Aging and Medicine, Huashan Hospital, Fudan University, Shanghai, P. R. China
- Department of Hand and Upper Extremity Surgery, Jing'an District Central Hospital, Fudan University, Shanghai, P.R. China
| | - Liang Li
- Center of Brain Sciences, Beijing Institute of Basic Medical Sciences, Beijing, P. R. China
| | - Gang Wang
- Center of Brain Sciences, Beijing Institute of Basic Medical Sciences, Beijing, P. R. China
| | - Xiaofen Mo
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Gengfeng Zheng
- Laboratory of Advanced Materials, Department of Chemistry, Fudan University, Shanghai, P. R. China
| | - Aihua Chen
- Key Laboratory of Brain Functional Genomics (Ministry of Education), East China Normal University, Shanghai, P. R. China
| | - Xingtao Zhou
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China
| | - Chunhui Jiang
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China.
| | - Yuanzhi Yuan
- Department of Ophthalmology, Zhongshan Hospital, Fudan University, Shanghai, P. R. China.
- Zhongshan Hospital (Xiamen), Fudan University, Xiamen, P.R. China.
| | - Biao Yan
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China.
| | - Jiayi Zhang
- State Key Laboratory of Medical Neurobiology, MOE Frontiers Center for Brain Science, Institutes of Brain Science, Institute for Medical and Engineering Innovation, Department of Ophthalmology, Eye & ENT Hospital, Fudan University, Shanghai, P. R. China.
| |
Collapse
|
4
|
Watanabe T, Sasaki Y, Ogawa D, Shibata K. Unsupervised learning as a computational principle works in visual learning of natural scenes, but not of artificial stimuli. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.07.31.605957. [PMID: 39211147 PMCID: PMC11361125 DOI: 10.1101/2024.07.31.605957] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/04/2024]
Abstract
The question of whether we learn exposed visual features remains a subject of controversy. A prevalent computational model suggests that visual features frequently exposed to observers in natural environments are likely to be learned. However, this unsupervised learning model appears to be contradicted by the significant body of experimental results with human participants that indicates visual perceptual learning (VPL) of visible task-irrelevant features does not occur with frequent exposure. Here, we demonstrate a resolution to this controversy with a new finding: Exposure to a dominant global orientation as task-irrelevant leads to VPL of the orientation, particularly when the orientation is derived from natural scene images, whereas VPL did not occur with artificial images even with matched distributions of local orientations and spatial frequencies to natural scene images. Further investigation revealed that this disparity arises from the presence of higher-order statistics derived from natural scene images-global structures such as correlations between different local orientation and spatial frequency channels. Moreover, behavioral and neuroimaging results indicate that the dominant orientation from these higher-order statistics undergoes less attentional suppression than that from artificial images, which may facilitate VPL. Our results contribute to resolving the controversy by affirming the validity of unsupervised learning models for natural scenes but not for artificial stimuli. They challenge the assumption that VPL occurring in everyday life can be predicted by laws governing VPL for conventionally used artificial stimuli.
Collapse
|
5
|
Castellotti S, Del Viva MM. Neural Substrates for Early Data Reduction in Fast Vision: A Psychophysical Investigation. Brain Sci 2024; 14:753. [PMID: 39199448 PMCID: PMC11352587 DOI: 10.3390/brainsci14080753] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2024] [Revised: 07/23/2024] [Accepted: 07/25/2024] [Indexed: 09/01/2024] Open
Abstract
To ensure survival, the visual system must rapidly extract the most important elements from a large stream of information. This necessity clashes with the computational limitations of the human brain, so a strong early data reduction is required to efficiently process information in fast vision. A theoretical early vision model, recently developed to preserve maximum information using minimal computational resources, allows efficient image data reduction by extracting simplified sketches containing only optimally informative, salient features. Here, we investigate the neural substrates of this mechanism for optimal encoding of information, possibly located in early visual structures. We adopted a flicker adaptation paradigm, which has been demonstrated to specifically impair the contrast sensitivity of the magnocellular pathway. We compared flicker-induced contrast threshold changes in three different tasks. The results indicate that, after adapting to a uniform flickering field, thresholds for image discrimination using briefly presented sketches increase. Similar threshold elevations occur for motion discrimination, a task typically targeting the magnocellular system. Instead, contrast thresholds for orientation discrimination, a task typically targeting the parvocellular system, do not change with flicker adaptation. The computation performed by this early data reduction mechanism seems thus consistent with magnocellular processing.
Collapse
Affiliation(s)
- Serena Castellotti
- Department of Translational Research on New Technologies in Medicine and Surgery, University of Pisa, 56126 Pisa, Italy
- Department of Neurosciences, Psychology, Drug Research and Child Health (NEUROFARBA), University of Florence, 50135 Florence, Italy;
| | - Maria Michela Del Viva
- Department of Neurosciences, Psychology, Drug Research and Child Health (NEUROFARBA), University of Florence, 50135 Florence, Italy;
| |
Collapse
|
6
|
Merscher AS, Gamer M. Can I see it in the eyes? An investigation of freezing-like motion patterns in response to avoidable threat. Psychophysiology 2024; 61:e14567. [PMID: 38469631 DOI: 10.1111/psyp.14567] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2023] [Revised: 02/23/2024] [Accepted: 03/02/2024] [Indexed: 03/13/2024]
Abstract
Freezing is one of the most extensively studied defensive behaviors in rodents. Both reduced body and gaze movements during anticipation of threat also occur in humans and have been discussed as translational indicators of freezing but their relationship remains unclear. We thus set out to elucidate body and eye movements and concomitant autonomic dynamics in anticipation of avoidable threat. Specifically, 50 participants viewed naturalistic pictures that were preceded by a colored fixation cross, signaling them whether to expect an inevitable (shock), no (safety), or a potential shock (flight) that could be avoided by a quick button press. Body sway, eye movements, the heart rate and skin conductance were recorded. We replicated previously described reductions in body sway, gaze dispersion, and the heart rate, and a skin conductance increase in flight trials. Stronger reductions in gaze but not in body sway predicted faster motor reactions on a trial-wise basis, highlighting their functional role in action preparation. We failed to find a trait-like relationship between body and gaze movements across participants, but their temporal profiles were positively related within individuals, suggesting that both metrics partly reflect the same construct. However, future research is desirable to assess these response patterns in naturalistic environments. A more ethological examination of different movement dynamics upon threat would not only warrant better comparability between rodent and human research but also help determine whether and how eye-tracking could be implemented as a proxy for fear-related movements in restricted brain imaging environments.
Collapse
Affiliation(s)
- Alma-Sophia Merscher
- Experimental Clinical Psychology, Department of Psychology, University of Würzburg, Würzburg, Germany
| | - Matthias Gamer
- Experimental Clinical Psychology, Department of Psychology, University of Würzburg, Würzburg, Germany
| |
Collapse
|
7
|
Kanth ST, Ray S. Gamma Responses to Colored Natural Stimuli Can Be Predicted from Local Low-Level Stimulus Features. eNeuro 2024; 11:ENEURO.0417-23.2024. [PMID: 39054054 PMCID: PMC11277289 DOI: 10.1523/eneuro.0417-23.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2023] [Revised: 05/17/2024] [Accepted: 07/01/2024] [Indexed: 07/27/2024] Open
Abstract
The role of gamma rhythm (30-80 Hz) in visual processing is debated; stimuli like gratings and hue patches generate strong gamma, but many natural images do not. Could image gamma responses be predicted by approximating images as gratings or hue patches? Surprisingly, this question remains unanswered, since the joint dependence of gamma on multiple features is poorly understood. We recorded local field potentials and electrocorticogram from two female monkeys while presenting natural images and parametric stimuli varying along several feature dimensions. Gamma responses to different grating/hue features were separable, allowing for a multiplicative model based on individual features. By fitting a hue patch to the image around the receptive field, this simple model could predict gamma responses to chromatic images across scales with reasonably high accuracy. Our results provide a simple "baseline" model to predict gamma from local image properties, against which more complex models of natural vision can be tested.
Collapse
Affiliation(s)
- Sidrat Tasawoor Kanth
- IISc Mathematics Initiative, Indian Institute of Science, Bangalore 560012, India
- Center for Neuroscience, Indian Institute of Science, Bangalore 560012, India
| | - Supratim Ray
- IISc Mathematics Initiative, Indian Institute of Science, Bangalore 560012, India
- Center for Neuroscience, Indian Institute of Science, Bangalore 560012, India
| |
Collapse
|
8
|
Hansen T, Conway BR. The color of fruits in photographs and still life paintings. J Vis 2024; 24:1. [PMID: 38691088 PMCID: PMC11077907 DOI: 10.1167/jov.24.5.1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2023] [Accepted: 03/14/2024] [Indexed: 05/03/2024] Open
Abstract
Still life paintings comprise a wealth of data on visual perception. Prior work has shown that the color statistics of objects show a marked bias for warm colors. Here, we ask about the relative chromatic contrast of these object-associated colors compared with background colors in still life paintings. We reasoned that, owing to the memory color effect, where the color of familiar objects is perceived more saturated, warm colors will be relatively more saturated than cool colors in still life paintings as compared with photographs. We analyzed color in 108 slides of still life paintings of fruit from the teaching slide collection of the Fogg University Art Museum and 41 color-calibrated photographs of fruit from the McGill data set. The results show that the relatively higher chromatic contrast of warm colors was greater for paintings compared with photographs, consistent with the hypothesis.
Collapse
Affiliation(s)
- Thorsten Hansen
- Department of Psychology, Justus Liebig University Giessen, Giessen, Germany
| | - Bevil R Conway
- Laboratory of Sensorimotor Research, National Institutes of Health, Bethesda, MD, USA
| |
Collapse
|
9
|
Nguyen P, Sooriyaarachchi J, Huang Q, Baker CL. Estimating receptive fields of simple and complex cells in early visual cortex: A convolutional neural network model with parameterized rectification. PLoS Comput Biol 2024; 20:e1012127. [PMID: 38820562 PMCID: PMC11168683 DOI: 10.1371/journal.pcbi.1012127] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2023] [Revised: 06/12/2024] [Accepted: 05/01/2024] [Indexed: 06/02/2024] Open
Abstract
Neurons in the primary visual cortex respond selectively to simple features of visual stimuli, such as orientation and spatial frequency. Simple cells, which have phase-sensitive responses, can be modeled by a single receptive field filter in a linear-nonlinear model. However, it is challenging to analyze phase-invariant complex cells, which require more elaborate models having a combination of nonlinear subunits. Estimating parameters of these models is made additionally more difficult by cortical neurons' trial-to-trial response variability. We develop a simple convolutional neural network method to estimate receptive field models for both simple and complex visual cortex cells from their responses to natural images. The model consists of a spatiotemporal filter, a parameterized rectifier unit (PReLU), and a two-dimensional Gaussian "map" of the receptive field envelope. A single model parameter determines the simple vs. complex nature of the receptive field, capturing complex cell responses as a summation of homogeneous subunits, and collapsing to a linear-nonlinear model for simple type cells. The convolutional method predicts simple and complex cell responses to natural image stimuli as well as grating tuning curves. The fitted models yield a continuum of values for the PReLU parameter across the sampled neurons, showing that the simple/complex nature of cells can vary in a continuous manner. We demonstrate that complex-like cells respond less reliably than simple-like cells. However, compensation for this unreliability with noise ceiling analysis reveals predictive performance for complex cells proportionately closer to that for simple cells. Most spatial receptive field structures are well fit by Gabor functions, whose parameters confirm well-known properties of cat A17/18 receptive fields.
Collapse
Affiliation(s)
- Philippe Nguyen
- Department of Biomedical Engineering, McGill University, Montreal, Quebec, Canada
| | | | - Qianyu Huang
- Department of Biology, McGill University, Montreal, Quebec, Canada
| | - Curtis L. Baker
- Department of Ophthalmology and Visual Sciences, McGill University, Montreal, Quebec, Canada
| |
Collapse
|
10
|
Rennig J, Langenberger C, Karnath HO. Beyond visual integration: sensitivity of the temporal-parietal junction for objects, places, and faces. BEHAVIORAL AND BRAIN FUNCTIONS : BBF 2024; 20:8. [PMID: 38637870 PMCID: PMC11027340 DOI: 10.1186/s12993-024-00233-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/12/2023] [Accepted: 03/24/2024] [Indexed: 04/20/2024]
Abstract
One important role of the TPJ is the contribution to perception of the global gist in hierarchically organized stimuli where individual elements create a global visual percept. However, the link between clinical findings in simultanagnosia and neuroimaging in healthy subjects is missing for real-world global stimuli, like visual scenes. It is well-known that hierarchical, global stimuli activate TPJ regions and that simultanagnosia patients show deficits during the recognition of hierarchical stimuli and real-world visual scenes. However, the role of the TPJ in real-world scene processing is entirely unexplored. In the present study, we first localized TPJ regions significantly responding to the global gist of hierarchical stimuli and then investigated the responses to visual scenes, as well as single objects and faces as control stimuli. All three stimulus classes evoked significantly positive univariate responses in the previously localized TPJ regions. In a multivariate analysis, we were able to demonstrate that voxel patterns of the TPJ were classified significantly above chance level for all three stimulus classes. These results demonstrate a significant involvement of the TPJ in processing of complex visual stimuli that is not restricted to visual scenes and that the TPJ is sensitive to different classes of visual stimuli with a specific signature of neuronal activations.
Collapse
Affiliation(s)
- Johannes Rennig
- Division of Neuropsychology, Center of Neurology, Hertie-Institute for Clinical Brain Research, University of Tübingen, D-72076, Tübingen, Germany.
| | - Christina Langenberger
- Division of Neuropsychology, Center of Neurology, Hertie-Institute for Clinical Brain Research, University of Tübingen, D-72076, Tübingen, Germany
| | - Hans-Otto Karnath
- Division of Neuropsychology, Center of Neurology, Hertie-Institute for Clinical Brain Research, University of Tübingen, D-72076, Tübingen, Germany
- Department of Psychology, University of South Carolina, Columbia, USA
| |
Collapse
|
11
|
Samiei T, Zou Z, Imani M, Nozari E. Optimal decoding of neural dynamics occurs at mesoscale spatial and temporal resolutions. Front Cell Neurosci 2024; 18:1287123. [PMID: 38419658 PMCID: PMC10899419 DOI: 10.3389/fncel.2024.1287123] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2023] [Accepted: 01/23/2024] [Indexed: 03/02/2024] Open
Abstract
Introduction Understanding the neural code has been one of the central aims of neuroscience research for decades. Spikes are commonly referred to as the units of information transfer, but multi-unit activity (MUA) recordings are routinely analyzed in aggregate forms such as binned spike counts, peri-stimulus time histograms, firing rates, or population codes. Various forms of averaging also occur in the brain, from the spatial averaging of spikes within dendritic trees to their temporal averaging through synaptic dynamics. However, how these forms of averaging are related to each other or to the spatial and temporal units of information representation within the neural code has remained poorly understood. Materials and methods In this work we developed NeuroPixelHD, a symbolic hyperdimensional model of MUA, and used it to decode the spatial location and identity of static images shown to n = 9 mice in the Allen Institute Visual Coding-NeuroPixels dataset from large-scale MUA recordings. We parametrically varied the spatial and temporal resolutions of the MUA data provided to the model, and compared its resulting decoding accuracy. Results For almost all subjects, we found 125ms temporal resolution to maximize decoding accuracy for both the spatial location of Gabor patches (81 classes for patches presented over a 9×9 grid) as well as the identity of natural images (118 classes corresponding to 118 images) across the whole brain. This optimal temporal resolution nevertheless varied greatly between different regions, followed a sensory-associate hierarchy, and was significantly modulated by the central frequency of theta-band oscillations across different regions. Spatially, the optimal resolution was at either of two mesoscale levels for almost all mice: the area level, where the spiking activity of all neurons within each brain area are combined, and the population level, where neuronal spikes within each area are combined across fast spiking (putatively inhibitory) and regular spiking (putatively excitatory) neurons, respectively. We also observed an expected interplay between optimal spatial and temporal resolutions, whereby increasing the amount of averaging across one dimension (space or time) decreases the amount of averaging that is optimal across the other dimension, and vice versa. Discussion Our findings corroborate existing empirical practices of spatiotemporal binning and averaging in MUA data analysis, and provide a rigorous computational framework for optimizing the level of such aggregations. Our findings can also synthesize these empirical practices with existing knowledge of the various sources of biological averaging in the brain into a new theory of neural information processing in which the unit of information varies dynamically based on neuronal signal and noise correlations across space and time.
Collapse
Affiliation(s)
- Toktam Samiei
- Department of Mechanical Engineering, University of California, Riverside, Riverside, CA, United States
| | - Zhuowen Zou
- Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| | - Mohsen Imani
- Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| | - Erfan Nozari
- Department of Mechanical Engineering, University of California, Riverside, Riverside, CA, United States
- Department of Electrical and Computer Engineering, University of California, Riverside, Riverside, CA, United States
- Department of Bioengineering, University of California, Riverside, Riverside, CA, United States
| |
Collapse
|
12
|
Samonds JM, Szinte M, Barr C, Montagnini A, Masson GS, Priebe NJ. Mammals Achieve Common Neural Coverage of Visual Scenes Using Distinct Sampling Behaviors. eNeuro 2024; 11:ENEURO.0287-23.2023. [PMID: 38164577 PMCID: PMC10860624 DOI: 10.1523/eneuro.0287-23.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Revised: 10/24/2023] [Accepted: 10/30/2023] [Indexed: 01/03/2024] Open
Abstract
Most vertebrates use head and eye movements to quickly change gaze orientation and sample different portions of the environment with periods of stable fixation. Visual information must be integrated across fixations to construct a complete perspective of the visual environment. In concert with this sampling strategy, neurons adapt to unchanging input to conserve energy and ensure that only novel information from each fixation is processed. We demonstrate how adaptation recovery times and saccade properties interact and thus shape spatiotemporal tradeoffs observed in the motor and visual systems of mice, cats, marmosets, macaques, and humans. These tradeoffs predict that in order to achieve similar visual coverage over time, animals with smaller receptive field sizes require faster saccade rates. Indeed, we find comparable sampling of the visual environment by neuronal populations across mammals when integrating measurements of saccadic behavior with receptive field sizes and V1 neuronal density. We propose that these mammals share a common statistically driven strategy of maintaining coverage of their visual environment over time calibrated to their respective visual system characteristics.
Collapse
Affiliation(s)
- Jason M Samonds
- Center for Learning and Memory and the Institute for Neuroscience, The University of Texas at Austin, Austin 78712, Texas
| | - Martin Szinte
- Institut de Neurosciences de la Timone (UMR 7289), Centre National de la Recherche Scientifique and Aix-Marseille Université, 13385 Marseille, France
| | - Carrie Barr
- Center for Learning and Memory and the Institute for Neuroscience, The University of Texas at Austin, Austin 78712, Texas
| | - Anna Montagnini
- Institut de Neurosciences de la Timone (UMR 7289), Centre National de la Recherche Scientifique and Aix-Marseille Université, 13385 Marseille, France
| | - Guillaume S Masson
- Institut de Neurosciences de la Timone (UMR 7289), Centre National de la Recherche Scientifique and Aix-Marseille Université, 13385 Marseille, France
| | - Nicholas J Priebe
- Center for Learning and Memory and the Institute for Neuroscience, The University of Texas at Austin, Austin 78712, Texas
| |
Collapse
|
13
|
Viejo G, Levenstein D, Skromne Carrasco S, Mehrotra D, Mahallati S, Vite GR, Denny H, Sjulson L, Battaglia FP, Peyrache A. Pynapple, a toolbox for data analysis in neuroscience. eLife 2023; 12:RP85786. [PMID: 37843985 PMCID: PMC10578930 DOI: 10.7554/elife.85786] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2023] Open
Abstract
Datasets collected in neuroscientific studies are of ever-growing complexity, often combining high-dimensional time series data from multiple data acquisition modalities. Handling and manipulating these various data streams in an adequate programming environment is crucial to ensure reliable analysis, and to facilitate sharing of reproducible analysis pipelines. Here, we present Pynapple, the PYthon Neural Analysis Package, a lightweight python package designed to process a broad range of time-resolved data in systems neuroscience. The core feature of this package is a small number of versatile objects that support the manipulation of any data streams and task parameters. The package includes a set of methods to read common data formats and allows users to easily write their own. The resulting code is easy to read and write, avoids low-level data processing and other error-prone steps, and is open source. Libraries for higher-level analyses are developed within the Pynapple framework but are contained within a collaborative repository of specialized and continuously updated analysis routines. This provides flexibility while ensuring long-term stability of the core package. In conclusion, Pynapple provides a common framework for data analysis in neuroscience.
Collapse
Affiliation(s)
- Guillaume Viejo
- Montreal Neurological Institute and Hospital, McGill UniversityMontrealCanada
- Flatiron Institute, Center for Computational NeuroscienceNew YorkUnited States
| | - Daniel Levenstein
- Montreal Neurological Institute and Hospital, McGill UniversityMontrealCanada
- MILA – Quebec IA InstituteMontrealCanada
| | | | - Dhruv Mehrotra
- Montreal Neurological Institute and Hospital, McGill UniversityMontrealCanada
| | - Sara Mahallati
- Montreal Neurological Institute and Hospital, McGill UniversityMontrealCanada
| | - Gilberto R Vite
- Montreal Neurological Institute and Hospital, McGill UniversityMontrealCanada
| | - Henry Denny
- Montreal Neurological Institute and Hospital, McGill UniversityMontrealCanada
| | - Lucas Sjulson
- Departments of Psychiatry and Neuroscience, Albert Einstein College of MedicineBronxUnited States
| | - Francesco P Battaglia
- Donders Institute for Brain, Cognition and Behaviour, Radboud UniversityNijmegenNetherlands
| | - Adrien Peyrache
- Montreal Neurological Institute and Hospital, McGill UniversityMontrealCanada
| |
Collapse
|
14
|
Castellotti S, Szinte M, Del Viva MM, Montagnini A. Saccadic trajectories deviate toward or away from optimally informative visual features. iScience 2023; 26:107282. [PMID: 37520738 PMCID: PMC10371840 DOI: 10.1016/j.isci.2023.107282] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2022] [Revised: 03/17/2023] [Accepted: 06/30/2023] [Indexed: 08/01/2023] Open
Abstract
The saccades' path is influenced by visual distractors, making their trajectory curve away or toward them. Previous research suggested that the more salient the distractor, the more pronounced is the curvature. We investigate the saliency of spatial visual features, predicted by a constrained maximum entropy model to be optimal or non-optimal information carriers in fast vision, by using them as distractors in a saccadic task. Their effect was compared to that of luminance-based control distractors. Optimal features evoke a larger saccadic curvature compared to non-optimal features, and the magnitude and direction of deviation change as a function of the delay between distractor and saccade onset. Effects were similar to those found with high-luminance versus low-luminance distractors. Therefore, model-predicted information optimal features interfere with target-oriented saccades, following a dynamic attraction-repulsion pattern. This suggests that the visuo-oculomotor system rapidly and automatically processes optimally informative features while programming visually guided eye movements.
Collapse
Affiliation(s)
| | - Martin Szinte
- Institut de Neurosciences de la Timone, CNRS and Aix-Marseille Université, Marseille, France
| | | | - Anna Montagnini
- Institut de Neurosciences de la Timone, CNRS and Aix-Marseille Université, Marseille, France
| |
Collapse
|
15
|
DiMattina C. Second-order boundaries segment more easily when they are density-defined rather than feature-defined. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.07.10.548431. [PMID: 37502940 PMCID: PMC10369903 DOI: 10.1101/2023.07.10.548431] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/29/2023]
Abstract
Previous studies have demonstrated that density is an important perceptual aspect of textural appearance to which the visual system is highly attuned. Furthermore, it is known that density cues not only influence texture segmentation, but can enable segmentation by themselves, in the absence of other cues. A popular computational model of texture segmentation known as the "Filter-Rectify-Filter" (FRF) model predicts that density should be a second-order cue enabling segmentation. For a compound texture boundary defined by superimposing two single-micropattern density boundaries, a version of the FRF model in which different micropattern-specific channels are analyzed separately by different second-stage filters makes the prediction that segmentation thresholds should be identical in two cases: (1) Compound boundaries with an equal number of micropatterns on each side but different relative proportions of each variety (compound feature boundaries) and (2) Compound boundaries with different numbers of micropatterns on each side, but with each side having an identical number of each variety (compound density boundaries). We directly tested this prediction by comparing segmentation thresholds for second-order compound feature and density boundaries, comprised of two superimposed single-micropattern density boundaries comprised of complementary micropattern pairs differing either in orientation or contrast polarity. In both cases, we observed lower segmentation thresholds for compound density boundaries than compound feature boundaries, with identical results when the compound density boundaries were equated for RMS contrast. In a second experiment, we considered how two varieties of micropatterns summate for compound boundary segmentation. In the case where two single micro-pattern density boundaries are superimposed to form a compound density boundary, we find that the two channels combine via probability summation. By contrast, when they are superimposed to form a compound feature boundary, segmentation performance is worse than for either channel alone. From these findings, we conclude that density segmentation may rely on neural mechanisms different from those which underlie feature segmentation, consistent with recent findings suggesting that density comprises a separate psychophysical 'channel'.
Collapse
Affiliation(s)
- Christopher DiMattina
- Computational Perception Laboratory, Florida Gulf Coast University, Fort Myers, FL, USA 33965-6565
- Department of Psychology, Florida Gulf Coast University, Fort Myers, FL, USA 33965-6565
| |
Collapse
|
16
|
Samonds JM, Szinte M, Barr C, Montagnini A, Masson GS, Priebe NJ. Mammals achieve common neural coverage of visual scenes using distinct sampling behaviors. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.03.20.533210. [PMID: 36993477 PMCID: PMC10055212 DOI: 10.1101/2023.03.20.533210] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/31/2023]
Abstract
Most vertebrates use head and eye movements to quickly change gaze orientation and sample different portions of the environment with periods of stable fixation. Visual information must be integrated across several fixations to construct a more complete perspective of the visual environment. In concert with this sampling strategy, neurons adapt to unchanging input to conserve energy and ensure that only novel information from each fixation is processed. We demonstrate how adaptation recovery times and saccade properties interact, and thus shape spatiotemporal tradeoffs observed in the motor and visual systems of different species. These tradeoffs predict that in order to achieve similar visual coverage over time, animals with smaller receptive field sizes require faster saccade rates. Indeed, we find comparable sampling of the visual environment by neuronal populations across mammals when integrating measurements of saccadic behavior with receptive field sizes and V1 neuronal density. We propose that these mammals share a common statistically driven strategy of maintaining coverage of their visual environment over time calibrated to their respective visual system characteristics.
Collapse
|
17
|
Zhang Y, Motoyoshi I. Perceiving the representative surface color of real-world materials. Sci Rep 2023; 13:6300. [PMID: 37072618 PMCID: PMC10111332 DOI: 10.1038/s41598-023-33563-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2022] [Accepted: 04/14/2023] [Indexed: 05/03/2023] Open
Abstract
Natural surfaces such as soil, grass, and skin usually involve far more complex and heterogenous structures than the perfectly uniform surfaces assumed in studies on color and material perception. Despite this, we can easily perceive the representative color of these surfaces. Here, we investigated the visual mechanisms underlying the perception of representative surface color using 120 natural images of diverse materials and their statistically synthesized images. Our matching experiments indicated that the perceived representative color revealed was not significantly different from the Portilla-Simoncelli-synthesized images or phase-randomized images except for one sample, even though the perceived shape and material properties were greatly impaired in the synthetic stimuli. The results also showed that the matched representative colors were predictable from the saturation-enhanced color of the brightest point in the image, excluding the high-intensity outliers. The results support the notion that humans judge the representative color and lightness of real-world surfaces depending on simple image measurements.
Collapse
Affiliation(s)
- Yan Zhang
- Department of Life Sciences, The University of Tokyo, Tokyo, Japan
| | - Isamu Motoyoshi
- Department of Life Sciences, The University of Tokyo, Tokyo, Japan.
| |
Collapse
|
18
|
St-Amand D, Baker CL. Model-Based Approach Shows ON Pathway Afferents Elicit a Transient Decrease of V1 Responses. J Neurosci 2023; 43:1920-1932. [PMID: 36759194 PMCID: PMC10027028 DOI: 10.1523/jneurosci.1220-22.2023] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2022] [Revised: 01/29/2023] [Accepted: 01/30/2023] [Indexed: 02/11/2023] Open
Abstract
Neurons in the primary visual cortex (V1) receive excitation and inhibition from distinct parallel pathways processing lightness (ON) and darkness (OFF). V1 neurons overall respond more strongly to dark than light stimuli, consistent with a preponderance of darker regions in natural images, as well as human psychophysics. However, it has been unclear whether this "dark-dominance" is because of more excitation from the OFF pathway or more inhibition from the ON pathway. To understand the mechanisms behind dark-dominance, we record electrophysiological responses of individual simple-type V1 neurons to natural image stimuli and then train biologically inspired convolutional neural networks to predict the neurons' responses. Analyzing a sample of 71 neurons (in anesthetized, paralyzed cats of either sex) has revealed their responses to be more driven by dark than light stimuli, consistent with previous investigations. We show that this asymmetry is predominantly because of slower inhibition to dark stimuli rather than to stronger excitation from the thalamocortical OFF pathway. Consistent with dark-dominant neurons having faster responses than light-dominant neurons, we find dark-dominance to solely occur in the early latencies of neurons' responses. Neurons that are strongly dark-dominated also tend to be less orientation-selective. This novel approach gives us new insight into the dark-dominance phenomenon and provides an avenue to address new questions about excitatory and inhibitory integration in cortical neurons.SIGNIFICANCE STATEMENT Neurons in the early visual cortex respond on average more strongly to dark than to light stimuli, but the mechanisms behind this bias have been unclear. Here we address this issue by combining single-unit electrophysiology with a novel machine learning model to analyze neurons' responses to natural image stimuli in primary visual cortex. Using these techniques, we find slower inhibition to light than to dark stimuli to be the leading mechanism behind stronger dark responses. This slower inhibition to light might help explain other empirical findings, such as why orientation selectivity is weaker at earlier response latencies. These results demonstrate how imbalances in excitation versus inhibition can give rise to response asymmetries in cortical neuron responses.
Collapse
Affiliation(s)
- David St-Amand
- McGill Vision Research Unit, Department of Ophthalmology & Visual Sciences, McGill University, Montreal, Quebec H3G 1A4, Canada
| | - Curtis L Baker
- McGill Vision Research Unit, Department of Ophthalmology & Visual Sciences, McGill University, Montreal, Quebec H3G 1A4, Canada
| |
Collapse
|
19
|
Castellotti S, D’Agostino O, Del Viva MM. Fast discrimination of fragmentary images: the role of local optimal information. Front Hum Neurosci 2023; 17:1049615. [PMID: 36845876 PMCID: PMC9945129 DOI: 10.3389/fnhum.2023.1049615] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2022] [Accepted: 01/18/2023] [Indexed: 02/11/2023] Open
Abstract
In naturalistic conditions, objects in the scene may be partly occluded and the visual system has to recognize the whole image based on the little information contained in some visible fragments. Previous studies demonstrated that humans can successfully recognize severely occluded images, but the underlying mechanisms occurring in the early stages of visual processing are still poorly understood. The main objective of this work is to investigate the contribution of local information contained in a few visible fragments to image discrimination in fast vision. It has been already shown that a specific set of features, predicted by a constrained maximum-entropy model to be optimal carriers of information (optimal features), are used to build simplified early visual representations (primal sketch) that are sufficient for fast image discrimination. These features are also considered salient by the visual system and can guide visual attention when presented isolated in artificial stimuli. Here, we explore whether these local features also play a significant role in more natural settings, where all existing features are kept, but the overall available information is drastically reduced. Indeed, the task requires discrimination of naturalistic images based on a very brief presentation (25 ms) of a few small visible image fragments. In the main experiment, we reduced the possibility to perform the task based on global-luminance positional cues by presenting randomly inverted-contrast images, and we measured how much observers' performance relies on the local features contained in the fragments or on global information. The size and the number of fragments were determined in two preliminary experiments. Results show that observers are very skilled in fast image discrimination, even when a drastic occlusion is applied. When observers cannot rely on the position of global-luminance information, the probability of correct discrimination increases when the visible fragments contain a high number of optimal features. These results suggest that such optimal local information contributes to the successful reconstruction of naturalistic images even in challenging conditions.
Collapse
|
20
|
Nigam S, Milton R, Pojoga S, Dragoi V. Adaptive coding across visual features during free-viewing and fixation conditions. Nat Commun 2023; 14:87. [PMID: 36604422 PMCID: PMC9816177 DOI: 10.1038/s41467-022-35656-w] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2022] [Accepted: 12/15/2022] [Indexed: 01/07/2023] Open
Abstract
Theoretical studies have long proposed that adaptation allows the brain to effectively use the limited response range of sensory neurons to encode widely varying natural inputs. However, despite this influential view, experimental studies have exclusively focused on how the neural code adapts to a range of stimuli lying along a single feature axis, such as orientation or contrast. Here, we performed electrical recordings in macaque visual cortex (area V4) to reveal significant adaptive changes in the neural code of single cells and populations across multiple feature axes. Both during free viewing and passive fixation, populations of cells improved their ability to encode image features after rapid exposure to stimuli lying on orthogonal feature axes even in the absence of initial tuning to these stimuli. These results reveal a remarkable adaptive capacity of visual cortical populations to improve network computations relevant for natural viewing despite the modularity of the functional cortical architecture.
Collapse
Affiliation(s)
- Sunny Nigam
- Department of Neurobiology and Anatomy McGovern Medical School, University of Texas at Houston, Houston, TX, 77030, US.
| | - Russell Milton
- Department of Neurobiology and Anatomy McGovern Medical School, University of Texas at Houston, Houston, TX, 77030, US
| | - Sorin Pojoga
- Department of Neurobiology and Anatomy McGovern Medical School, University of Texas at Houston, Houston, TX, 77030, US
| | - Valentin Dragoi
- Department of Neurobiology and Anatomy McGovern Medical School, University of Texas at Houston, Houston, TX, 77030, US.
- Department of Electrical and Computer Engineering, Rice University, Houston, TX, 77005, US.
| |
Collapse
|
21
|
Berlijn AM, Hildebrandt LK, Gamer M. Idiosyncratic viewing patterns of social scenes reflect individual preferences. J Vis 2022; 22:10. [PMID: 36583910 PMCID: PMC9807181 DOI: 10.1167/jov.22.13.10] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2022] [Accepted: 11/16/2022] [Indexed: 12/31/2022] Open
Abstract
In general, humans preferentially look at conspecifics in naturalistic images. However, such group-based effects might conceal systematic individual differences concerning the preference for social information. Here, we investigated to what degree fixations on social features occur consistently within observers and whether this preference generalizes to other measures of social prioritization in the laboratory as well as the real world. Participants carried out a free viewing task, a relevance taps task that required them to actively select image regions that are crucial for understanding a given scene, and they were asked to freely take photographs outside the laboratory that were later classified regarding their social content. We observed stable individual differences in the fixation and active selection of human heads and faces that were correlated across tasks and partly predicted the social content of self-taken photographs. Such relationship was not observed for human bodies indicating that different social elements need to be dissociated. These findings suggest that idiosyncrasies in the visual exploration and interpretation of social features exist and predict real-world behavior. Future studies should further characterize these preferences and elucidate how they shape perception and interpretation of social contexts in healthy participants and patients with mental disorders that affect social functioning.
Collapse
Affiliation(s)
- Adam M Berlijn
- Department of Experimental Psychology, Heinrich-Heine-University Düsseldorf, Düsseldorf, Germany
- Institute of Clinical Neuroscience and Medical Psychology, Medical Faculty, University Hospital Düsseldorf, Heinrich-Heine University Düsseldorf, Düsseldorf, Germany
- Institute of Neuroscience and Medicine (INM-1), Research Centre Jülich, Jülich, Germany
- Department of Psychology, Julius-Maximilians-University Würzburg, Würzburg, Germany
| | - Lea K Hildebrandt
- Department of Psychology, Julius-Maximilians-University Würzburg, Würzburg, Germany
| | - Matthias Gamer
- Department of Psychology, Julius-Maximilians-University Würzburg, Würzburg, Germany
| |
Collapse
|
22
|
Abstract
To obtain a single percept of the world, the visual system must combine inputs from the two eyes. Understanding the principles that govern this binocular combination process has important real-world clinical and technological applications. However, most research examining binocular combination has relied on relatively simple visual stimuli and it is unclear how well the findings apply to real-world scenarios. For example, it is well-known that, when the two eyes view sine wave gratings with differing contrast (dichoptic stimuli), the binocular percept often matches the higher contrast grating. Does this winner-take-all property of binocular contrast combination apply to more naturalistic imagery, which include broadband structure and spatially varying contrast? To better understand binocular combination during naturalistic viewing, we conducted psychophysical experiments characterizing binocular contrast perception for a range of visual stimuli. In two experiments, we measured the binocular contrast perception of dichoptic sine wave gratings and naturalistic stimuli, and asked how the contrast of the surrounding context affected percepts. Binocular contrast percepts were close to winner-take-all across many of the stimuli when the surrounding context was the average contrast of the two eyes. However, we found that changing the surrounding context modulated the binocular percept of some patterns and not others. We show evidence that this contextual effect may be due to the spatial orientation structure of the stimuli. These findings provide a step toward understanding binocular combination in the natural world and highlight the importance of considering the effect of the spatial interactions in complex stimuli.
Collapse
Affiliation(s)
- Minqi Wang
- Herbert Wertheim School of Optometry & Vision Science, University of California, Berkeley, CA USA
| | - Jian Ding
- Herbert Wertheim School of Optometry & Vision Science, University of California, Berkeley, CA USA
- Helen Wills Neuroscience Institute, University of California, Berkeley, CA, USA
| | - Dennis M. Levi
- Herbert Wertheim School of Optometry & Vision Science, University of California, Berkeley, CA USA
- Helen Wills Neuroscience Institute, University of California, Berkeley, CA, USA
| | - Emily A. Cooper
- Herbert Wertheim School of Optometry & Vision Science, University of California, Berkeley, CA USA
- Helen Wills Neuroscience Institute, University of California, Berkeley, CA, USA
| |
Collapse
|
23
|
Ng CJ, Sabesan R, Barbot A, Banks MS, Yoon G. Suprathreshold Contrast Perception Is Altered by Long-term Adaptation to Habitual Optical Blur. Invest Ophthalmol Vis Sci 2022; 63:6. [PMID: 36223102 PMCID: PMC9583751 DOI: 10.1167/iovs.63.11.6] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/26/2021] [Accepted: 09/11/2022] [Indexed: 01/16/2023] Open
Abstract
Purpose To investigate whether visual experience with habitual blur alters the neural processing of suprathreshold contrast in emmetropic and highly aberrated eyes. Methods A large stroke adaptive optics system was used to correct ocular aberrations. Contrast constancy was assessed psychophysically in emmetropic and keratoconic eyes using a contrast matching paradigm. Participants adjusted the contrasts of gratings at various spatial frequencies to match the contrast perception of a reference grating at 4 c/deg. Matching was done both with fully corrected and uncorrected ocular aberrations. Optical correction allowed keratoconus patients to perceive high spatial frequencies that they have not experienced for some time. Results Emmetropic observers exhibited contrast constancy both with their native aberrations and when their aberrations were corrected. Keratoconus patients exhibited contrast constancy with their uncorrected, native optics but they did not exhibit constancy during adaptive optics correction. Instead. they exhibited striking underconstancy: they required more contrast at high spatial frequencies than the contrast of the 4-c/deg stimulus to make them seem to have the same contrast. Conclusions The presence of contrast constancy in emmetropes and keratoconus patients viewing with their native optics suggests that they have learned to amplify neural signals to offset the effects of habitual optical aberrations. The fact that underconstancy was observed in keratoconus patients when their optics were corrected suggests that they were unable to learn the appropriate neural amplification because they did not have experience with fine spatial detail. These results show that even adults can learn neural amplification to counteract the effects of their own optical aberrations.
Collapse
Affiliation(s)
- Cherlyn J. Ng
- College of Optometry, University of Houston, Houston, TX, United States
| | - Ramkumar Sabesan
- Department of Ophthalmology, University of Washington, Seattle, WA, United States
| | - Antoine Barbot
- Department of Ophthalmology, University of Rochester, Rochester, NY, United States
| | - Martin S. Banks
- School of Optometry, University of California, Berkeley, CA, United States
| | - Geunyoung Yoon
- College of Optometry, University of Houston, Houston, TX, United States
| |
Collapse
|
24
|
Abstract
The primary visual cortex signals the onset of light and dark stimuli with ON and OFF cortical pathways. Here, we demonstrate that both pathways generate similar response increments to large homogeneous surfaces and their response average increases with surface brightness. We show that, in cat visual cortex, response dominance from ON or OFF pathways is bimodally distributed when stimuli are smaller than one receptive field center but unimodally distributed when they are larger. Moreover, whereas small bright stimuli drive opposite responses from ON and OFF pathways (increased versus suppressed activity), large bright surfaces drive similar response increments. We show that this size-brightness relation emerges because strong illumination increases the size of light surfaces in nature and both ON and OFF cortical neurons receive input from ON thalamic pathways. We conclude that visual scenes are perceived as brighter when the average response increments from ON and OFF cortical pathways become stronger. Mazade et al. find that the visual cortex encodes brightness differently for small than large stimuli. Bright small stimuli drive cortical pathways signaling lights and suppress cortical pathways signaling darks. Conversely, large surfaces drive response increments from both pathways and appear brightest when the response average is strongest.
Collapse
|
25
|
DiMattina C, Burnham JJ, Guner BN, Yerxa HB. Distinguishing shadows from surface boundaries using local achromatic cues. PLoS Comput Biol 2022; 18:e1010473. [PMID: 36103558 PMCID: PMC9512248 DOI: 10.1371/journal.pcbi.1010473] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2022] [Revised: 09/26/2022] [Accepted: 08/05/2022] [Indexed: 11/24/2022] Open
Abstract
In order to accurately parse the visual scene into distinct surfaces, it is essential to determine whether a local luminance edge is caused by a boundary between two surfaces or a shadow cast across a single surface. Previous studies have demonstrated that local chromatic cues may help to distinguish edges caused by shadows from those caused by surface boundaries, but the information potentially available in local achromatic cues like contrast, texture, and penumbral blur remains poorly understood. In this study, we develop and analyze a large database of hand-labeled achromatic shadow edges to better understand what image properties distinguish them from occlusion edges. We find that both the highest contrast as well as the lowest contrast edges are more likely to be occlusions than shadows, extending previous observations based on a more limited image set. We also find that contrast cues alone can reliably distinguish the two edge categories with nearly 70% accuracy at 40x40 resolution. Logistic regression on a Gabor Filter bank (GFB) modeling a population of V1 simple cells separates the categories with nearly 80% accuracy, and furthermore exhibits tuning to penumbral blur. A Filter-Rectify Filter (FRF) style neural network extending the GFB model performed at better than 80% accuracy, and exhibited blur tuning and greater sensitivity to texture differences. We compare human performance on our edge classification task to that of the FRF and GFB models, finding the best human observers attaining the same performance as the machine classifiers. Several analyses demonstrate both classifiers exhibit significant positive correlation with human behavior, although we find a slightly better agreement on an image-by-image basis between human performance and the FRF model than the GFB model, suggesting an important role for texture. Distinguishing edges caused by changes in illumination from edges caused by surface boundaries is an essential computation for accurately parsing the visual scene. Previous psychophysical investigations examining the utility of various locally available cues to classify edges as shadows or surface boundaries have primarily focused on color, as surface boundaries often give rise to more significant change in color than shadows. However, even in grayscale images we can readily distinguish shadows from surface boundaries, suggesting an important role for achromatic cues in addition to color. We demonstrate using statistical analysis of natural shadow and surface boundary edges that locally available achromatic cues can be exploited by machine classifiers to reliably distinguish these two edge categories. These classifiers exhibit sensitivity to blur and local texture differences, and exhibit reasonably good agreement with humans classifying edges as shadows or surface boundaries. As trichromatic vision is relatively rare in the animal kingdom, our work suggests how organisms lacking rich color vision can still exploit other cues to avoid mistaking illumination changes for surface changes.
Collapse
Affiliation(s)
- Christopher DiMattina
- Computational Perception Laboratory, FGCU Computational Facility, & Department of Psychology, Florida Gulf Coast University, Fort Myers, Florida, United States of America
- * E-mail:
| | - Josiah J. Burnham
- Computational Perception Laboratory & Department of Software Engineering, Florida Gulf Coast University, Fort Myers, Florida, United States of America
| | - Betul N. Guner
- Computational Perception Laboratory & Department of Psychology, Florida Gulf Coast University, Fort Myers, Florida, United States of America
| | - Haley B. Yerxa
- Computational Perception Laboratory & Department of Software Engineering, Florida Gulf Coast University, Fort Myers, Florida, United States of America
| |
Collapse
|
26
|
Amrutha E, Arivazhagan S, Jebarani WSL. Novel color image steganalysis method based on RGB channel empirical modes to expose stego images with diverse payloads. Pattern Anal Appl 2022. [DOI: 10.1007/s10044-022-01102-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/14/2022]
|
27
|
Castellotti S, Montagnini A, Del Viva MM. Information-optimal local features automatically attract covert and overt attention. Sci Rep 2022; 12:9994. [PMID: 35705616 PMCID: PMC9200825 DOI: 10.1038/s41598-022-14262-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2022] [Accepted: 06/03/2022] [Indexed: 11/09/2022] Open
Abstract
In fast vision, local spatial properties of the visual scene can automatically capture the observer's attention. We used specific local features, predicted by a constrained maximum-entropy model to be optimal information-carriers, as candidate "salient features''. Previous studies showed that participants choose these optimal features as "more salient" if explicitly asked. Here, we investigated the implicit saliency of these optimal features in two attentional tasks. In a covert-attention experiment, we measured the luminance-contrast threshold for discriminating the orientation of a peripheral gabor. In a gaze-orienting experiment, we analyzed latency and direction of saccades towards a peripheral target. In both tasks, two brief peripheral cues, differing in saliency according to the model, preceded the target, presented on the same (valid trials) or the opposite side (invalid trials) of the optimal cue. Results showed reduced contrast thresholds, saccadic latencies, and direction errors in valid trials, and the opposite in invalid trials, compared to baseline values obtained with equally salient cues. Also, optimal features triggered more anticipatory saccades. Similar effects emerged in a luminance-control condition. Overall, in fast vision, optimal features automatically attract covert and overt attention, suggesting that saliency is determined by information maximization criteria coupled with computational limitations.
Collapse
Affiliation(s)
| | - Anna Montagnini
- Institut de Neurosciences de la Timone, CNRS and Aix-Marseille Universitè, Marseilles, France
| | | |
Collapse
|
28
|
Abstract
An ultimate goal in retina science is to understand how the neural circuit of the retina processes natural visual scenes. Yet most studies in laboratories have long been performed with simple, artificial visual stimuli such as full-field illumination, spots of light, or gratings. The underlying assumption is that the features of the retina thus identified carry over to the more complex scenario of natural scenes. As the application of corresponding natural settings is becoming more commonplace in experimental investigations, this assumption is being put to the test and opportunities arise to discover processing features that are triggered by specific aspects of natural scenes. Here, we review how natural stimuli have been used to probe, refine, and complement knowledge accumulated under simplified stimuli, and we discuss challenges and opportunities along the way toward a comprehensive understanding of the encoding of natural scenes. Expected final online publication date for the Annual Review of Vision Science, Volume 8 is September 2022. Please see http://www.annualreviews.org/page/journal/pubdates for revised estimates.
Collapse
Affiliation(s)
- Dimokratis Karamanlis
- Department of Ophthalmology, University Medical Center Göttingen, Göttingen, Germany.,Bernstein Center for Computational Neuroscience Göttingen, Göttingen, Germany.,International Max Planck Research School for Neurosciences, Göttingen, Germany
| | - Helene Marianne Schreyer
- Department of Ophthalmology, University Medical Center Göttingen, Göttingen, Germany.,Bernstein Center for Computational Neuroscience Göttingen, Göttingen, Germany
| | - Tim Gollisch
- Department of Ophthalmology, University Medical Center Göttingen, Göttingen, Germany.,Bernstein Center for Computational Neuroscience Göttingen, Göttingen, Germany.,Cluster of Excellence "Multiscale Bioimaging: from Molecular Machines to Networks of Excitable Cells" (MBExC), University of Göttingen, Göttingen, Germany
| |
Collapse
|
29
|
Merscher AS, Tovote P, Pauli P, Gamer M. Centralized gaze as an adaptive component of defensive states in humans. Proc Biol Sci 2022; 289:20220405. [PMID: 35582796 PMCID: PMC9114933 DOI: 10.1098/rspb.2022.0405] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/29/2023] Open
Abstract
Adequate defensive responding is crucial for mental health but scientifically not well understood. Specifically, it seems difficult to dissociate defense and approach states based on autonomic response patterns. We thus explored the robustness and threat-specificity of recently described oculomotor dynamics upon threat in anticipation of either threatening or rewarding stimuli in humans. While visually exploring naturalistic images, participants (50 per experiment) expected an inevitable, no, or avoidable shock (Experiment 1) or a guaranteed, no, or achievable reward (Experiment 2) that could be averted or gained by a quick behavioural response. We observed reduced heart rate (bradycardia), increased skin conductance, pupil dilation and globally centralized gaze when shocks were inevitable but, more pronouncedly, when they were avoidable. Reward trials were not associated with globally narrowed visual exploration, but autonomic responses resembled characteristics of the threat condition. While bradycardia and concomitant sympathetic activation reflect not only threat-related but also action-preparatory states independent of valence, global centralization of gaze seems a robust phenomenon during the anticipation of avoidable threat. Thus, instead of relying on single readouts, translational research in animals and humans should consider the multi-dimensionality of states in aversive and rewarding contexts, especially when investigating ambivalent, conflicting situations.
Collapse
Affiliation(s)
- Alma-Sophia Merscher
- Department of Psychology, University of Würzburg, Marcusstr. 9-11, 97070 Würzburg, Germany
| | - Philip Tovote
- Systems Neurobiology, Institute of Clinical Neurobiology, University Hospital Würzburg, Versbacher Str. 5, 97078 Würzburg, Germany
| | - Paul Pauli
- Department of Psychology, University of Würzburg, Marcusstr. 9-11, 97070 Würzburg, Germany
| | - Matthias Gamer
- Department of Psychology, University of Würzburg, Marcusstr. 9-11, 97070 Würzburg, Germany
| |
Collapse
|
30
|
Latency shortening with enhanced sparseness and responsiveness in V1 during active visual sensing. Sci Rep 2022; 12:6021. [PMID: 35410997 PMCID: PMC9001710 DOI: 10.1038/s41598-022-09405-4] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2021] [Accepted: 03/21/2022] [Indexed: 12/03/2022] Open
Abstract
In natural vision, neuronal responses to visual stimuli occur due to self-initiated eye movements. Here, we compare single-unit activity in the primary visual cortex (V1) of non-human primates to flashed natural scenes (passive vision condition) to when they freely explore the images by self-initiated eye movements (active vision condition). Active vision enhances the number of neurons responding, and the response latencies become shorter and less variable across neurons. The increased responsiveness and shortened latency during active vision were not explained by increased visual contrast. While the neuronal activities in all layers of V1 show enhanced responsiveness and shortened latency, a significant increase in lifetime sparseness during active vision is observed only in the supragranular layer. These findings demonstrate that the neuronal responses become more distinct in active vision than passive vision, interpreted as consequences of top-down predictive mechanisms.
Collapse
|
31
|
Liu JK, Karamanlis D, Gollisch T. Simple model for encoding natural images by retinal ganglion cells with nonlinear spatial integration. PLoS Comput Biol 2022; 18:e1009925. [PMID: 35259159 PMCID: PMC8932571 DOI: 10.1371/journal.pcbi.1009925] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2021] [Revised: 03/18/2022] [Accepted: 02/14/2022] [Indexed: 01/05/2023] Open
Abstract
A central goal in sensory neuroscience is to understand the neuronal signal processing involved in the encoding of natural stimuli. A critical step towards this goal is the development of successful computational encoding models. For ganglion cells in the vertebrate retina, the development of satisfactory models for responses to natural visual scenes is an ongoing challenge. Standard models typically apply linear integration of visual stimuli over space, yet many ganglion cells are known to show nonlinear spatial integration, in particular when stimulated with contrast-reversing gratings. We here study the influence of spatial nonlinearities in the encoding of natural images by ganglion cells, using multielectrode-array recordings from isolated salamander and mouse retinas. We assess how responses to natural images depend on first- and second-order statistics of spatial patterns inside the receptive field. This leads us to a simple extension of current standard ganglion cell models. We show that taking not only the weighted average of light intensity inside the receptive field into account but also its variance over space can partly account for nonlinear integration and substantially improve response predictions of responses to novel images. For salamander ganglion cells, we find that response predictions for cell classes with large receptive fields profit most from including spatial contrast information. Finally, we demonstrate how this model framework can be used to assess the spatial scale of nonlinear integration. Our results underscore that nonlinear spatial stimulus integration translates to stimulation with natural images. Furthermore, the introduced model framework provides a simple, yet powerful extension of standard models and may serve as a benchmark for the development of more detailed models of the nonlinear structure of receptive fields. For understanding how sensory systems operate in the natural environment, an important goal is to develop models that capture neuronal responses to natural stimuli. For retinal ganglion cells, which connect the eye to the brain, current standard models often fail to capture responses to natural visual scenes. This shortcoming is at least partly rooted in the fact that ganglion cells may combine visual signals over space in a nonlinear fashion. We here show that a simple model, which not only considers the average light intensity inside a cell’s receptive field but also the variance of light intensity over space, can partly account for these nonlinearities and thereby improve current standard models. This provides an easy-to-obtain benchmark for modeling ganglion cell responses to natural images.
Collapse
Affiliation(s)
- Jian K. Liu
- University Medical Center Göttingen, Department of Ophthalmology, Göttingen, Germany
- Bernstein Center for Computational Neuroscience Göttingen, Göttingen, Germany
- School of Computing, University of Leeds, Leeds, United Kingdom
| | - Dimokratis Karamanlis
- University Medical Center Göttingen, Department of Ophthalmology, Göttingen, Germany
- Bernstein Center for Computational Neuroscience Göttingen, Göttingen, Germany
- International Max Planck Research School for Neurosciences, Göttingen, Germany
| | - Tim Gollisch
- University Medical Center Göttingen, Department of Ophthalmology, Göttingen, Germany
- Bernstein Center for Computational Neuroscience Göttingen, Göttingen, Germany
- Cluster of Excellence “Multiscale Bioimaging: from Molecular Machines to Networks of Excitable Cells” (MBExC), University of Göttingen, Göttingen, Germany
- * E-mail:
| |
Collapse
|
32
|
Jia S, Li X, Huang T, Liu JK, Yu Z. Representing the dynamics of high-dimensional data with non-redundant wavelets. PATTERNS 2022; 3:100424. [PMID: 35510192 PMCID: PMC9058841 DOI: 10.1016/j.patter.2021.100424] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/26/2021] [Revised: 09/22/2021] [Accepted: 12/09/2021] [Indexed: 11/19/2022]
Abstract
A crucial question in data science is to extract meaningful information embedded in high-dimensional data into a low-dimensional set of features that can represent the original data at different levels. Wavelet analysis is a pervasive method for decomposing time-series signals into a few levels with detailed temporal resolution. However, obtained wavelets are intertwined and over-represented across levels for each sample and across different samples within one population. Here, using neuroscience data of simulated spikes, experimental spikes, calcium imaging signals, and human electrocorticography signals, we leveraged conditional mutual information between wavelets for feature selection. The meaningfulness of selected features was verified to decode stimulus or condition with high accuracy yet using only a small set of features. These results provide a new way of wavelet analysis for extracting essential features of the dynamics of spatiotemporal neural data, which then enables to support novel model design of machine learning with representative features. WCMI can extract meaningful information from high-dimensional data Extracted features from neural signals are non-redundant Simple decoders can read out these features with superb accuracy
One of the essential questions in data science is to extract meaningful information from high-dimensional data. A useful approach is to represent data using a few features that maintain the crucial information. The leading property of spatiotemporal data is foremost ever-changing dynamics in time. Wavelet analysis, as a classical method for disentangling time series, can capture temporal dynamics with detail. Here, we leveraged conditional mutual information between wavelets to select a small subset of non-redundant features. We demonstrated the efficiency and effectiveness of features using various types of neuroscience data with different sampling frequencies at the level of the single cell, cell population, and coarse-scale brain activity. Our results shed new insights into representing the dynamics of spatiotemporal data using a few fundamental features extracted by wavelet analysis, which may have wide implications to other types of data with rich temporal dynamics.
Collapse
|
33
|
Kancharla P, Channappayya SS. Completely Blind Quality Assessment of User Generated Video Content. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 31:263-274. [PMID: 34855597 DOI: 10.1109/tip.2021.3130541] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
In this work, we address the challenging problem of completely blind video quality assessment (BVQA) of user generated content (UGC). The challenge is twofold since the quality prediction model is oblivious of human opinion scores, and there are no well-defined distortion models for UGC content. Our solution is inspired by a recent computational neuroscience model which hypothesizes that the human visual system (HVS) transforms a natural video input to follow a straighter temporal trajectory in the perceptual domain. A bandpass filter based computational model of the lateral geniculate nucleus (LGN) and V1 regions of the HVS was used to validate the perceptual straightening hypothesis. We hypothesize that distortions in natural videos lead to loss in straightness (or increased curvature) in their transformed representations in the HVS. We provide extensive empirical evidence to validate our hypothesis. We quantify the loss in straightness as a measure of temporal quality, and show that this measure delivers acceptable quality prediction performance on its own. Further, the temporal quality measure is combined with a state-of-the-art blind spatial (image) quality metric to design a blind video quality predictor that we call STraightness Evaluation Metric (STEM). STEM is shown to deliver state-of-the-art performance over the class of BVQA algorithms on five UGC VQA datasets including KoNViD-1K, LIVE-Qualcomm, LIVE-VQC, CVD and YouTube-UGC. Importantly, our solution is completely blind i.e., training-free, generalizes very well, is explainable, has few tunable parameters, and is simple and easy to implement.
Collapse
|
34
|
O'Hare L, Hird E, Whybrow M. Steady-state visual evoked potential responses predict visual discomfort judgements. Eur J Neurosci 2021; 54:7575-7598. [PMID: 34661322 DOI: 10.1111/ejn.15492] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2021] [Revised: 10/07/2021] [Accepted: 10/08/2021] [Indexed: 11/29/2022]
Abstract
It has been suggested that aesthetically pleasing stimuli are processed efficiently by the visual system, whereas uncomfortable stimuli are processed inefficiently. This study consists of a series of three experiments investigating this idea using a range of images of abstract artworks, photographs of natural scenes, and computer-generated stimuli previously shown to be uncomfortable. Subjective judgements and neural correlates were measured using electroencephalogram (EEG) (steady-state visual evoked potentials, SSVEPs). In addition, global image statistics (contrast, Fourier amplitude spectral slope and fractal dimension) were taken into account. When effects of physical image contrast were controlled, fractal dimension predicted discomfort judgements, suggesting the SSVEP response is more likely to be influenced by distribution of edges than the spectral slope. Importantly, when effects of physical contrast and fractal dimension were accounted for using linear mixed effects modelling, SSVEP responses predicted subjective judgements of images. Specifically, when stimuli were not matched for perceived contrast, there was a positive relationship between SSVEP responses and how pleasing a stimulus was judged to be, and conversely a negative relationship between discomfort and SSVEP response. This is significant as it shows that the neural responses in early visual areas contribute to the subjective (un)pleasantness of images, although the results of this study do not provide clear support for the theory of efficient coding as the cause of perceived pleasantness or discomfort of images, and so other explanations need to be considered.
Collapse
Affiliation(s)
- Louise O'Hare
- School of Psychology, University of Lincoln, Lincoln, UK.,Department of Psychology, Nottingham Trent University, Nottingham, UK
| | - Emily Hird
- School of Psychology, University of Lincoln, Lincoln, UK
| | | |
Collapse
|
35
|
Van Zuijlen MJP, Lin H, Bala K, Pont SC, Wijntjes MWA. Materials In Paintings (MIP): An interdisciplinary dataset for perception, art history, and computer vision. PLoS One 2021; 16:e0255109. [PMID: 34437544 PMCID: PMC8389402 DOI: 10.1371/journal.pone.0255109] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2020] [Accepted: 07/09/2021] [Indexed: 11/18/2022] Open
Abstract
In this paper, we capture and explore the painterly depictions of materials to enable the study of depiction and perception of materials through the artists' eye. We annotated a dataset of 19k paintings with 200k+ bounding boxes from which polygon segments were automatically extracted. Each bounding box was assigned a coarse material label (e.g., fabric) and half was also assigned a fine-grained label (e.g., velvety, silky). The dataset in its entirety is available for browsing and downloading at materialsinpaintings.tudelft.nl. We demonstrate the cross-disciplinary utility of our dataset by presenting novel findings across human perception, art history and, computer vision. Our experiments include a demonstration of how painters create convincing depictions using a stylized approach. We further provide an analysis of the spatial and probabilistic distributions of materials depicted in paintings, in which we for example show that strong patterns exists for material presence and location. Furthermore, we demonstrate how paintings could be used to build more robust computer vision classifiers by learning a more perceptually relevant feature representation. Additionally, we demonstrate that training classifiers on paintings could be used to uncover hidden perceptual cues by visualizing the features used by the classifiers. We conclude that our dataset of painterly material depictions is a rich source for gaining insights into the depiction and perception of materials across multiple disciplines and hope that the release of this dataset will drive multidisciplinary research.
Collapse
Affiliation(s)
| | - Hubert Lin
- Computer Science Department, Cornell University, Ithaca, New York, United States of America
| | - Kavita Bala
- Computer Science Department, Cornell University, Ithaca, New York, United States of America
| | - Sylvia C Pont
- Perceptual Intelligence Lab, Delft University of Technology, Delft, The Netherlands
| | - Maarten W A Wijntjes
- Perceptual Intelligence Lab, Delft University of Technology, Delft, The Netherlands
| |
Collapse
|
36
|
Klimmasch L, Schneider J, Lelais A, Fronius M, Shi BE, Triesch J. The development of active binocular vision under normal and alternate rearing conditions. eLife 2021; 10:e56212. [PMID: 34402429 PMCID: PMC8445622 DOI: 10.7554/elife.56212] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2020] [Accepted: 08/04/2021] [Indexed: 12/18/2022] Open
Abstract
The development of binocular vision is an active learning process comprising the development of disparity tuned neurons in visual cortex and the establishment of precise vergence control of the eyes. We present a computational model for the learning and self-calibration of active binocular vision based on the Active Efficient Coding framework, an extension of classic efficient coding ideas to active perception. Under normal rearing conditions with naturalistic input, the model develops disparity tuned neurons and precise vergence control, allowing it to correctly interpret random dot stereograms. Under altered rearing conditions modeled after neurophysiological experiments, the model qualitatively reproduces key experimental findings on changes in binocularity and disparity tuning. Furthermore, the model makes testable predictions regarding how altered rearing conditions impede the learning of precise vergence control. Finally, the model predicts a surprising new effect that impaired vergence control affects the statistics of orientation tuning in visual cortical neurons.
Collapse
Affiliation(s)
- Lukas Klimmasch
- Frankfurt Institute for Advanced Studies (FIAS)Frankfurt am MainGermany
| | - Johann Schneider
- Frankfurt Institute for Advanced Studies (FIAS)Frankfurt am MainGermany
| | - Alexander Lelais
- Frankfurt Institute for Advanced Studies (FIAS)Frankfurt am MainGermany
| | - Maria Fronius
- Department of Ophthalmology, Child Vision Research Unit, Goethe UniversityFrankfurt am MainGermany
| | - Bertram Emil Shi
- Department of Electronic and Computer Engineering, Hong Kong University of Science and TechnologyHong KongChina
| | - Jochen Triesch
- Frankfurt Institute for Advanced Studies (FIAS)Frankfurt am MainGermany
| |
Collapse
|
37
|
Okada K, Motoyoshi I. Human Texture Vision as Multi-Order Spectral Analysis. Front Comput Neurosci 2021; 15:692334. [PMID: 34381346 PMCID: PMC8349988 DOI: 10.3389/fncom.2021.692334] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/08/2021] [Accepted: 06/28/2021] [Indexed: 11/13/2022] Open
Abstract
Texture information plays a critical role in the rapid perception of scenes, objects, and materials. Here, we propose a novel model in which visual texture perception is essentially determined by the 1st-order (2D-luminance) and 2nd-order (4D-energy) spectra. This model is an extension of the dimensionality of the Filter-Rectify-Filter (FRF) model, and it also corresponds to the frequency representation of the Portilla-Simoncelli (PS) statistics. We show that preserving two spectra and randomizing phases of a natural texture image result in a perceptually similar texture, strongly supporting the model. Based on only two single spectral spaces, this model provides a simpler framework to describe and predict texture representations in the primate visual system. The idea of multi-order spectral analysis is consistent with the hierarchical processing principle of the visual cortex, which is approximated by a multi-layer convolutional network.
Collapse
Affiliation(s)
- Kosuke Okada
- Department of Life Sciences, The University of Tokyo, Tokyo, Japan
| | - Isamu Motoyoshi
- Department of Life Sciences, The University of Tokyo, Tokyo, Japan
| |
Collapse
|
38
|
Siegle JH, Ledochowitsch P, Jia X, Millman DJ, Ocker GK, Caldejon S, Casal L, Cho A, Denman DJ, Durand S, Groblewski PA, Heller G, Kato I, Kivikas S, Lecoq J, Nayan C, Ngo K, Nicovich PR, North K, Ramirez TK, Swapp J, Waughman X, Williford A, Olsen SR, Koch C, Buice MA, de Vries SEJ. Reconciling functional differences in populations of neurons recorded with two-photon imaging and electrophysiology. eLife 2021; 10:e69068. [PMID: 34270411 PMCID: PMC8285106 DOI: 10.7554/elife.69068] [Citation(s) in RCA: 21] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2021] [Accepted: 07/02/2021] [Indexed: 11/20/2022] Open
Abstract
Extracellular electrophysiology and two-photon calcium imaging are widely used methods for measuring physiological activity with single-cell resolution across large populations of cortical neurons. While each of these two modalities has distinct advantages and disadvantages, neither provides complete, unbiased information about the underlying neural population. Here, we compare evoked responses in visual cortex recorded in awake mice under highly standardized conditions using either imaging of genetically expressed GCaMP6f or electrophysiology with silicon probes. Across all stimulus conditions tested, we observe a larger fraction of responsive neurons in electrophysiology and higher stimulus selectivity in calcium imaging, which was partially reconciled by applying a spikes-to-calcium forward model to the electrophysiology data. However, the forward model could only reconcile differences in responsiveness when restricted to neurons with low contamination and an event rate above a minimum threshold. This work established how the biases of these two modalities impact functional metrics that are fundamental for characterizing sensory-evoked responses.
Collapse
Affiliation(s)
| | | | - Xiaoxuan Jia
- MindScope Program, Allen InstituteSeattleUnited States
| | | | | | | | - Linzy Casal
- MindScope Program, Allen InstituteSeattleUnited States
| | - Andy Cho
- MindScope Program, Allen InstituteSeattleUnited States
| | - Daniel J Denman
- Allen Institute for Brain Science, Allen InstituteSeattleUnited States
| | | | | | - Gregg Heller
- MindScope Program, Allen InstituteSeattleUnited States
| | - India Kato
- MindScope Program, Allen InstituteSeattleUnited States
| | - Sara Kivikas
- MindScope Program, Allen InstituteSeattleUnited States
| | - Jérôme Lecoq
- MindScope Program, Allen InstituteSeattleUnited States
| | - Chelsea Nayan
- MindScope Program, Allen InstituteSeattleUnited States
| | - Kiet Ngo
- Allen Institute for Brain Science, Allen InstituteSeattleUnited States
| | - Philip R Nicovich
- Allen Institute for Brain Science, Allen InstituteSeattleUnited States
| | - Kat North
- MindScope Program, Allen InstituteSeattleUnited States
| | | | - Jackie Swapp
- MindScope Program, Allen InstituteSeattleUnited States
| | - Xana Waughman
- MindScope Program, Allen InstituteSeattleUnited States
| | - Ali Williford
- MindScope Program, Allen InstituteSeattleUnited States
| | - Shawn R Olsen
- MindScope Program, Allen InstituteSeattleUnited States
| | - Christof Koch
- MindScope Program, Allen InstituteSeattleUnited States
| | | | | |
Collapse
|
39
|
Goutcher R, Barrington C, Hibbard PB, Graham B. Binocular vision supports the development of scene segmentation capabilities: Evidence from a deep learning model. J Vis 2021; 21:13. [PMID: 34289490 PMCID: PMC8300045 DOI: 10.1167/jov.21.7.13] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2021] [Accepted: 06/16/2021] [Indexed: 11/24/2022] Open
Abstract
The application of deep learning techniques has led to substantial progress in solving a number of critical problems in machine vision, including fundamental problems of scene segmentation and depth estimation. Here, we report a novel deep neural network model, capable of simultaneous scene segmentation and depth estimation from a pair of binocular images. By manipulating the arrangement of binocular image pairs, presenting the model with standard left-right image pairs, identical image pairs or swapped left-right images, we show that performance levels depend on the presence of appropriate binocular image arrangements. Segmentation and depth estimation performance are both impaired when images are swapped. Segmentation performance levels are maintained, however, for identical image pairs, despite the absence of binocular disparity information. Critically, these performance levels exceed those found for an equivalent, monocularly trained, segmentation model. These results provide evidence that binocular image differences support both the direct recovery of depth and segmentation information, and the enhanced learning of monocular segmentation signals. This finding suggests that binocular vision may play an important role in visual development. Better understanding of this role may hold implications for the study and treatment of developmentally acquired perceptual impairments.
Collapse
Affiliation(s)
- Ross Goutcher
- Psychology Division, Faculty of Natural Sciences, University of Stirling, Stirling, UK
| | - Christian Barrington
- Psychology Division, Faculty of Natural Sciences, University of Stirling, Stirling, UK
- Computing Science and Mathematics Division, Faculty of Natural Sciences, University of Stirling, Stirling, UK
| | - Paul B Hibbard
- Department of Psychology, University of Essex, Colchester, UK
| | - Bruce Graham
- Computing Science and Mathematics Division, Faculty of Natural Sciences, University of Stirling, Stirling, UK
| |
Collapse
|
40
|
Kuroki 黒木 忍 S, Sawayama 澤山 正貴 M, Nishida 西田 眞也 S. The roles of lower- and higher-order surface statistics in tactile texture perception. J Neurophysiol 2021; 126:95-111. [PMID: 34038163 DOI: 10.1152/jn.00577.2020] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Humans can haptically discriminate surface textures when there is a significant difference in the statistics of the surface profile. Previous studies on tactile texture discrimination have emphasized the perceptual effects of lower-order statistical features such as carving depth, inter-ridge distance, and anisotropy, which can be characterized by local amplitude spectra or spatial-frequency/orientation subband histograms. However, the real-world surfaces we encounter in everyday life also differ in the higher-order statistics, such as statistics about correlations of nearby spatial-frequencies/orientations. For another modality, vision, the human brain has the ability to use the textural differences in both higher- and lower-order image statistics. In this work, we examined whether the haptic texture perception can use higher-order surface statistics as visual texture perception does, by three-dimensional (3-D)-printing textured surfaces transcribed from different "photos" of natural scenes such as stones and leaves. Even though the maximum carving depth was well above the haptic detection threshold, some texture pairs were hard to discriminate. Specifically, those texture pairs with similar amplitude spectra were difficult to discriminate, which suggests that the lower-order statistics have the dominant effect on tactile texture discrimination. To directly test the poor sensitivity of the tactile texture perception to higher-order surface statistics, we matched the lower-order statistics across different textures using a texture synthesis algorithm and found that haptic discrimination of the matched textures was nearly impossible unless the stimuli contained salient local features. We found no evidence for the ability of the human tactile system to use higher-order surface statistics for texture discrimination.NEW & NOTEWORTHY Humans can discriminate subtle spatial patterns differences in the surrounding world through their hands, but the underlying computation remains poorly understood. Here, we 3-D-printed textured surfaces and analyzed the tactile discrimination performance regarding the sensitivity to surface statistics. The results suggest that observers have sensitivity to lower-order statistics whereas not to higher-order statistics. That is, touch differs from vision not only in spatiotemporal resolution but also in (in)sensitivity to high-level surface statistics.
Collapse
Affiliation(s)
| | - Masataka Sawayama 澤山 正貴
- NTT Communication Science Laboratories, NTT Corporation, Atsugi, Japan.,Inria, Bordeaux, France
| | - Shin'ya Nishida 西田 眞也
- NTT Communication Science Laboratories, NTT Corporation, Atsugi, Japan.,Department of Intelligence Science and Technology, Graduate School of Informatics, Kyoto University, Kyoto, Japan
| |
Collapse
|
41
|
Castellotti S, Montagnini A, Del Viva MM. Early Visual Saliency Based on Isolated Optimal Features. Front Neurosci 2021; 15:645743. [PMID: 33994923 PMCID: PMC8120310 DOI: 10.3389/fnins.2021.645743] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2020] [Accepted: 04/06/2021] [Indexed: 12/02/2022] Open
Abstract
Under fast viewing conditions, the visual system extracts salient and simplified representations of complex visual scenes. Saccadic eye movements optimize such visual analysis through the dynamic sampling of the most informative and salient regions in the scene. However, a general definition of saliency, as well as its role for natural active vision, is still a matter for discussion. Following the general idea that visual saliency may be based on the amount of local information, a recent constrained maximum-entropy model of early vision, applied to natural images, extracts a set of local optimal information-carriers, as candidate salient features. These optimal features proved to be more informative than others in fast vision, when embedded in simplified sketches of natural images. In the present study, for the first time, these features were presented in isolation, to investigate whether they can be visually more salient than other non-optimal features, even in the absence of any meaningful global arrangement (contour, line, etc.). In four psychophysics experiments, fast discriminability of a compound of optimal features (target) in comparison with a similar compound of non-optimal features (distractor) was measured as a function of their number and contrast. Results showed that the saliency predictions from the constrained maximum-entropy model are well verified in the data, even when the optimal features are presented in smaller numbers or at lower contrast. In the eye movements experiment, the target and the distractor compounds were presented in the periphery at different angles. Participants were asked to perform a simple choice-saccade task. Results showed that saccades can select informative optimal features spatially interleaved with non-optimal features even at the shortest latencies. Saccades’ choice accuracy and landing position precision improved with SNR. In conclusion, the optimal features predicted by the reference model, turn out to be more salient than others, despite the lack of any clues coming from a global meaningful structure, suggesting that they get preferential treatment during fast image analysis. Also, peripheral fast visual processing of these informative local features is able to guide gaze orientation. We speculate that active vision is efficiently adapted to maximize information in natural visual scenes.
Collapse
Affiliation(s)
| | - Anna Montagnini
- Institut de Neurosciences de la Timone (UMR 7289), CNRS and Aix-Marseille Université, Marseille, France
| | | |
Collapse
|
42
|
Tsekouras GE, Rigos A, Chatzistamatis S, Tsimikas J, Kotis K, Caridakis G, Anagnostopoulos CN. A Novel Approach to Image Recoloring for Color Vision Deficiency. SENSORS (BASEL, SWITZERLAND) 2021; 21:2740. [PMID: 33924510 PMCID: PMC8069325 DOI: 10.3390/s21082740] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/15/2021] [Revised: 04/07/2021] [Accepted: 04/09/2021] [Indexed: 11/16/2022]
Abstract
In this paper, a novel method to modify color images for the protanopia and deuteranopia color vision deficiencies is proposed. The method admits certain criteria, such as preserving image naturalness and color contrast enhancement. Four modules are employed in the process. First, fuzzy clustering-based color segmentation extracts key colors (which are the cluster centers) of the input image. Second, the key colors are mapped onto the CIE 1931 chromaticity diagram. Then, using the concept of confusion line (i.e., loci of colors confused by the color-blind), a sophisticated mechanism translates (i.e., removes) key colors lying on the same confusion line to different confusion lines so that they can be discriminated by the color-blind. In the third module, the key colors are further adapted by optimizing a regularized objective function that combines the aforementioned criteria. Fourth, the recolored image is obtained by color transfer that involves the adapted key colors and the associated fuzzy clusters. Three related methods are compared with the proposed one, using two performance indices, and evaluated by several experiments over 195 natural images and six digitized art paintings. The main outcomes of the comparative analysis are as follows. (a) Quantitative evaluation based on nonparametric statistical analysis is conducted by comparing the proposed method to each one of the other three methods for protanopia and deuteranopia, and for each index. In most of the comparisons, the Bonferroni adjusted p-values are <0.015, favoring the superiority of the proposed method. (b) Qualitative evaluation verifies the aesthetic appearance of the recolored images.
Collapse
Affiliation(s)
- George E. Tsekouras
- Department of Cultural Technology and Communications, University of the Aegean, 811 00 Mitilini, Greece; (A.R.); (S.C.); (K.K.); (G.C.); (C.-N.A.)
| | - Anastasios Rigos
- Department of Cultural Technology and Communications, University of the Aegean, 811 00 Mitilini, Greece; (A.R.); (S.C.); (K.K.); (G.C.); (C.-N.A.)
| | - Stamatis Chatzistamatis
- Department of Cultural Technology and Communications, University of the Aegean, 811 00 Mitilini, Greece; (A.R.); (S.C.); (K.K.); (G.C.); (C.-N.A.)
| | - John Tsimikas
- Department of Statistics and Actuarial-Financial Mathematics, University of the Aegean, 811 00 Mitilini, Greece;
| | - Konstantinos Kotis
- Department of Cultural Technology and Communications, University of the Aegean, 811 00 Mitilini, Greece; (A.R.); (S.C.); (K.K.); (G.C.); (C.-N.A.)
| | - George Caridakis
- Department of Cultural Technology and Communications, University of the Aegean, 811 00 Mitilini, Greece; (A.R.); (S.C.); (K.K.); (G.C.); (C.-N.A.)
| | - Christos-Nikolaos Anagnostopoulos
- Department of Cultural Technology and Communications, University of the Aegean, 811 00 Mitilini, Greece; (A.R.); (S.C.); (K.K.); (G.C.); (C.-N.A.)
| |
Collapse
|
43
|
Nonlinear Spatial Integration Underlies the Diversity of Retinal Ganglion Cell Responses to Natural Images. J Neurosci 2021; 41:3479-3498. [PMID: 33664129 PMCID: PMC8051676 DOI: 10.1523/jneurosci.3075-20.2021] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2020] [Revised: 02/05/2021] [Accepted: 02/09/2021] [Indexed: 02/06/2023] Open
Abstract
How neurons encode natural stimuli is a fundamental question for sensory neuroscience. In the early visual system, standard encoding models assume that neurons linearly filter incoming stimuli through their receptive fields, but artificial stimuli, such as contrast-reversing gratings, often reveal nonlinear spatial processing. We investigated to what extent such nonlinear processing is relevant for the encoding of natural images in retinal ganglion cells in mice of either sex. How neurons encode natural stimuli is a fundamental question for sensory neuroscience. In the early visual system, standard encoding models assume that neurons linearly filter incoming stimuli through their receptive fields, but artificial stimuli, such as contrast-reversing gratings, often reveal nonlinear spatial processing. We investigated to what extent such nonlinear processing is relevant for the encoding of natural images in retinal ganglion cells in mice of either sex. We found that standard linear receptive field models yielded good predictions of responses to flashed natural images for a subset of cells but failed to capture the spiking activity for many others. Cells with poor model performance displayed pronounced sensitivity to fine spatial contrast and local signal rectification as the dominant nonlinearity. By contrast, sensitivity to high-frequency contrast-reversing gratings, a classical test for nonlinear spatial integration, was not a good predictor of model performance and thus did not capture the variability of nonlinear spatial integration under natural images. In addition, we also observed a class of nonlinear ganglion cells with inverse tuning for spatial contrast, responding more strongly to spatially homogeneous than to spatially structured stimuli. These findings highlight the diversity of receptive field nonlinearities as a crucial component for understanding early sensory encoding in the context of natural stimuli. SIGNIFICANCE STATEMENT Experiments with artificial visual stimuli have revealed that many types of retinal ganglion cells pool spatial input signals nonlinearly. However, it is still unclear how relevant this nonlinear spatial integration is when the input signals are natural images. Here we analyze retinal responses to natural scenes in large populations of mouse ganglion cells. We show that nonlinear spatial integration strongly influences responses to natural images for some ganglion cells, but not for others. Cells with nonlinear spatial integration were sensitive to spatial structure inside their receptive fields, and a small group of cells displayed a surprising sensitivity to spatially homogeneous stimuli. Traditional analyses with contrast-reversing gratings did not predict this variability of nonlinear spatial integration under natural images.
Collapse
|
44
|
Nigam S, Pojoga S, Dragoi V. A distinct population of heterogeneously color-tuned neurons in macaque visual cortex. SCIENCE ADVANCES 2021; 7:7/8/eabc5837. [PMID: 33608266 PMCID: PMC7895441 DOI: 10.1126/sciadv.abc5837] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/01/2020] [Accepted: 01/06/2021] [Indexed: 06/12/2023]
Abstract
Color is a key feature of natural environments that higher mammals routinely use to detect food, avoid predators, and interpret social signals. The distribution of color signals in natural scenes is widely variable, ranging from uniform patches to highly nonuniform regions in which different colors lie in close proximity. Whether individual neurons are tuned to this high degree of variability of color signals is unknown. Here, we identified a distinct population of cells in macaque visual cortex (area V4) that have a heterogeneous receptive field (RF) structure in which individual subfields are tuned to different colors even though the full RF is only weakly tuned. This spatial heterogeneity in color tuning indicates a higher degree of complexity of color-encoding mechanisms in visual cortex than previously believed to efficiently extract chromatic information from the environment.
Collapse
Affiliation(s)
- Sunny Nigam
- Department of Neurobiology and Anatomy, McGovern Medical School, University of Texas at Houston, Houston, TX 77030, USA
| | - Sorin Pojoga
- Department of Neurobiology and Anatomy, McGovern Medical School, University of Texas at Houston, Houston, TX 77030, USA
| | - Valentin Dragoi
- Department of Neurobiology and Anatomy, McGovern Medical School, University of Texas at Houston, Houston, TX 77030, USA.
- Department of Electrical and Computer Engineering, Rice University, Houston, TX 77005, USA
| |
Collapse
|
45
|
Siegle JH, Jia X, Durand S, Gale S, Bennett C, Graddis N, Heller G, Ramirez TK, Choi H, Luviano JA, Groblewski PA, Ahmed R, Arkhipov A, Bernard A, Billeh YN, Brown D, Buice MA, Cain N, Caldejon S, Casal L, Cho A, Chvilicek M, Cox TC, Dai K, Denman DJ, de Vries SEJ, Dietzman R, Esposito L, Farrell C, Feng D, Galbraith J, Garrett M, Gelfand EC, Hancock N, Harris JA, Howard R, Hu B, Hytnen R, Iyer R, Jessett E, Johnson K, Kato I, Kiggins J, Lambert S, Lecoq J, Ledochowitsch P, Lee JH, Leon A, Li Y, Liang E, Long F, Mace K, Melchior J, Millman D, Mollenkopf T, Nayan C, Ng L, Ngo K, Nguyen T, Nicovich PR, North K, Ocker GK, Ollerenshaw D, Oliver M, Pachitariu M, Perkins J, Reding M, Reid D, Robertson M, Ronellenfitch K, Seid S, Slaughterbeck C, Stoecklin M, Sullivan D, Sutton B, Swapp J, Thompson C, Turner K, Wakeman W, Whitesell JD, Williams D, Williford A, Young R, Zeng H, Naylor S, Phillips JW, Reid RC, Mihalas S, Olsen SR, Koch C. Survey of spiking in the mouse visual system reveals functional hierarchy. Nature 2021; 592:86-92. [PMID: 33473216 PMCID: PMC10399640 DOI: 10.1038/s41586-020-03171-x] [Citation(s) in RCA: 181] [Impact Index Per Article: 60.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2019] [Accepted: 12/09/2020] [Indexed: 12/14/2022]
Abstract
The anatomy of the mammalian visual system, from the retina to the neocortex, is organized hierarchically1. However, direct observation of cellular-level functional interactions across this hierarchy is lacking due to the challenge of simultaneously recording activity across numerous regions. Here we describe a large, open dataset-part of the Allen Brain Observatory2-that surveys spiking from tens of thousands of units in six cortical and two thalamic regions in the brains of mice responding to a battery of visual stimuli. Using cross-correlation analysis, we reveal that the organization of inter-area functional connectivity during visual stimulation mirrors the anatomical hierarchy from the Allen Mouse Brain Connectivity Atlas3. We find that four classical hierarchical measures-response latency, receptive-field size, phase-locking to drifting gratings and response decay timescale-are all correlated with the hierarchy. Moreover, recordings obtained during a visual task reveal that the correlation between neural activity and behavioural choice also increases along the hierarchy. Our study provides a foundation for understanding coding and signal propagation across hierarchically organized cortical and thalamic visual areas.
Collapse
Affiliation(s)
| | - Xiaoxuan Jia
- Allen Institute for Brain Science, Seattle, WA, USA.
| | | | - Sam Gale
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | - Nile Graddis
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | | | - Hannah Choi
- Allen Institute for Brain Science, Seattle, WA, USA.,Department of Applied Mathematics, University of Washington, Seattle, WA, USA
| | | | | | | | | | - Amy Bernard
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | - Dillan Brown
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | - Nicolas Cain
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | - Linzy Casal
- Allen Institute for Brain Science, Seattle, WA, USA
| | - Andrew Cho
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | - Timothy C Cox
- University of Missouri-Kansas City School of Dentistry, Kansas City, MO, USA
| | - Kael Dai
- Allen Institute for Brain Science, Seattle, WA, USA
| | - Daniel J Denman
- Allen Institute for Brain Science, Seattle, WA, USA.,The University of Colorado Denver, Anschutz Medical Campus, Aurora, CO, USA
| | | | | | | | | | - David Feng
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | | | | | | | | | | | - Brian Hu
- Allen Institute for Brain Science, Seattle, WA, USA
| | - Ross Hytnen
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | | | | | - India Kato
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | | | - Jerome Lecoq
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | | | - Arielle Leon
- Allen Institute for Brain Science, Seattle, WA, USA
| | - Yang Li
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | - Fuhui Long
- Allen Institute for Brain Science, Seattle, WA, USA
| | - Kyla Mace
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | | | | | | | - Lydia Ng
- Allen Institute for Brain Science, Seattle, WA, USA
| | - Kiet Ngo
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | | | - Kat North
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | | | | | | | - Jed Perkins
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | - David Reid
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | | | - Sam Seid
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | | | | | - Ben Sutton
- Allen Institute for Brain Science, Seattle, WA, USA
| | - Jackie Swapp
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | | | | | | | | | | | - Rob Young
- Allen Institute for Brain Science, Seattle, WA, USA
| | - Hongkui Zeng
- Allen Institute for Brain Science, Seattle, WA, USA
| | - Sarah Naylor
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | - R Clay Reid
- Allen Institute for Brain Science, Seattle, WA, USA
| | | | - Shawn R Olsen
- Allen Institute for Brain Science, Seattle, WA, USA.
| | | |
Collapse
|
46
|
De A, Horwitz GD. Spatial receptive field structure of double-opponent cells in macaque V1. J Neurophysiol 2021; 125:843-857. [PMID: 33405995 DOI: 10.1152/jn.00547.2020] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022] Open
Abstract
The spatial processing of color is important for visual perception. Double-opponent (DO) cells likely contribute to this processing by virtue of their spatially opponent and cone-opponent receptive fields (RFs). However, the representation of visual features by DO cells in the primary visual cortex of primates is unclear because the spatial structure of their RFs has not been fully characterized. To fill this gap, we mapped the RFs of DO cells in awake macaques with colorful, dynamic white noise patterns. The spatial RF of each neuron was fitted with a Gabor function and three versions of the difference of Gaussians (DoG) function. The Gabor function provided the more accurate description for most DO cells, a result that is incompatible with a center-surround RF organization. A nonconcentric version of the DoG function, in which the RFs have a circular center and a crescent-shaped surround, performed nearly as well as the Gabor model thus reconciling results from previous reports. For comparison, we also measured the RFs of simple cells. We found that the superiority of the Gabor fits over DoG fits was slightly more decisive for simple cells than for DO cells. The implications of these results on biological image processing and visual perception are discussed.NEW & NOTEWORTHY Double-opponent cells in macaque area V1 respond to spatial chromatic contrast in visual scenes. What information they carry is debated because their receptive field organization has not been characterized thoroughly. Using white noise analysis and statistical model comparisons, De and Horwitz show that many double-opponent receptive fields can be captured by either a Gabor model or a center-with-an-asymmetric-surround model but not by a difference of Gaussians model.
Collapse
Affiliation(s)
- Abhishek De
- Systems Neurobiology Laboratories, Salk Institute for Biological Studies, La Jolla, California.,Department of Physiology and Biophysics, Washington National Primate Research Center, University of Washington, Seattle, Washington
| | - Gregory D Horwitz
- Department of Physiology and Biophysics, Washington National Primate Research Center, University of Washington, Seattle, Washington
| |
Collapse
|
47
|
Rifai K, Habtegiorgis SW, Erlenwein C, Wahl S. Motion-form interaction: Motion and form aftereffects induced by distorted static natural scenes. J Vis 2020; 20:10. [PMID: 33325995 PMCID: PMC7745598 DOI: 10.1167/jov.20.13.10] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Spatially varying distortions (SVDs) are common artifacts of spectacles like progressive additional lenses (PALs). To habituate to distortions of PALs, the visual system has to adapt to distortion-induced image alterations, termed skew adaptation. But how this visual adjustment is achieved is largely unknown. This study examines the properties of visual adaptation to distortions of PALs in natural scenes. The visual adaptation in response to altered form and motion features of the natural stimuli were probed in two different psychophysical experiments. Observers were exposed to distortions in natural images, and form and motion aftereffects were tested subsequently in a constant stimuli procedure where subjects were asked to judge the skew, or the motion direction of an according test stimulus. Exposure to skewed natural stimuli induced a shift in perceived undistorted form as well as motion direction, when viewing distorted dynamic natural scenes, and also after exposure to static distorted natural images. Therefore, skew adaptation occurred in form and motion for dynamic visual scenes as well as static images. Thus, specifically in the condition of static skewed images and the test feature of motion direction, cortical interactions between motion-form processing presumably contributed to the adaptation process. In a nutshell, interfeature cortical interactions constituted the adaptation process to distortion of PALs. Thus, comprehensive investigation of adaptation to distortions of PALs would benefit from taking into account content richness of the stimuli to be used, like natural images.
Collapse
Affiliation(s)
- Katharina Rifai
- Institute for Ophthalmic Research, University of Tübingen, Tübingen, Germany.,Carl Zeiss Vision International GmbH, Aalen, Germany.,
| | | | - Caroline Erlenwein
- Institute for Ophthalmic Research, University of Tübingen, Tübingen, Germany.,
| | - Siegfried Wahl
- Institute for Ophthalmic Research, University of Tübingen, Tübingen, Germany.,Carl Zeiss Vision International GmbH, Aalen, Germany.,
| |
Collapse
|
48
|
Mazade R, Jin J, Pons C, Alonso JM. Functional Specialization of ON and OFF Cortical Pathways for Global-Slow and Local-Fast Vision. Cell Rep 2020; 27:2881-2894.e5. [PMID: 31167135 DOI: 10.1016/j.celrep.2019.05.007] [Citation(s) in RCA: 26] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2018] [Revised: 03/07/2019] [Accepted: 04/30/2019] [Indexed: 12/20/2022] Open
Abstract
Visual information is processed in the cortex by ON and OFF pathways that respond to light and dark stimuli. Responses to darks are stronger, faster, and driven by a larger number of cortical neurons than responses to lights. Here, we demonstrate that these light-dark cortical asymmetries reflect a functional specialization of ON and OFF pathways for different stimulus properties. We show that large long-lasting stimuli drive stronger cortical responses when they are light, whereas small fast stimuli drive stronger cortical responses when they are dark. Moreover, we show that these light-dark asymmetries are preserved under a wide variety of luminance conditions that range from photopic to low mesopic light. Our results suggest that ON and OFF pathways extract different spatiotemporal information from visual scenes, making OFF local-fast signals better suited to maximize visual acuity and ON global-slow signals better suited to guide the eye movements needed for retinal image stabilization.
Collapse
Affiliation(s)
- Reece Mazade
- Department of Biological and Visual Sciences, SUNY College of Optometry, New York, NY 10036, USA
| | - Jianzhong Jin
- Department of Biological and Visual Sciences, SUNY College of Optometry, New York, NY 10036, USA
| | - Carmen Pons
- Department of Biological and Visual Sciences, SUNY College of Optometry, New York, NY 10036, USA
| | - Jose-Manuel Alonso
- Department of Biological and Visual Sciences, SUNY College of Optometry, New York, NY 10036, USA.
| |
Collapse
|
49
|
Seijdel N, Jahfari S, Groen IIA, Scholte HS. Low-level image statistics in natural scenes influence perceptual decision-making. Sci Rep 2020; 10:10573. [PMID: 32601499 PMCID: PMC7324621 DOI: 10.1038/s41598-020-67661-8] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2020] [Accepted: 06/08/2020] [Indexed: 11/10/2022] Open
Abstract
A fundamental component of interacting with our environment is gathering and interpretation of sensory information. When investigating how perceptual information influences decision-making, most researchers have relied on manipulated or unnatural information as perceptual input, resulting in findings that may not generalize to real-world scenes. Unlike simplified, artificial stimuli, real-world scenes contain low-level regularities that are informative about the structural complexity, which the brain could exploit. In this study, participants performed an animal detection task on low, medium or high complexity scenes as determined by two biologically plausible natural scene statistics, contrast energy (CE) or spatial coherence (SC). In experiment 1, stimuli were sampled such that CE and SC both influenced scene complexity. Diffusion modelling showed that the speed of information processing was affected by low-level scene complexity. Experiment 2a/b refined these observations by showing how isolated manipulation of SC resulted in weaker but comparable effects, with an additional change in response boundary, whereas manipulation of only CE had no effect. Overall, performance was best for scenes with intermediate complexity. Our systematic definition quantifies how natural scene complexity interacts with decision-making. We speculate that CE and SC serve as an indication to adjust perceptual decision-making based on the complexity of the input.
Collapse
Affiliation(s)
- Noor Seijdel
- Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands. .,Amsterdam Brain and Cognition (ABC) Center, University of Amsterdam, Amsterdam, The Netherlands.
| | - Sara Jahfari
- Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands.,Spinoza Centre for Neuroimaging, Royal Netherlands Academy of Arts and Sciences (KNAW), Amsterdam, The Netherlands
| | - Iris I A Groen
- Department of Psychology, New York University, New York, USA
| | - H Steven Scholte
- Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands.,Amsterdam Brain and Cognition (ABC) Center, University of Amsterdam, Amsterdam, The Netherlands
| |
Collapse
|
50
|
Otero-Millan J, Langston RE, Costela F, Macknik SL, Martinez-Conde S. Microsaccade generation requires a foveal anchor. J Eye Mov Res 2020; 12. [PMID: 33828756 PMCID: PMC7962683 DOI: 10.16910/jemr.12.6.14] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Visual scene characteristics can affect various aspects of saccade and microsaccade dynamics. For example, blank visual scenes are known to elicit diminished saccade and microsaccade production, compared to natural scenes. Similarly, microsaccades are less frequent in the dark. Yet, the extent to which foveal versus peripheral visual information contribute to microsaccade production remains unclear: because microsaccade directions are biased towards covert attention locations, it follows that peripheral visual stimulation could suffice to produce regular microsaccade dynamics, even without foveal stimulation being present. Here we determined the characteristics of microsaccades as a function of foveal and/or peripheral visual stimulation, while human subjects conducted four types of oculomotor tasks (fixation, free viewing, guided viewing and passive viewing). Foveal information was either available, or made unavailable, by the presentation of simulated scotomas. We found foveal stimulation to be critical for microsaccade production, and peripheral stimulation, by itself, to be insufficient to yield normal microsaccades. In each oculomotor task, microsaccade production decreased when scotomas blocked foveal stimulation. Across comparable foveal stimulation conditions, the type of peripheral stimulation (static versus dynamic) moreover affected microsaccade production, with dynamic backgrounds resulting in lower microsaccadic rates than static backgrounds. These results indicate that a foveal visual anchor is necessary for normal microsaccade generation. Whereas peripheral visual stimulation, on its own, does not suffice for normal microsaccade production, it can nevertheless modulate microsaccadic characteristics. These findings extend our current understanding of the links between visual input and ocular motor control, and may therefore help improve the diagnosis and treatment of ophthalmic conditions that degrade central vision, such as age-related macular degeneration.
Collapse
|