1
|
Deng Z, Gao J, Li T, Chen Y, Gao B, Fang F, Culham JC, Chen J. Viewpoint adaptation revealed potential representational differences between 2D images and 3D objects. Cognition 2024; 251:105903. [PMID: 39126975 DOI: 10.1016/j.cognition.2024.105903] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2023] [Revised: 07/17/2024] [Accepted: 07/22/2024] [Indexed: 08/12/2024]
Abstract
For convenience and experimental control, cognitive science has relied largely on images as stimuli rather than the real, tangible objects encountered in the real world. Recent evidence suggests that the cognitive processing of images may differ from real objects, especially in the processing of spatial locations and actions, thought to be mediated by the dorsal visual stream. Perceptual and semantic processing in the ventral visual stream, however, has been assumed to be largely unaffected by the realism of objects. Several studies have found that one key difference accounting for differences between real objects and images is actability; however, less research has investigated another potential difference - the three-dimensional nature of real objects as conveyed by cues like binocular disparity. To investigate the extent to which perception is affected by the realism of a stimulus, we compared viewpoint adaptation when stimuli (a face or a kettle) were 2D (flat images without binocular disparity) vs. 3D (i.e., real, tangible objects or stereoscopic images with binocular disparity). For both faces and kettles, adaptation to 3D stimuli induced stronger viewpoint aftereffects than adaptation to 2D images when the adapting orientation was rightward. A computational model suggested that the difference in aftereffects could be explained by broader viewpoint tuning for 3D compared to 2D stimuli. Overall, our finding narrowed the gap between understanding the neural processing of visual images and real-world objects by suggesting that compared to 2D images, real and simulated 3D objects evoke more broadly tuned neural representations, which may result in stronger viewpoint invariance.
Collapse
Affiliation(s)
- Zhiqing Deng
- Center for the Study of Applied Psychology, Guangdong Key Laboratory of Mental Health and Cognitive Science, and the School of Psychology, South China Normal University, Guangzhou, Guangdong Province 510631, China
| | - Jie Gao
- Center for the Study of Applied Psychology, Guangdong Key Laboratory of Mental Health and Cognitive Science, and the School of Psychology, South China Normal University, Guangzhou, Guangdong Province 510631, China
| | - Toni Li
- Division of Emergency Medicine, Department of Medicine, University of Toronto, Toronto M5S 3H2, Canada
| | - Yan Chen
- Center for the Study of Applied Psychology, Guangdong Key Laboratory of Mental Health and Cognitive Science, and the School of Psychology, South China Normal University, Guangzhou, Guangdong Province 510631, China
| | - BoYu Gao
- College of Information Science and Technology/Cyber Security, Jinan University, Guangzhou 510632, China
| | - Fang Fang
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing 100871, China; Key Laboratory of Machine Perception (Ministry of Education), Peking University, Beijing 100871, People's Republic of China; Peking-Tsinghua Center for Life Sciences, Peking University, Beijing 100871, China; IDG/McGovern Institute for Brain Research, Peking University, Beijing 100871, China
| | - Jody C Culham
- Department of Psychology, The University of Western Ontario, London, ON N6A 5C2, Canada
| | - Juan Chen
- Center for the Study of Applied Psychology, Guangdong Key Laboratory of Mental Health and Cognitive Science, and the School of Psychology, South China Normal University, Guangzhou, Guangdong Province 510631, China; Key Laboratory of Brain, Cognition and Education Sciences (South China Normal University), Ministry of Education, China.
| |
Collapse
|
2
|
Babu AS, Scotti PS, Golomb JD. The dominance of spatial information in object identity judgments: A persistent congruency bias even amidst conflicting statistical regularities. J Exp Psychol Hum Percept Perform 2023; 49:672-686. [PMID: 37261773 PMCID: PMC10298748 DOI: 10.1037/xhp0001104] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
Previous studies have posited that spatial location plays a special role in object recognition. Notably, the "spatial congruency bias (SCB)" is a tendency to report objects as the same identity if they are presented at the same location, compared to different locations. Here we found that even when statistical regularities were manipulated in the opposite direction (objects in the same location were three times more likely to be different identities), subjects still exhibited a robust SCB (more likely to report them as the same identity). We replicated this finding across two preregistered experiments. Only in a third experiment where we explicitly informed subjects of the manipulation did the SCB disappear, though the lack of a significantly reversed bias suggests the ingrained congruency bias was not completely overcome. The inclusion of catch trials where the second object was completely masked further bolsters previous evidence that the congruency bias is perceptual, not simply a guessing strategy. These results reinforce the dominant role of spatial information during object recognition and present the SCB as a strong perceptual phenomenon that is incredibly hard to overcome even in the face of opposing regularities and explicit instruction. (PsycInfo Database Record (c) 2023 APA, all rights reserved).
Collapse
Affiliation(s)
| | | | - Julie D. Golomb
- Department of Neuroscience, The Ohio State University
- Department of Psychology, The Ohio State University
| |
Collapse
|
3
|
Meyer EM, Reynolds MR. Multidimensional Scaling of Cognitive Ability and Academic Achievement Scores. J Intell 2022; 10:jintelligence10040117. [PMID: 36547504 PMCID: PMC9785841 DOI: 10.3390/jintelligence10040117] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2022] [Revised: 11/18/2022] [Accepted: 11/21/2022] [Indexed: 12/03/2022] Open
Abstract
Multidimensional scaling (MDS) was used as an alternate multivariate procedure for investigating intelligence and academic achievement test score correlations. Correlation coefficients among Wechsler Intelligence Scale for Children, Fifth Edition (WISC-5) and Wechsler Individual Achievement Test, Third Edition (WIAT-III) validity sample scores and among Kaufman Assessment Battery for Children, Second Edition (KABC-II) and Kaufman Test of Educational Achievement, Second Edition (KTEA-2) co-norming sample scores were analyzed using multidimensional scaling (MDS). Three-dimensional MDS configurations were the best fit for interpretation in both datasets. Subtests were more clearly organized by CHC ability and academic domain instead of complexity. Auditory-linguistic, figural-visual, reading-writing, and quantitative-numeric regions were visible in all models. Results were mostly similar across different grade levels. Additional analysis with WISC-V and WIAT-III tests showed that content (verbal, numeric, figural) and response process facets (verbal, manual, paper-pencil) were also useful in explaining test locations. Two implications from this study are that caution may be needed when interpreting fluency scores across academic areas, and MDS provides more empirically based validity evidence regarding content and response mode processes.
Collapse
Affiliation(s)
- Em M. Meyer
- Department of Counseling, School Psychology and Family Science, College of Education, University of Nebraska, Kearney, NE 68849, USA
- Correspondence:
| | - Matthew R. Reynolds
- Department of Educational Psychology, School of Education and Human Sciences, University of Kansas, Lawrence, KS 66045, USA
| |
Collapse
|
4
|
Xue C, Tang Y, Wang C, Yang H, Li L. The Effects of Normal Aging, Subjective Cognitive Decline, Mild Cognitive Impairment, or Alzheimer’s Disease on Visual Search. J Alzheimers Dis 2022; 88:1639-1650. [PMID: 35811526 DOI: 10.3233/jad-220209] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Background: Alzheimer’s disease (AD) has been confirmed as an influencing factor of visual impairment, but potential concomitant effects on visual and cognitive performance are not well understood. Objective: To provide a new method for early screening of Alzheimer’s disease and further explore the theoretical mechanism of the decline of whole visual and cognitive performance in AD. Methods: We studied 60 individuals without dementia as normal control (NC), 74 individuals with subjective cognitive decline (SCD), 60 individuals with amnesia mild cognitive impairment (aMCI), and 75 patients with AD on a battery of tests designed to measure multiple aspects of basic and higher-order visual perception and cognition. All subjects performed on same visual and cognitive test batteries. Results: The results showed both of four groups, with the stimulus-presentation time being longer, the visual-search performance improved, and both the eye interest-area first fixation duration and the interest-area-fixation count increased. Particularly under the noise-masking condition, the AD group performed the worst at stimulus-presentation times between 300 and 900 ms. The aMCI group, but not the SCD group, performed worse than the NC group at the stimulus-presentation time of either 300 or 500 ms. The interest-area-fixation count was higher in all the patient groups than that in the NC group, and distinguishable between participants with AD and those with SCD or aMCI. Conclusion: The visual-search performance combined with eye-movement tracking under the noise-masking condition can be used for distinguishing AD from normal aging, SCD, and aMCI.
Collapse
Affiliation(s)
- Chuanwei Xue
- Beijing Institute for Brain Disorders, Capital Medical University, Beijing, China
- Collaborative Innovation Center for Brain Disorders, Capital Medical University, Beijing, China
- Fuwai Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Yi Tang
- Xuan Wu Hospital, Capital Medical University, Beijing, China
| | - Changming Wang
- Xuan Wu Hospital, Capital Medical University, Beijing, China
| | - Haibo Yang
- Academy of Psychology and Behavior, Tianjin Normal University, Tianjin, China
| | - Liang Li
- Beijing Institute for Brain Disorders, Capital Medical University, Beijing, China
- Collaborative Innovation Center for Brain Disorders, Capital Medical University, Beijing, China
- School of Psychological and Cognitive Sciences, Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, China
- Speech and Hearing Research Center, Key Laboratory on Machine Perception (Ministry of Education), Peking University, Beijing, China
| |
Collapse
|
5
|
Chaisilprungraung T, Park S. "Scene" from inside: The representation of Observer's space in high-level visual cortex. Neuropsychologia 2021; 161:108010. [PMID: 34454940 DOI: 10.1016/j.neuropsychologia.2021.108010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2021] [Revised: 07/30/2021] [Accepted: 08/23/2021] [Indexed: 10/20/2022]
Abstract
Human observers are remarkably adept at perceiving and interacting with visual stimuli around them. Compared to visual stimuli like objects or faces, scenes are unique in that they provide enclosures for observers. An observer looks at a scene by being physically inside the scene. The current research explored this unique observer-scene relationship by studying the neural representation of scenes' spatial boundaries. Previous studies hypothesized that scenes' boundaries were processed in sets of high-level visual cortices. Notably, the parahippocampal place area (PPA), exhibited neural sensitivity to scenes that had closed vs. open spatial boundaries (Kravitz et al., 2011; Park et al., 2011). We asked whether this sensitivity reflected the openness of landscape (e.g., forest vs. beach), or the openness of the environment immediately surrounding the observer (i.e., whether a scene was viewed from inside vs. outside a room). Across two human fMRI experiments, we found that the PPA, as well as another well-known navigation-processing area, the occipital place area (OPA), processed scenes' boundaries according to the observer's space rather than the landscape. Moreover, we found that the PPA's activation pattern was susceptible to manipulations involving mid-level perceptual properties of scenes (e.g., rectilinear pattern of window frames), while the OPA's response was not. Our results have important implications for research in visual scene processing and suggest an important role of an observer's location in representing the spatial boundary, beyond the low-level visual input of a landscape.
Collapse
Affiliation(s)
| | - Soojin Park
- Department of Psychology, Yonsei University, Seoul, South Korea.
| |
Collapse
|
6
|
Snow JC, Culham JC. The Treachery of Images: How Realism Influences Brain and Behavior. Trends Cogn Sci 2021; 25:506-519. [PMID: 33775583 PMCID: PMC10149139 DOI: 10.1016/j.tics.2021.02.008] [Citation(s) in RCA: 39] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2020] [Revised: 02/08/2021] [Accepted: 02/22/2021] [Indexed: 10/21/2022]
Abstract
Although the cognitive sciences aim to ultimately understand behavior and brain function in the real world, for historical and practical reasons, the field has relied heavily on artificial stimuli, typically pictures. We review a growing body of evidence that both behavior and brain function differ between image proxies and real, tangible objects. We also propose a new framework for immersive neuroscience to combine two approaches: (i) the traditional build-up approach of gradually combining simplified stimuli, tasks, and processes; and (ii) a newer tear-down approach that begins with reality and compelling simulations such as virtual reality to determine which elements critically affect behavior and brain processing.
Collapse
Affiliation(s)
- Jacqueline C Snow
- Department of Psychology, University of Nevada Reno, Reno, NV 89557, USA
| | - Jody C Culham
- Department of Psychology, University of Western Ontario, London, Ontario, N6A 5C2, Canada; Brain and Mind Institute, Western Interdisciplinary Research Building, University of Western Ontario, London, Ontario, N6A 3K7, Canada.
| |
Collapse
|
7
|
Leptourgos P, Bouttier V, Jardri R, Denève S. A functional theory of bistable perception based on dynamical circular inference. PLoS Comput Biol 2020; 16:e1008480. [PMID: 33315961 PMCID: PMC7769606 DOI: 10.1371/journal.pcbi.1008480] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2020] [Revised: 12/28/2020] [Accepted: 10/30/2020] [Indexed: 12/31/2022] Open
Abstract
When we face ambiguous images, the brain cannot commit to a single percept; instead, it switches between mutually exclusive interpretations every few seconds, a phenomenon known as bistable perception. While neuromechanistic models, e.g., adapting neural populations with lateral inhibition, may account for the dynamics of bistability, a larger question remains unresolved: how this phenomenon informs us on generic perceptual processes in less artificial contexts. Here, we propose that bistable perception is due to our prior beliefs being reverberated in the cortical hierarchy and corrupting the sensory evidence, a phenomenon known as “circular inference”. Such circularity could occur in a hierarchical brain where sensory responses trigger activity in higher-level areas but are also modulated by feedback projections from these same areas. We show that in the face of ambiguous sensory stimuli, circular inference can change the dynamics of the perceptual system and turn what should be an integrator of inputs into a bistable attractor switching between two highly trusted interpretations. The model captures various aspects of bistability, including Levelt’s laws and the stabilizing effects of intermittent presentation of the stimulus. Since it is related to the generic perceptual inference and belief updating mechanisms, this approach can be used to predict the tendency of individuals to form aberrant beliefs from their bistable perception behavior. Overall, we suggest that feedforward/feedback information loops in hierarchical neural networks, a phenomenon that could lead to psychotic symptoms when overly strong, could also underlie perception in nonclinical populations. In cases of high ambiguity, our perceptual system cannot commit to a single percept and switches between different interpretations, giving rise to bistable perception. In this paper we outline a computational model of bistability based on the notion of circular inference, i.e. a form of suboptimal hierarchical inference in which priors and / or sensory inputs are reverberated and over-counted. We suggest that descending loops (i.e. reverberated priors) transform our perceptual system from a simple accumulator of sensory inputs into a bistable attractor, that switches between two highly-trusted interpretations. Using analytical methods we derive the necessary conditions for bistable perception to occur. We show that our dynamical circular inference model is able to capture many features of bistability, such as Levelt’s laws and the stabilizing effects of intermittent presentation of the stimulus. Finally we make novel predictions about the behavior of psychotic patients.
Collapse
Affiliation(s)
- Pantelis Leptourgos
- Department of Psychiatry, Connecticut Mental Health Center, Yale University, New Haven, Connecticut, United States of America
- * E-mail: (PL); (RJ)
| | - Vincent Bouttier
- Laboratoire de Neurosciences Cognitives & Computationnelles, ENS, INSERM U-960, PSL Research University, Paris, France
- Univ Lille, INSERM U-1172, Lille Neuroscience & Cognition Centre, Plasticity & SubjectivitY (PSY) team, Lille, France
| | - Renaud Jardri
- Laboratoire de Neurosciences Cognitives & Computationnelles, ENS, INSERM U-960, PSL Research University, Paris, France
- Univ Lille, INSERM U-1172, Lille Neuroscience & Cognition Centre, Plasticity & SubjectivitY (PSY) team, Lille, France
- CHU Lille, Fontan Hospital, CURE platform, Psychiatric Clinical Investigation Centre, Lille, France
- * E-mail: (PL); (RJ)
| | - Sophie Denève
- Laboratoire de Neurosciences Cognitives & Computationnelles, ENS, INSERM U-960, PSL Research University, Paris, France
| |
Collapse
|
8
|
3D shape estimation in a constraint optimization neural network. Vision Res 2020; 177:118-129. [PMID: 33152591 DOI: 10.1016/j.visres.2020.08.010] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2020] [Revised: 07/30/2020] [Accepted: 08/11/2020] [Indexed: 11/21/2022]
Abstract
One of the most important aspects of visual perception is the inference of 3D shape from a 2D retinal image of the outside world. The existence of several valid mapping functions from object to data makes this inverse problem ill-posed and therefore computationally difficult. In human vision, the retinal image is a 2D projection of the 3D world. The visual system imposes certain constraints on the family of solutions in order to uniquely and efficiently solve this inverse problem. This work specifically focused on the minimization of standard deviations of 3D angles (MSDA) for 3D perception. Our goal was to use a Deep Convolutional Neural Network based on biological principles derived from visual area V4 to achieve 3D reconstruction using constrained minimization of MSDA. We conducted an experiment using novel shapes with human subjects to collect data and test the model. The performance of the network largely agreed with how humans estimated novel 3D shapes. The results show that the constraint of MSDA in 3D shape can be implemented in a neural network and produce human-like results. Additional visual constraints can be added to the network in the future to fully test the theory of visual constraints as a basis of 3D shape perception.
Collapse
|
9
|
Starks MD, Shafer-Skelton A, Paradiso M, Martinez AM, Golomb JD. The influence of spatial location on same-different judgments of facial identity and expression. J Exp Psychol Hum Percept Perform 2020; 46:2020-78982-001. [PMID: 33090835 PMCID: PMC8641643 DOI: 10.1037/xhp0000872] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
The "spatial congruency bias" is a behavioral phenomenon where 2 objects presented sequentially are more likely to be judged as being the same object if they are presented in the same location (Golomb, Kupitz, & Thiemann, 2014), suggesting that irrelevant spatial location information may be bound to object representations. Here, we examine whether the spatial congruency bias extends to higher-level object judgments of facial identity and expression. On each trial, 2 real-world faces were sequentially presented in variable screen locations, and subjects were asked to make same-different judgments on the facial expression (Experiments 1-2) or facial identity (Experiment 3) of the stimuli. We observed a robust spatial congruency bias for judgments of facial identity, yet a more fragile one for judgments of facial expression. Subjects were more likely to judge 2 faces as displaying the same expression if they were presented in the same location (compared to in different locations), but only when the faces shared the same identity. On the other hand, a spatial congruency bias was found when subjects made judgments on facial identity, even across faces displaying different facial expressions. These findings suggest a possible difference between the binding of facial identity and facial expression to spatial location. (PsycInfo Database Record (c) 2020 APA, all rights reserved).
Collapse
Affiliation(s)
| | - Anna Shafer-Skelton
- Department of Psychology, The Ohio State University
- Department of Psychology, University of California, San Diego
| | | | - Aleix M. Martinez
- Department of Electrical and Computer Engineering, The Ohio State University
| | - Julie D. Golomb
- Department of Psychology, The Ohio State University
- Department of Neuroscience, The Ohio State University
| |
Collapse
|
10
|
Abstract
Working memory is considered as a cognitive memory buffer for temporarily holding, processing, and manipulating information. Although working memory for verbal and visual information has been studied extensively in the past literature, few studies have systematically investigated how depth information is stored in working memory. Here, we show that the memory performance for detecting changes in stereoscopic depth is low when there is no change in relative depth order, and the performance is reliably better when depth order is changed. Increasing the magnitude of change only improves memory performance when depth order is kept constant. However, if depth order is changed, the performance remains high, even with a small change magnitude. Our findings suggest that relative depth order is a better indicator for working memory performance than absolute metric depth. The memory representation for individual depth is not independent, but inherently relational, revealing a fundamental organizing principle for depth information in the visual system.
Collapse
|
11
|
Coding of Navigational Distance and Functional Constraint of Boundaries in the Human Scene-Selective Cortex. J Neurosci 2020; 40:3621-3630. [PMID: 32209608 DOI: 10.1523/jneurosci.1991-19.2020] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2019] [Revised: 02/28/2020] [Accepted: 03/05/2020] [Indexed: 11/21/2022] Open
Abstract
For visually guided navigation, the use of environmental cues is essential. Particularly, detecting local boundaries that impose limits to locomotion and estimating their location is crucial. In a series of three fMRI experiments, we investigated whether there is a neural coding of navigational distance in the human visual cortex (both female and male). We used virtual reality software to systematically manipulate the distance from a viewer perspective to different types of a boundary. Using a multivoxel pattern classification employing a linear support vector machine, we found that the occipital place area (OPA) is sensitive to the navigational distance restricted by the transparent glass wall. Further, the OPA was sensitive to a non-crossable boundary only, suggesting an importance of the functional constraint of a boundary. Together, we propose the OPA as a perceptual source of external environmental features relevant for navigation.SIGNIFICANCE STATEMENT One of major goals in cognitive neuroscience has been to understand the nature of visual scene representation in human ventral visual cortex. An aspect of scene perception that has been overlooked despite its ecological importance is the analysis of space for navigation. One of critical computation necessary for navigation is coding of distance to environmental boundaries that impose limit on navigator's movements. This paper reports the first empirical evidence for coding of navigational distance in the human visual cortex and its striking sensitivity to functional constraint of environmental boundaries. Such finding links the paper to previous neurological and behavioral works that emphasized the distance to boundaries as a crucial geometric property for reorientation behavior of children and other animal species.
Collapse
|
12
|
Working memory for stereoscopic depth is limited and imprecise-evidence from a change detection task. Psychon Bull Rev 2020; 26:1657-1665. [PMID: 31388836 DOI: 10.3758/s13423-019-01640-7] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Most studies on visual working memory (VWM) and spatial working memory (SWM) have employed visual stimuli presented at the fronto-parallel plane and few have involved depth perception. VWM is often considered as a memory buffer for temporarily holding and manipulating visual information that relates to visual features of an object, and SWM for holding and manipulating spatial information that concerns the spatial location of an object. Although previous research has investigated the effect of stereoscopic depth on VWM, the question of how depth positions are stored in working memory has not been systematically investigated, leaving gaps in the existing literature on working memory. Here, we explore working memory for depth by using a change detection task. The memory items were presented at various stereoscopic depth planes perpendicular to the line of sight, with one item per depth plane. Participants were asked to make judgments on whether the depth position of the target (one of the memory items) had changed. The results showed a conservative response bias that observers tended to make 'no change' responses when detecting changes in depth. In addition, we found that similar to VWM, the change detection accuracy degraded with the number of memory items presented, but the accuracy was much lower than that reported for VWM, suggesting that the storage for depth information is severely limited and less precise than that for visual information. The detection sensitivity was higher for the nearest and farthest depths and was better when the probe was presented along with the other items originally in the memory array, indicating that how well the to-be-stored depth can be stored in working memory depends on its relation with the other depth positions.
Collapse
|
13
|
Liu C, Li Y, Song S, Zhang J. Decoding disparity categories in 3-dimensional images from fMRI data using functional connectivity patterns. Cogn Neurodyn 2019; 14:169-179. [PMID: 32226560 DOI: 10.1007/s11571-019-09557-6] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2018] [Revised: 09/05/2019] [Accepted: 09/29/2019] [Indexed: 02/02/2023] Open
Abstract
Humans use binocular disparity to extract depth information from two-dimensional retinal images in a process called stereopsis. Previous studies usually introduce the standard univariate analysis to describe the correlation between disparity level and brain activity within a given brain region based on functional magnetic resonance imaging (fMRI) data. Recently, multivariate pattern analysis has been developed to extract activity patterns across multiple voxels for deciphering categories of binocular disparity. However, the functional connectivity (FC) of patterns based on regions of interest or voxels and their mapping onto disparity category perception remain unknown. The present study extracted functional connectivity patterns for three disparity conditions (crossed disparity, uncrossed disparity, and zero disparity) at distinct spatial scales to decode the binocular disparity. Results of 27 subjects' fMRI data demonstrate that FC features are more discriminatory than traditional voxel activity features in binocular disparity classification. The average binary classification of the whole brain and visual areas are respectively 87% and 79% at single subject level, and thus above the chance level (50%). Our research highlights the importance of exploring functional connectivity patterns to achieve a novel understanding of 3D image processing.
Collapse
Affiliation(s)
- Chunyu Liu
- 1College of Information Science and Technology, Beijing Normal University, Beijing, China
| | - Yuan Li
- 2School of Electrical and Information Engineering, Tianjin University, Tianjin, China
| | - Sutao Song
- 3School of Education and Psychology, University of Jinan, Jinan, China
| | - Jiacai Zhang
- 1College of Information Science and Technology, Beijing Normal University, Beijing, China
| |
Collapse
|
14
|
Multivariate Analysis of BOLD Activation Patterns Recovers Graded Depth Representations in Human Visual and Parietal Cortex. eNeuro 2019; 6:ENEURO.0362-18.2019. [PMID: 31285275 PMCID: PMC6709213 DOI: 10.1523/eneuro.0362-18.2019] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2018] [Revised: 06/24/2019] [Accepted: 06/26/2019] [Indexed: 11/21/2022] Open
Abstract
Navigating through natural environments requires localizing objects along three distinct spatial axes. Information about position along the horizontal and vertical axes is available from an object’s position on the retina, while position along the depth axis must be inferred based on second-order cues such as the disparity between the images cast on the two retinae. Past work has revealed that object position in two-dimensional (2D) retinotopic space is robustly represented in visual cortex and can be robustly predicted using a multivariate encoding model, in which an explicit axis is modeled for each spatial dimension. However, no study to date has used an encoding model to estimate a representation of stimulus position in depth. Here, we recorded BOLD fMRI while human subjects viewed a stereoscopic random-dot sphere at various positions along the depth (z) and the horizontal (x) axes, and the stimuli were presented across a wider range of disparities (out to ∼40 arcmin) compared to previous neuroimaging studies. In addition to performing decoding analyses for comparison to previous work, we built encoding models for depth position and for horizontal position, allowing us to directly compare encoding between these dimensions. Our results validate this method of recovering depth representations from retinotopic cortex. Furthermore, we find convergent evidence that depth is encoded most strongly in dorsal area V3A.
Collapse
|
15
|
Nag S, Berman D, Golomb JD. Category-selective areas in human visual cortex exhibit preferences for stimulus depth. Neuroimage 2019; 196:289-301. [PMID: 30978498 DOI: 10.1016/j.neuroimage.2019.04.025] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2018] [Revised: 03/21/2019] [Accepted: 04/07/2019] [Indexed: 12/01/2022] Open
Abstract
Multiple regions in the human brain are dedicated to accomplish the feat of object recognition; yet our brains must also compute the 2D and 3D locations of the objects we encounter in order to make sense of our visual environments. A number of studies have explored how various object category-selective regions are sensitive to and have preferences for specific 2D spatial locations in addition to processing their preferred-stimulus categories, but there is no survey of how these regions respond to depth information. In a blocked functional MRI experiment, subjects viewed a series of category-specific (i.e., faces, objects, scenes) and unspecific (e.g., random moving dots) stimuli with red/green anaglyph glasses. Critically, these stimuli were presented at different depth planes such that they appeared in front of, behind, or at the same (i.e., middle) depth plane as the fixation point (Experiment 1) or simultaneously in front of and behind fixation (i.e., mixed depth; Experiment 2). Comparisons of mean response magnitudes between back, middle, and front depth planes reveal that face and object regions OFA and LOC exhibit a preference for front depths, and motion area MT+ exhibits a strong linear preference for front, followed by middle, followed by back depth planes. In contrast, scene-selective regions PPA and OPA prefer front and/or back depth planes (relative to middle). Moreover, the occipital place area demonstrates a strong preference for "mixed" depth above and beyond back alone, raising potential implications about its particular role in scene perception. Crucially, the observed depth preferences in nearly all areas were evoked irrespective of the semantic stimulus category being viewed. These results reveal that the object category-selective regions may play a role in processing or incorporating depth information that is orthogonal to their primary processing of object category information.
Collapse
Affiliation(s)
- Samoni Nag
- Department of Psychology, Center for Cognitive & Brain Sciences, The Ohio State University, USA; Department of Psychology, The George Washington University, USA
| | - Daniel Berman
- Department of Psychology, Center for Cognitive & Brain Sciences, The Ohio State University, USA
| | - Julie D Golomb
- Department of Psychology, Center for Cognitive & Brain Sciences, The Ohio State University, USA.
| |
Collapse
|
16
|
Chunharas C, Rademaker RL, Sprague TC, Brady TF, Serences JT. Separating memoranda in depth increases visual working memory performance. J Vis 2019; 19:4. [PMID: 30634185 PMCID: PMC6333109 DOI: 10.1167/19.1.4] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2018] [Accepted: 11/03/2018] [Indexed: 11/24/2022] Open
Abstract
Visual working memory is the mechanism supporting the continued maintenance of information after sensory inputs are removed. Although the capacity of visual working memory is limited, memoranda that are spaced farther apart on a 2-D display are easier to remember, potentially because neural representations are more distinct within retinotopically organized areas of visual cortex during memory encoding, maintenance, or retrieval. The impact on memory of spatial separability in depth is less clear, even though depth information is essential to guiding interactions with objects in the environment. On one account, separating memoranda in depth may facilitate performance if interference between items is reduced. However, depth information must be inferred indirectly from the 2-D retinal image, and less is known about how visual cortex represents depth. Thus, an alternative possibility is that separation in depth does not attenuate between-items interference; it may even impair performance, as attention must be distributed across a larger volume of 3-D space. We tested these alternatives using a stereo display while participants remembered the colors of stimuli presented either near or far in the 2-D plane or in depth. Increasing separation in-plane and in depth both enhanced performance. Furthermore, participants who were better able to utilize stereo depth cues showed larger benefits when memoranda were separated in depth, particularly for large memory arrays. The observation that spatial separation in the inferred 3-D structure of the environment improves memory performance, as is the case in 2-D environments, suggests that separating memoranda in depth might reduce neural competition by utilizing cortically separable resources.
Collapse
Affiliation(s)
- Chaipat Chunharas
- Psychology Department, University of California San Diego, La Jolla, CA, USA
- King Chulalongkorn Memorial Hospital, Chulalongkorn University, Bangkok, Thailand
| | - Rosanne L Rademaker
- Psychology Department, University of California San Diego, La Jolla, CA, USA
- Donders Institute for Brain, Cognition and Behavior, Radboud University, Nijmegen, the Netherlands
| | - Thomas C Sprague
- Department of Psychology, University of California Santa Barbara, Santa Barbara, CA, USA
| | - Timothy F Brady
- Psychology Department, University of California San Diego, La Jolla, CA, USA
| | - John T Serences
- Psychology Department, University of California San Diego, La Jolla, CA, USA
- Neurosciences Graduate Program, University of California San Diego, La Jolla, CA, USA
- Kavli Institute for Brain and Mind, University of California, San Diego, La Jolla, CA, USA
| |
Collapse
|
17
|
Abstract
Visual cognition in our 3D world requires understanding how we accurately localize objects in 2D and depth, and what influence both types of location information have on visual processing. Spatial location is known to play a special role in visual processing, but most of these findings have focused on the special role of 2D location. One such phenomena is the spatial congruency bias (Golomb, Kupitz, & Thiemann, 2014), where 2D location biases judgments of object features but features do not bias location judgments. This paradigm has recently been used to compare different types of location information in terms of how much they bias different types of features. Here we used this paradigm to ask a related question: whether 2D and depth-from-disparity location bias localization judgments for each other. We found that presenting two objects in the same 2D location biased position-in-depth judgments, but presenting two objects at the same depth (disparity) did not bias 2D location judgments. We conclude that an object's 2D location may be automatically incorporated into perception of its depth location, but not vice versa, which is consistent with a fundamentally special role for 2D location in visual processing.
Collapse
Affiliation(s)
- Nonie J. Finlayson
- Department of Psychology, Center for Cognitive & Brain Sciences, The Ohio State University, Columbus, OH 43210, USA
| | | |
Collapse
|