1
|
Baltaretu BR, Schuetz I, Võ MLH, Fiehler K. Scene semantics affects allocentric spatial coding for action in naturalistic (virtual) environments. Sci Rep 2024; 14:15549. [PMID: 38969745 PMCID: PMC11226608 DOI: 10.1038/s41598-024-66428-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2024] [Accepted: 07/01/2024] [Indexed: 07/07/2024] Open
Abstract
Interacting with objects in our environment requires determining their locations, often with respect to surrounding objects (i.e., allocentrically). According to the scene grammar framework, these usually small, local objects are movable within a scene and represent the lowest level of a scene's hierarchy. How do higher hierarchical levels of scene grammar influence allocentric coding for memory-guided actions? Here, we focused on the effect of large, immovable objects (anchors) on the encoding of local object positions. In a virtual reality study, participants (n = 30) viewed one of four possible scenes (two kitchens or two bathrooms), with two anchors connected by a shelf, onto which were presented three local objects (congruent with one anchor) (Encoding). The scene was re-presented (Test) with 1) local objects missing and 2) one of the anchors shifted (Shift) or not (No shift). Participants, then, saw a floating local object (target), which they grabbed and placed back on the shelf in its remembered position (Response). Eye-tracking data revealed that both local objects and anchors were fixated, with preference for local objects. Additionally, anchors guided allocentric coding of local objects, despite being task-irrelevant. Overall, anchors implicitly influence spatial coding of local object locations for memory-guided actions within naturalistic (virtual) environments.
Collapse
Affiliation(s)
- Bianca R Baltaretu
- Department of Experimental Psychology, Justus Liebig University Giessen, Otto-Behaghel-Strasse 10F, 35394, Giessen, Hesse, Germany.
| | - Immo Schuetz
- Department of Experimental Psychology, Justus Liebig University Giessen, Otto-Behaghel-Strasse 10F, 35394, Giessen, Hesse, Germany
| | - Melissa L-H Võ
- Department of Psychology, Goethe University Frankfurt, 60323, Frankfurt am Main, Hesse, Germany
| | - Katja Fiehler
- Department of Experimental Psychology, Justus Liebig University Giessen, Otto-Behaghel-Strasse 10F, 35394, Giessen, Hesse, Germany
| |
Collapse
|
2
|
Ma AC, Cameron AD, Wiener M. Memorability shapes perceived time (and vice versa). Nat Hum Behav 2024; 8:1296-1308. [PMID: 38649460 DOI: 10.1038/s41562-024-01863-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2023] [Accepted: 03/13/2024] [Indexed: 04/25/2024]
Abstract
Visual stimuli are known to vary in their perceived duration. Some visual stimuli are also known to linger for longer in memory. Yet, whether these two features of visual processing are linked is unknown. Despite early assumptions that time is an extracted or higher-order feature of perception, more recent work over the past two decades has demonstrated that timing may be instantiated within sensory modality circuits. A primary location for many of these studies is the visual system, where duration-sensitive responses have been demonstrated. Furthermore, visual stimulus features have been observed to shift perceived duration. These findings suggest that visual circuits mediate or construct perceived time. Here we present evidence across a series of experiments that perceived time is affected by the image properties of scene size, clutter and memorability. More specifically, we observe that scene size and memorability dilate time, whereas clutter contracts it. Furthermore, the durations of more memorable images are also perceived more precisely. Conversely, the longer the perceived duration of an image, the more memorable it is. To explain these findings, we applied a recurrent convolutional neural network model of the ventral visual system, in which images are progressively processed over time. We find that more memorable images are processed faster, and that this increase in processing speed predicts both the lengthening and the increased precision of perceived durations. These findings provide evidence for a link between image features, time perception and memory that can be further explored with models of visual processing.
Collapse
Affiliation(s)
- Alex C Ma
- Department of Psychology, George Mason University, Fairfax, VA, USA
| | - Ayana D Cameron
- Department of Psychology, George Mason University, Fairfax, VA, USA
| | - Martin Wiener
- Department of Psychology, George Mason University, Fairfax, VA, USA.
| |
Collapse
|
3
|
Park J, Soucy E, Segawa J, Mair R, Konkle T. Immersive scene representation in human visual cortex with ultra-wide-angle neuroimaging. Nat Commun 2024; 15:5477. [PMID: 38942766 PMCID: PMC11213904 DOI: 10.1038/s41467-024-49669-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2023] [Accepted: 06/13/2024] [Indexed: 06/30/2024] Open
Abstract
While human vision spans 220°, traditional functional MRI setups display images only up to central 10-15°. Thus, it remains unknown how the brain represents a scene perceived across the full visual field. Here, we introduce a method for ultra-wide angle display and probe signatures of immersive scene representation. An unobstructed view of 175° is achieved by bouncing the projected image off angled-mirrors onto a custom-built curved screen. To avoid perceptual distortion, scenes are created with wide field-of-view from custom virtual environments. We find that immersive scene representation drives medial cortex with far-peripheral preferences, but shows minimal modulation in classic scene regions. Further, scene and face-selective regions maintain their content preferences even with extreme far-periphery stimulation, highlighting that not all far-peripheral information is automatically integrated into scene regions computations. This work provides clarifying evidence on content vs. peripheral preferences in scene representation and opens new avenues to research immersive vision.
Collapse
Affiliation(s)
- Jeongho Park
- Department of Psychology, Harvard University, Cambridge, MA, USA.
| | - Edward Soucy
- Center for Brain Science, Harvard University, Cambridge, MA, USA
| | - Jennifer Segawa
- Center for Brain Science, Harvard University, Cambridge, MA, USA
| | - Ross Mair
- Center for Brain Science, Harvard University, Cambridge, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
- Department of Radiology, Massachusetts General Hospital, Boston, MA, USA
| | - Talia Konkle
- Department of Psychology, Harvard University, Cambridge, MA, USA
- Center for Brain Science, Harvard University, Cambridge, MA, USA
- Kempner Institute for Biological and Artificial Intelligence, Harvard University, Boston, MA, USA
| |
Collapse
|
4
|
Du J, DiNicola LM, Angeli PA, Saadon-Grosman N, Sun W, Kaiser S, Ladopoulou J, Xue A, Yeo BTT, Eldaief MC, Buckner RL. Organization of the human cerebral cortex estimated within individuals: networks, global topography, and function. J Neurophysiol 2024; 131:1014-1082. [PMID: 38489238 PMCID: PMC11383390 DOI: 10.1152/jn.00308.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2023] [Revised: 01/18/2024] [Accepted: 02/16/2024] [Indexed: 03/17/2024] Open
Abstract
The cerebral cortex is populated by specialized regions that are organized into networks. Here we estimated networks from functional MRI (fMRI) data in intensively sampled participants. The procedure was developed in two participants (scanned 31 times) and then prospectively applied to 15 participants (scanned 8-11 times). Analysis of the networks revealed a global organization. Locally organized first-order sensory and motor networks were surrounded by spatially adjacent second-order networks that linked to distant regions. Third-order networks possessed regions distributed widely throughout association cortex. Regions of distinct third-order networks displayed side-by-side juxtapositions with a pattern that repeated across multiple cortical zones. We refer to these as supra-areal association megaclusters (SAAMs). Within each SAAM, two candidate control regions were adjacent to three separate domain-specialized regions. Response properties were explored with task data. The somatomotor and visual networks responded to body movements and visual stimulation, respectively. Second-order networks responded to transients in an oddball detection task, consistent with a role in orienting to salient events. The third-order networks, including distinct regions within each SAAM, showed two levels of functional specialization. Regions linked to candidate control networks responded to working memory load across multiple stimulus domains. The remaining regions dissociated across language, social, and spatial/episodic processing domains. These results suggest that progressively higher-order networks nest outward from primary sensory and motor cortices. Within the apex zones of association cortex, there is specialization that repeatedly divides domain-flexible from domain-specialized regions. We discuss implications of these findings, including how repeating organizational motifs may emerge during development.NEW & NOTEWORTHY The organization of cerebral networks was estimated within individuals with intensive, repeat sampling of fMRI data. A hierarchical organization emerged in each individual that delineated first-, second-, and third-order cortical networks. Regions of distinct third-order association networks consistently exhibited side-by-side juxtapositions that repeated across multiple cortical zones, with clear and robust functional specialization among the embedded regions.
Collapse
Affiliation(s)
- Jingnan Du
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States
| | - Lauren M DiNicola
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States
| | - Peter A Angeli
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States
| | - Noam Saadon-Grosman
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States
| | - Wendy Sun
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States
| | - Stephanie Kaiser
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States
| | - Joanna Ladopoulou
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States
| | - Aihuiping Xue
- Centre for Sleep & Cognition and Department of Electrical and Computer Engineering, National University of Singapore, Singapore, Singapore
| | - B T Thomas Yeo
- Centre for Sleep & Cognition and Department of Electrical and Computer Engineering, National University of Singapore, Singapore, Singapore
| | - Mark C Eldaief
- Department of Psychiatry, Massachusetts General Hospital, Charlestown, Massachusetts, United States
| | - Randy L Buckner
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States
- Department of Psychiatry, Massachusetts General Hospital, Charlestown, Massachusetts, United States
- Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Charlestown, Massachusetts, United States
| |
Collapse
|
5
|
Park J, Soucy E, Segawa J, Mair R, Konkle T. Immersive scene representation in human visual cortex with ultra-wide angle neuroimaging. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2023.05.14.540275. [PMID: 37292806 PMCID: PMC10245572 DOI: 10.1101/2023.05.14.540275] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
While humans experience the visual environment in a panoramic 220° view, traditional functional MRI setups are limited to display images like postcards in the central 10-15° of the visual field. Thus, it remains unknown how a scene is represented in the brain when perceived across the full visual field. Here, we developed a novel method for ultra-wide angle visual presentation and probed for signatures of immersive scene representation. To accomplish this, we bounced the projected image off angled-mirrors directly onto a custom-built curved screen, creating an unobstructed view of 175°. Scene images were created from custom-built virtual environments with a compatible wide field-of-view to avoid perceptual distortion. We found that immersive scene representation drives medial cortex with far-peripheral preferences, but surprisingly had little effect on classic scene regions. That is, scene regions showed relatively minimal modulation over dramatic changes of visual size. Further, we found that scene and face-selective regions maintain their content preferences even under conditions of central scotoma, when only the extreme far-peripheral visual field is stimulated. These results highlight that not all far-peripheral information is automatically integrated into the computations of scene regions, and that there are routes to high-level visual areas that do not require direct stimulation of the central visual field. Broadly, this work provides new clarifying evidence on content vs. peripheral preferences in scene representation, and opens new neuroimaging research avenues to understand immersive visual representation.
Collapse
Affiliation(s)
| | | | | | - Ross Mair
- Center for Brain Science, Harvard University
- Department of Radiology, Harvard Medical School
- Department of Radiology, Massachusetts General Hospital
| | - Talia Konkle
- Department of Psychology, Harvard University
- Center for Brain Science, Harvard University
- Kempner Institute for Biological and Artificial Intelligence, Harvard University
| |
Collapse
|
6
|
Park J, Josephs E, Konkle T. Systematic transition from boundary extension to contraction along an object-to-scene continuum. J Vis 2024; 24:9. [PMID: 38252521 PMCID: PMC10810016 DOI: 10.1167/jov.24.1.9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2023] [Accepted: 11/23/2023] [Indexed: 01/24/2024] Open
Abstract
After viewing a picture of an environment, our memory of it typically extends beyond what was presented, a phenomenon referred to as boundary extension. But, sometimes memory errors show the opposite pattern-boundary contraction-and the relationship between these phenomena is controversial. We constructed virtual three-dimensional environments and created a series of views at different distances, from object close-ups to wide-angle indoor views, and tested for memory errors along this object-to-scene continuum. Boundary extension was evident for close-scale views and transitioned parametrically to boundary contraction for far-scale views. However, this transition point was not tied to a specific position in the environment (e.g., the point of reachability). Instead, it tracked with judgments of the best-looking view of the environment, in both rich-object and low-object environments. We offer a dynamic-tension account, where competition between object-based and scene-based affordances determines whether a view will extend or contract in memory. This study demonstrates that boundary extension and boundary contraction are not two separate phenomena but rather two parts of a continuum, suggesting a common underlying mechanism. The transition point between the two is not fixed but depends on the observer's judgment of the best-looking view of the environment. These findings provide new insights into how we perceive and remember a view of environment.
Collapse
Affiliation(s)
- Jeongho Park
- Department of Psychology, Harvard University, Cambridge, MA, USA
| | - Emilie Josephs
- Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Talia Konkle
- Department of Psychology, Harvard University, Cambridge, MA, USA
| |
Collapse
|
7
|
Peelen MV, Berlot E, de Lange FP. Predictive processing of scenes and objects. NATURE REVIEWS PSYCHOLOGY 2024; 3:13-26. [PMID: 38989004 PMCID: PMC7616164 DOI: 10.1038/s44159-023-00254-0] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 10/25/2023] [Indexed: 07/12/2024]
Abstract
Real-world visual input consists of rich scenes that are meaningfully composed of multiple objects which interact in complex, but predictable, ways. Despite this complexity, we recognize scenes, and objects within these scenes, from a brief glance at an image. In this review, we synthesize recent behavioral and neural findings that elucidate the mechanisms underlying this impressive ability. First, we review evidence that visual object and scene processing is partly implemented in parallel, allowing for a rapid initial gist of both objects and scenes concurrently. Next, we discuss recent evidence for bidirectional interactions between object and scene processing, with scene information modulating the visual processing of objects, and object information modulating the visual processing of scenes. Finally, we review evidence that objects also combine with each other to form object constellations, modulating the processing of individual objects within the object pathway. Altogether, these findings can be understood by conceptualizing object and scene perception as the outcome of a joint probabilistic inference, in which "best guesses" about objects act as priors for scene perception and vice versa, in order to concurrently optimize visual inference of objects and scenes.
Collapse
Affiliation(s)
- Marius V Peelen
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
| | - Eva Berlot
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
| | - Floris P de Lange
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
| |
Collapse
|
8
|
DiNicola LM, Sun W, Buckner RL. Side-by-side regions in dorsolateral prefrontal cortex estimated within the individual respond differentially to domain-specific and domain-flexible processes. J Neurophysiol 2023; 130:1602-1615. [PMID: 37937340 PMCID: PMC11068361 DOI: 10.1152/jn.00277.2023] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2023] [Revised: 10/06/2023] [Accepted: 10/28/2023] [Indexed: 11/09/2023] Open
Abstract
A recurring debate concerns whether regions of primate prefrontal cortex (PFC) support domain-flexible or domain-specific processes. Here we tested the hypothesis with functional MRI (fMRI) that side-by-side PFC regions, within distinct parallel association networks, differentially support domain-flexible and domain-specialized processing. Individuals (N = 9) were intensively sampled, and all effects were estimated within their own idiosyncratic anatomy. Within each individual, we identified PFC regions linked to distinct networks, including a dorsolateral PFC (DLPFC) region coupled to the medial temporal lobe (MTL) and an extended region associated with the canonical multiple-demand network. We further identified an inferior PFC region coupled to the language network. Exploration in separate task data, collected within the same individuals, revealed a robust functional triple dissociation. The DLPFC region linked to the MTL was recruited during remembering and imagining the future, distinct from juxtaposed regions that were modulated in a domain-flexible manner during working memory. The inferior PFC region linked to the language network was recruited during sentence processing. Detailed analysis of the trial-level responses further revealed that the DLPFC region linked to the MTL specifically tracked processes associated with scene construction. These results suggest that the DLPFC possesses a domain-specialized region that is small and easily confused with nearby (larger) regions associated with cognitive control. The newly described region is domain specialized for functions traditionally associated with the MTL. We discuss the implications of these findings in relation to convergent anatomical analysis in the monkey.NEW & NOTEWORTHY Competing hypotheses link regions of prefrontal cortex (PFC) to domain-flexible or domain-specific processes. Here, using a precision neuroimaging approach, we identify a domain-specialized region in dorsolateral PFC, coupled to the medial temporal lobe and recruited for scene construction. This region is juxtaposed to, but distinct from, broader PFC regions recruited flexibly for cognitive control. Region distinctions align with broader network differences, suggesting that PFC regions gain dissociable processing properties via segregated anatomical projections.
Collapse
Affiliation(s)
- Lauren M DiNicola
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States
| | - Wendy Sun
- Division of Medical Sciences, Harvard Medical School, Boston, Massachusetts, United States
| | - Randy L Buckner
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States
- Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Charlestown, Massachusetts, United States
- Department of Psychiatry, Massachusetts General Hospital, Charlestown, Massachusetts, United States
| |
Collapse
|
9
|
Snow JC, Gomez MA, Compton MT. Human memory for real-world solid objects is not predicted by responses to image displays. J Exp Psychol Gen 2023; 152:2703-2712. [PMID: 37079829 PMCID: PMC10587360 DOI: 10.1037/xge0001387] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/22/2023]
Abstract
In experimental psychology and neuroscience, computerized image stimuli are typically used as artificial proxies for real-world objects to understand brain and behavior. Here, in a series of five experiments (n = 165), we studied human memory for objects presented as tangible solids versus computerized images. We found that recall for solids was superior to images, both immediately after learning, and after a 24-hr delay. A "realness advantage" was also evident relative to three-dimensional (3-D) stereoscopic images, and when solids were viewed monocularly, arguing against explanations based on the presence of binocular depth cues in the stimulus. Critically, memory for solids was modulated by physical distance, with superior recall for objects positioned within versus outside of observers' reach, whereas recall for images was unaffected by distance. We conclude that solids are processed quantitatively and qualitatively differently in episodic memory than are images, suggesting caution in assuming that artifice can always substitute for reality. (PsycInfo Database Record (c) 2023 APA, all rights reserved).
Collapse
Affiliation(s)
| | - Michael A. Gomez
- The University of Nevada Reno, Reno, Nevada, USA
- Clovis Community College, Fresno, CA
| | | |
Collapse
|
10
|
Du J, DiNicola LM, Angeli PA, Saadon-Grosman N, Sun W, Kaiser S, Ladopoulou J, Xue A, Yeo BTT, Eldaief MC, Buckner RL. Within-Individual Organization of the Human Cerebral Cortex: Networks, Global Topography, and Function. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.08.08.552437. [PMID: 37609246 PMCID: PMC10441314 DOI: 10.1101/2023.08.08.552437] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/24/2023]
Abstract
The human cerebral cortex is populated by specialized regions that are organized into networks. Here we estimated networks using a Multi-Session Hierarchical Bayesian Model (MS-HBM) applied to intensively sampled within-individual functional MRI (fMRI) data. The network estimation procedure was initially developed and tested in two participants (each scanned 31 times) and then prospectively applied to 15 new participants (each scanned 8 to 11 times). Detailed analysis of the networks revealed a global organization. Locally organized first-order sensory and motor networks were surrounded by spatially adjacent second-order networks that also linked to distant regions. Third-order networks each possessed regions distributed widely throughout association cortex. Moreover, regions of distinct third-order networks displayed side-by-side juxtapositions with a pattern that repeated similarly across multiple cortical zones. We refer to these as Supra-Areal Association Megaclusters (SAAMs). Within each SAAM, two candidate control regions were typically adjacent to three separate domain-specialized regions. Independent task data were analyzed to explore functional response properties. The somatomotor and visual first-order networks responded to body movements and visual stimulation, respectively. A subset of the second-order networks responded to transients in an oddball detection task, consistent with a role in orienting to salient or novel events. The third-order networks, including distinct regions within each SAAM, showed two levels of functional specialization. Regions linked to candidate control networks responded to working memory load across multiple stimulus domains. The remaining regions within each SAAM did not track working memory load but rather dissociated across language, social, and spatial / episodic processing domains. These results support a model of the cerebral cortex in which progressively higher-order networks nest outwards from primary sensory and motor cortices. Within the apex zones of association cortex there is specialization of large-scale networks that divides domain-flexible from domain-specialized regions repeatedly across parietal, temporal, and prefrontal cortices. We discuss implications of these findings including how repeating organizational motifs may emerge during development.
Collapse
Affiliation(s)
- Jingnan Du
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, MA 02138, USA
| | - Lauren M DiNicola
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, MA 02138, USA
| | - Peter A Angeli
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, MA 02138, USA
| | - Noam Saadon-Grosman
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, MA 02138, USA
| | - Wendy Sun
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, MA 02138, USA
| | - Stephanie Kaiser
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, MA 02138, USA
| | - Joanna Ladopoulou
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, MA 02138, USA
| | - Aihuiping Xue
- Centre for Sleep & Cognition & Department of Electrical and Computer Engineering, National University of Singapore, Singapore, Singapore
| | - B T Thomas Yeo
- Centre for Sleep & Cognition & Department of Electrical and Computer Engineering, National University of Singapore, Singapore, Singapore
| | - Mark C Eldaief
- Department of Psychiatry, Massachusetts General Hospital, Charlestown, MA 02129, USA
| | - Randy L Buckner
- Department of Psychology, Center for Brain Science, Harvard University, Cambridge, MA 02138, USA
- Department of Psychiatry, Massachusetts General Hospital, Charlestown, MA 02129, USA
- Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Charlestown, MA 02129, USA
| |
Collapse
|
11
|
Josephs EL, Hebart MN, Konkle T. Dimensions underlying human understanding of the reachable world. Cognition 2023; 234:105368. [PMID: 36641868 DOI: 10.1016/j.cognition.2023.105368] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2022] [Revised: 12/20/2022] [Accepted: 01/09/2023] [Indexed: 01/15/2023]
Abstract
Near-scale environments, like work desks, restaurant place settings or lab benches, are the interface of our hand-based interactions with the world. How are our conceptual representations of these environments organized? What properties distinguish among reachspaces, and why? We obtained 1.25 million similarity judgments on 990 reachspace images, and generated a 30-dimensional embedding which accurately predicts these judgments. Examination of the embedding dimensions revealed key properties underlying these judgments, such as reachspace layout, affordance, and visual appearance. Clustering performed over the embedding revealed four distinct interpretable classes of reachspaces, distinguishing among spaces related to food, electronics, analog activities, and storage or display. Finally, we found that reachspace similarity ratings were better predicted by the function of the spaces than their locations, suggesting that reachspaces are largely conceptualized in terms of the actions they support. Altogether, these results reveal the behaviorally-relevant principles that structure our internal representations of reach-relevant environments.
Collapse
Affiliation(s)
- Emilie L Josephs
- Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, USA; Psychology Department, Harvard University, Cambridge, USA.
| | - Martin N Hebart
- Vision and Computational Cognition Group, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.
| | - Talia Konkle
- Psychology Department, Harvard University, Cambridge, USA.
| |
Collapse
|
12
|
Barker M, Rehrig G, Ferreira F. Speakers prioritise affordance-based object semantics in scene descriptions. LANGUAGE, COGNITION AND NEUROSCIENCE 2023; 38:1045-1067. [PMID: 37841974 PMCID: PMC10572038 DOI: 10.1080/23273798.2023.2190136] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/18/2022] [Accepted: 02/18/2023] [Indexed: 10/17/2023]
Abstract
This work investigates the linearisation strategies used by speakers when describing real-world scenes to better understand production plans for multi-utterance sequences. In this study, 30 participants described real-world scenes aloud. To investigate which semantic features of scenes predict order of mention, we quantified three features (meaning, graspability, and interactability) using two techniques (whole-object ratings and feature map values). We found that object-level semantic features, namely those affordance-based, predicted order of mention in a scene description task. Our findings provide the first evidence for an object-related semantic feature that guides linguistic ordering decisions and offer theoretical support for the role of object semantics in scene viewing and description.
Collapse
Affiliation(s)
- M. Barker
- Department of Psychology, University of California, Davis
| | - G. Rehrig
- Department of Psychology, University of California, Davis
| | - F. Ferreira
- Department of Psychology, University of California, Davis
| |
Collapse
|
13
|
Abstract
Visual cortex contains regions of selectivity for domains of ecological importance. Food is an evolutionarily critical category whose visual heterogeneity may make the identification of selectivity more challenging. We investigate neural responsiveness to food using natural images combined with large-scale human fMRI. Leveraging the improved sensitivity of modern designs and statistical analyses, we identify two food-selective regions in the ventral visual cortex. Our results are robust across 8 subjects from the Natural Scenes Dataset (NSD), multiple independent image sets and multiple analysis methods. We then test our findings of food selectivity in an fMRI "localizer" using grayscale food images. These independent results confirm the existence of food selectivity in ventral visual cortex and help illuminate why earlier studies may have failed to do so. Our identification of food-selective regions stands alongside prior findings of functional selectivity and adds to our understanding of the organization of knowledge within the human visual system.
Collapse
|
14
|
Bracci S, Op de Beeck HP. Understanding Human Object Vision: A Picture Is Worth a Thousand Representations. Annu Rev Psychol 2023; 74:113-135. [PMID: 36378917 DOI: 10.1146/annurev-psych-032720-041031] [Citation(s) in RCA: 13] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Objects are the core meaningful elements in our visual environment. Classic theories of object vision focus upon object recognition and are elegant and simple. Some of their proposals still stand, yet the simplicity is gone. Recent evolutions in behavioral paradigms, neuroscientific methods, and computational modeling have allowed vision scientists to uncover the complexity of the multidimensional representational space that underlies object vision. We review these findings and propose that the key to understanding this complexity is to relate object vision to the full repertoire of behavioral goals that underlie human behavior, running far beyond object recognition. There might be no such thing as core object recognition, and if it exists, then its importance is more limited than traditionally thought.
Collapse
Affiliation(s)
- Stefania Bracci
- Center for Mind/Brain Sciences, University of Trento, Rovereto, Italy;
| | - Hans P Op de Beeck
- Leuven Brain Institute, Research Unit Brain & Cognition, KU Leuven, Leuven, Belgium;
| |
Collapse
|
15
|
Moskowitz JB, Berger SA, Fooken J, Castelhano MS, Gallivan JP, Flanagan JR. The influence of movement-related costs when searching to act and acting to search. J Neurophysiol 2023; 129:115-130. [PMID: 36475897 DOI: 10.1152/jn.00305.2022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2022] Open
Abstract
Real-world search behavior often involves limb movements, either during search or after search. Here we investigated whether movement-related costs influence search behavior in two kinds of search tasks. In our visual search tasks, participants made saccades to find a target object among distractors and then moved a cursor, controlled by the handle of a robotic manipulandum, to the target. In our manual search tasks, participants moved the cursor to perform the search, placing it onto objects to reveal their identity as either a target or a distractor. In all tasks, there were multiple targets. Across experiments, we manipulated either the effort or time costs associated with movement such that these costs varied across the search space. We varied effort by applying different resistive forces to the handle, and we varied time costs by altering the speed of the cursor. Our analysis of cursor and eye movements during manual and visual search, respectively, showed that effort influenced manual search but did not influence visual search. In contrast, time costs influenced both visual and manual search. Our results demonstrate that, in addition to perceptual and cognitive factors, movement-related costs can also influence search behavior.NEW & NOTEWORTHY Numerous studies have investigated the perceptual and cognitive factors that influence decision making about where to look, or move, in search tasks. However, little is known about how search is influenced by movement-related costs associated with acting on an object once it has been visually located or acting during manual search. In this article, we show that movement time costs can bias visual and manual search and that movement effort costs bias manual search.
Collapse
Affiliation(s)
- Joshua B Moskowitz
- Centre for Neuroscience Studies, Queen's University, Kingston, Ontario, Canada.,Department of Psychology, Queen's University, Kingston, Ontario, Canada
| | - Sarah A Berger
- Department of Psychology, Queen's University, Kingston, Ontario, Canada
| | - Jolande Fooken
- Centre for Neuroscience Studies, Queen's University, Kingston, Ontario, Canada
| | - Monica S Castelhano
- Centre for Neuroscience Studies, Queen's University, Kingston, Ontario, Canada.,Department of Psychology, Queen's University, Kingston, Ontario, Canada
| | - Jason P Gallivan
- Centre for Neuroscience Studies, Queen's University, Kingston, Ontario, Canada.,Department of Psychology, Queen's University, Kingston, Ontario, Canada.,Department of Biomedical and Molecular Sciences, Queen's University, Kingston, Ontario, Canada
| | - J Randall Flanagan
- Centre for Neuroscience Studies, Queen's University, Kingston, Ontario, Canada.,Department of Psychology, Queen's University, Kingston, Ontario, Canada
| |
Collapse
|
16
|
Park J, Josephs E, Konkle T. Ramp-shaped neural tuning supports graded population-level representation of the object-to-scene continuum. Sci Rep 2022; 12:18081. [PMID: 36302932 PMCID: PMC9613906 DOI: 10.1038/s41598-022-21768-2] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2022] [Accepted: 09/30/2022] [Indexed: 01/24/2023] Open
Abstract
We can easily perceive the spatial scale depicted in a picture, regardless of whether it is a small space (e.g., a close-up view of a chair) or a much larger space (e.g., an entire class room). How does the human visual system encode this continuous dimension? Here, we investigated the underlying neural coding of depicted spatial scale, by examining the voxel tuning and topographic organization of brain responses. We created naturalistic yet carefully-controlled stimuli by constructing virtual indoor environments, and rendered a series of snapshots to smoothly sample between a close-up view of the central object and far-scale view of the full environment (object-to-scene continuum). Human brain responses were measured to each position using functional magnetic resonance imaging. We did not find evidence for a smooth topographic mapping for the object-to-scene continuum on the cortex. Instead, we observed large swaths of cortex with opposing ramp-shaped profiles, with highest responses to one end of the object-to-scene continuum or the other, and a small region showing a weak tuning to intermediate scale views. However, when we considered the population code of the entire ventral occipito-temporal cortex, we found smooth and linear representation of the object-to-scene continuum. Our results together suggest that depicted spatial scale information is encoded parametrically in large-scale population codes across the entire ventral occipito-temporal cortex.
Collapse
Affiliation(s)
- Jeongho Park
- Department of Psychology, Harvard University, Cambridge, USA.
| | - Emilie Josephs
- Computer Science & Artificial Intelligence Lab, Massachusetts Institute of Technology, Cambridge, USA
| | - Talia Konkle
- Department of Psychology, Harvard University, Cambridge, USA
| |
Collapse
|
17
|
Rehrig G, Barker M, Peacock CE, Hayes TR, Henderson JM, Ferreira F. Look at what I can do: Object affordances guide visual attention while speakers describe potential actions. Atten Percept Psychophys 2022; 84:1583-1610. [PMID: 35484443 PMCID: PMC9246959 DOI: 10.3758/s13414-022-02467-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 02/23/2022] [Indexed: 11/08/2022]
Abstract
As we act on the world around us, our eyes seek out objects we plan to interact with. A growing body of evidence suggests that overt visual attention selects objects in the environment that could be interacted with, even when the task precludes physical interaction. In previous work, objects that afford grasping interactions influenced attention when static scenes depicted reachable spaces, and attention was otherwise better explained by general informativeness. Because grasping is but one of many object interactions, previous work may have downplayed the influence of object affordances on attention. The current study investigated the relationship between overt visual attention and object affordances versus broadly construed semantic information in scenes as speakers describe or memorize scenes. In addition to meaning and grasp maps-which capture informativeness and grasping object affordances in scenes, respectively-we introduce interact maps, which capture affordances more broadly. In a mixed-effects analysis of 5 eyetracking experiments, we found that meaning predicted fixated locations in a general description task and during scene memorization. Grasp maps marginally predicted fixated locations during action description for scenes that depicted reachable spaces only. Interact maps predicted fixated regions in description experiments alone. Our findings suggest observers allocate attention to scene regions that could be readily interacted with when talking about the scene, while general informativeness preferentially guides attention when the task does not encourage careful consideration of objects in the scene. The current study suggests that the influence of object affordances on visual attention in scenes is mediated by task demands.
Collapse
Affiliation(s)
- Gwendolyn Rehrig
- Department of Psychology, University of California, Davis, Davis, CA, 95616, USA.
| | - Madison Barker
- Department of Psychology, University of California, Davis, Davis, CA, 95616, USA
| | - Candace E Peacock
- Department of Psychology and Center for Mind and Brain, University of California, Davis, Davis, CA, USA
| | - Taylor R Hayes
- Center for Mind and Brain, University of California, Davis, Davis, CA, USA
| | - John M Henderson
- Department of Psychology and Center for Mind and Brain, University of California, Davis, Davis, CA, USA
| | - Fernanda Ferreira
- Department of Psychology, University of California, Davis, Davis, CA, 95616, USA
| |
Collapse
|
18
|
Ayzenberg V, Behrmann M. The Dorsal Visual Pathway Represents Object-Centered Spatial Relations for Object Recognition. J Neurosci 2022; 42:4693-4710. [PMID: 35508386 PMCID: PMC9186804 DOI: 10.1523/jneurosci.2257-21.2022] [Citation(s) in RCA: 17] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/12/2021] [Revised: 04/19/2022] [Accepted: 04/21/2022] [Indexed: 11/21/2022] Open
Abstract
Although there is mounting evidence that input from the dorsal visual pathway is crucial for object processes in the ventral pathway, the specific functional contributions of dorsal cortex to these processes remain poorly understood. Here, we hypothesized that dorsal cortex computes the spatial relations among an object's parts, a process crucial for forming global shape percepts, and transmits this information to the ventral pathway to support object categorization. Using fMRI with human participants (females and males), we discovered regions in the intraparietal sulcus (IPS) that were selectively involved in computing object-centered part relations. These regions exhibited task-dependent functional and effective connectivity with ventral cortex, and were distinct from other dorsal regions, such as those representing allocentric relations, 3D shape, and tools. In a subsequent experiment, we found that the multivariate response of posterior (p)IPS, defined on the basis of part-relations, could be used to decode object category at levels comparable to ventral object regions. Moreover, mediation and multivariate effective connectivity analyses further suggested that IPS may account for representations of part relations in the ventral pathway. Together, our results highlight specific contributions of the dorsal visual pathway to object recognition. We suggest that dorsal cortex is a crucial source of input to the ventral pathway and may support the ability to categorize objects on the basis of global shape.SIGNIFICANCE STATEMENT Humans categorize novel objects rapidly and effortlessly. Such categorization is achieved by representing an object's global shape structure, that is, the relations among object parts. Yet, despite their importance, it is unclear how part relations are represented neurally. Here, we hypothesized that object-centered part relations may be computed by the dorsal visual pathway, which is typically implicated in visuospatial processing. Using fMRI, we identified regions selective for the part relations in dorsal cortex. We found that these regions can support object categorization, and even mediate representations of part relations in the ventral pathway, the region typically thought to support object categorization. Together, these findings shed light on the broader network of brain regions that support object categorization.
Collapse
Affiliation(s)
- Vladislav Ayzenberg
- Neuroscience Institute and Psychology Department, Carnegie Mellon University, Pittsburgh, PA 15213
| | - Marlene Behrmann
- Neuroscience Institute and Psychology Department, Carnegie Mellon University, Pittsburgh, PA 15213
| |
Collapse
|
19
|
Marrazzo G, Vaessen MJ, de Gelder B. Decoding the difference between explicit and implicit body expression representation in high level visual, prefrontal and inferior parietal cortex. Neuroimage 2021; 243:118545. [PMID: 34478822 DOI: 10.1016/j.neuroimage.2021.118545] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2020] [Revised: 08/30/2021] [Accepted: 08/31/2021] [Indexed: 11/28/2022] Open
Abstract
Recent studies provide an increasing understanding of how visual objects categories like faces or bodies are represented in the brain and also raised the question whether a category based or more dynamic network inspired models are more powerful. Two important and so far sidestepped issues in this debate are, first, how major category attributes like the emotional expression directly influence category representation and second, whether category and attribute representation are sensitive to task demands. This study investigated the impact of a crucial category attribute like emotional expression on category area activity and whether this varies with the participants' task. Using (fMRI) we measured BOLD responses while participants viewed whole body expressions and performed either an explicit (emotion) or an implicit (shape) recognition task. Our results based on multivariate methods show that the type of task is the strongest determinant of brain activity and can be decoded in EBA, VLPFC and IPL. Brain activity was higher for the explicit task condition in VLPFC and was not emotion specific. This pattern suggests that during explicit recognition of the body expression, body category representation may be strengthened, and emotion and action related activity suppressed. Taken together these results stress the importance of the task and of the role of category attributes for understanding the functional organization of high level visual cortex.
Collapse
Affiliation(s)
- Giuseppe Marrazzo
- Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, Limburg 6200 MD, Maastricht, the Netherlands
| | - Maarten J Vaessen
- Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, Limburg 6200 MD, Maastricht, the Netherlands
| | - Beatrice de Gelder
- Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, Limburg 6200 MD, Maastricht, the Netherlands; Department of Computer Science, University College London, London WC1E 6BT, United Kingdom.
| |
Collapse
|
20
|
de Gelder B, Poyo Solanas M. A computational neuroethology perspective on body and expression perception. Trends Cogn Sci 2021; 25:744-756. [PMID: 34147363 DOI: 10.1016/j.tics.2021.05.010] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2020] [Revised: 04/22/2021] [Accepted: 05/24/2021] [Indexed: 01/17/2023]
Abstract
Survival prompts organisms to prepare adaptive behavior in response to environmental and social threat. However, what are the specific features of the appearance of a conspecific that trigger such adaptive behaviors? For social species, the prime candidates for triggering defense systems are the visual features of the face and the body. We propose a novel approach for studying the ability of the brain to gather survival-relevant information from seeing conspecific body features. Specifically, we propose that behaviorally relevant information from bodies and body expressions is coded at the levels of midlevel features in the brain. These levels are relatively independent from higher-order cognitive and conscious perception of bodies and emotions. Instead, our approach is embedded in an ethological framework and mobilizes computational models for feature discovery.
Collapse
Affiliation(s)
- Beatrice de Gelder
- Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, Maastricht, Limburg 6200, MD, The Netherlands; Department of Computer Science, University College London, London WC1E 6BT, UK.
| | - Marta Poyo Solanas
- Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, Maastricht, Limburg 6200, MD, The Netherlands
| |
Collapse
|
21
|
Josephs EL, Zhao H, Konkle T. The world within reach: An image database of reach-relevant environments. J Vis 2021; 21:14. [PMID: 34289491 PMCID: PMC8300055 DOI: 10.1167/jov.21.7.14] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Near-scale spaces are a key component of our visual experience: Whether for work or for leisure, we spend much of our days immersed in, and acting upon, the world within reach. Here, we present the Reachspace Database, a novel stimulus set containing over 10,000 images depicting first person, motor-relevant views at an approximated reachable scale (hereafter “reachspaces”), which reflect the visual input that an agent would experience while performing a task with her hands. These images are divided into over 350 categories, based on a taxonomy we developed, which captures information relating to the identity of each reachspace, including the broader setting and room it is found in, the locus of interaction (e.g., kitchen counter, desk), and the specific action it affords. Summary analyses of the taxonomy labels in the database suggest a tight connection between activities and the spaces that support them: While a small number of rooms and interaction loci afford many diverse actions (e.g., workshops, tables), most reachspaces were relatively specialized, typically affording only one main activity (e.g., gas station pump, airplane cockpit, kitchen cutting board). Overall, this Reachspace Database represents a large sampling of reachable environments and provides a new resource to support behavioral and neural research into the visual representation of reach-relevant environments. The database is available for download on the Open Science Framework (osf.io/bfyxk/).
Collapse
Affiliation(s)
- Emilie L Josephs
- Department of Psychology, Harvard University, Cambridge, MA, USA.,
| | - Haoyun Zhao
- Department of Psychology, Harvard University, Cambridge, MA, USA.,
| | - Talia Konkle
- Department of Psychology, Harvard University, Cambridge, MA, USA.,
| |
Collapse
|
22
|
Shiffrin RM, Bassett DS, Kriegeskorte N, Tenenbaum JB. The brain produces mind by modeling. Proc Natl Acad Sci U S A 2020; 117:29299-29301. [PMID: 33229525 PMCID: PMC7703556 DOI: 10.1073/pnas.1912340117] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/28/2023] Open
Affiliation(s)
- Richard M Shiffrin
- Psychological and Brain Sciences Department, Indiana University, Bloomington, IN 47405;
| | - Danielle S Bassett
- Department of Bioengineering, University of Pennsylvania, Philadelphia, PA 19104
| | | | - Joshua B Tenenbaum
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139-4307
| |
Collapse
|