1
|
Ritchie JB, Andrews ST, Vaziri-Pashkam M, Baker CI. Graspable foods and tools elicit similar responses in visual cortex. Cereb Cortex 2024; 34:bhae383. [PMID: 39319569 DOI: 10.1093/cercor/bhae383] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2024] [Revised: 08/28/2024] [Accepted: 09/04/2024] [Indexed: 09/26/2024] Open
Abstract
The extrastriatal visual cortex is known to exhibit distinct response profiles to complex stimuli of varying ecological importance (e.g. faces, scenes, and tools). Although food is primarily distinguished from other objects by its edibility, not its appearance, recent evidence suggests that there is also food selectivity in human visual cortex. Food is also associated with a common behavior, eating, and food consumption typically also involves the manipulation of food, often with hands. In this context, food items share many properties with tools: they are graspable objects that we manipulate in self-directed and stereotyped forms of action. Thus, food items may be preferentially represented in extrastriatal visual cortex in part because of these shared affordance properties, rather than because they reflect a wholly distinct kind of category. We conducted functional MRI and behavioral experiments to test this hypothesis. We found that graspable food items and tools were judged to be similar in their action-related properties and that the location, magnitude, and patterns of neural responses for images of graspable food items were similar in profile to the responses for tool stimuli. Our findings suggest that food selectivity may reflect the behavioral affordances of food items rather than a distinct form of category selectivity.
Collapse
Affiliation(s)
- John Brendan Ritchie
- The Laboratory of Brain and Cognition, The National Institute of Mental Health, 10 Center Drive, Bethesda, MD 20982, United States
| | - Spencer T Andrews
- The Laboratory of Brain and Cognition, The National Institute of Mental Health, 10 Center Drive, Bethesda, MD 20982, United States
- Harvard Law School, Harvard University, 1585 Massachusetts Ave, Cambridge, MA 02138, United States
| | - Maryam Vaziri-Pashkam
- The Laboratory of Brain and Cognition, The National Institute of Mental Health, 10 Center Drive, Bethesda, MD 20982, United States
- Department of Psychological and Brain Sciences, University of Delaware, 434 Wolf Hall, Newark, DE 19716, United States
| | - Chris I Baker
- The Laboratory of Brain and Cognition, The National Institute of Mental Health, 10 Center Drive, Bethesda, MD 20982, United States
| |
Collapse
|
2
|
Ritchie JB, Andrews S, Vaziri-Pashkam M, Baker CI. Graspable foods and tools elicit similar responses in visual cortex. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.02.20.581258. [PMID: 38529495 PMCID: PMC10962699 DOI: 10.1101/2024.02.20.581258] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/27/2024]
Abstract
Extrastriatal visual cortex is known to exhibit distinct response profiles to complex stimuli of varying ecological importance (e.g., faces, scenes, and tools). The dominant interpretation of these effects is that they reflect activation of distinct "category-selective" brain regions specialized to represent these and other stimulus categories. We sought to explore an alternative perspective: that the response to these stimuli is determined less by whether they form distinct categories, and more by their relevance to different forms of natural behavior. In this regard, food is an interesting test case, since it is primarily distinguished from other objects by its edibility, not its appearance, and there is evidence of food-selectivity in human visual cortex. Food is also associated with a common behavior, eating, and food consumption typically also involves the manipulation of food, often with the hands. In this context, food items share many properties in common with tools: they are graspable objects that we manipulate in self-directed and stereotyped forms of action. Thus, food items may be preferentially represented in extrastriatal visual cortex in part because of these shared affordance properties, rather than because they reflect a wholly distinct kind of category. We conducted fMRI and behavioral experiments to test this hypothesis. We found that behaviorally graspable food items and tools were judged to be similar in their action-related properties, and that the location, magnitude, and patterns of neural responses for images of graspable food items were similar in profile to the responses for tool stimuli. Our findings suggest that food-selectivity may reflect the behavioral affordances of food items rather than a distinct form of category-selectivity.
Collapse
Affiliation(s)
- J. Brendan Ritchie
- The Laboratory of Brain and Cognition, The National Institute of Mental Health, MD, USA
| | - Spencer Andrews
- The Laboratory of Brain and Cognition, The National Institute of Mental Health, MD, USA
| | - Maryam Vaziri-Pashkam
- The Laboratory of Brain and Cognition, The National Institute of Mental Health, MD, USA
- Department of Psychological and Brain Sciences, University of Delaware, Newark, DE, USA
| | - Christopher I. Baker
- The Laboratory of Brain and Cognition, The National Institute of Mental Health, MD, USA
| |
Collapse
|
3
|
Schultz J, Frith CD. Animacy and the prediction of behaviour. Neurosci Biobehav Rev 2022; 140:104766. [DOI: 10.1016/j.neubiorev.2022.104766] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2022] [Revised: 06/24/2022] [Accepted: 07/01/2022] [Indexed: 10/17/2022]
|
4
|
Katayama R, Yoshida W, Ishii S. Confidence modulates the decodability of scene prediction during partially-observable maze exploration in humans. Commun Biol 2022; 5:367. [PMID: 35440615 PMCID: PMC9018866 DOI: 10.1038/s42003-022-03314-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/14/2021] [Accepted: 03/23/2022] [Indexed: 11/23/2022] Open
Abstract
Prediction ability often involves some degree of uncertainty-a key determinant of confidence. Here, we sought to assess whether predictions are decodable in partially-observable environments where one's state is uncertain, and whether this information is sensitive to confidence produced by such uncertainty. We used functional magnetic resonance imaging-based, partially-observable maze navigation tasks in which subjects predicted upcoming scenes and reported their confidence regarding these predictions. Using a multi-voxel pattern analysis, we successfully decoded both scene predictions and subjective confidence from activities in the localized parietal and prefrontal regions. We also assessed confidence in their beliefs about where they were in the maze. Importantly, prediction decodability varied according to subjective scene confidence in the superior parietal lobule and state confidence estimated by the behavioral model in the inferior parietal lobule. These results demonstrate that prediction in uncertain environments depends on the prefrontal-parietal network within which prediction and confidence interact.
Collapse
Affiliation(s)
- Risa Katayama
- Graduate School of Informatics, Kyoto University, Kyoto, Kyoto, 606-8501, Japan.
| | - Wako Yoshida
- Nuffield Department of Clinical Neuroscience, University of Oxford, Oxford, OX3 9DU, UK
- Department of Neural Computation for Decision-making, Advanced Telecommunications Research Institute International, Soraku-gun, Kyoto, 619-0288, Japan
| | - Shin Ishii
- Graduate School of Informatics, Kyoto University, Kyoto, Kyoto, 606-8501, Japan
- Neural Information Analysis Laboratories, Advanced Telecommunications Research Institute International, Soraku-gun, Kyoto, 619-0288, Japan
- International Research Center for Neurointelligence, The University of Tokyo, Bunkyo-ku, Tokyo, 113-0033, Japan
| |
Collapse
|
5
|
|
6
|
Human Reaction Times: Linking Individual and Collective Behaviour Through Physics Modeling. Symmetry (Basel) 2021. [DOI: 10.3390/sym13030451] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/10/2023] Open
Abstract
An individual’s reaction time data to visual stimuli have usually been represented in Experimental Psychology by means of an ex-Gaussian function. In most previous works, researchers have mainly aimed at finding a meaning for the parameters of the ex-Gaussian function which are known to correlate with cognitive disorders. Based on the recent evidence of correlations between the reaction time series to visual stimuli produced by different individuals within a group, we go beyond and propose a Physics-inspired model to represent the reaction time data of a coetaneous group of individuals. In doing so, a Maxwell–Boltzmann-like distribution appeared, the same distribution as for the velocities of the molecules in an Ideal Gas model. We describe step by step the methodology we use to go from the individual reaction times to the distribution of the individuals response within the coetaneous group. In practical terms, by means of this model we also provide a simple entropy-based methodology for the classification of the individuals within the collective they belong to with no need for an external reference which can be applicable in diverse areas of social sciences.
Collapse
|
7
|
Reaction times predict dynamic brain representations measured with MEG for only some object categorisation tasks. Neuropsychologia 2020; 151:107687. [PMID: 33212137 DOI: 10.1016/j.neuropsychologia.2020.107687] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2020] [Revised: 09/29/2020] [Accepted: 11/10/2020] [Indexed: 11/21/2022]
Abstract
Behavioural categorisation reaction times (RTs) provide a useful way to link behaviour to brain representations measured with neuroimaging. In this framework, objects are assumed to be represented in a multidimensional activation space, with the distances between object representations indicating their degree of neural similarity. Faster RTs have been reported to correlate with greater distances from a classification decision boundary for animacy. Objects inherently belong to more than one category, yet it is not known whether the RT-distance relationship, and its evolution over the time-course of the neural response, is similar across different categories. Here we used magnetoencephalography (MEG) to address this question. Our stimuli included typically animate and inanimate objects, as well as more ambiguous examples (i.e., robots and toys). We conducted four semantic categorisation tasks on the same stimulus set assessing animacy, living, moving, and human-similarity concepts, and linked the categorisation RTs to MEG time-series decoding data. Our results show a sustained RT-distance relationship throughout the time course of object processing for not only animacy, but also categorisation according to human-similarity. Interestingly, this sustained RT-distance relationship was not observed for the living and moving category organisations, despite comparable classification accuracy of the MEG data across all four category organisations. Our findings show that behavioural RTs predict representational distance for an organisational principle other than animacy, however further research is needed to determine why this relationship is observed only for some category organisations and not others.
Collapse
|
8
|
Machinery Failure Approach and Spectral Analysis to Study the Reaction Time Dynamics over Consecutive Visual Stimuli: An Entropy-Based Model. MATHEMATICS 2020. [DOI: 10.3390/math8111979] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/10/2023]
Abstract
The reaction times of individuals over consecutive visual stimuli have been studied using an entropy-based model and a failure machinery approach. The used tools include the fast Fourier transform and a spectral entropy analysis. The results indicate that the reaction times produced by the independently responding individuals to visual stimuli appear to be correlated. The spectral analysis and the entropy of the spectrum yield that there are features of similarity in the response times of each participant and among them. Furthermore, the analysis of the mistakes made by the participants during the reaction time experiments concluded that they follow a behavior which is consistent with the MTBF (Mean Time Between Failures) model, widely used in industry for the predictive diagnosis of electrical machines and equipment.
Collapse
|
9
|
Tovar DA, Murray MM, Wallace MT. Selective Enhancement of Object Representations through Multisensory Integration. J Neurosci 2020; 40:5604-5615. [PMID: 32499378 PMCID: PMC7363464 DOI: 10.1523/jneurosci.2139-19.2020] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/27/2019] [Revised: 04/17/2020] [Accepted: 05/21/2020] [Indexed: 11/21/2022] Open
Abstract
Objects are the fundamental building blocks of how we create a representation of the external world. One major distinction among objects is between those that are animate versus those that are inanimate. In addition, many objects are specified by more than a single sense, yet the nature by which multisensory objects are represented by the brain remains poorly understood. Using representational similarity analysis of male and female human EEG signals, we show enhanced encoding of audiovisual objects when compared with their corresponding visual and auditory objects. Surprisingly, we discovered that the often-found processing advantages for animate objects were not evident under multisensory conditions. This was due to a greater neural enhancement of inanimate objects-which are more weakly encoded under unisensory conditions. Further analysis showed that the selective enhancement of inanimate audiovisual objects corresponded with an increase in shared representations across brain areas, suggesting that the enhancement was mediated by multisensory integration. Moreover, a distance-to-bound analysis provided critical links between neural findings and behavior. Improvements in neural decoding at the individual exemplar level for audiovisual inanimate objects predicted reaction time differences between multisensory and unisensory presentations during a Go/No-Go animate categorization task. Links between neural activity and behavioral measures were most evident at intervals of 100-200 ms and 350-500 ms after stimulus presentation, corresponding to time periods associated with sensory evidence accumulation and decision-making, respectively. Collectively, these findings provide key insights into a fundamental process the brain uses to maximize the information it captures across sensory systems to perform object recognition.SIGNIFICANCE STATEMENT Our world is filled with ever-changing sensory information that we are able to seamlessly transform into a coherent and meaningful perceptual experience. We accomplish this feat by combining different stimulus features into objects. However, despite the fact that these features span multiple senses, little is known about how the brain combines the various forms of sensory information into object representations. Here, we used EEG and machine learning to study how the brain processes auditory, visual, and audiovisual objects. Surprisingly, we found that nonliving (i.e., inanimate) objects, which are more difficult to process with one sense alone, benefited the most from engaging multiple senses.
Collapse
Affiliation(s)
- David A Tovar
- School of Medicine, Vanderbilt University, Nashville, Tennessee 37240
- Vanderbilt Brain Institute, Vanderbilt University, Nashville, Tennessee 37240
| | - Micah M Murray
- The Laboratory for Investigative Neurophysiology (The LINE), Department of Radiology, Lausanne University Hospital and University of Lausanne (CHUV-UNIL), 1011 Lausanne, Switzerland
- Sensory, Cognitive and Perceptual Neuroscience Section, Center for Biomedical Imaging (CIBM) of Lausanne and Geneva, 1015 Lausanne, Switzerland
- Department of Ophthalmology, Fondation Asile des aveugles and University of Lausanne, 1002 Lausanne, Switzerland
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, Tennessee 37240
| | - Mark T Wallace
- School of Medicine, Vanderbilt University, Nashville, Tennessee 37240
- Vanderbilt Brain Institute, Vanderbilt University, Nashville, Tennessee 37240
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, Tennessee 37240
- Department of Psychology, Vanderbilt University, Nashville, Tennessee 37240
- Department of Psychiatry and Behavioral Sciences, Vanderbilt University Medical Center, Nashville, Tennessee 37240
- Department of Pharmacology, Vanderbilt University, Nashville, Tennessee 37240
| |
Collapse
|
10
|
|
11
|
Zeman AA, Ritchie JB, Bracci S, Op de Beeck H. Orthogonal Representations of Object Shape and Category in Deep Convolutional Neural Networks and Human Visual Cortex. Sci Rep 2020; 10:2453. [PMID: 32051467 PMCID: PMC7016009 DOI: 10.1038/s41598-020-59175-0] [Citation(s) in RCA: 35] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2019] [Accepted: 01/22/2020] [Indexed: 11/16/2022] Open
Abstract
Deep Convolutional Neural Networks (CNNs) are gaining traction as the benchmark model of visual object recognition, with performance now surpassing humans. While CNNs can accurately assign one image to potentially thousands of categories, network performance could be the result of layers that are tuned to represent the visual shape of objects, rather than object category, since both are often confounded in natural images. Using two stimulus sets that explicitly dissociate shape from category, we correlate these two types of information with each layer of multiple CNNs. We also compare CNN output with fMRI activation along the human visual ventral stream by correlating artificial with neural representations. We find that CNNs encode category information independently from shape, peaking at the final fully connected layer in all tested CNN architectures. Comparing CNNs with fMRI brain data, early visual cortex (V1) and early layers of CNNs encode shape information. Anterior ventral temporal cortex encodes category information, which correlates best with the final layer of CNNs. The interaction between shape and category that is found along the human visual ventral pathway is echoed in multiple deep networks. Our results suggest CNNs represent category information independently from shape, much like the human visual system.
Collapse
Affiliation(s)
- Astrid A Zeman
- Department of Brain and Cognition & Leuven Brain Institute, KU Leuven, Leuven, Belgium.
| | - J Brendan Ritchie
- Department of Brain and Cognition & Leuven Brain Institute, KU Leuven, Leuven, Belgium
| | - Stefania Bracci
- Department of Brain and Cognition & Leuven Brain Institute, KU Leuven, Leuven, Belgium
- Center for Mind/Brain Sciences, University of Trento, Rovereto, Italy
| | - Hans Op de Beeck
- Department of Brain and Cognition & Leuven Brain Institute, KU Leuven, Leuven, Belgium
| |
Collapse
|