101
|
De Cesarei A, Cavicchi S, Micucci A, Codispoti M. Categorization Goals Modulate the Use of Natural Scene Statistics. J Cogn Neurosci 2019; 31:109-125. [DOI: 10.1162/jocn_a_01333] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Understanding natural scenes involves the contribution of bottom–up analysis and top–down modulatory processes. However, the interaction of these processes during the categorization of natural scenes is not well understood. In the current study, we approached this issue using ERPs and behavioral and computational data. We presented pictures of natural scenes and asked participants to categorize them in response to different questions (Is it an animal/vehicle? Is it indoors/outdoors? Are there one/two foreground elements?). ERPs for target scenes requiring a “yes” response began to differ from those of nontarget scenes, beginning at 250 msec from picture onset, and this ERP difference was unmodulated by the categorization questions. Earlier ERPs showed category-specific differences (e.g., between animals and vehicles), which were associated with the processing of scene statistics. From 180 msec after scene onset, these category-specific ERP differences were modulated by the categorization question that was asked. Categorization goals do not modulate only later stages associated with target/nontarget decision but also earlier perceptual stages, which are involved in the processing of scene statistics.
Collapse
|
102
|
Greene MR. The information content of scene categories. PSYCHOLOGY OF LEARNING AND MOTIVATION 2019. [DOI: 10.1016/bs.plm.2019.03.004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
|
103
|
Wang X, Gu J, Xu J, Li X, Geng J, Wang B, Liu B. Decoding natural scenes based on sounds of objects within scenes using multivariate pattern analysis. Neurosci Res 2018; 148:9-18. [PMID: 30513353 DOI: 10.1016/j.neures.2018.11.009] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2018] [Revised: 11/21/2018] [Accepted: 11/30/2018] [Indexed: 10/27/2022]
Abstract
Scene recognition plays an important role in spatial navigation and scene classification. It remains unknown whether the occipitotemporal cortex could represent the semantic association between the scenes and sounds of objects within the scenes. In this study, we used the functional magnetic resonance imaging (fMRI) technique and multivariate pattern analysis to assess whether diff ; ;erent scenes could be discriminated based on the patterns evoked by sounds of objects within the scenes. We found that patterns evoked by scenes could be predicted with patterns evoked by sounds of objects within the scenes in the posterior fusiform area (pF), lateral occipital area (LO) and superior temporal sulcus (STS). The further functional connectivity analysis suggested significant correlations between pF, LO and parahippocampal place area (PPA) except that between STS and other three regions under the scene and sound conditions. A distinct network in processing scenes and sounds was discovered using a seed-to-voxel analysis with STS as the seed. This study may provide a cross-modal channel of scene decoding through the sounds of objects within the scenes in the occipitotemporal cortex, which could complement the single-modal channel of scene decoding based on the global scene properties or objects within the scenes.
Collapse
Affiliation(s)
- Xiaojing Wang
- College of Intelligence and Computing, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin, 300350, China
| | - Jin Gu
- College of Intelligence and Computing, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin, 300350, China
| | - Junhai Xu
- College of Intelligence and Computing, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin, 300350, China
| | - Xianglin Li
- Medical Imaging Research Institute, Binzhou Medical University, Yantai, Shandong 264003, China
| | - Junzu Geng
- Department of Radiology, Yantai Affiliated Hospital of Binzhou Medical University, Yantai, Shandong, 264003, China
| | - Bin Wang
- Medical Imaging Research Institute, Binzhou Medical University, Yantai, Shandong 264003, China
| | - Baolin Liu
- College of Intelligence and Computing, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin, 300350, China; State Key Laboratory of Intelligent Technology and Systems, National Laboratory for Information Science and Technology, Tsinghua University, Beijing, 100084, China.
| |
Collapse
|
104
|
Liesefeld HR, Liesefeld AM, Müller HJ. Two good reasons to say 'change!' - ensemble representations as well as item representations impact standard measures of VWM capacity. Br J Psychol 2018; 110:328-356. [PMID: 30506907 DOI: 10.1111/bjop.12359] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2018] [Revised: 09/18/2018] [Indexed: 11/26/2022]
Abstract
Visual working memory (VWM) is a central bottleneck in human information processing. Its capacity is most often measured in terms of how many individual-item representations VWM can hold (k). In the standard task employed to estimate k, an array of highly discriminable colour patches is maintained and, after a short retention interval, compared to a test display (change detection). Recent research has shown that with more complex, structured displays, change-detection performance is, in addition to individual-item representations, supported by ensemble representations formed as a result of spatial subgroupings. Here, by asking participants to additionally localize the change, we reveal indication for an influence of ensemble representations even in the very simple, unstructured displays of the colour-patch change-detection task. Critically, pure-item models from which standard formulae of k are derived do not consider ensemble representations and, therefore, potentially overestimate k. To gauge this overestimation, we develop an item-plus-ensemble model of change detection and change localization. Estimates of k from this new model are about 1 item (~30%) lower than the estimates from traditional pure-item models, even if derived from the same data sets.
Collapse
Affiliation(s)
- Heinrich René Liesefeld
- Department Psychologie, Ludwig-Maximilians-Universität München, Germany.,Graduate School of Systemic Neurosciences, Ludwig-Maximilians-Universität München, Germany
| | - Anna M Liesefeld
- Department Psychologie, Ludwig-Maximilians-Universität München, Germany
| | - Hermann J Müller
- Department Psychologie, Ludwig-Maximilians-Universität München, Germany.,Department of Psychological Sciences, Birkbeck College, University of London, UK
| |
Collapse
|
105
|
Lescroart MD, Gallant JL. Human Scene-Selective Areas Represent 3D Configurations of Surfaces. Neuron 2018; 101:178-192.e7. [PMID: 30497771 DOI: 10.1016/j.neuron.2018.11.004] [Citation(s) in RCA: 63] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2017] [Revised: 08/01/2018] [Accepted: 11/02/2018] [Indexed: 10/27/2022]
Abstract
It has been argued that scene-selective areas in the human brain represent both the 3D structure of the local visual environment and low-level 2D features (such as spatial frequency) that provide cues for 3D structure. To evaluate the degree to which each of these hypotheses explains variance in scene-selective areas, we develop an encoding model of 3D scene structure and test it against a model of low-level 2D features. We fit the models to fMRI data recorded while subjects viewed visual scenes. The fit models reveal that scene-selective areas represent the distance to and orientation of large surfaces, at least partly independent of low-level features. Principal component analysis of the model weights reveals that the most important dimensions of 3D structure are distance and openness. Finally, reconstructions of the stimuli based on the model weights demonstrate that our model captures unprecedented detail about the local visual environment from scene-selective areas.
Collapse
Affiliation(s)
- Mark D Lescroart
- Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, CA 94720, USA
| | - Jack L Gallant
- Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, CA 94720, USA; Department of Psychology, University of California, Berkeley, Berkeley, CA 94720, USA.
| |
Collapse
|
106
|
Dissociable Neural Systems for Recognizing Places and Navigating through Them. J Neurosci 2018; 38:10295-10304. [PMID: 30348675 DOI: 10.1523/jneurosci.1200-18.2018] [Citation(s) in RCA: 28] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2018] [Revised: 09/19/2018] [Accepted: 09/24/2018] [Indexed: 02/05/2023] Open
Abstract
When entering an environment, we can use the present visual information from the scene to either recognize the kind of place it is (e.g., a kitchen or a bedroom) or navigate through it. Here we directly test the hypothesis that these two processes, what we call "scene categorization" and "visually-guided navigation", are supported by dissociable neural systems. Specifically, we manipulated task demands by asking human participants (male and female) to perform a scene categorization, visually-guided navigation, and baseline task on images of scenes, and measured both the average univariate responses and multivariate spatial pattern of responses within two scene-selective cortical regions, the parahippocampal place area (PPA) and occipital place area (OPA), hypothesized to be separably involved in scene categorization and visually-guided navigation, respectively. As predicted, in the univariate analysis, PPA responded significantly more during the categorization task than during both the navigation and baseline tasks, whereas OPA showed the complete opposite pattern. Similarly, in the multivariate analysis, a linear support vector machine achieved above-chance classification for the categorization task, but not the navigation task in PPA. By contrast, above-chance classification was achieved for both the navigation and categorization tasks in OPA. However, above-chance classification for both tasks was also found in early visual cortex and hence not specific to OPA, suggesting that the spatial patterns of responses in OPA are merely inherited from early vision, and thus may be epiphenomenal to behavior. Together, these results are evidence for dissociable neural systems involved in recognizing places and navigating through them.SIGNIFICANCE STATEMENT It has been nearly three decades since Goodale and Milner demonstrated that recognizing objects and manipulating them involve distinct neural processes. Today we show the same is true of our interactions with our environment: recognizing places and navigating through them are neurally dissociable. More specifically, we found that a scene-selective region, the parahippocampal place area, is active when participants are asked to categorize a scene, but not when asked to imagine navigating through it, whereas another scene-selective region, the occipital place area, shows the exact opposite pattern. This double dissociation is evidence for dissociable neural systems within scene processing, similar to the bifurcation of object processing described by Goodale and Milner (1992).
Collapse
|
107
|
Lauer T, Cornelissen THW, Draschkow D, Willenbockel V, Võ MLH. The role of scene summary statistics in object recognition. Sci Rep 2018; 8:14666. [PMID: 30279431 PMCID: PMC6168578 DOI: 10.1038/s41598-018-32991-1] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2017] [Accepted: 09/19/2018] [Indexed: 12/04/2022] Open
Abstract
Objects that are semantically related to the visual scene context are typically better recognized than unrelated objects. While context effects on object recognition are well studied, the question which particular visual information of an object's surroundings modulates its semantic processing is still unresolved. Typically, one would expect contextual influences to arise from high-level, semantic components of a scene but what if even low-level features could modulate object processing? Here, we generated seemingly meaningless textures of real-world scenes, which preserved similar summary statistics but discarded spatial layout information. In Experiment 1, participants categorized such textures better than colour controls that lacked higher-order scene statistics while original scenes resulted in the highest performance. In Experiment 2, participants recognized briefly presented consistent objects on scenes significantly better than inconsistent objects, whereas on textures, consistent objects were recognized only slightly more accurately. In Experiment 3, we recorded event-related potentials and observed a pronounced mid-central negativity in the N300/N400 time windows for inconsistent relative to consistent objects on scenes. Critically, inconsistent objects on textures also triggered N300/N400 effects with a comparable time course, though less pronounced. Our results suggest that a scene's low-level features contribute to the effective processing of objects in complex real-world environments.
Collapse
Affiliation(s)
- Tim Lauer
- Scene Grammar Lab, Department of Psychology, Goethe University Frankfurt, Frankfurt am Main, Germany.
| | - Tim H W Cornelissen
- Scene Grammar Lab, Department of Psychology, Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Dejan Draschkow
- Scene Grammar Lab, Department of Psychology, Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Verena Willenbockel
- Scene Grammar Lab, Department of Psychology, Goethe University Frankfurt, Frankfurt am Main, Germany
- Department of Psychology, University of Victoria, Victoria, BC, Canada
| | - Melissa L-H Võ
- Scene Grammar Lab, Department of Psychology, Goethe University Frankfurt, Frankfurt am Main, Germany
| |
Collapse
|
108
|
Stronger shared taste for natural aesthetic domains than for artifacts of human culture. Cognition 2018; 179:121-131. [DOI: 10.1016/j.cognition.2018.06.009] [Citation(s) in RCA: 61] [Impact Index Per Article: 8.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2017] [Revised: 06/05/2018] [Accepted: 06/12/2018] [Indexed: 11/20/2022]
|
109
|
Gong M, Xuan Y, Smart LJ, Olzak LA. The extraction of natural scene gist in visual crowding. Sci Rep 2018; 8:14073. [PMID: 30232470 PMCID: PMC6145949 DOI: 10.1038/s41598-018-32455-6] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2018] [Accepted: 09/10/2018] [Indexed: 12/02/2022] Open
Abstract
The gist of natural scenes can be extracted very rapidly and even without focal attention. However, it is unclear whether and to what extent the gist of natural scenes can break through the bottleneck of crowding, a phenomenon in which object recognition will be immensely impaired. In the first two experiments, a target scene, either presented alone or surrounded by four flankers, was categorized at basic (Experiment 1) or global levels (Experiment 2). It was showed that the elimination of high-level semantic information of flankers greatly alleviated the crowding effect, demonstrating that high-level information played an important role in crowding of scene gist. More importantly, participants were able to categorize the scenes in crowding at rather high accuracies, suggesting that the extraction of scene gist might be a prioritized process. To test this hypothesis, in Experiment 3 we compared the crowding effect of three types of stimuli, namely, scenes, facial expressions and letter "E"s. The results showed that scenes could be better categorized than the other two types of stimuli in the crowding condition. This scene gist advantage thus supported our hypothesis. Together, the present studies suggest that scene gist is highly recognizable in crowding, probably due to its prioritization in visual processing.
Collapse
Affiliation(s)
- Mingliang Gong
- Department of Psychology, Miami University, Oxford, OH, USA.
| | - Yuming Xuan
- State Key Laboratory of Brain and Cognitive Science, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.
- Department of Psychology, University of the Chinese Academy of Sciences, Beijing, China.
| | - L James Smart
- Department of Psychology, Miami University, Oxford, OH, USA
| | - Lynn A Olzak
- Department of Psychology, Miami University, Oxford, OH, USA
| |
Collapse
|
110
|
Miao Q, Zhang G, Yan W, Liu B. Investigating the Brain Neural Mechanism when Signature Objects were Masked during a Scene Categorization Task using Functional MRI. Neuroscience 2018; 388:248-262. [PMID: 30056114 DOI: 10.1016/j.neuroscience.2018.07.030] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2018] [Revised: 07/17/2018] [Accepted: 07/18/2018] [Indexed: 11/17/2022]
Abstract
Objects play vital roles in scene categorization. Although a number of studies have researched on the neural responses during object and object-based scene recognition, few studies have investigated the neural mechanism underlying object-masked scene categorization. Here, we used functional magnetic resonance imaging (fMRI) to measure the changes in brain activations and functional connectivity (FC) while subjects performed a visual scene-categorization task with different numbers of 'signature objects' masked. The object-selective region in the lateral occipital complex (LOC) showed a decrease in activations and changes in FC with the default mode network (DMN), indicating changes in object attention after the masking of signature objects. Changes in top-down modulation effect were revealed in the FC from the dorsolateral prefrontal cortex (DLPFC) to LOC and the extrastriate visual cortex, possibly participating in conscious object recognition. The whole-brain analyses showed the participation of fronto-parietal network (FPN) in scene categorization judgment, and right DLPFC served as the core hub in this network. Another core hub was found in left middle temporal gyrus (MTG) and its connection with middle cingulate cortex (MCC), supramarginal gyrus (SMG) and insula might serve in the processing of motor response and the semantic relations between objects and scenes. Brain-behavior correlation analysis substantiated the contributions of the FC to the different processes in the object-masked scene-categorization tasks. Altogether, the results suggest that masking of objects significantly affected the object attention, cognitive demand, top-down modulation effect, and semantic judgment.
Collapse
Affiliation(s)
- Qiaomu Miao
- School of Computer Science and Technology, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin 300350, PR China
| | - Gaoyan Zhang
- School of Computer Science and Technology, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin 300350, PR China
| | - Weiran Yan
- School of Computer Science and Technology, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin 300350, PR China
| | - Baolin Liu
- School of Computer Science and Technology, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin 300350, PR China; State Key Laboratory of Intelligent Technology and Systems, National Laboratory for Information Science and Technology, Tsinghua University, Beijing 100084, PR China.
| |
Collapse
|
111
|
Ryan G, Mosca A, Chang R, Wu E. At a Glance: Pixel Approximate Entropy as a Measure of Line Chart Complexity. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2018; 25:872-881. [PMID: 30137006 DOI: 10.1109/tvcg.2018.2865264] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
When inspecting information visualizations under time critical settings, such as emergency response or monitoring the heart rate in a surgery room, the user only has a small amount of time to view the visualization "at a glance". In these settings, it is important to provide a quantitative measure of the visualization to understand whether or not the visualization is too "complex" to accurately judge at a glance. This paper proposes Pixel Approximate Entropy (PAE), which adapts the approximate entropy statistical measure commonly used to quantify regularity and unpredictability in time-series data, as a measure of visual complexity for line charts. We show that PAE is correlated with user-perceived chart complexity, and that increased chart PAE correlates with reduced judgement accuracy. 'We also find that the correlation between PAE values and participants' judgment increases when the user has less time to examine the line charts.
Collapse
|
112
|
van den Bosch KAM, Welch D, Andringa TC. The Evolution of Soundscape Appraisal Through Enactive Cognition. Front Psychol 2018; 9:1129. [PMID: 30038591 PMCID: PMC6046435 DOI: 10.3389/fpsyg.2018.01129] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2017] [Accepted: 06/13/2018] [Indexed: 11/13/2022] Open
Abstract
We propose a framework based on evolutionary principles and the theory of enactive cognition ("being by doing"), that addresses the foundation of key results and central questions of soundscape research. We hypothesize that the two main descriptors (measures of how people perceive the acoustic environment) of soundscape appraisal ('pleasantness' and 'eventfulness'), reflect evolutionarily old motivational and affective systems that promote survival through preferences for certain environments and avoidance of others. Survival is aimed at ending or avoiding existential threats and protecting viability in a deficient environment. On the other hand, flourishing occurs whenever survival is not an immediate concern and aims to improve the agent's viability and by co-creating ever better conditions for existence. As such, survival is experienced as unpleasant, and deals with immediate problems to be ended or avoided, while flourishing is enjoyable, and therefore to be aimed for and maintained. Therefore, the simplest, safety-relevant meaning attributable to soundscapes (audible safety) should be key to understanding soundscape appraisal. To strengthen this, we show that the auditory nervous system is intimately connected to the parts of our brains associated with arousal and emotions. Furthermore, our theory demonstrates that 'complexity' and 'affordance content' of the perceived environment are important underlying soundscape indicators (measures used to predict the value of a soundscape descriptor). Consideration of these indicators allows the same soundscape to be viewed from a second perspective; one driven more by meaning attribution characteristics than merely emotional appraisal. The synthesis of both perspectives of the same person-environment interaction thus consolidates the affective, informational, and even the activity related perspectives on soundscape appraisal. Furthermore, we hypothesize that our current habitats are not well matched to our, evolutionarily old, auditory warning systems, and that we consequently have difficulty establishing audible safety. This leads to more negative and aroused moods and emotions, with stress-related symptoms as a result.
Collapse
Affiliation(s)
| | - David Welch
- School of Population Health, University of Auckland, Auckland, New Zealand
| | - Tjeerd C Andringa
- SoundAppraisal Ltd., Groningen, Netherlands.,University College Groningen, University of Groningen, Groningen, Netherlands
| |
Collapse
|
113
|
Lowe MX, Rajsic J, Ferber S, Walther DB. Discriminating scene categories from brain activity within 100 milliseconds. Cortex 2018; 106:275-287. [PMID: 30037637 DOI: 10.1016/j.cortex.2018.06.006] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2017] [Revised: 02/26/2018] [Accepted: 06/01/2018] [Indexed: 10/28/2022]
Abstract
Humans have the ability to make sense of the world around them in only a single glance. This astonishing feat requires the visual system to extract information from our environment with remarkable speed. How quickly does this process unfold across time, and what visual information contributes to our understanding of the visual world? We address these questions by directly measuring the temporal dynamics of the perception of colour photographs and line drawings of scenes with electroencephalography (EEG) during a scene-memorization task. Within a fraction of a second, event-related potentials (ERPs) show dissociable response patterns for global scene properties of content (natural versus manmade) and layout (open versus closed). Subsequent detailed analyses of within-category versus between-category discriminations found significant dissociations of basic-level scene categories (e.g., forest; city) within the first 100 msec of perception. The similarity of this neural activity with feature-based discriminations suggests low-level image statistics may be foundational for this rapid categorization. Interestingly, our results also suggest that the structure preserved in line drawings may form a primary and necessary basis for visual processing, whereas surface information may further enhance category selectivity in later-stage processing. Critically, these findings provide evidence that the distinction of both basic-level categories and global properties of scenes from neural signals occurs within 100 msec.
Collapse
Affiliation(s)
| | - Jason Rajsic
- Psychology Department, University of Toronto, Canada
| | - Susanne Ferber
- Psychology Department, University of Toronto, Canada; Rotman Research Institute, Baycrest, Toronto, Canada
| | - Dirk B Walther
- Psychology Department, University of Toronto, Canada; Rotman Research Institute, Baycrest, Toronto, Canada
| |
Collapse
|
114
|
Brand J, Johnson AP. The effects of distributed and focused attention on rapid scene categorization. VISUAL COGNITION 2018. [DOI: 10.1080/13506285.2018.1485808] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
Affiliation(s)
- John Brand
- Department of Epidemiology, Geisel School of Medicine Dartmouth College, Hanover, USA
| | - Aaron P. Johnson
- Department of Psychology, Concordia University, Montreal, Canada
| |
Collapse
|
115
|
Gianni E, De Zorzi L, Lee SA. The developing role of transparent surfaces in children's spatial representation. Cogn Psychol 2018; 105:39-52. [PMID: 29920399 DOI: 10.1016/j.cogpsych.2018.05.003] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2018] [Revised: 04/17/2018] [Accepted: 05/30/2018] [Indexed: 11/28/2022]
Abstract
Children adeptly use environmental boundaries to navigate. But how do they represent surfaces as boundaries, and how does this change over development? To investigate the effects of boundaries as visual and physical barriers, we tested spatial reorientation in 160 children (2-7 year-olds) in a transparent rectangular arena (Condition 1). In contrast with their consistent success using opaque surfaces (Condition 2), children only succeeded at using transparent surfaces at 5-7 years of age. These results suggest a critical role of visually opaque surfaces in early spatial coding and a developmental change around the age of five in representing locations with respect to transparent surfaces. In application, these findings may inform our usage of windows and glass surfaces in designing and building environments occupied by young children.
Collapse
Affiliation(s)
- Eugenia Gianni
- Center for Mind/Brain Sciences, University of Trento, Corso Bettini 31, Rovereto, Italy
| | - Laura De Zorzi
- Department of Psychology and Cognitive Science, Corso Bettini 84, Rovereto, Italy
| | - Sang Ah Lee
- Department of Bio and Brain Engineering, Korea Advanced Institute of Science and Technology, Daehak-ro 291, Daejeon, Republic of Korea; Center for Mind/Brain Sciences, University of Trento, Corso Bettini 31, Rovereto, Italy.
| |
Collapse
|
116
|
Establishing reference scales for scene naturalness and openness : Naturalness and openness scales. Behav Res Methods 2018; 51:1179-1186. [PMID: 29845553 DOI: 10.3758/s13428-018-1053-4] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
A key question in the field of scene perception is what information people use when making decisions about images of scenes. A significant body of evidence has indicated the importance of global properties of a scene image. Ideally, well-controlled, real-world images would be used to examine the influence of these properties on perception. Unfortunately, real-world images are generally complex and impractical to control. In the current research, we elicit ratings of naturalness and openness from a large number of subjects using Amazon Mechanic Turk. Subjects were asked to indicate which of a randomly chosen pair of scene images was more representative of a global property. A score and rank for each image was then estimated based on those comparisons using the Bradley-Terry-Luce model. These ranked images offer the opportunity to exercise control over the global scene properties in stimulus set drawn from complex real-world images. This will allow a deeper exploration of the relationship between global scene properties and behavioral and neural responses.
Collapse
|
117
|
Jahanian A, Keshvari S, Rosenholtz R. Web pages: What can you see in a single fixation? COGNITIVE RESEARCH-PRINCIPLES AND IMPLICATIONS 2018; 3:14. [PMID: 29774229 PMCID: PMC5945715 DOI: 10.1186/s41235-018-0099-2] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/18/2017] [Accepted: 03/23/2018] [Indexed: 11/10/2022]
Abstract
Research in human vision suggests that in a single fixation, humans can extract a significant amount of information from a natural scene, e.g. the semantic category, spatial layout, and object identities. This ability is useful, for example, for quickly determining location, navigating around obstacles, detecting threats, and guiding eye movements to gather more information. In this paper, we ask a new question: What can we see at a glance at a web page – an artificial yet complex “real world” stimulus? Is it possible to notice the type of website, or where the relevant elements are, with only a glimpse? We find that observers, fixating at the center of a web page shown for only 120 milliseconds, are well above chance at classifying the page into one of ten categories. Furthermore, this ability is supported in part by text that they can read at a glance. Users can also understand the spatial layout well enough to reliably localize the menu bar and to detect ads, even though the latter are often camouflaged among other graphical elements. We discuss the parallels between web page gist and scene gist, and the implications of our findings for both vision science and human-computer interaction.
Collapse
Affiliation(s)
- Ali Jahanian
- Department: Computer Science and Artificial Intelligence Laboratory (CSAIL), Institution: Massachusetts Institute of Technology, Cambridge, MA USA
| | - Shaiyan Keshvari
- Department: Computer Science and Artificial Intelligence Laboratory (CSAIL), Institution: Massachusetts Institute of Technology, Cambridge, MA USA
| | - Ruth Rosenholtz
- Department: Computer Science and Artificial Intelligence Laboratory (CSAIL), Institution: Massachusetts Institute of Technology, Cambridge, MA USA
| |
Collapse
|
118
|
Hansen NE, Noesen BT, Nador JD, Harel A. The influence of behavioral relevance on the processing of global scene properties: An ERP study. Neuropsychologia 2018; 114:168-180. [PMID: 29729276 DOI: 10.1016/j.neuropsychologia.2018.04.040] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2017] [Revised: 04/27/2018] [Accepted: 04/30/2018] [Indexed: 12/01/2022]
Abstract
Recent work studying the temporal dynamics of visual scene processing (Harel et al., 2016) has found that global scene properties (GSPs) modulate the amplitude of early Event-Related Potentials (ERPs). It is still not clear, however, to what extent the processing of these GSPs is influenced by their behavioral relevance, determined by the goals of the observer. To address this question, we investigated how behavioral relevance, operationalized by the task context impacts the electrophysiological responses to GSPs. In a set of two experiments we recorded ERPs while participants viewed images of real-world scenes, varying along two GSPs, naturalness (manmade/natural) and spatial expanse (open/closed). In Experiment 1, very little attention to scene content was required as participants viewed the scenes while performing an orthogonal fixation-cross task. In Experiment 2 participants saw the same scenes but now had to actively categorize them, based either on their naturalness or spatial expense. We found that task context had very little impact on the early ERP responses to the naturalness and spatial expanse of the scenes: P1, N1, and P2 could distinguish between open and closed scenes and between manmade and natural scenes across both experiments. Further, the specific effects of naturalness and spatial expanse on the ERP components were largely unaffected by their relevance for the task. A task effect was found at the N1 and P2 level, but this effect was manifest across all scene dimensions, indicating a general effect rather than an interaction between task context and GSPs. Together, these findings suggest that the extraction of global scene information reflected in the early ERP components is rapid and very little influenced by top-down observer-based goals.
Collapse
Affiliation(s)
- Natalie E Hansen
- Department of Psychology, Wright State University, Dayton, OH, United States
| | - Birken T Noesen
- Department of Psychology, Wright State University, Dayton, OH, United States
| | - Jeffrey D Nador
- Department of Psychology, Wright State University, Dayton, OH, United States
| | - Assaf Harel
- Department of Psychology, Wright State University, Dayton, OH, United States.
| |
Collapse
|
119
|
Schwabe K, Menzel C, Mullin C, Wagemans J, Redies C. Gist Perception of Image Composition in Abstract Artworks. Iperception 2018; 9:2041669518780797. [PMID: 29977489 PMCID: PMC6024551 DOI: 10.1177/2041669518780797] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2018] [Accepted: 05/11/2018] [Indexed: 11/27/2022] Open
Abstract
Most recent studies in experimental aesthetics have focused on the cognitive processing of visual artworks. In contrast, the perception of formal compositional features of artworks has been studied less extensively. Here, we investigated whether fast and automatic processing of artistic image composition can lead to a stable and consistent aesthetic evaluation when cognitive processing is minimized or absent. To this aim, we compared aesthetic ratings on abstract artworks and their shuffled counterparts in a gist experiment. Results show that exposure times as short as 50 ms suffice for the participants to reach a stable and consistent rating on how ordered and harmonious the abstract stimuli were. Moreover, the rating scores for the 50 ms exposure time exhibited similar dependencies on image type and self-similarity and a similar pattern of correlations between different rating terms, as the rating scores for the long exposure time (3,000 ms). Ratings were less consistent for the term interesting and inconsistent for the term pleasing. Our results are compatible with a model of aesthetic experience, in which the early perceptual processing of the formal aspects of visual artworks can lead to a consistent aesthetic judgment, even if there is no cognitive contribution to this judgment.
Collapse
Affiliation(s)
- Kana Schwabe
- Experimental Aesthetics Group, Institute of Anatomy I, University of Jena School of Medicine, Germany
| | - Claudia Menzel
- Experimental Aesthetics Group, Institute of Anatomy I, University of Jena School of Medicine, Germany
| | - Caitlin Mullin
- Laboratory of Experimental Psychology, Brain & Cognition, University of Leuven (KU Leuven), Belgium
| | - Johan Wagemans
- Laboratory of Experimental Psychology, Brain & Cognition, University of Leuven (KU Leuven), Belgium
| | - Christoph Redies
- Experimental Aesthetics Group, Institute of Anatomy I, University of Jena School of Medicine, Germany; Laboratory of Experimental Psychology, Brain & Cognition, University of Leuven (KU Leuven), Belgium
| |
Collapse
|
120
|
Bonner MF, Epstein RA. Computational mechanisms underlying cortical responses to the affordance properties of visual scenes. PLoS Comput Biol 2018; 14:e1006111. [PMID: 29684011 PMCID: PMC5933806 DOI: 10.1371/journal.pcbi.1006111] [Citation(s) in RCA: 47] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2017] [Revised: 05/03/2018] [Accepted: 03/31/2018] [Indexed: 11/24/2022] Open
Abstract
Biologically inspired deep convolutional neural networks (CNNs), trained for computer vision tasks, have been found to predict cortical responses with remarkable accuracy. However, the internal operations of these models remain poorly understood, and the factors that account for their success are unknown. Here we develop a set of techniques for using CNNs to gain insights into the computational mechanisms underlying cortical responses. We focused on responses in the occipital place area (OPA), a scene-selective region of dorsal occipitoparietal cortex. In a previous study, we showed that fMRI activation patterns in the OPA contain information about the navigational affordances of scenes; that is, information about where one can and cannot move within the immediate environment. We hypothesized that this affordance information could be extracted using a set of purely feedforward computations. To test this idea, we examined a deep CNN with a feedforward architecture that had been previously trained for scene classification. We found that responses in the CNN to scene images were highly predictive of fMRI responses in the OPA. Moreover the CNN accounted for the portion of OPA variance relating to the navigational affordances of scenes. The CNN could thus serve as an image-computable candidate model of affordance-related responses in the OPA. We then ran a series of in silico experiments on this model to gain insights into its internal operations. These analyses showed that the computation of affordance-related features relied heavily on visual information at high-spatial frequencies and cardinal orientations, both of which have previously been identified as low-level stimulus preferences of scene-selective visual cortex. These computations also exhibited a strong preference for information in the lower visual field, which is consistent with known retinotopic biases in the OPA. Visualizations of feature selectivity within the CNN suggested that affordance-based responses encoded features that define the layout of the spatial environment, such as boundary-defining junctions and large extended surfaces. Together, these results map the sensory functions of the OPA onto a fully quantitative model that provides insights into its visual computations. More broadly, they advance integrative techniques for understanding visual cortex across multiple level of analysis: from the identification of cortical sensory functions to the modeling of their underlying algorithms. How does visual cortex compute behaviorally relevant properties of the local environment from sensory inputs? For decades, computational models have been able to explain only the earliest stages of biological vision, but recent advances in deep neural networks have yielded a breakthrough in the modeling of high-level visual cortex. However, these models are not explicitly designed for testing neurobiological theories, and, like the brain itself, their internal operations remain poorly understood. We examined a deep neural network for insights into the cortical representation of navigational affordances in visual scenes. In doing so, we developed a set of high-throughput techniques and statistical tools that are broadly useful for relating the internal operations of neural networks with the information processes of the brain. Our findings demonstrate that a deep neural network with purely feedforward computations can account for the processing of navigational layout in high-level visual cortex. We next performed a series of experiments and visualization analyses on this neural network. These analyses characterized a set of stimulus input features that may be critical for computing navigationally related cortical representations, and they identified a set of high-level, complex scene features that may serve as a basis set for the cortical coding of navigational layout. These findings suggest a computational mechanism through which high-level visual cortex might encode the spatial structure of the local navigational environment, and they demonstrate an experimental approach for leveraging the power of deep neural networks to understand the visual computations of the brain.
Collapse
Affiliation(s)
- Michael F. Bonner
- Department of Psychology, University of Pennsylvania, Philadelphia, PA, United States of America
- * E-mail:
| | - Russell A. Epstein
- Department of Psychology, University of Pennsylvania, Philadelphia, PA, United States of America
| |
Collapse
|
121
|
Carrigan AJ, Wardle SG, Rich AN. Finding cancer in mammograms: if you know it's there, do you know where? Cogn Res Princ Implic 2018; 3:10. [PMID: 29707615 PMCID: PMC5904219 DOI: 10.1186/s41235-018-0096-5] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2017] [Accepted: 02/27/2018] [Indexed: 11/23/2022] Open
Abstract
Humans can extract considerable information from scenes, even when these are presented extremely quickly. The ability of an experienced radiologist to rapidly detect an abnormality on a mammogram may build upon this general capacity. Although radiologists have been shown to be able to detect an abnormality 'above chance' at short durations, the extent to which abnormalities can be localised at brief presentations is less clear. Extending previous work, we presented radiologists with unilateral mammograms, 50% containing a mass, for 250 or 1000 ms. As the female breast varies with respect to the level of normal fibroglandular tissue, the images were categorised into high and low density (50% of each), resulting in difficult and easy searches, respectively. Participants were asked to decide whether there was an abnormality (detection) and then to locate the mass on a blank outline of the mammogram (localisation). We found both detection and localisation information for all conditions. Although there may be a dissociation between detection and localisation on a small proportion of trials, we find a number of factors that lead to the underestimation of localisation including stimulus variability, response imprecision and participant guesses. We emphasise the importance of taking these factors into account when interpreting results. The effect of density on detection and localisation highlights the importance of considering breast density in medical screening.
Collapse
Affiliation(s)
- Ann J. Carrigan
- Perception in Action Research Centre & Department of Cognitive Science, Macquarie University, Sydney, Australia
- ARC Centre of Excellence in Cognition & Its Disorders, Macquarie University, Sydney, Australia
- Centre for Elite Performance, Expertise, and Training, Macquarie University, Sydney, Australia
| | - Susan G. Wardle
- Perception in Action Research Centre & Department of Cognitive Science, Macquarie University, Sydney, Australia
- ARC Centre of Excellence in Cognition & Its Disorders, Macquarie University, Sydney, Australia
| | - Anina N. Rich
- Perception in Action Research Centre & Department of Cognitive Science, Macquarie University, Sydney, Australia
- ARC Centre of Excellence in Cognition & Its Disorders, Macquarie University, Sydney, Australia
- Centre for Elite Performance, Expertise, and Training, Macquarie University, Sydney, Australia
| |
Collapse
|
122
|
Sensitivity to stimulus similarity is associated with greater sustained attention ability. Atten Percept Psychophys 2018; 80:1390-1408. [DOI: 10.3758/s13414-018-1504-0] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
123
|
Jorge L, Canário N, Castelhano J, Castelo-Branco M. Processing of performance-matched visual object categories: faces and places are related to lower processing load in the frontoparietal executive network than other objects. Eur J Neurosci 2018; 47:938-946. [DOI: 10.1111/ejn.13892] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2017] [Revised: 02/22/2018] [Accepted: 02/23/2018] [Indexed: 11/27/2022]
Affiliation(s)
- Lília Jorge
- CIBIT, CNC.IBILI - Center for Biomedical Imaging and Translational Research; Faculty of Medicine; University of Coimbra; Coimbra Portugal
- ICNAS - Institute for Nuclear Sciences Applied to Health; Brain Imaging Network of Portugal; Coimbra Portugal
| | - Nádia Canário
- CIBIT, CNC.IBILI - Center for Biomedical Imaging and Translational Research; Faculty of Medicine; University of Coimbra; Coimbra Portugal
- ICNAS - Institute for Nuclear Sciences Applied to Health; Brain Imaging Network of Portugal; Coimbra Portugal
| | - João Castelhano
- CIBIT, CNC.IBILI - Center for Biomedical Imaging and Translational Research; Faculty of Medicine; University of Coimbra; Coimbra Portugal
- ICNAS - Institute for Nuclear Sciences Applied to Health; Brain Imaging Network of Portugal; Coimbra Portugal
| | - Miguel Castelo-Branco
- CIBIT, CNC.IBILI - Center for Biomedical Imaging and Translational Research; Faculty of Medicine; University of Coimbra; Coimbra Portugal
- ICNAS - Institute for Nuclear Sciences Applied to Health; Brain Imaging Network of Portugal; Coimbra Portugal
- Laboratório de Neurociências da Visão - IBILI; FMUC; Azinhaga Santa Comba; Celas Coimbra 3000-548 Portugal
| |
Collapse
|
124
|
Hafri A, Trueswell JC, Strickland B. Encoding of event roles from visual scenes is rapid, spontaneous, and interacts with higher-level visual processing. Cognition 2018; 175:36-52. [PMID: 29459238 DOI: 10.1016/j.cognition.2018.02.011] [Citation(s) in RCA: 36] [Impact Index Per Article: 5.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2017] [Revised: 02/06/2018] [Accepted: 02/08/2018] [Indexed: 11/24/2022]
Abstract
A crucial component of event recognition is understanding event roles, i.e. who acted on whom: boy hitting girl is different from girl hitting boy. We often categorize Agents (i.e. the actor) and Patients (i.e. the one acted upon) from visual input, but do we rapidly and spontaneously encode such roles even when our attention is otherwise occupied? In three experiments, participants observed a continuous sequence of two-person scenes and had to search for a target actor in each (the male/female or red/blue-shirted actor) by indicating with a button press whether the target appeared on the left or the right. Critically, although role was orthogonal to gender and shirt color, and was never explicitly mentioned, participants responded more slowly when the target's role switched from trial to trial (e.g., the male went from being the Patient to the Agent). In a final experiment, we demonstrated that this effect cannot be fully explained by differences in posture associated with Agents and Patients. Our results suggest that extraction of event structure from visual scenes is rapid and spontaneous.
Collapse
Affiliation(s)
- Alon Hafri
- Department of Psychology, University of Pennsylvania, 425 S. University Avenue, Philadelphia, PA 19104, USA.
| | - John C Trueswell
- Department of Psychology, University of Pennsylvania, 425 S. University Avenue, Philadelphia, PA 19104, USA
| | - Brent Strickland
- Département d'Etudes Cognitives, Ecole Normale Supérieure, PSL Research University, Institut Jean Nicod, (ENS, EHESS, CNRS), 75005 Paris, France
| |
Collapse
|
125
|
Wang S, Cao L, Xu J, Zhang G, Lou Y, Liu B. Revealing the Semantic Association between Perception of Scenes and Significant Objects by Representational Similarity Analysis. Neuroscience 2018; 372:87-96. [DOI: 10.1016/j.neuroscience.2017.12.043] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2017] [Revised: 12/19/2017] [Accepted: 12/23/2017] [Indexed: 11/29/2022]
|
126
|
Neumann MF, Ng R, Rhodes G, Palermo R. Ensemble coding of face identity is not independent of the coding of individual identity. Q J Exp Psychol (Hove) 2018; 71:1357-1366. [DOI: 10.1080/17470218.2017.1318409] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
Information about a group of similar objects can be summarized into a compressed code, known as ensemble coding. Ensemble coding of simple stimuli (e.g., groups of circles) can occur in the absence of detailed exemplar coding, suggesting dissociable processes. Here, we investigate whether a dissociation would still be apparent when coding facial identity, where individual exemplar information is much more important. We examined whether ensemble coding can occur when exemplar coding is difficult, as a result of large sets or short viewing times, or whether the two types of coding are positively associated. We found a positive association, whereby both ensemble and exemplar coding were reduced for larger groups and shorter viewing times. There was no evidence for ensemble coding in the absence of exemplar coding. At longer presentation times, there was an unexpected dissociation, where exemplar coding increased yet ensemble coding decreased, suggesting that robust information about face identity might suppress ensemble coding. Thus, for face identity, we did not find the classic dissociation—of access to ensemble information in the absence of detailed exemplar information—that has been used to support claims of distinct mechanisms for ensemble and exemplar coding.
Collapse
Affiliation(s)
- Markus F Neumann
- ARC Centre of Excellence in Cognition and Its Disorders, School of Psychological Science, The University of Western Australia, Crawley, WA, Australia
| | - Ryan Ng
- ARC Centre of Excellence in Cognition and Its Disorders, School of Psychological Science, The University of Western Australia, Crawley, WA, Australia
| | - Gillian Rhodes
- ARC Centre of Excellence in Cognition and Its Disorders, School of Psychological Science, The University of Western Australia, Crawley, WA, Australia
| | - Romina Palermo
- ARC Centre of Excellence in Cognition and Its Disorders, School of Psychological Science, The University of Western Australia, Crawley, WA, Australia
| |
Collapse
|
127
|
Abstract
The human visual system has the remarkable ability to rapidly detect meaning from visual stimuli. Potter, Wyble, Hagmann, and McCourt (Attention, Perception, & Psychophysics, 76, 270-279, 2014) tested the minimum viewing time required to obtain meaning from a stream of pictures shown in a rapid serial visual presentation (RSVP) sequence containing either six or 12 pictures. They reported that observers could detect the presence of a target picture specified by name (e.g., smiling couple) even when the pictures in the sequence were presented for just 13 ms each. Potter et al. claimed that this was insufficient time for feedback processing to occur, so feedforward processing alone must be able to generate conscious awareness of the target pictures. A potential confound in their study is that the pictures in the RSVP sequence sometime contained areas with no high-contrast edges, and so may not have adequately masked each other. Consequently, iconic memories of portions of the target pictures may have persisted in the visual system, thereby increasing the effective presentation time. Our study addressed this issue by redoing the Potter et al. study, but using four different types of masks. We found that when adequate masking was used, no evidence emerged that observers could detect the presence of a specific target picture, even when each picture in the RSVP sequence was presented for 27 ms. On the basis of these findings, we cannot rule out the possibility that feedback processing is necessary for individual pictures to be recognized.
Collapse
|
128
|
Pinto Y, de Haan EH, Lamme VA. The Split-Brain Phenomenon Revisited: A Single Conscious Agent with Split Perception. Trends Cogn Sci 2017; 21:835-851. [DOI: 10.1016/j.tics.2017.09.003] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2017] [Revised: 08/24/2017] [Accepted: 09/05/2017] [Indexed: 11/16/2022]
|
129
|
Gong M, Xuan Y, Xu X, Fu X. The Effect of Consistency on Short-Term Memory for Scenes. Front Psychol 2017; 8:1712. [PMID: 29046654 PMCID: PMC5632670 DOI: 10.3389/fpsyg.2017.01712] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2017] [Accepted: 09/19/2017] [Indexed: 11/13/2022] Open
Abstract
Which is more detectable, the change of a consistent or an inconsistent object in a scene? This question has been debated for decades. We noted that the change of objects in scenes might simultaneously be accompanied with gist changes. In the present study we aimed to examine how the alteration of gist, as well as the consistency of the changed objects, modulated change detection. In Experiment 1, we manipulated the semantic content by either keeping or changing the consistency of the scene. Results showed that the changes of consistent and inconsistent scenes were equally detected. More importantly, the changes were more accurately detected when scene consistency changed than when the consistency remained unchanged, regardless of the consistency of the memory scenes. A phase-scrambled version of stimuli was adopted in Experiment 2 to decouple the possible confounding effect of low-level factors. The results of Experiment 2 demonstrated that the effect found in Experiment 1 was indeed due to the change of high-level semantic consistency rather than the change of low-level physical features. Together, the study suggests that the change of consistency plays an important role in scene short-term memory, which might be attributed to the sensitivity to the change of semantic content.
Collapse
Affiliation(s)
- Mingliang Gong
- State Key Laboratory of Brain and Cognitive Sciences, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.,Department of Psychology, Miami University, Oxford, OH, United States
| | - Yuming Xuan
- State Key Laboratory of Brain and Cognitive Sciences, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.,Department of Psychology, University of the Chinese Academy of Sciences, Beijing, China
| | - Xinwen Xu
- State Key Laboratory of Brain and Cognitive Sciences, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.,Department of Psychology, University of the Chinese Academy of Sciences, Beijing, China
| | - Xiaolan Fu
- State Key Laboratory of Brain and Cognitive Sciences, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.,Department of Psychology, University of the Chinese Academy of Sciences, Beijing, China
| |
Collapse
|
130
|
Affiliation(s)
- Miguel P. Eckstein
- Department of Psychological and Brain Sciences, University of California, Santa Barbara, California 93106-9660
| |
Collapse
|
131
|
Eckstein MP, Koehler K, Welbourne LE, Akbas E. Humans, but Not Deep Neural Networks, Often Miss Giant Targets in Scenes. Curr Biol 2017; 27:2827-2832.e3. [PMID: 28889976 DOI: 10.1016/j.cub.2017.07.068] [Citation(s) in RCA: 35] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2017] [Revised: 07/21/2017] [Accepted: 07/28/2017] [Indexed: 11/26/2022]
Abstract
Even with great advances in machine vision, animals are still unmatched in their ability to visually search complex scenes. Animals from bees [1, 2] to birds [3] to humans [4-12] learn about the statistical relations in visual environments to guide and aid their search for targets. Here, we investigate a novel manner in which humans utilize rapidly acquired information about scenes by guiding search toward likely target sizes. We show that humans often miss targets when their size is inconsistent with the rest of the scene, even when the targets were made larger and more salient and observers fixated the target. In contrast, we show that state-of-the-art deep neural networks do not exhibit such deficits in finding mis-scaled targets but, unlike humans, can be fooled by target-shaped distractors that are inconsistent with the expected target's size within the scene. Thus, it is not a human deficiency to miss targets when they are inconsistent in size with the scene; instead, it is a byproduct of a useful strategy that the brain has implemented to rapidly discount potential distractors.
Collapse
Affiliation(s)
- Miguel P Eckstein
- Department of Psychological and Brain Sciences, University of California, Santa Barbara, Santa Barbara, CA, 93103, USA; Institute for Collaborative Biotechnologies, University of California, Santa Barbara, Santa Barbara, CA, 93103, USA.
| | - Kathryn Koehler
- Department of Psychological and Brain Sciences, University of California, Santa Barbara, Santa Barbara, CA, 93103, USA
| | - Lauren E Welbourne
- Department of Psychological and Brain Sciences, University of California, Santa Barbara, Santa Barbara, CA, 93103, USA; Institute for Collaborative Biotechnologies, University of California, Santa Barbara, Santa Barbara, CA, 93103, USA
| | - Emre Akbas
- Department of Psychological and Brain Sciences, University of California, Santa Barbara, Santa Barbara, CA, 93103, USA; Department of Computer Engineering, Middle East Technical University, 06800 Ankara, Turkey
| |
Collapse
|
132
|
Bahle B, Matsukura M, Hollingworth A. Contrasting gist-based and template-based guidance during real-world visual search. J Exp Psychol Hum Percept Perform 2017; 44:367-386. [PMID: 28795834 DOI: 10.1037/xhp0000468] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Visual search through real-world scenes is guided both by a representation of target features and by knowledge of the sematic properties of the scene (derived from scene gist recognition). In 3 experiments, we compared the relative roles of these 2 sources of guidance. Participants searched for a target object in the presence of a critical distractor object. The color of the critical distractor either matched or mismatched (a) the color of an item maintained in visual working memory for a secondary task (Experiment 1), or (b) the color of the target, cued by a picture before search commenced (Experiments 2 and 3). Capture of gaze by a matching distractor served as an index of template guidance. There were 4 main findings: (a) The distractor match effect was observed from the first saccade on the scene, (b) it was independent of the availability of scene-level gist-based guidance, (c) it was independent of whether the distractor appeared in a plausible location for the target, and (d) it was preserved even when gist-based guidance was available before scene onset. Moreover, gist-based, semantic guidance of gaze to target-plausible regions of the scene was delayed relative to template-based guidance. These results suggest that feature-based template guidance is not limited to plausible scene regions after an initial, scene-level analysis. (PsycINFO Database Record
Collapse
Affiliation(s)
- Brett Bahle
- Department of Psychological and Brain Sciences, The University of Iowa
| | - Michi Matsukura
- Department of Psychological and Brain Sciences, The University of Iowa
| | | |
Collapse
|
133
|
Are allocentric spatial reference frames compatible with theories of Enactivism? PSYCHOLOGICAL RESEARCH 2017; 83:498-513. [PMID: 28770385 DOI: 10.1007/s00426-017-0899-x] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2016] [Accepted: 07/25/2017] [Indexed: 10/19/2022]
Abstract
Theories of Enactivism propose an action-oriented approach to understand human cognition. So far, however, empirical evidence supporting these theories has been sparse. Here, we investigate whether spatial navigation based on allocentric reference frames that are independent of the observer's physical body can be understood within an action-oriented approach. Therefore, we performed three experiments testing the knowledge of the absolute orientation of houses and streets towards north, the relative orientation of two houses and two streets, respectively, and the location of houses towards each other in a pointing task. Our results demonstrate that under time pressure, the relative orientation of two houses can be retrieved more accurately than the absolute orientation of single houses. With infinite time for cognitive reasoning, the performance of the task using house stimuli increased greatly for the absolute orientation and surpassed the slightly improved performance in the relative orientation task. In contrast, with streets as stimuli participants performed under time pressure better in the absolute orientation task. Overall, pointing from one house to another house yielded the best performance. This suggests, first, that orientation and location information about houses are primarily coded in house-to-house relations, whereas cardinal information is deduced via cognitive reasoning. Second, orientation information for streets is preferentially coded in absolute orientations. Thus, our results suggest that spatial information about house and street orientation is coded differently and that house orientation and location is primarily learned in an action-oriented way, which is in line with an enactive framework for human cognition.
Collapse
|
134
|
|
135
|
Caddigan E, Choo H, Fei-Fei L, Beck DM. Categorization influences detection: A perceptual advantage for representative exemplars of natural scene categories. J Vis 2017; 17:21. [PMID: 28114496 PMCID: PMC5852945 DOI: 10.1167/17.1.21] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Traditional models of recognition and categorization proceed from registering low-level features, perceptually organizing that input, and linking it with stored representations. Recent evidence, however, suggests that this serial model may not be accurate, with object and category knowledge affecting rather than following early visual processing. Here, we show that the degree to which an image exemplifies its category influences how easily it is detected. Participants performed a two-alternative forced-choice task in which they indicated whether a briefly presented image was an intact or phase-scrambled scene photograph. Critically, the category of the scene is irrelevant to the detection task. We nonetheless found that participants “see” good images better, more accurately discriminating them from phase-scrambled images than bad scenes, and this advantage is apparent regardless of whether participants are asked to consider category during the experiment or not. We then demonstrate that good exemplars are more similar to same-category images than bad exemplars, influencing behavior in two ways: First, prototypical images are easier to detect, and second, intact good scenes are more likely than bad to have been primed by a previous trial.
Collapse
Affiliation(s)
- Eamon Caddigan
- Department of Psychology, University of Illinois, Champaign, IL, USA
| | - Heeyoung Choo
- Beckman Institute, University of Illinois, Urbana, IL, USA
| | | | - Diane M Beck
- Department of Psychology and Beckman Institute, University of Illinois, Champaign, IL, USA
| |
Collapse
|
136
|
Watson DM, Andrews TJ, Hartley T. A data driven approach to understanding the organization of high-level visual cortex. Sci Rep 2017; 7:3596. [PMID: 28620238 PMCID: PMC5472563 DOI: 10.1038/s41598-017-03974-5] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/22/2017] [Accepted: 05/08/2017] [Indexed: 11/16/2022] Open
Abstract
The neural representation in scene-selective regions of human visual cortex, such as the PPA, has been linked to the semantic and categorical properties of the images. However, the extent to which patterns of neural response in these regions reflect more fundamental organizing principles is not yet clear. Existing studies generally employ stimulus conditions chosen by the experimenter, potentially obscuring the contribution of more basic stimulus dimensions. To address this issue, we used a data-driven approach to describe a large database of scenes (>100,000 images) in terms of their visual properties (orientation, spatial frequency, spatial location). K-means clustering was then used to select images from distinct regions of this feature space. Images in each cluster did not correspond to typical scene categories. Nevertheless, they elicited distinct patterns of neural response in the PPA. Moreover, the similarity of the neural response to different clusters in the PPA could be predicted by the similarity in their image properties. Interestingly, the neural response in the PPA was also predicted by perceptual responses to the scenes, but not by their semantic properties. These findings provide an image-based explanation for the emergence of higher-level representations in scene-selective regions of the human brain.
Collapse
Affiliation(s)
- David M Watson
- Department of Psychology and York Neuroimaging Centre, University of York, York, YO10 5DD, United Kingdom. .,School of Psychology, The University of Nottingham, Nottingham, NG7 2RD, United Kingdom.
| | - Timothy J Andrews
- Department of Psychology and York Neuroimaging Centre, University of York, York, YO10 5DD, United Kingdom
| | - Tom Hartley
- Department of Psychology and York Neuroimaging Centre, University of York, York, YO10 5DD, United Kingdom
| |
Collapse
|
137
|
Abstract
A central component of spatial navigation is determining where one can and cannot go in the immediate environment. We used fMRI to test the hypothesis that the human visual system solves this problem by automatically identifying the navigational affordances of the local scene. Multivoxel pattern analyses showed that a scene-selective region of dorsal occipitoparietal cortex, known as the occipital place area, represents pathways for movement in scenes in a manner that is tolerant to variability in other visual features. These effects were found in two experiments: One using tightly controlled artificial environments as stimuli, the other using a diverse set of complex, natural scenes. A reconstruction analysis demonstrated that the population codes of the occipital place area could be used to predict the affordances of novel scenes. Taken together, these results reveal a previously unknown mechanism for perceiving the affordance structure of navigable space.
Collapse
|
138
|
Groen IIA, Silson EH, Baker CI. Contributions of low- and high-level properties to neural processing of visual scenes in the human brain. Philos Trans R Soc Lond B Biol Sci 2017; 372:rstb.2016.0102. [PMID: 28044013 DOI: 10.1098/rstb.2016.0102] [Citation(s) in RCA: 96] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/20/2016] [Indexed: 11/12/2022] Open
Abstract
Visual scene analysis in humans has been characterized by the presence of regions in extrastriate cortex that are selectively responsive to scenes compared with objects or faces. While these regions have often been interpreted as representing high-level properties of scenes (e.g. category), they also exhibit substantial sensitivity to low-level (e.g. spatial frequency) and mid-level (e.g. spatial layout) properties, and it is unclear how these disparate findings can be united in a single framework. In this opinion piece, we suggest that this problem can be resolved by questioning the utility of the classical low- to high-level framework of visual perception for scene processing, and discuss why low- and mid-level properties may be particularly diagnostic for the behavioural goals specific to scene perception as compared to object recognition. In particular, we highlight the contributions of low-level vision to scene representation by reviewing (i) retinotopic biases and receptive field properties of scene-selective regions and (ii) the temporal dynamics of scene perception that demonstrate overlap of low- and mid-level feature representations with those of scene category. We discuss the relevance of these findings for scene perception and suggest a more expansive framework for visual scene analysis.This article is part of the themed issue 'Auditory and visual scene analysis'.
Collapse
Affiliation(s)
- Iris I A Groen
- Laboratory of Brain and Cognition, National Institutes of Health, 10 Center Drive 10-3N228, Bethesda, MD, USA
| | - Edward H Silson
- Laboratory of Brain and Cognition, National Institutes of Health, 10 Center Drive 10-3N228, Bethesda, MD, USA
| | - Chris I Baker
- Laboratory of Brain and Cognition, National Institutes of Health, 10 Center Drive 10-3N228, Bethesda, MD, USA
| |
Collapse
|
139
|
Abstract
Attention readily facilitates the detection and discrimination of objects, but it is not known whether it helps to form the vast volume of visual space that contains the objects and where actions are implemented. Conventional wisdom suggests not, given the effortless ease with which we perceive three-dimensional (3D) scenes on opening our eyes. Here, we show evidence to the contrary. In Experiment 1, the observer judged the location of a briefly presented target, placed either on the textured ground or ceiling surface. Judged location was more accurate for a target on the ground, provided that the ground was visible and that the observer directed attention to the lower visual field, not the upper field. This reveals that attention facilitates space perception with reference to the ground. Experiment 2 showed that judged location of a target in mid-air, with both ground and ceiling surfaces present, was more accurate when the observer directed their attention to the lower visual field; this indicates that the attention effect extends to visual space above the ground. These findings underscore the role of attention in anchoring visual orientation in space, which is arguably a primal event that enhances one's ability to interact with objects and surface layouts within the visual space. The fact that the effect of attention was contingent on the ground being visible suggests that our terrestrial visual system is best served by its ecological niche.
Collapse
Affiliation(s)
- Liu Zhou
- Key Laboratory of Brain Functional Genomics (MOE & STCSM), Institute of Cognitive Neuroscience, School of Psychology and Cognitive Science, East China Normal University, Shanghai 200062, China
| | - Chenglong Deng
- Key Laboratory of Brain Functional Genomics (MOE & STCSM), Institute of Cognitive Neuroscience, School of Psychology and Cognitive Science, East China Normal University, Shanghai 200062, China
| | - Teng Leng Ooi
- College of Optometry, The Ohio State University, Columbus, Ohio 43210, USA
| | - Zijiang J He
- Key Laboratory of Brain Functional Genomics (MOE & STCSM), Institute of Cognitive Neuroscience, School of Psychology and Cognitive Science, East China Normal University, Shanghai 200062, China.,Department of Psychological and Brain Sciences, University of Louisville, Louisville, Kentucky 40292, USA.,CAS Key Laboratory of Behavioral Science, Institute of Psychology, Chinese Academy of Sciences, Beijing 100101, China
| |
Collapse
|
140
|
Ghodrati M, Ghodousi M, Yoonessi A. Low-Level Contrast Statistics of Natural Images Can Modulate the Frequency of Event-Related Potentials (ERP) in Humans. Front Hum Neurosci 2016; 10:630. [PMID: 28018197 PMCID: PMC5145888 DOI: 10.3389/fnhum.2016.00630] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2016] [Accepted: 11/25/2016] [Indexed: 11/20/2022] Open
Abstract
Humans are fast and accurate in categorizing complex natural images. It is, however, unclear what features of visual information are exploited by brain to perceive the images with such speed and accuracy. It has been shown that low-level contrast statistics of natural scenes can explain the variance of amplitude of event-related potentials (ERP) in response to rapidly presented images. In this study, we investigated the effect of these statistics on frequency content of ERPs. We recorded ERPs from human subjects, while they viewed natural images each presented for 70 ms. Our results showed that Weibull contrast statistics, as a biologically plausible model, explained the variance of ERPs the best, compared to other image statistics that we assessed. Our time-frequency analysis revealed a significant correlation between these statistics and ERPs' power within theta frequency band (~3–7 Hz). This is interesting, as theta band is believed to be involved in context updating and semantic encoding. This correlation became significant at ~110 ms after stimulus onset, and peaked at 138 ms. Our results show that not only the amplitude but also the frequency of neural responses can be modulated with low-level contrast statistics of natural images and highlights their potential role in scene perception.
Collapse
Affiliation(s)
- Masoud Ghodrati
- Department of Physiology, Monash UniversityClayton, VIC, Australia; Neuroscience Program, Biomedicine Discovery Institute, Monash UniversityClayton, VIC, Australia
| | - Mahrad Ghodousi
- Department of Neuroscience, School of Advanced Technologies in Medicine, Tehran University of Medical Sciences Tehran, Iran
| | - Ali Yoonessi
- Department of Neuroscience, School of Advanced Technologies in Medicine, Tehran University of Medical Sciences Tehran, Iran
| |
Collapse
|
141
|
Effects of varying presentation time on long-term recognition memory for scenes: Verbatim and gist representations. Mem Cognit 2016; 45:390-403. [DOI: 10.3758/s13421-016-0672-1] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
142
|
Making Sense of Real-World Scenes. Trends Cogn Sci 2016; 20:843-856. [PMID: 27769727 DOI: 10.1016/j.tics.2016.09.003] [Citation(s) in RCA: 87] [Impact Index Per Article: 9.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2016] [Revised: 09/06/2016] [Accepted: 09/06/2016] [Indexed: 11/23/2022]
Abstract
To interact with the world, we have to make sense of the continuous sensory input conveying information about our environment. A recent surge of studies has investigated the processes enabling scene understanding, using increasingly complex stimuli and sophisticated analyses to highlight the visual features and brain regions involved. However, there are two major challenges to producing a comprehensive framework for scene understanding. First, scene perception is highly dynamic, subserving multiple behavioral goals. Second, a multitude of different visual properties co-occur across scenes and may be correlated or independent. We synthesize the recent literature and argue that for a complete view of scene understanding, it is necessary to account for both differing observer goals and the contribution of diverse scene properties.
Collapse
|
143
|
Affiliation(s)
- Ruth Rosenholtz
- Department of Brain and Cognitive Sciences, CSAIL, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139;
| |
Collapse
|
144
|
The Temporal Dynamics of Scene Processing: A Multifaceted EEG Investigation. eNeuro 2016; 3:eN-NWR-0139-16. [PMID: 27699208 PMCID: PMC5037322 DOI: 10.1523/eneuro.0139-16.2016] [Citation(s) in RCA: 39] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2016] [Revised: 08/12/2016] [Accepted: 09/06/2016] [Indexed: 11/25/2022] Open
Abstract
Our remarkable ability to process complex visual scenes is supported by a network of scene-selective cortical regions. Despite growing knowledge about the scene representation in these regions, much less is known about the temporal dynamics with which these representations emerge. We conducted two experiments aimed at identifying and characterizing the earliest markers of scene-specific processing. In the first experiment, human participants viewed images of scenes, faces, and everyday objects while event-related potentials (ERPs) were recorded. We found that the first ERP component to evince a significantly stronger response to scenes than the other categories was the P2, peaking ∼220 ms after stimulus onset. To establish that the P2 component reflects scene-specific processing, in the second experiment, we recorded ERPs while the participants viewed diverse real-world scenes spanning the following three global scene properties: spatial expanse (open/closed), relative distance (near/far), and naturalness (man-made/natural). We found that P2 amplitude was sensitive to these scene properties at both the categorical level, distinguishing between open and closed natural scenes, as well as at the single-image level, reflecting both computationally derived scene statistics and behavioral ratings of naturalness and spatial expanse. Together, these results establish the P2 as an ERP marker for scene processing, and demonstrate that scene-specific global information is available in the neural response as early as 220 ms.
Collapse
|
145
|
Juliani AW, Bies AJ, Boydston CR, Taylor RP, Sereno ME. Navigation performance in virtual environments varies with fractal dimension of landscape. JOURNAL OF ENVIRONMENTAL PSYCHOLOGY 2016; 47:155-165. [PMID: 27346905 PMCID: PMC4918639 DOI: 10.1016/j.jenvp.2016.05.011] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Fractal geometry has been used to describe natural and built environments, but has yet to be studied in navigational research. In order to establish a relationship between the fractal dimension (D) of a natural environment and humans' ability to navigate such spaces, we conducted two experiments using virtual environments that simulate the fractal properties of nature. In Experiment 1, participants completed a goal-driven search task either with or without a map in landscapes that varied in D. In Experiment 2, participants completed a map-reading and location-judgment task in separate sets of fractal landscapes. In both experiments, task performance was highest at the low-to-mid range of D, which was previously reported as most preferred and discriminable in studies of fractal aesthetics and discrimination, respectively, supporting a theory of visual fluency. The applicability of these findings to architecture, urban planning, and the general design of constructed spaces is discussed.
Collapse
Affiliation(s)
- Arthur W. Juliani
- Department of Psychology, University of Oregon, Eugene OR 97403, USA
| | - Alexander J. Bies
- Department of Psychology, University of Oregon, Eugene OR 97403, USA
| | | | | | | |
Collapse
|
146
|
Kreindel E, Intraub H. Anticipatory scene representation in preschool children's recall and recognition memory. Dev Sci 2016; 20. [PMID: 27582346 DOI: 10.1111/desc.12444] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/27/2015] [Accepted: 03/23/2016] [Indexed: 11/26/2022]
Abstract
Behavioral and neuroscience research on boundary extension (false memory beyond the edges of a view of a scene) has provided new insights into the constructive nature of scene representation, and motivates questions about development. Early research with children (as young as 6-7 years) was consistent with boundary extension, but relied on an analysis of spatial errors in drawings which are open to alternative explanations (e.g. drawing ability). Experiment 1 replicated and extended prior drawing results with 4-5-year-olds and adults. In Experiment 2, a new, forced-choice immediate recognition memory test was implemented with the same children. On each trial, a card (photograph of a simple scene) was immediately replaced by a test card (identical view and either a closer or more wide-angle view) and participants indicated which one matched the original view. Error patterns supported boundary extension; identical photographs were more frequently rejected when the closer view was the original view, than vice versa. This asymmetry was not attributable to a selection bias (guessing tasks; Experiments 3-5). In Experiment 4, working memory load was increased by presenting more expansive views of more complex scenes. Again, children exhibited boundary extension, but now adults did not, unless stimulus duration was reduced to 5 s (limiting time to implement strategies; Experiment 5). We propose that like adults, children interpret photographs as views of places in the world; they extrapolate the anticipated continuation of the scene beyond the view and misattribute it to having been seen. Developmental differences in source attribution decision processes provide an explanation for the age-related differences observed.
Collapse
Affiliation(s)
- Erica Kreindel
- Department of Psychological and Brain Sciences, University of Delaware, USA
| | - Helene Intraub
- Department of Psychological and Brain Sciences, University of Delaware, USA
| |
Collapse
|
147
|
Two scenes or not two scenes: The effects of stimulus repetition and view-similarity on scene categorization from brief displays. Mem Cognit 2016; 45:49-62. [DOI: 10.3758/s13421-016-0640-9] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
148
|
Griffiths TL, Abbott JT, Hsu AS. Exploring Human Cognition Using Large Image Databases. Top Cogn Sci 2016; 8:569-88. [DOI: 10.1111/tops.12209] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2013] [Revised: 01/18/2015] [Accepted: 01/18/2015] [Indexed: 11/30/2022]
Affiliation(s)
| | | | - Anne S. Hsu
- School of Electronic Engineering and Computer Science Queen Mary, University of London
| |
Collapse
|
149
|
Greene MR, Baldassano C, Esteva A, Beck DM, Fei-Fei L. Visual scenes are categorized by function. J Exp Psychol Gen 2016; 145:82-94. [PMID: 26709590 DOI: 10.1037/xge0000129] [Citation(s) in RCA: 43] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
How do we know that a kitchen is a kitchen by looking? Traditional models posit that scene categorization is achieved through recognizing necessary and sufficient features and objects, yet there is little consensus about what these may be. However, scene categories should reflect how we use visual information. Therefore, we test the hypothesis that scene categories reflect functions, or the possibilities for actions within a scene. Our approach is to compare human categorization patterns with predictions made by both functions and alternative models. We collected a large-scale scene category distance matrix (5 million trials) by asking observers to simply decide whether 2 images were from the same or different categories. Using the actions from the American Time Use Survey, we mapped actions onto each scene (1.4 million trials). We found a strong relationship between ranked category distance and functional distance (r = .50, or 66% of the maximum possible correlation). The function model outperformed alternative models of object-based distance (r = .33), visual features from a convolutional neural network (r = .39), lexical distance (r = .27), and models of visual features. Using hierarchical linear regression, we found that functions captured 85.5% of overall explained variance, with nearly half of the explained variance captured only by functions, implying that the predictive power of alternative models was because of their shared variance with the function-based model. These results challenge the dominant school of thought that visual features and objects are sufficient for scene categorization, suggesting instead that a scene's category may be determined by the scene's function.
Collapse
Affiliation(s)
| | | | - Andre Esteva
- Department of Electrical Engineering, Stanford University
| | - Diane M Beck
- Department of Psychology, University of Illinois at Urbana-Champaign
| | - Li Fei-Fei
- Department of Computer Science, Stanford University
| |
Collapse
|
150
|
Vanmarcke S, Wagemans J. Individual differences in spatial frequency processing in scene perception: the influence of autism-related traits. VISUAL COGNITION 2016. [DOI: 10.1080/13506285.2016.1199625] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|