1
|
Kondyli V, Bhatt M, Levin D, Suchan J. How do drivers mitigate the effects of naturalistic visual complexity? : On attentional strategies and their implications under a change blindness protocol. Cogn Res Princ Implic 2023; 8:54. [PMID: 37556047 PMCID: PMC10412523 DOI: 10.1186/s41235-023-00501-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2022] [Accepted: 07/07/2023] [Indexed: 08/10/2023] Open
Abstract
How do the limits of high-level visual processing affect human performance in naturalistic, dynamic settings of (multimodal) interaction where observers can draw on experience to strategically adapt attention to familiar forms of complexity? In this backdrop, we investigate change detection in a driving context to study attentional allocation aimed at overcoming environmental complexity and temporal load. Results indicate that visuospatial complexity substantially increases change blindness but also that participants effectively respond to this load by increasing their focus on safety-relevant events, by adjusting their driving, and by avoiding non-productive forms of attentional elaboration, thereby also controlling "looked-but-failed-to-see" errors. Furthermore, analyses of gaze patterns reveal that drivers occasionally, but effectively, limit attentional monitoring and lingering for irrelevant changes. Overall, the experimental outcomes reveal how drivers exhibit effective attentional compensation in highly complex situations. Our findings uncover implications for driving education and development of driving skill-testing methods, as well as for human-factors guided development of AI-based driving assistance systems.
Collapse
Affiliation(s)
- Vasiliki Kondyli
- CoDesign Lab EU - codesign-lab.org, Örebro University, Örebro, Sweden.
| | - Mehul Bhatt
- CoDesign Lab EU - codesign-lab.org, Örebro University, Örebro, Sweden
| | | | - Jakob Suchan
- German Aerospace Center - DLR, Institute of Systems Engineering for Future Mobility, Oldenburg, Germany
| |
Collapse
|
2
|
Loh Z, Hall EH, Cronin D, Henderson JM. Working memory control predicts fixation duration in scene-viewing. PSYCHOLOGICAL RESEARCH 2023; 87:1143-1154. [PMID: 35879564 PMCID: PMC11129724 DOI: 10.1007/s00426-022-01694-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2021] [Accepted: 06/02/2022] [Indexed: 11/28/2022]
Abstract
When viewing scenes, observers differ in how long they linger at each fixation location and how far they move their eyes between fixations. What factors drive these differences in eye-movement behaviors? Previous work suggests individual differences in working memory capacity may influence fixation durations and saccade amplitudes. In the present study, participants (N = 98) performed two scene-viewing tasks, aesthetic judgment and memorization, while viewing 100 photographs of real-world scenes. Working memory capacity, working memory processing ability, and fluid intelligence were assessed with an operation span task, a memory updating task, and Raven's Advanced Progressive Matrices, respectively. Across participants, we found significant effects of task on both fixation durations and saccade amplitudes. At the level of each individual participant, we also found a significant relationship between memory updating task performance and participants' fixation duration distributions. However, we found no effect of fluid intelligence and no effect of working memory capacity on fixation duration or saccade amplitude distributions, inconsistent with previous findings. These results suggest that the ability to flexibly maintain and update working memory is strongly related to fixation duration behavior.
Collapse
Affiliation(s)
- Zoe Loh
- Management of Complex Systems Department, University of California Merced, Merced, CA, 95343, USA
- Center for Mind and Brain, University of California Davis, Davis, CA, 95618, USA
| | - Elizabeth H Hall
- Center for Mind and Brain, University of California Davis, Davis, CA, 95618, USA.
- Department of Psychology, University of California Davis, Davis, CA, 95616, USA.
| | - Deborah Cronin
- Center for Mind and Brain, University of California Davis, Davis, CA, 95618, USA
- Department of Psychology, Drake University, Des Moines, IA, 50311, USA
| | - John M Henderson
- Center for Mind and Brain, University of California Davis, Davis, CA, 95618, USA
- Department of Psychology, University of California Davis, Davis, CA, 95616, USA
| |
Collapse
|
3
|
Han NX, Eckstein MP. Head and body cues guide eye movements and facilitate target search in real-world videos. J Vis 2023; 23:5. [PMID: 37294703 DOI: 10.1167/jov.23.6.5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/11/2023] Open
Abstract
Static gaze cues presented in central vision result in observer shifts of covert attention and eye movements, and benefits in perceptual performance in the detection of simple targets. Less is known about how dynamic gazer behaviors with head and body motion influence search eye movements and performance in perceptual tasks in real-world scenes. Participants searched for a target person (yes/no task, 50% presence), whereas watching videos of one to three gazers looking at a designated person (50% valid gaze cue, looking at the target). To assess the contributions of different body parts, we digitally erase parts of the gazers in the videos to create three different body parts/whole conditions for gazers: floating heads (only head movements), headless bodies (only lower body movements), and the baseline condition with intact head and body. We show that valid dynamic gaze cues guided participants' eye movements (up to 3 fixations) closer to the target, speeded the time to foveate the target, reduced fixations to the gazers, and improved target detection. The effect of gaze cues in guiding eye movements to the search target was the smallest when the gazer's head was removed from the videos. To assess the inherent information about gaze goal location for each body parts/whole condition, we collected perceptual judgments estimating gaze goals by a separate group of observers with unlimited time. Observers' perceptual judgments showed larger estimate errors when the gazer's head was removed. This suggests that the reduced eye movement guidance from lower body cueing is related to observers' difficulty extracting gaze information without the presence of the head. Together, the study extends previous work by evaluating the impact of dynamic gazer behaviors on search with videos of real-world cluttered scenes.
Collapse
Affiliation(s)
- Nicole X Han
- Department of Psychological and Brain Sciences, Institute for Collaborative Biotechnologies, University of California, Santa Barbara, CA, USA
| | - Miguel P Eckstein
- Department of Psychological and Brain Sciences, Institute for Collaborative Biotechnologies, University of California, Santa Barbara, CA, USA
| |
Collapse
|
4
|
Active visual search in naturalistic environments reflects individual differences in classic visual search performance. Sci Rep 2023; 13:631. [PMID: 36635491 PMCID: PMC9837148 DOI: 10.1038/s41598-023-27896-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2022] [Accepted: 01/10/2023] [Indexed: 01/13/2023] Open
Abstract
Visual search is a ubiquitous activity in real-world environments. Yet, traditionally, visual search is investigated in tightly controlled paradigms, where head-restricted participants locate a minimalistic target in a cluttered array that is presented on a computer screen. Do traditional visual search tasks predict performance in naturalistic settings, where participants actively explore complex, real-world scenes? Here, we leverage advances in virtual reality technology to test the degree to which classic and naturalistic search are limited by a common factor, set size, and the degree to which individual differences in classic search behavior predict naturalistic search behavior in a large sample of individuals (N = 75). In a naturalistic search task, participants looked for an object within their environment via a combination of head-turns and eye-movements using a head-mounted display. Then, in a classic search task, participants searched for a target within a simple array of colored letters using only eye-movements. In each task, we found that participants' search performance was impacted by increases in set size-the number of items in the visual display. Critically, we observed that participants' efficiency in classic search tasks-the degree to which set size slowed performance-indeed predicted efficiency in real-world scenes. These results demonstrate that classic, computer-based visual search tasks are excellent models of active, real-world search behavior.
Collapse
|
5
|
Nicholson DA, Prinz AA. Could simplified stimuli change how the brain performs visual search tasks? A deep neural network study. J Vis 2022; 22:3. [PMID: 35675057 PMCID: PMC9187944 DOI: 10.1167/jov.22.7.3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Visual search is a complex behavior influenced by many factors. To control for these factors, many studies use highly simplified stimuli. However, the statistics of these stimuli are very different from the statistics of the natural images that the human visual system is optimized by evolution and experience to perceive. Could this difference change search behavior? If so, simplified stimuli may contribute to effects typically attributed to cognitive processes, such as selective attention. Here we use deep neural networks to test how optimizing models for the statistics of one distribution of images constrains performance on a task using images from a different distribution. We train four deep neural network architectures on one of three source datasets—natural images, faces, and x-ray images—and then adapt them to a visual search task using simplified stimuli. This adaptation produces models that exhibit performance limitations similar to humans, whereas models trained on the search task alone exhibit no such limitations. However, we also find that deep neural networks trained to classify natural images exhibit similar limitations when adapted to a search task that uses a different set of natural images. Therefore, the distribution of data alone cannot explain this effect. We discuss how future work might integrate an optimization-based approach into existing models of visual search behavior.
Collapse
Affiliation(s)
- David A Nicholson
- Emory University, Department of Biology, O. Wayne Rollins Research Center, Atlanta, Georgia.,
| | - Astrid A Prinz
- Emory University, Department of Biology, O. Wayne Rollins Research Center, Atlanta, Georgia.,
| |
Collapse
|
6
|
Miuccio MT, Zelinsky GJ, Schmidt J. Are all real-world objects created equal? Estimating the "set-size" of the search target in visual working memory. Psychophysiology 2022; 59:e13998. [PMID: 35001411 PMCID: PMC8957527 DOI: 10.1111/psyp.13998] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2021] [Revised: 11/23/2021] [Accepted: 12/16/2021] [Indexed: 11/30/2022]
Abstract
Are all real-world objects created equal? Visual search difficulty increases with the number of targets and as target-related visual working memory (VWM) load increases. Our goal was to investigate the load imposed by individual real-world objects held in VWM in the context of search. Measures of visual clutter attempt to quantify real-world set-size in the context of scenes. We applied one of these measures, the number of proto-objects, to individual real-world objects and used contralateral delay activity (CDA) to measure the resulting VWM load. The current study presented a real-world object as a target cue, followed by a delay where CDA was measured. This was followed by a four-object search array. We compared CDA and later search performance from target cues containing a high or low number of proto-objects. High proto-object target cues resulted in greater CDA, longer search RTs, target dwell times, and reduced search guidance, relative to low proto-object targets. These findings demonstrate that targets with more proto-objects result in a higher VWM load and reduced search performance. This shows that the number of proto-objects contained within individual objects produce set-size like effects in VWM and suggests proto-objects may be a viable unit of measure of real-world VWM load. Importantly, this demonstrates that not all real-world objects are created equal.
Collapse
Affiliation(s)
- Michael T Miuccio
- Department of Psychology, University of Central Florida, Orlando, FLorida, USA
| | - Gregory J Zelinsky
- Department of Psychology, Stony Brook University, Stony Brook, New York, USA.,Department of Computer Science, Stony Brook University, Stony Brook, New York, USA
| | - Joseph Schmidt
- Department of Psychology, University of Central Florida, Orlando, FLorida, USA
| |
Collapse
|
7
|
Nuthmann A, Canas-Bajo T. Visual search in naturalistic scenes from foveal to peripheral vision: A comparison between dynamic and static displays. J Vis 2022; 22:10. [PMID: 35044436 PMCID: PMC8802022 DOI: 10.1167/jov.22.1.10] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2021] [Accepted: 12/03/2021] [Indexed: 11/24/2022] Open
Abstract
How important foveal, parafoveal, and peripheral vision are depends on the task. For object search and letter search in static images of real-world scenes, peripheral vision is crucial for efficient search guidance, whereas foveal vision is relatively unimportant. Extending this research, we used gaze-contingent Blindspots and Spotlights to investigate visual search in complex dynamic and static naturalistic scenes. In Experiment 1, we used dynamic scenes only, whereas in Experiments 2 and 3, we directly compared dynamic and static scenes. Each scene contained a static, contextually irrelevant target (i.e., a gray annulus). Scene motion was not predictive of target location. For dynamic scenes, the search-time results from all three experiments converge on the novel finding that neither foveal nor central vision was necessary to attain normal search proficiency. Since motion is known to attract attention and gaze, we explored whether guidance to the target was equally efficient in dynamic as compared to static scenes. We found that the very first saccade was guided by motion in the scene. This was not the case for subsequent saccades made during the scanning epoch, representing the actual search process. Thus, effects of task-irrelevant motion were fast-acting and short-lived. Furthermore, when motion was potentially present (Spotlights) or absent (Blindspots) in foveal or central vision only, we observed differences in verification times for dynamic and static scenes (Experiment 2). When using scenes with greater visual complexity and more motion (Experiment 3), however, the differences between dynamic and static scenes were much reduced.
Collapse
Affiliation(s)
- Antje Nuthmann
- Institute of Psychology, Kiel University, Kiel, Germany
- Psychology Department, School of Philosophy, Psychology and Language Sciences, University of Edinburgh, Edinburgh, UK
- http://orcid.org/0000-0003-3338-3434
| | - Teresa Canas-Bajo
- Vision Science Graduate Group, University of California, Berkeley, Berkeley, CA, USA
- Psychology Department, School of Philosophy, Psychology and Language Sciences, University of Edinburgh, Edinburgh, UK
| |
Collapse
|
8
|
Essex C, Gliga T, Singh M, Smith TJ. Understanding the differential impact of children's TV on executive functions: a narrative-processing analysis. Infant Behav Dev 2021; 66:101661. [PMID: 34784571 DOI: 10.1016/j.infbeh.2021.101661] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2021] [Revised: 10/25/2021] [Accepted: 10/28/2021] [Indexed: 11/28/2022]
Abstract
Evidence from multiple empirical studies suggests children's Executive Functions are depleted immediately after viewing some types of TV content but not others. Correlational evidence suggests any such effects may be most problematic during the pre-school years. To establish whether "screen-time" is developmentally appropriate at this age we believe a nuanced approach must be taken to the analysis of individual pieces of media and their potential demands on viewer cognition. To this end we apply a cognitive theory of visual narrative processing, the Scene Perception and Event Comprehension Theory (SPECT; Loschky, Larson, Smith, & Magliano, 2020) to the analysis of TV shows previously used to investigate short-term effects of TV viewing. A theoretical formalisation of individual content properties, together with a quantitative content-based analysis of previously used children's content (Lillard & Peterson, 2011; Lillard et al., 2015b) is presented. This analysis found a pattern of greater stimulus saliency, increased situational change and a greater combined presence of cognitively demanding features for videos previously shown to reduce children's EF after viewing. Limitations of this pilot application of SPECT are presented and proposals for future empirical investigations of the psychological mechanisms activated by specific TV viewing content are considered.
Collapse
Affiliation(s)
- Claire Essex
- Centre for Brain and Cognitive Development, Birkbeck, University of London, UK.
| | | | - Maninda Singh
- School of Psychology, University of Surrey, Guildford, Surrey, UK
| | - Tim J Smith
- Department of Psychological Sciences, Birkbeck, University of London, UK
| |
Collapse
|
9
|
Kanaan M, Moacdieh NM. How do we react to cluttered displays? Evidence from the first seconds of visual search in websites. ERGONOMICS 2021; 64:1452-1464. [PMID: 33957850 DOI: 10.1080/00140139.2021.1927200] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/27/2019] [Accepted: 05/03/2021] [Indexed: 06/12/2023]
Abstract
Display clutter is known to degrade search performance and lead to differences in eye movement measures in different contexts. The goal of this study was to determine whether these differences in eye movements could be detected in the first few seconds of a search task using a realistic display, both with or without time pressure. Participants were asked to search for image or word targets in 40 website screenshots. Time pressure was introduced for half the trials. Clutter algorithms were used to classify the websites as low- or high-clutter. Performance, subjective, and eye-tracking metrics were collected. Results showed that people's attention allocation within the first 3 s of search is different when viewing high-clutter websites. In particular, people's spread of attention was larger in high-clutter websites. The results can be used to detect whether a person is struggling with clutter early on after they view a display. Practitioner summary: Eye-tracking metrics showed that people react differently to a cluttered website in a variety of conditions. These differences were evident within the first 3 s of the search. The eye-tracking metrics identified can be used to detect people struggling with clutter as soon as they look at a website.
Collapse
Affiliation(s)
- Malk Kanaan
- Department of Industrial Engineering and Management, American University of Beirut, Beirut, Lebanon
| | - Nadine Marie Moacdieh
- Department of Industrial Engineering and Management, American University of Beirut, Beirut, Lebanon
| |
Collapse
|
10
|
Markant J, Amso D. Context and attention control determine whether attending to competing information helps or hinders learning in school-aged children. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2021; 13:e1577. [PMID: 34498382 DOI: 10.1002/wcs.1577] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/28/2021] [Revised: 08/05/2021] [Accepted: 08/18/2021] [Indexed: 11/07/2022]
Abstract
Attention control regulates efficient processing of goal-relevant information by suppressing interference from irrelevant competing inputs while also flexibly allocating attention across relevant inputs according to task demands. Research has established that developing attention control skills promote effective learning by minimizing distractions from task-irrelevant competing information. Additional research also suggests that competing contextual information can provide meaningful input for learning and should not always be ignored. Instead, attending to competing information that is relevant to task goals can facilitate and broaden the scope of children's learning. We review this past research examining effects of attending to task-relevant and task-irrelevant competing information on learning outcomes, focusing on relations between visual attention and learning in childhood. We then present a synthesis argument that complex interactions across learning goals, the contexts of learning environments and tasks, and developing attention control mechanisms will determine whether attending to competing information helps or hinders learning. This article is categorized under: Psychology > Attention Psychology > Learning Psychology > Development and Aging.
Collapse
Affiliation(s)
- Julie Markant
- Department of Psychology, Tulane University, New Orleans, Louisiana, USA
- Tulane Brain Institute, Tulane University, New Orleans, Louisiana, USA
| | - Dima Amso
- Department of Psychology, Columbia University, New York, New York, USA
| |
Collapse
|
11
|
Schweitzer R, Rolfs M. Intrasaccadic motion streaks jump-start gaze correction. SCIENCE ADVANCES 2021; 7:eabf2218. [PMID: 34301596 PMCID: PMC8302125 DOI: 10.1126/sciadv.abf2218] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/12/2020] [Accepted: 04/22/2021] [Indexed: 05/09/2023]
Abstract
Rapid eye movements (saccades) incessantly shift objects across the retina. To establish object correspondence, the visual system is thought to match surface features of objects across saccades. Here, we show that an object's intrasaccadic retinal trace-a signal previously considered unavailable to visual processing-facilitates this match making. Human observers made saccades to a cued target in a circular stimulus array. Using high-speed visual projection, we swiftly rotated this array during the eyes' flight, displaying continuous intrasaccadic target motion. Observers' saccades landed between the target and a distractor, prompting secondary saccades. Independently of the availability of object features, which we controlled tightly, target motion increased the rate and reduced the latency of gaze-correcting saccades to the initial presaccadic target, in particular when the target's stimulus features incidentally gave rise to efficient motion streaks. These results suggest that intrasaccadic visual information informs the establishment of object correspondence and jump-starts gaze correction.
Collapse
Affiliation(s)
- Richard Schweitzer
- Department of Psychology, Humboldt-Universität zu Berlin, Berlin, Germany.
- Exzellenzcluster Science of Intelligence, Technische Universität Berlin, Berlin, Germany
- Berlin School of Mind and Brain, Humboldt-Universität zu Berlin, Berlin, Germany
- Bernstein Center for Computational Neuroscience, Berlin, Germany
| | - Martin Rolfs
- Department of Psychology, Humboldt-Universität zu Berlin, Berlin, Germany
- Exzellenzcluster Science of Intelligence, Technische Universität Berlin, Berlin, Germany
- Berlin School of Mind and Brain, Humboldt-Universität zu Berlin, Berlin, Germany
- Bernstein Center for Computational Neuroscience, Berlin, Germany
| |
Collapse
|
12
|
Sauter M, Stefani M, Mack W. Towards Interactive Search: Investigating Visual Search in a Novel Real-World Paradigm. Brain Sci 2020; 10:E927. [PMID: 33271888 PMCID: PMC7761395 DOI: 10.3390/brainsci10120927] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2020] [Revised: 11/27/2020] [Accepted: 11/30/2020] [Indexed: 12/01/2022] Open
Abstract
An overwhelming majority of studies on visual search and selective attention were conducted using computer screens. There are arguably shortcomings in transferring knowledge from computer-based studies to real-world search behavior as findings are based on viewing static pictures on computer screens. This does not go well with the dynamic and interactive nature of vision in the real world. It is crucial to take visual search research to the real world in order to study everyday visual search processes. The aim of the present study was to develop an interactive search paradigm that can serve as a "bridge" between classical computerized search and everyday interactive search. We based our search paradigm on simple LEGO® bricks arranged on tabletop trays to ensure comparability with classical computerized visual search studies while providing room for easily increasing the complexity of the search environment. We found that targets were grasped slower when there were more distractors (Experiment 1) and there were sizable differences between various search conditions (Experiment 2), largely in line with classical visual search research and revealing similarities to research in natural scenes. Therefore, our paradigm can be seen as a valuable asset complementing visual search research in an environment between computerized search and everyday search.
Collapse
Affiliation(s)
- Marian Sauter
- General Psychology, Bundeswehr University Munich, 85579 Neubiberg, Germany; (M.S.); (W.M.)
- General Psychology, Ulm University, Albert-Einstein-Allee 47, 89081 Ulm, Germany
| | - Maximilian Stefani
- General Psychology, Bundeswehr University Munich, 85579 Neubiberg, Germany; (M.S.); (W.M.)
| | - Wolfgang Mack
- General Psychology, Bundeswehr University Munich, 85579 Neubiberg, Germany; (M.S.); (W.M.)
| |
Collapse
|
13
|
Henderson JM, Goold JE, Choi W, Hayes TR. Neural Correlates of Fixated Low- and High-level Scene Properties during Active Scene Viewing. J Cogn Neurosci 2020; 32:2013-2023. [PMID: 32573384 PMCID: PMC11164273 DOI: 10.1162/jocn_a_01599] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
During real-world scene perception, viewers actively direct their attention through a scene in a controlled sequence of eye fixations. During each fixation, local scene properties are attended, analyzed, and interpreted. What is the relationship between fixated scene properties and neural activity in the visual cortex? Participants inspected photographs of real-world scenes in an MRI scanner while their eye movements were recorded. Fixation-related fMRI was used to measure activation as a function of lower- and higher-level scene properties at fixation, operationalized as edge density and meaning maps, respectively. We found that edge density at fixation was most associated with activation in early visual areas, whereas semantic content at fixation was most associated with activation along the ventral visual stream including core object and scene-selective areas (lateral occipital complex, parahippocampal place area, occipital place area, and retrosplenial cortex). The observed activation from semantic content was not accounted for by differences in edge density. The results are consistent with active vision models in which fixation gates detailed visual analysis for fixated scene regions, and this gating influences both lower and higher levels of scene analysis.
Collapse
Affiliation(s)
| | | | - Wonil Choi
- Gwangju Institute of Science and Technology
| | | |
Collapse
|
14
|
Ognjanovic S, Thüring M, Murphy RO, Hölscher C. Display clutter and its effects on visual attention distribution and financial risk judgment. APPLIED ERGONOMICS 2019; 80:168-174. [PMID: 31280801 DOI: 10.1016/j.apergo.2019.05.008] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/01/2017] [Revised: 12/06/2018] [Accepted: 05/20/2019] [Indexed: 06/09/2023]
Abstract
Display clutter is a widely studied phenomenon in ergonomics, where information density and other properties of task-relevant visualizations are related to effective user performance and visual attention. This paper examines the impact of clutter in the context of financial stock visualizations. Depending on their expertise, traders can use a variety of different cues to judge the current and future value of a stock and to assess its riskiness. In our study, two groups of participants (novices and experts) judge the riskiness of 28 pairs of stocks under two clutter conditions (low and high). Consistency of judgments and group concordance serve as measures for judgment performance, while mean fixation duration, fixation frequency, and transition matrix density are employed to capture visual attention. Our results reveal significant effects of display clutter and expertise on both the performance measures as well as the visual attention measures.
Collapse
Affiliation(s)
- Svetlana Ognjanovic
- ETH Zurich, Chair of Cognitive Science, Clausiusstrasse 59, 8092, Zürich, Switzerland.
| | - Manfred Thüring
- Technische Universität Berlin, Faculty V of Mechanical Engineering and Transport Systems, Department of Psychology and Ergonomics, Chair of Cognitive Psychology and Cognitive Ergonomics, Marchstraße 23, D-10587, Berlin, Germany
| | - Ryan O Murphy
- Morningstar Investment Management, Head of Decision Sciences, 22 W. Washington St, Chicago, IL, 60602, USA
| | - Christoph Hölscher
- ETH Zurich, Chair of Cognitive Science, Clausiusstrasse 59, 8092, Zürich, Switzerland
| |
Collapse
|
15
|
Williams CC, Castelhano MS. The Changing Landscape: High-Level Influences on Eye Movement Guidance in Scenes. Vision (Basel) 2019; 3:E33. [PMID: 31735834 PMCID: PMC6802790 DOI: 10.3390/vision3030033] [Citation(s) in RCA: 17] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2019] [Revised: 06/20/2019] [Accepted: 06/24/2019] [Indexed: 11/16/2022] Open
Abstract
The use of eye movements to explore scene processing has exploded over the last decade. Eye movements provide distinct advantages when examining scene processing because they are both fast and spatially measurable. By using eye movements, researchers have investigated many questions about scene processing. Our review will focus on research performed in the last decade examining: (1) attention and eye movements; (2) where you look; (3) influence of task; (4) memory and scene representations; and (5) dynamic scenes and eye movements. Although typically addressed as separate issues, we argue that these distinctions are now holding back research progress. Instead, it is time to examine the intersections of these seemingly separate influences and examine the intersectionality of how these influences interact to more completely understand what eye movements can tell us about scene processing.
Collapse
Affiliation(s)
- Carrick C. Williams
- Department of Psychology, California State University San Marcos, San Marcos, CA 92069, USA
| | | |
Collapse
|
16
|
Do target detection and target localization always go together? Extracting information from briefly presented displays. Atten Percept Psychophys 2019; 81:2685-2699. [PMID: 31218599 DOI: 10.3758/s13414-019-01782-9] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
Abstract
The human visual system is capable of processing an enormous amount of information in a short time. Although rapid target detection has been explored extensively, less is known about target localization. Here we used natural scenes and explored the relationship between being able to detect a target (present vs. absent) and being able to localize it. Across four presentation durations (~ 33-199 ms), participants viewed scenes taken from two superordinate categories (natural and manmade), each containing exemplars from four basic scene categories. In a two-interval forced choice task, observers were asked to detect a Gabor target inserted in one of the two scenes. This was followed by one of two different localization tasks. Participants were asked either to discriminate whether the target was on the left or the right side of the display or to click on the exact location where they had seen the target. Targets could be detected and localized at our shortest exposure duration (~ 33 ms), with a predictable improvement in performance with increasing exposure duration. We saw some evidence at this shortest duration of detection without localization, but further analyses demonstrated that these trials typically reflected coarse or imprecise localization information, rather than its complete absence. Experiment 2 replicated our main findings while exploring the effect of the level of "openness" in the scene. Our results are consistent with the notion that when we are able to extract what objects are present in a scene, we also have information about where each object is, which provides crucial guidance for our goal-directed actions.
Collapse
|
17
|
Semizer Y, Michel MM. Natural image clutter degrades overt search performance independently of set size. J Vis 2019; 19:1. [PMID: 30933237 DOI: 10.1167/19.4.1] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Although studies of visual search have repeatedly demonstrated that visual clutter impairs search performance in natural scenes, these studies have not attempted to disentangle the effects of search set size from those of clutter per se. Here, we investigate the effect of natural image clutter on performance in an overt search for categorical targets when the search set size is controlled. Observers completed a search task that required detecting and localizing common objects in a set of natural images. The images were sorted into high- and low-clutter conditions based on the clutter metric by Bravo and Farid (2008). The search set size was varied independently by fixing the number and positions of potential targets across set size conditions within a block of trials. Within each fixed set size condition, search times increased as a function of increasing clutter, suggesting that clutter degrades overt search performance independently of set size.
Collapse
Affiliation(s)
- Yelda Semizer
- Department of Psychology, Rutgers University, New Brunswick, NJ, USA
| | - Melchi M Michel
- Department of Psychology, Rutgers University, New Brunswick, NJ, USA
| |
Collapse
|
18
|
Developing attentional control in naturalistic dynamic road crossing situations. Sci Rep 2019; 9:4176. [PMID: 30862845 PMCID: PMC6414534 DOI: 10.1038/s41598-019-39737-7] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2018] [Accepted: 01/24/2019] [Indexed: 11/09/2022] Open
Abstract
In the last 20 years, there has been increasing interest in studying visual attentional processes under more natural conditions. In the present study, we propose to determine the critical age at which children show similar to adult performance and attentional control in a visually guided task; in a naturalistic dynamic and socially relevant context: road crossing. We monitored visual exploration and crossing decisions in adults and children aged between 5 and 15 while they watched road traffic videos containing a range of traffic densities with or without pedestrians. 5–10 year old (y/o) children showed less systematic gaze patterns. More specifically, adults and 11–15 y/o children look mainly at the vehicles’ appearing point, which is an optimal location to sample diagnostic information for the task. In contrast, 5–10 y/os look more at socially relevant stimuli and attend to moving vehicles further down the trajectory when the traffic density is high. Critically, 5-10 y/o children also make an increased number of crossing decisions compared to 11–15 y/os and adults. Our findings reveal a critical shift around 10 y/o in attentional control and crossing decisions in a road crossing task.
Collapse
|
19
|
van Marlen T, van Wermeskerken M, van Gog T. Effects of visual complexity and ambiguity of verbal instructions on target identification. JOURNAL OF COGNITIVE PSYCHOLOGY 2019. [DOI: 10.1080/20445911.2018.1552700] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
Affiliation(s)
- Tim van Marlen
- Department of Education, Utrecht University Utrecht, Netherlands
| | | | - Tamara van Gog
- Department of Education, Utrecht University Utrecht, Netherlands
| |
Collapse
|
20
|
Still JD. Web page visual hierarchy: Examining Faraday's guidelines for entry points. COMPUTERS IN HUMAN BEHAVIOR 2018. [DOI: 10.1016/j.chb.2018.03.014] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
21
|
|
22
|
Pankok C, Kaber DB. An integrated measure of display clutter based on feature content, user knowledge and attention allocation factors. ERGONOMICS 2018; 61:682-696. [PMID: 29099336 DOI: 10.1080/00140139.2017.1401669] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/31/2017] [Accepted: 10/31/2017] [Indexed: 06/07/2023]
Abstract
Existing measures of display clutter in the literature generally exhibit weak correlations with task performance, which limits their utility in safety-critical domains. A literature review led to formulation of an integrated display data- and user knowledge-driven measure of display clutter. A driving simulation experiment was conducted in which participants were asked to search 'high' and 'low' clutter displays for navigation information. Data-driven measures and subjective perceptions of clutter were collected along with patterns of visual attention allocation and driving performance responses during time periods in which participants searched the navigation display for information. The new integrated measure was more strongly correlated with driving performance than other, previously developed measures of clutter, particularly in the case of low-clutter displays. Integrating display data and user knowledge factors with patterns of visual attention allocation shows promise for measuring display clutter and correlation with task performance, particularly for low-clutter displays. Practitioner Summary: A novel measure of display clutter was formulated, accounting for display data content, user knowledge states and patterns of visual attention allocation. The measure was evaluated in terms of correlations with driver performance in a safety-critical driving simulation study. The measure exhibited stronger correlations with task performance than previously defined measures.
Collapse
Affiliation(s)
- Carl Pankok
- a Industrial & Systems Engineering , North Carolina State University , Raleigh , NC , USA
| | - David B Kaber
- a Industrial & Systems Engineering , North Carolina State University , Raleigh , NC , USA
| |
Collapse
|
23
|
Cutting JE, DeLong JE, Brunick KL. Temporal fractals in movies and mind. Cogn Res Princ Implic 2018; 3:8. [PMID: 29577071 PMCID: PMC5849648 DOI: 10.1186/s41235-018-0091-x] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2017] [Accepted: 01/19/2018] [Indexed: 11/28/2022] Open
Abstract
Fractal patterns are seemingly everywhere. They can be analyzed through Fourier and power analyses, and other methods. Cutting, DeLong, and Nothelfer (2010) analyzed as time-series data the fluctuations of shot durations in 150 popular movies released over 70 years. They found that these patterns had become increasingly fractal-like and concluded that they might be linked to those found in the results of psychological tasks involving attention. To explore this possibility further, we began by analyzing the shot patterns of almost twice as many movies released over a century. The increasing fractal-like nature of shot patterns is affirmed, as determined by both a slope measure and a long-range dependence measure, neither of which is sensitive to the vector lengths of their inputs within the ranges explored here. But the main reason for increased long-range dependence is related to, but not caused by, the increasing vector length of the shot-series samples. It appears that, in generating increasingly fractal-like patterns, filmmakers have systematically explored dimensions that are important for holding our attention-shot durations, scene durations, motion, and sound amplitude-and have crafted fluctuations in them like those of our endogenous attention patterns. Other dimensions-luminance, clutter, and shot scale-are important to film style but their variations seem not to be important to holding viewers' moment-to-moment attention and have not changed in their fractional dimension over time.
Collapse
Affiliation(s)
- James E. Cutting
- Department of Psychology, Uris Hall, Cornell University, 109 Tower Road, Ithaca, NY 14853-7601 USA
| | | | | |
Collapse
|
24
|
Stainer MJ, Scott-Brown KC, Tatler BW. On the Factors Causing Processing Difficulty of Multiple-Scene Displays. Iperception 2017; 8:2041669516689572. [PMID: 28540027 PMCID: PMC5433555 DOI: 10.1177/2041669516689572] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/02/2022] Open
Abstract
Multiplex viewing of static or dynamic scenes is an increasing feature of screen media. Most existing multiplex experiments have examined detection across increasing scene numbers, but currently no systematic evaluation of the factors that might produce difficulty in processing multiplexes exists. Across five experiments we provide such an evaluation. Experiment 1 characterises difficulty in change detection when the number of scenes is increased. Experiment 2 reveals that the increased difficulty across multiple-scene displays is caused by the total amount of visual information accounts for differences in change detection times, regardless of whether this information is presented across multiple scenes, or contained in one scene. Experiment 3 shows that whether quadrants of a display were drawn from the same, or different scenes did not affect change detection performance. Experiment 4 demonstrates that knowing which scene the change will occur in means participants can perform at monoplex level. Finally, Experiment 5 finds that changes of central interest in multiplexed scenes are detected far easier than marginal interest changes to such an extent that a centrally interesting object removal in nine screens is detected more rapidly than a marginally interesting object removal in four screens. Processing multiple-screen displays therefore seems dependent on the amount of information, and the importance of that information to the task, rather than simply the number of scenes in the display. We discuss the theoretical and applied implications of these findings.
Collapse
|
25
|
Fixation durations in scene viewing: Modeling the effects of local image features, oculomotor parameters, and task. Psychon Bull Rev 2017; 24:370-392. [PMID: 27480268 PMCID: PMC5390002 DOI: 10.3758/s13423-016-1124-4] [Citation(s) in RCA: 49] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
Abstract
Scene perception requires the orchestration of image- and task-related processes with oculomotor constraints. The present study was designed to investigate how these factors influence how long the eyes remain fixated on a given location. Linear mixed models (LMMs) were used to test whether local image statistics (including luminance, luminance contrast, edge density, visual clutter, and the number of homogeneous segments), calculated for 1° circular regions around fixation locations, modulate fixation durations, and how these effects depend on task-related control. Fixation durations and locations were recorded from 72 participants, each viewing 135 scenes under three different viewing instructions (memorization, preference judgment, and search). Along with the image-related predictors, the LMMs simultaneously considered a number of oculomotor and spatiotemporal covariates, including the amplitudes of the previous and next saccades, and viewing time. As a key finding, the local image features around the current fixation predicted this fixation’s duration. For instance, greater luminance was associated with shorter fixation durations. Such immediacy effects were found for all three viewing tasks. Moreover, in the memorization and preference tasks, some evidence for successor effects emerged, such that some image characteristics of the upcoming location influenced how long the eyes stayed at the current location. In contrast, in the search task, scene processing was not distributed across fixation durations within the visual span. The LMM-based framework of analysis, applied to the control of fixation durations in scenes, suggests important constraints for models of scene perception and search, and for visual attention in general.
Collapse
|
26
|
Abstract
Many studies in reading have shown the enhancing effect of context on the
processing of a word before it is directly fixated (parafoveal processing of
words). Here, we examined whether scene context influences the parafoveal
processing of objects and enhances the extraction of object information. Using a
modified boundary paradigm called the Dot-Boundary paradigm, participants
fixated on a suddenly onsetting cue before the preview object would onset 4°
away. The preview object could be identical to the target, visually similar,
visually dissimilar or a control (black rectangle). The preview changed to the
target object once a saccade toward the object was made. Critically, the objects
were presented on either a consistent or an inconsistent scene background.
Results revealed that there was a greater processing benefit for consistent than
inconsistent scene backgrounds and that identical and visually similar previews
produced greater processing benefits than other previews. In the second
experiment, we added an additional context condition in which the target
location was inconsistent, but the scene semantics remained consistent. We found
that changing the location of the target object disrupted the processing benefit
derived from the consistent context. Most importantly, across both experiments,
the effect of preview was not enhanced by scene context. Thus, preview
information and scene context appear to independently boost the parafoveal
processing of objects without any interaction from object–scene congruency.
Collapse
Affiliation(s)
| | - Effie J Pereira
- Department of Psychology, McGill University, Kingston, ON, Canada
| |
Collapse
|
27
|
Segraves MA, Kuo E, Caddigan S, Berthiaume EA, Kording KP. Predicting rhesus monkey eye movements during natural-image search. J Vis 2017; 17:12. [PMID: 28355625 PMCID: PMC5373813 DOI: 10.1167/17.3.12] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
There are three prominent factors that can predict human visual-search behavior in natural scenes: the distinctiveness of a location (salience), similarity to the target (relevance), and features of the environment that predict where the object might be (context). We do not currently know how well these factors are able to predict macaque visual search, which matters because it is arguably the most popular model for asking how the brain controls eye movements. Here we trained monkeys to perform the pedestrian search task previously used for human subjects. Salience, relevance, and context models were all predictive of monkey eye fixations and jointly about as precise as for humans. We attempted to disrupt the influence of scene context on search by testing the monkeys with an inverted set of the same images. Surprisingly, the monkeys were able to locate the pedestrian at a rate similar to that for upright images. The best predictions of monkey fixations in searching inverted images were obtained by rotating the results of the model predictions for the original image. The fact that the same models can predict human and monkey search behavior suggests that the monkey can be used as a good model for understanding how the human brain enables natural-scene search.
Collapse
Affiliation(s)
- Mark A Segraves
- Department of Neurobiology, Weinberg College of Arts and Sciences, Northwestern University, Evanston, IL, USA
| | - Emory Kuo
- Department of Neurobiology, Weinberg College of Arts and Sciences, Northwestern University, Evanston, IL, USA
| | - Sara Caddigan
- Department of Neurobiology, Weinberg College of Arts and Sciences, Northwestern University, Evanston, IL, USA
| | - Emily A Berthiaume
- Department of Neurobiology, Weinberg College of Arts and Sciences, Northwestern University, Evanston, IL, USA
| | - Konrad P Kording
- Departments of Physical Medicine and Rehabilitation and Physiology, Feinberg School of Medicine, Northwestern University, Chicago, IL, USA
| |
Collapse
|
28
|
Detecting distortions of peripherally presented letter stimuli under crowded conditions. Atten Percept Psychophys 2017; 79:850-862. [DOI: 10.3758/s13414-016-1245-x] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
29
|
Henderson JM. Gaze Control as Prediction. Trends Cogn Sci 2017; 21:15-23. [DOI: 10.1016/j.tics.2016.11.003] [Citation(s) in RCA: 98] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2016] [Revised: 11/07/2016] [Accepted: 11/08/2016] [Indexed: 11/25/2022]
|
30
|
Cutting JE, Armstrong KL. Facial expression, size, and clutter: Inferences from movie structure to emotion judgments and back. Atten Percept Psychophys 2016; 78:891-901. [PMID: 26728045 PMCID: PMC4819543 DOI: 10.3758/s13414-015-1003-5] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
Abstract
The perception of facial expressions and objects at a distance are entrenched psychological research venues, but their intersection is not. We were motivated to study them together because of their joint importance in the physical composition of popular movies-shots that show a larger image of a face typically have shorter durations than those in which the face is smaller. For static images, we explore the time it takes viewers to categorize the valence of different facial expressions as a function of their visual size. In two studies, we find that smaller faces take longer to categorize than those that are larger, and this pattern interacts with local background clutter. More clutter creates crowding and impedes the interpretation of expressions for more distant faces but not proximal ones. Filmmakers at least tacitly know this. In two other studies, we show that contemporary movies lengthen shots that show smaller faces, and even more so with increased clutter.
Collapse
Affiliation(s)
- James E Cutting
- Department of Psychology, Cornell University, Uris Hall, Ithaca, NY, 14853-7601, USA.
| | - Kacie L Armstrong
- Department of Psychology, Cornell University, Uris Hall, Ithaca, NY, 14853-7601, USA
| |
Collapse
|
31
|
Bruce ND, Wloka C, Frosst N, Rahman S, Tsotsos JK. On computational modeling of visual saliency: Examining what’s right, and what’s left. Vision Res 2015; 116:95-112. [DOI: 10.1016/j.visres.2015.01.010] [Citation(s) in RCA: 54] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2014] [Revised: 12/16/2014] [Accepted: 01/19/2015] [Indexed: 11/26/2022]
|
32
|
Coco MI, Keller F, Malcolm GL. Anticipation in Real-World Scenes: The Role of Visual Context and Visual Memory. Cogn Sci 2015; 40:1995-2024. [DOI: 10.1111/cogs.12313] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2014] [Revised: 05/06/2015] [Accepted: 07/07/2015] [Indexed: 11/26/2022]
|
33
|
Koolen R, Krahmer E, Swerts M. How Distractor Objects Trigger Referential Overspecification: Testing the Effects of Visual Clutter and Distractor Distance. Cogn Sci 2015; 40:1617-1647. [PMID: 26432277 DOI: 10.1111/cogs.12297] [Citation(s) in RCA: 48] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2013] [Revised: 06/19/2015] [Accepted: 07/02/2015] [Indexed: 11/28/2022]
Abstract
In two experiments, we investigate to what extent various visual saliency cues in realistic visual scenes cause speakers to overspecify their definite object descriptions with a redundant color attribute. The results of the first experiment demonstrate that speakers are more likely to redundantly mention color when visual clutter is present in a scene as compared to when this is not the case. In the second experiment, we found that distractor type and distractor color affect redundant color use: Speakers are most likely to overspecify if there is at least one distractor object present that has the same type, but a different color than the target referent. Reliable effects of distractor distance were not found. Taken together, our results suggest that certain visual saliency cues guide speakers in determining which objects in a visual scene are relevant distractors, and which not. We argue that this is problematic for algorithms that aim to generate human-like descriptions of objects (such as the Incremental Algorithm), since these generally select properties that help to distinguish a target from all objects that are present in a scene.
Collapse
Affiliation(s)
- Ruud Koolen
- Tilburg Center for Cognition and Communication (TiCC), School of Humanities, Tilburg University.
| | - Emiel Krahmer
- Tilburg Center for Cognition and Communication (TiCC), School of Humanities, Tilburg University
| | - Marc Swerts
- Tilburg Center for Cognition and Communication (TiCC), School of Humanities, Tilburg University
| |
Collapse
|
34
|
Baltaretu A, Krahmer E, Maes A. Improving Route Directions: The Role of Intersection Type and Visual Clutter for Spatial Reference. APPLIED COGNITIVE PSYCHOLOGY 2015. [DOI: 10.1002/acp.3145] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Affiliation(s)
- Adriana Baltaretu
- Tilburg Center for Cognition and Communication (TiCC); Tilburg University; Tilburg The Netherlands
| | - Emiel Krahmer
- Tilburg Center for Cognition and Communication (TiCC); Tilburg University; Tilburg The Netherlands
| | - Alfons Maes
- Tilburg Center for Cognition and Communication (TiCC); Tilburg University; Tilburg The Netherlands
| |
Collapse
|
35
|
Adamo SH, Cain MS, Mitroff SR. Targets Need Their Own Personal Space: Effects of Clutter on Multiple-Target Search Accuracy. Perception 2015; 44:1203-14. [PMID: 26562889 DOI: 10.1177/0301006615594921] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Visual search is an essential task for many lifesaving professions; airport security personnel search baggage X-ray images for dangerous items and radiologists examine radiographs for tumors. Accuracy is critical for such searches; however, there are potentially negative influences that can affect performance; for example, the displays can be cluttered and can contain multiple targets. Previous research has demonstrated that clutter can hurt search performance and a second target is less likely to be detected in a multiple-target search after a first target has been found, which raises a concern-how does clutter affect multiple-target search performance? The current study explored clutter in a multiple-target search paradigm, where there could be one or two targets present, and targets appeared in varying levels of clutter. There was a significant interaction between clutter and target number: Increasing levels of clutter did not affect single-target detection but did reduce detection of a second target. Multiple-target search accuracy is known to be sensitive to contextual influences, and the current results reveal a specific effect wherein clutter disproportionally affected multiple-target search accuracy. These results suggest that the detection and processing of a first target might enhance the masking effects of clutter around a second target.
Collapse
Affiliation(s)
- Stephen H Adamo
- Department of Psychology and Neuroscience, Center for Cognitive Neuroscience, Duke University, NC, USA
| | - Matthew S Cain
- U.S. Army Natick Soldier Research, Development, and Engineering Center, MA, USA
| | - Stephen R Mitroff
- Department of Psychology and Neuroscience, Center for Cognitive Neuroscience, Duke University, NC, USA
| |
Collapse
|
36
|
Moacdieh N, Sarter N. Clutter in electronic medical records: examining its performance and attentional costs using eye tracking. HUMAN FACTORS 2015; 57:591-606. [PMID: 25850110 DOI: 10.1177/0018720814564594] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/22/2014] [Accepted: 11/21/2014] [Indexed: 06/04/2023]
Abstract
OBJECTIVE The objective was to use eye tracking to trace the underlying changes in attention allocation associated with the performance effects of clutter, stress, and task difficulty in visual search and noticing tasks. BACKGROUND Clutter can degrade performance in complex domains, yet more needs to be known about the associated changes in attention allocation, particularly in the presence of stress and for different tasks. Frequently used and relatively simple eye tracking metrics do not effectively capture the various effects of clutter, which is critical for comprehensively analyzing clutter and developing targeted, real-time countermeasures. METHOD Electronic medical records (EMRs) were chosen as the application domain for this research. Clutter, stress, and task difficulty were manipulated, and physicians' performance on search and noticing tasks was recorded. Several eye tracking metrics were used to trace attention allocation throughout those tasks, and subjective data were gathered via a debriefing questionnaire. RESULTS Clutter degraded performance in terms of response time and noticing accuracy. These decrements were largely accentuated by high stress and task difficulty. Eye tracking revealed the underlying attentional mechanisms, and several display-independent metrics were shown to be significant indicators of the effects of clutter. CONCLUSION Eye tracking provides a promising means to understand in detail (offline) and prevent (in real time) major performance breakdowns due to clutter. APPLICATION Display designers need to be aware of the risks of clutter in EMRs and other complex displays and can use the identified eye tracking metrics to evaluate and/or adjust their display.
Collapse
|
37
|
Zelinsky GJ, Yu CP. Clutter perception is invariant to image size. Vision Res 2015; 116:142-51. [PMID: 25982717 DOI: 10.1016/j.visres.2015.04.017] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2014] [Revised: 04/07/2015] [Accepted: 04/15/2015] [Indexed: 10/23/2022]
Abstract
Two experiments evaluated the effect of retinal image size on the proto-object model of visual clutter perception. Experiment 1 had 20 participants order 90 small images of random-category real-world scenes from least to most cluttered. Aggregating these individual rankings into a single median clutter ranking and comparing it to a previously reported clutter ranking of larger versions of the identical scenes yielded a Spearman's ρ=.953 (p<.001), suggesting that relative clutter perception is largely invariant to image size. We then applied the proto-object model of clutter perception to these smaller images and obtained a clutter estimate for each. Correlating these estimates with the median behavioral ranking yielded a Spearman's ρ=.852 (p<.001), which we showed in a comparative analysis to be better than six other methods of estimating clutter. Experiment 2 intermixed large and small versions of the Experiment 1 scenes and had participants (n=18) again rank them for clutter. We found that median clutter rankings of these size-intermixed images were essentially the same as the small and large median rankings from Experiment 1, suggesting size invariance in absolute clutter perception. Moreover, the proto-object model again successfully captured this result. We conclude that both relative and absolute clutter perception is invariant to retinal image size. We further speculate that clutter perception is mediated by proto-objects-a preattentive level of visual representation between features and objects-and that using the proto-object model we may be able to glimpse into this pre-attentive world.
Collapse
Affiliation(s)
- Gregory J Zelinsky
- Department of Psychology, Stony Brook University, USA; Department of Computer Science, Stony Brook University, USA.
| | - Chen-Ping Yu
- Department of Computer Science, Stony Brook University, USA
| |
Collapse
|
38
|
Abstract
Eye movements depend on cognitive processes related to visual information processing. Much has been learned about the spatial selection of fixation locations, while the principles governing the temporal control (fixation durations) are less clear. Here, we review current theories for the control of fixation durations in tasks like visual search, scanning, scene perception, and reading and propose a new model for the control of fixation durations. We distinguish two local principles from one global principle of control. First, an autonomous saccade timer initiates saccades after random time intervals (local-I). Second, foveal inhibition permits immediate prolongation of fixation durations by ongoing processing (local-II). Third, saccade timing is adaptive, so that the mean timer value depends on task requirements and fixation history (Global). We demonstrate by numerical simulations that our model qualitatively reproduces patterns of mean fixation durations and fixation duration distributions observed in typical experiments. When combined with assumptions of saccade target selection and oculomotor control, the model accounts for both temporal and spatial aspects of eye movement control in two versions of a visual search task. We conclude that the model provides a promising framework for the control of fixation durations in saccadic tasks.
Collapse
|
39
|
Nuthmann A, Einhäuser W. A new approach to modeling the influence of image features on fixation selection in scenes. Ann N Y Acad Sci 2015; 1339:82-96. [PMID: 25752239 PMCID: PMC4402003 DOI: 10.1111/nyas.12705] [Citation(s) in RCA: 24] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
Abstract
Which image characteristics predict where people fixate when memorizing natural images? To answer this question, we introduce a new analysis approach that combines a novel scene-patch analysis with generalized linear mixed models (GLMMs). Our method allows for (1) directly describing the relationship between continuous feature value and fixation probability, and (2) assessing each feature's unique contribution to fixation selection. To demonstrate this method, we estimated the relative contribution of various image features to fixation selection: luminance and luminance contrast (low-level features); edge density (a mid-level feature); visual clutter and image segmentation to approximate local object density in the scene (higher-level features). An additional predictor captured the central bias of fixation. The GLMM results revealed that edge density, clutter, and the number of homogenous segments in a patch can independently predict whether image patches are fixated or not. Importantly, neither luminance nor contrast had an independent effect above and beyond what could be accounted for by the other predictors. Since the parcellation of the scene and the selection of features can be tailored to the specific research question, our approach allows for assessing the interplay of various factors relevant for fixation selection in scenes in a powerful and flexible manner.
Collapse
Affiliation(s)
- Antje Nuthmann
- Psychology Department, School of Philosophy, Psychology and Language Sciences, University of EdinburghUnited Kingdom
| | | |
Collapse
|
40
|
Abstract
The auditory sense of humans transforms intrinsically senseless pressure waveforms into spectacularly rich perceptual phenomena: the music of Bach or the Beatles, the poetry of Li Bai or Omar Khayyam, or more prosaically the sense of the world filled with objects emitting sounds that is so important for those of us lucky enough to have hearing. Whereas the early representations of sounds in the auditory system are based on their physical structure, higher auditory centers are thought to represent sounds in terms of their perceptual attributes. In this symposium, we will illustrate the current research into this process, using four case studies. We will illustrate how the spectral and temporal properties of sounds are used to bind together, segregate, categorize, and interpret sound patterns on their way to acquire meaning, with important lessons to other sensory systems as well.
Collapse
|
41
|
Moacdieh N, Sarter N. Display clutter: a review of definitions and measurement techniques. HUMAN FACTORS 2015; 57:61-100. [PMID: 25790571 DOI: 10.1177/0018720814541145] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
OBJECTIVE We aimed to synthesize the literature on display clutter by reviewing definitions and measurement techniques and to develop a comprehensive, ergonomics-oriented perspective on clutter. We provide guidance for the selection of measurement approaches that can capture the various aspects and effects of clutter on visual search in particular. BACKGROUND There is agreement that clutter may represent a significant problem to operators. The challenge is to determine the ideal middle ground between excessive data and insufficient information. However, definitions of clutter vary widely, which explains the range of measurement approaches, including image-processing algorithms, performance evaluation, subjective evaluation, and eye tracking. It is important to understand the affordances of each technique and provide guidance for their use. METHOD We provide a systematic review of clutter definitions and develop a performance-oriented perspective for ergonomics research. Next, we present a critical overview of clutter measurement approaches. The benefits and limitations of each technique are detailed, and recommendations for best practice are provided. RESULTS From an ergonomics perspective, clutter matters to the extent that it affects performance, particularly, visual search. A combination of measurement techniques can be employed in order to assess the performance costs stemming from the multiple aspects of clutter. CONCLUSION Display clutter is a multifaceted construct that is a problem when it results in performance and attentional costs. The assessment of these costs is critical and requires the use of appropriate and complementary techniques. APPLICATION It is important that the performance costs of clutter are detected reliably to be able to better support attention management.
Collapse
|
42
|
Krishnan L, Elhilali M, Shamma S. Segregating complex sound sources through temporal coherence. PLoS Comput Biol 2014; 10:e1003985. [PMID: 25521593 PMCID: PMC4270434 DOI: 10.1371/journal.pcbi.1003985] [Citation(s) in RCA: 53] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/07/2014] [Accepted: 10/14/2014] [Indexed: 11/18/2022] Open
Abstract
A new approach for the segregation of monaural sound mixtures is presented based on the principle of temporal coherence and using auditory cortical representations. Temporal coherence is the notion that perceived sources emit coherently modulated features that evoke highly-coincident neural response patterns. By clustering the feature channels with coincident responses and reconstructing their input, one may segregate the underlying source from the simultaneously interfering signals that are uncorrelated with it. The proposed algorithm requires no prior information or training on the sources. It can, however, gracefully incorporate cognitive functions and influences such as memories of a target source or attention to a specific set of its attributes so as to segregate it from its background. Aside from its unusual structure and computational innovations, the proposed model provides testable hypotheses of the physiological mechanisms of this ubiquitous and remarkable perceptual ability, and of its psychophysical manifestations in navigating complex sensory environments. Humans and many animals can effortlessly navigate complex sensory environments, segregating and attending to one desired target source while suppressing distracting and interfering others. In this paper, we present an algorithmic model that can accomplish this task with no prior information or training on complex signals such as speech mixtures, and speech in noise and music. The model accounts for this ability relying solely on the temporal coherence principle, the notion that perceived sources emit coherently modulated features that evoke coincident cortical response patterns. It further demonstrates how basic cortical mechanisms common to all sensory systems can implement the necessary representations, as well as the adaptive computations necessary to maintain continuity by tracking slowly changing characteristics of different sources in a scene.
Collapse
Affiliation(s)
- Lakshmi Krishnan
- Department of Electrical and Computer Engineering, University of Maryland, College Park, Maryland, United States of America
- * E-mail:
| | - Mounya Elhilali
- Department of Electrical and Computer Engineering, The Johns Hopkins University, Baltimore, Maryland, United States of America
| | - Shihab Shamma
- Department of Electrical and Computer Engineering, University of Maryland, College Park, Maryland, United States of America
- Department Etudes Cognitive, Ecole Normale Superieure, Paris, France
| |
Collapse
|
43
|
Gnanaseelan R, Gonzalez DA, Niechwiej-Szwedo E. Binocular advantage for prehension movements performed in visually enriched environments requiring visual search. Front Hum Neurosci 2014; 8:959. [PMID: 25506323 PMCID: PMC4246685 DOI: 10.3389/fnhum.2014.00959] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2014] [Accepted: 11/11/2014] [Indexed: 11/13/2022] Open
Abstract
The purpose of this study was to examine the role of binocular vision during a prehension task performed in a visually enriched environment where the target object was surrounded by distractors/obstacles. Fifteen adults reached and grasped for a cylindrical peg while eye movements and upper limb kinematics were recorded. The complexity of the visual environment was manipulated by varying the number of distractors and by varying the saliency of the target. Gaze behavior (i.e., the latency of the primary gaze shift and frequency of gaze shifts prior to reach initiation) was comparable between viewing conditions. In contrast, a binocular advantage was evident in performance accuracy. Specifically, participants picked up the wrong object twice as often during monocular viewing when the complexity of the environment increased. Reach performance was more efficient during binocular viewing, which was demonstrated by shorter reach reaction time and overall movement time. Reaching movements during the approach phase had higher peak velocity during binocular viewing. During monocular viewing reach trajectories exhibited a direction bias during the acceleration phase, which was leftward during left eye viewing and rightward during right eye viewing. This bias can be explained by the presence of esophoria in the covered eye. The grasping interval was also extended by ~20% during monocular viewing; however, the duration of the return phase after the target was picked up was comparable across viewing conditions. In conclusion, binocular vision provides important input for planning and execution of prehension movements in visually enriched environments. Binocular advantage was evident, regardless of set size or target saliency, indicating that adults plan their movements more cautiously during monocular viewing, even in relatively simple environments with a highly salient target. Nevertheless, in visually-normal adults monocular input provides sufficient information to engage in online control to correct the initial errors in movement planning.
Collapse
Affiliation(s)
- Roshani Gnanaseelan
- Visuomotor Neuroscience Lab, Department of Kinesiology, University of Waterloo Waterloo, ON, Canada
| | - Dave A Gonzalez
- Visuomotor Neuroscience Lab, Department of Kinesiology, University of Waterloo Waterloo, ON, Canada
| | - Ewa Niechwiej-Szwedo
- Visuomotor Neuroscience Lab, Department of Kinesiology, University of Waterloo Waterloo, ON, Canada
| |
Collapse
|
44
|
Integrating mechanisms of visual guidance in naturalistic language production. Cogn Process 2014; 16:131-50. [DOI: 10.1007/s10339-014-0642-0] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2014] [Accepted: 10/30/2014] [Indexed: 10/24/2022]
|
45
|
Caroux L, Le Bigot L, Vibert N. Impairment of shooting performance by background complexity and motion. Exp Psychol 2014; 62:98-109. [PMID: 25384639 DOI: 10.1027/1618-3169/a000277] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
Abstract
In many visual displays such as virtual environments, human tasks involve objects superimposed on both complex and moving backgrounds. However, most studies investigated the influence of background complexity or background motion in isolation. Two experiments were designed to investigate the joint influences of background complexity and lateral motion on a simple shooting task typical of video games. Participants had to perform the task on the moving and static versions of backgrounds of three levels of complexity, while their eye movements were recorded. The backgrounds displayed either an abstract (Experiment 1) or a naturalistic (Experiment 2) virtual environment. The results showed that performance was impaired by background motion in both experiments. The effects of motion and complexity were additive for the abstract background and multiplicative for the naturalistic background. Eye movement recordings showed that performance impairments reflected at least in part the impact of the background visual features on gaze control.
Collapse
Affiliation(s)
- Loïc Caroux
- Centre de Recherches sur la Cognition et l'Apprentissage (CeRCA), UMR 7295 - University of Poitiers, University of Tours, CNRS, France INRIA Bordeaux Sud-Ouest, Talence, France
| | - Ludovic Le Bigot
- Centre de Recherches sur la Cognition et l'Apprentissage (CeRCA), UMR 7295 - University of Poitiers, University of Tours, CNRS, France
| | - Nicolas Vibert
- Centre de Recherches sur la Cognition et l'Apprentissage (CeRCA), UMR 7295 - University of Poitiers, University of Tours, CNRS, France
| |
Collapse
|
46
|
Wass SV, Smith TJ. Individual Differences in Infant Oculomotor Behavior During the Viewing of Complex Naturalistic Scenes. INFANCY 2014; 19:352-384. [PMID: 25635173 PMCID: PMC4286103 DOI: 10.1111/infa.12049] [Citation(s) in RCA: 51] [Impact Index Per Article: 5.1] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Abstract
Little research hitherto has examined how individual differences in attention, as assessed using standard experimental paradigms, relate to individual differences in how attention is spontaneously allocated in more naturalistic contexts. Here, we analyzed the time intervals between refoveating eye movements (fixation durations) while typically developing 11-month-old infants viewed a 90-min battery ranging from complex dynamic to noncomplex static materials. The same infants also completed experimental assessments of cognitive control, psychomotor reaction times (RT), processing speed (indexed via peak look during habituation), and arousal (indexed via tonic pupil size). High test-retest reliability was found for fixation duration, across testing sessions and across types of viewing material. Increased cognitive control and increased arousal were associated with reduced variability in fixation duration. For fixations to dynamic stimuli, in which a large proportion of saccades may be exogenously cued, we found that psychomotor RT measures were most predictive of mean fixation duration; for fixations to static stimuli, in contrast, in which there is less exogenous attentional capture, we found that psychomotor RT did not predict performance, but that measures of cognitive control and arousal did. The implications of these findings for understanding the development of attentional control in naturalistic settings are discussed.
Collapse
Affiliation(s)
| | - Tim J Smith
- School of Psychological SciencesBirkbeck College, University of London
| |
Collapse
|
47
|
Yu CP, Samaras D, Zelinsky GJ. Modeling visual clutter perception using proto-object segmentation. J Vis 2014; 14:4. [PMID: 24904121 PMCID: PMC4528410 DOI: 10.1167/14.7.4] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2013] [Accepted: 02/23/2014] [Indexed: 11/24/2022] Open
Abstract
We introduce the proto-object model of visual clutter perception. This unsupervised model segments an image into superpixels, then merges neighboring superpixels that share a common color cluster to obtain proto-objects-defined here as spatially extended regions of coherent features. Clutter is estimated by simply counting the number of proto-objects. We tested this model using 90 images of realistic scenes that were ranked by observers from least to most cluttered. Comparing this behaviorally obtained ranking to a ranking based on the model clutter estimates, we found a significant correlation between the two (Spearman's ρ = 0.814, p < 0.001). We also found that the proto-object model was highly robust to changes in its parameters and was generalizable to unseen images. We compared the proto-object model to six other models of clutter perception and demonstrated that it outperformed each, in some cases dramatically. Importantly, we also showed that the proto-object model was a better predictor of clutter perception than an actual count of the number of objects in the scenes, suggesting that the set size of a scene may be better described by proto-objects than objects. We conclude that the success of the proto-object model is due in part to its use of an intermediate level of visual representation-one between features and objects-and that this is evidence for the potential importance of a proto-object representation in many common visual percepts and tasks.
Collapse
Affiliation(s)
- Chen-Ping Yu
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
| | - Dimitris Samaras
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
| | - Gregory J. Zelinsky
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
- Department of Psychology, Stony Brook University, Stony Brook, NY, USA
| |
Collapse
|
48
|
Edler D, Bestgen AK, Kuchinke L, Dickmann F. Grids in topographic maps reduce distortions in the recall of learned object locations. PLoS One 2014; 9:e98148. [PMID: 24869486 PMCID: PMC4037198 DOI: 10.1371/journal.pone.0098148] [Citation(s) in RCA: 36] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2013] [Accepted: 04/29/2014] [Indexed: 11/18/2022] Open
Abstract
To date, it has been shown that cognitive map representations based on cartographic visualisations are systematically distorted. The grid is a traditional element of map graphics that has rarely been considered in research on perception-based spatial distortions. Grids do not only support the map reader in finding coordinates or locations of objects, they also provide a systematic structure for clustering visual map information (“spatial chunks”). The aim of this study was to examine whether different cartographic kinds of grids reduce spatial distortions and improve recall memory for object locations. Recall performance was measured as both the percentage of correctly recalled objects (hit rate) and the mean distance errors of correctly recalled objects (spatial accuracy). Different kinds of grids (continuous lines, dashed lines, crosses) were applied to topographic maps. These maps were also varied in their type of characteristic areas (LANDSCAPE) and different information layer compositions (DENSITY) to examine the effects of map complexity. The study involving 144 participants shows that all experimental cartographic factors (GRID, LANDSCAPE, DENSITY) improve recall performance and spatial accuracy of learned object locations. Overlaying a topographic map with a grid significantly reduces the mean distance errors of correctly recalled map objects. The paper includes a discussion of a square grid's usefulness concerning object location memory, independent of whether the grid is clearly visible (continuous or dashed lines) or only indicated by crosses.
Collapse
Affiliation(s)
- Dennis Edler
- Department of Geography, Ruhr-University Bochum, Bochum, Germany
- * E-mail:
| | | | - Lars Kuchinke
- Department of Psychology, Ruhr-University Bochum, Bochum, Germany
| | - Frank Dickmann
- Department of Geography, Ruhr-University Bochum, Bochum, Germany
| |
Collapse
|
49
|
Abstract
A head camera was used to examine the visual correlates of object name learning by toddlers as they played with novel objects and as the parent spontaneously named those objects. The toddlers' learning of the object names was tested after play, and the visual properties of the head camera images during naming events associated with learned and unlearned object names were analyzed. Naming events associated with learning had a clear visual signature, one in which the visual information itself was clean and visual competition among objects was minimized. Moreover, for learned object names, the visual advantage of the named target over competitors was sustained, both before and after the heard name. The findings are discussed in terms of the visual and cognitive processes that may depend on clean sensory input for learning and also on the sensory-motor, cognitive, and social processes that may create these optimal visual moments for learning.
Collapse
|
50
|
Clarke ADF, Coco MI, Keller F. The impact of attentional, linguistic, and visual features during object naming. Front Psychol 2013; 4:927. [PMID: 24379792 PMCID: PMC3861867 DOI: 10.3389/fpsyg.2013.00927] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2013] [Accepted: 11/23/2013] [Indexed: 11/14/2022] Open
Abstract
Object detection and identification are fundamental to human vision, and there is mounting evidence that objects guide the allocation of visual attention. However, the role of objects in tasks involving multiple modalities is less clear. To address this question, we investigate object naming, a task in which participants have to verbally identify objects they see in photorealistic scenes. We report an eye-tracking study that investigates which features (attentional, visual, and linguistic) influence object naming. We find that the amount of visual attention directed toward an object, its position and saliency, along with linguistic factors such as word frequency, animacy, and semantic proximity, significantly influence whether the object will be named or not. We then ask how features from different modalities are combined during naming, and find significant interactions between saliency and position, saliency and linguistic features, and attention and position. We conclude that when the cognitive system performs tasks such as object naming, it uses input from one modality to constraint or enhance the processing of other modalities, rather than processing each input modality independently.
Collapse
Affiliation(s)
- Alasdair D. F. Clarke
- Institute for Language, Cognition and Computation, School of Informatics, University of EdinburghEdinburgh, UK
| | - Moreno I. Coco
- Institute for Language, Cognition and Computation, School of Informatics, University of EdinburghEdinburgh, UK
- Faculdade de Psicologia, Universidade de LisboaLisboa, Portugal
| | - Frank Keller
- Institute for Language, Cognition and Computation, School of Informatics, University of EdinburghEdinburgh, UK
| |
Collapse
|