1
|
Guo Y, Pannasch S, Helmert JR, Kaszowska A. Ambient and focal attention during complex problem-solving: preliminary evidence from real-world eye movement data. Front Psychol 2024; 15:1217106. [PMID: 38425554 PMCID: PMC10902451 DOI: 10.3389/fpsyg.2024.1217106] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2023] [Accepted: 01/31/2024] [Indexed: 03/02/2024] Open
Abstract
Time course analysis of eye movements during free exploration of real-world scenes often reveals an increase in fixation durations together with a decrease in saccade amplitudes, which has been explained within the two visual systems approach, i.e., a transition from ambient to focal. Short fixations and long saccades during early viewing periods are classified as ambient mode of vision, which is concerned with spatial orientation and is related to simple visual properties such as motion, contrast, and location. Longer fixations and shorter saccades during later viewing periods are classified as focal mode of vision, which is concentrated in the foveal projection and is capable of object identification and its semantic categorization. While these findings are mainly obtained in the context of image exploration, the present study endeavors to investigate whether the same pattern of interplay between ambient and focal visual attention is deployed when people work on complex real-world tasks-and if so, when? Based on a re-analysis of existing data that integrates concurrent think aloud and eye tracking protocols, the present study correlated participants' internal thinking models to the parameters of their eye movements when they planned solutions to an open-ended design problem in a real-world setting. We hypothesize that switching between ambient and focal attentional processing is useful when solvers encounter difficulty compelling them to shift their conceptual direction to adjust the solution path. Individuals may prefer different attentional strategies for information-seeking behavior, such as ambient-to-focal or focal-to-ambient. The observed increase in fixation durations and decrease in saccade amplitudes during the periods around shifts in conceptual direction lends support to the postulation of the ambient-to-focal processing; however, focal-to-ambient processing is not evident. Furthermore, our data demonstrate that the beginning of a shift in conceptual direction is observable in eye movement behavior with a significant prolongation of fixation. Our findings add to the conclusions drawn from laboratory settings by providing preliminary evidence for ambient and focal processing characteristics in real-world problem-solving.
Collapse
Affiliation(s)
- Yuxuan Guo
- Institute of Psychology III, Engineering Psychology and Applied Cognitive Research, Technische Universität Dresden, Dresden, Germany
| | - Sebastian Pannasch
- Institute of Psychology III, Engineering Psychology and Applied Cognitive Research, Technische Universität Dresden, Dresden, Germany
| | - Jens R. Helmert
- Institute of Psychology III, Engineering Psychology and Applied Cognitive Research, Technische Universität Dresden, Dresden, Germany
| | | |
Collapse
|
2
|
Gronau N. To Grasp the World at a Glance: The Role of Attention in Visual and Semantic Associative Processing. J Imaging 2021; 7:jimaging7090191. [PMID: 34564117 PMCID: PMC8470651 DOI: 10.3390/jimaging7090191] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2021] [Revised: 08/30/2021] [Accepted: 09/15/2021] [Indexed: 11/16/2022] Open
Abstract
Associative relations among words, concepts and percepts are the core building blocks of high-level cognition. When viewing the world ‘at a glance’, the associative relations between objects in a scene, or between an object and its visual background, are extracted rapidly. The extent to which such relational processing requires attentional capacity, however, has been heavily disputed over the years. In the present manuscript, I review studies investigating scene–object and object–object associative processing. I then present a series of studies in which I assessed the necessity of spatial attention to various types of visual–semantic relations within a scene. Importantly, in all studies, the spatial and temporal aspects of visual attention were tightly controlled in an attempt to minimize unintentional attention shifts from ‘attended’ to ‘unattended’ regions. Pairs of stimuli—either objects, scenes or a scene and an object—were briefly presented on each trial, while participants were asked to detect a pre-defined target category (e.g., an animal, a nonsense shape). Response times (RTs) to the target detection task were registered when visual attention spanned both stimuli in a pair vs. when attention was focused on only one of two stimuli. Among non-prioritized stimuli that were not defined as to-be-detected targets, findings consistently demonstrated rapid associative processing when stimuli were fully attended, i.e., shorter RTs to associated than unassociated pairs. Focusing attention on a single stimulus only, however, largely impaired this relational processing. Notably, prioritized targets continued to affect performance even when positioned at an unattended location, and their associative relations with the attended items were well processed and analyzed. Our findings portray an important dissociation between unattended task-irrelevant and task-relevant items: while the former require spatial attentional resources in order to be linked to stimuli positioned inside the attentional focus, the latter may influence high-level recognition and associative processes via feature-based attentional mechanisms that are largely independent of spatial attention.
Collapse
Affiliation(s)
- Nurit Gronau
- Department of Psychology and Department of Cognitive Science Studies, The Open University of Israel, Raanana 4353701, Israel
| |
Collapse
|
3
|
Hayes TR, Henderson JM. Deep saliency models learn low-, mid-, and high-level features to predict scene attention. Sci Rep 2021; 11:18434. [PMID: 34531484 PMCID: PMC8445969 DOI: 10.1038/s41598-021-97879-z] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2021] [Accepted: 08/31/2021] [Indexed: 02/08/2023] Open
Abstract
Deep saliency models represent the current state-of-the-art for predicting where humans look in real-world scenes. However, for deep saliency models to inform cognitive theories of attention, we need to know how deep saliency models prioritize different scene features to predict where people look. Here we open the black box of three prominent deep saliency models (MSI-Net, DeepGaze II, and SAM-ResNet) using an approach that models the association between attention, deep saliency model output, and low-, mid-, and high-level scene features. Specifically, we measured the association between each deep saliency model and low-level image saliency, mid-level contour symmetry and junctions, and high-level meaning by applying a mixed effects modeling approach to a large eye movement dataset. We found that all three deep saliency models were most strongly associated with high-level and low-level features, but exhibited qualitatively different feature weightings and interaction patterns. These findings suggest that prominent deep saliency models are primarily learning image features associated with high-level scene meaning and low-level image saliency and highlight the importance of moving beyond simply benchmarking performance.
Collapse
Affiliation(s)
- Taylor R Hayes
- Center for Mind and Brain, University of California, Davis, 95618, USA.
| | - John M Henderson
- Center for Mind and Brain, University of California, Davis, 95618, USA
- Department of Psychology, University of California, Davis, 95616, USA
| |
Collapse
|
4
|
Henderson JM, Hayes TR, Peacock CE, Rehrig G. Meaning maps capture the density of local semantic features in scenes: A reply to Pedziwiatr, Kümmerer, Wallis, Bethge & Teufel (2021). Cognition 2021; 214:104742. [PMID: 33892912 PMCID: PMC11166323 DOI: 10.1016/j.cognition.2021.104742] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2021] [Revised: 04/13/2021] [Accepted: 04/15/2021] [Indexed: 11/17/2022]
Abstract
Pedziwiatr, Kümmerer, Wallis, Bethge, & Teufel (2021) contend that Meaning Maps do not represent the spatial distribution of semantic features in scenes. We argue that Pesziwiatr et al. provide neither logical nor empirical support for that claim, and we conclude that Meaning Maps do what they were designed to do: represent the spatial distribution of meaning in scenes.
Collapse
Affiliation(s)
- John M Henderson
- Center for Mind and Brain, University of California, Davis, USA; Department of Psychology, University of California, Davis, USA.
| | - Taylor R Hayes
- Center for Mind and Brain, University of California, Davis, USA
| | - Candace E Peacock
- Center for Mind and Brain, University of California, Davis, USA; Department of Psychology, University of California, Davis, USA
| | | |
Collapse
|
5
|
Chen Y, Yang Z, Ahn S, Samaras D, Hoai M, Zelinsky G. COCO-Search18 fixation dataset for predicting goal-directed attention control. Sci Rep 2021; 11:8776. [PMID: 33888734 PMCID: PMC8062491 DOI: 10.1038/s41598-021-87715-9] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2020] [Accepted: 03/31/2021] [Indexed: 11/23/2022] Open
Abstract
Attention control is a basic behavioral process that has been studied for decades. The currently best models of attention control are deep networks trained on free-viewing behavior to predict bottom-up attention control - saliency. We introduce COCO-Search18, the first dataset of laboratory-quality goal-directed behavior large enough to train deep-network models. We collected eye-movement behavior from 10 people searching for each of 18 target-object categories in 6202 natural-scene images, yielding [Formula: see text] 300,000 search fixations. We thoroughly characterize COCO-Search18, and benchmark it using three machine-learning methods: a ResNet50 object detector, a ResNet50 trained on fixation-density maps, and an inverse-reinforcement-learning model trained on behavioral search scanpaths. Models were also trained/tested on images transformed to approximate a foveated retina, a fundamental biological constraint. These models, each having a different reliance on behavioral training, collectively comprise the new state-of-the-art in predicting goal-directed search fixations. Our expectation is that future work using COCO-Search18 will far surpass these initial efforts, finding applications in domains ranging from human-computer interactive systems that can anticipate a person's intent and render assistance to the potentially early identification of attention-related clinical disorders (ADHD, PTSD, phobia) based on deviation from neurotypical fixation behavior.
Collapse
Affiliation(s)
- Yupei Chen
- Department of Psychology, Stony Brook University, New York, USA
| | - Zhibo Yang
- Department of Computer Science, Stony Brook University, New York, USA
| | - Seoyoung Ahn
- Department of Psychology, Stony Brook University, New York, USA
| | - Dimitris Samaras
- Department of Computer Science, Stony Brook University, New York, USA
| | - Minh Hoai
- Department of Computer Science, Stony Brook University, New York, USA
| | - Gregory Zelinsky
- Department of Psychology, Stony Brook University, New York, USA.
- Department of Computer Science, Stony Brook University, New York, USA.
| |
Collapse
|
6
|
Nickel AE, Hopkins LS, Minor GN, Hannula DE. Attention capture by episodic long-term memory. Cognition 2020; 201:104312. [PMID: 32387722 DOI: 10.1016/j.cognition.2020.104312] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2019] [Revised: 03/16/2020] [Accepted: 04/19/2020] [Indexed: 10/24/2022]
Abstract
Everyday behavior depends upon the operation of concurrent cognitive processes. In visual search, studies that examine memory-attention interactions have indicated that long-term memory facilitates search for a target (e.g., contextual cueing), but the potential for memories to capture attention and decrease search efficiency has not been investigated. To address this gap in the literature, five experiments were conducted to examine whether task-irrelevant encoded objects might capture attention. In each experiment, participants encoded scene-object pairs. Then, in a visual search task, 6-object search displays were presented and participants were told to make a single saccade to targets defined by shape (e.g., diamond among differently colored circles; Experiments 1, 4, and 5) or by color (e.g., blue shape among differently shaped gray objects; Experiments 2 and 3). Sometimes, one of the distractors was from the encoded set, and occasionally the scene that had been paired with that object was presented prior to the search display. Results indicated that eye movements were made, in error, more often to encoded distractors than to baseline distractors, and that this effect was greatest when the corresponding scene was presented prior to search. When capture did occur, participants looked longer at encoded distractors if scenes had been presented, an effect that we attribute to the representational match between a retrieved associate and the identity of the encoded distractor in the search display. In addition, the presence of a scene resulted in slower saccade deployment when participants made first saccades to targets, as instructed. Experiments 4 and 5 suggest that this slowdown may be due to the relatively rare and therefore, surprising, appearance of visual stimulus information prior to search. Collectively, results suggest that information encoded into episodic memory can capture attention, which is consistent with the recent proposal that selection history can guide attentional selection.
Collapse
Affiliation(s)
- Allison E Nickel
- Department of Psychology, University of Wisconsin - Milwaukee, Milwaukee, WI, USA
| | - Lauren S Hopkins
- Department of Psychology, University of Wisconsin - Milwaukee, Milwaukee, WI, USA
| | - Greta N Minor
- Department of Psychology, University of Wisconsin - Milwaukee, Milwaukee, WI, USA
| | - Deborah E Hannula
- Department of Psychology, University of Wisconsin - Milwaukee, Milwaukee, WI, USA.
| |
Collapse
|
7
|
Henderson JM. Meaning and attention in scenes. PSYCHOLOGY OF LEARNING AND MOTIVATION 2020. [DOI: 10.1016/bs.plm.2020.08.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
8
|
Peacock CE, Hayes TR, Henderson JM. The role of meaning in attentional guidance during free viewing of real-world scenes. Acta Psychol (Amst) 2019; 198:102889. [PMID: 31302302 DOI: 10.1016/j.actpsy.2019.102889] [Citation(s) in RCA: 20] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2018] [Revised: 06/27/2019] [Accepted: 07/05/2019] [Indexed: 10/26/2022] Open
Abstract
In real-world vision, humans prioritize the most relevant visual information at the expense of other information via attentional selection. The current study sought to understand the role of semantic features and image features on attentional selection during free viewing of real-world scenes. We compared the ability of meaning maps generated from ratings of isolated, context-free image patches and saliency maps generated from the Graph-Based Visual Saliency model to predict the spatial distribution of attention in scenes as measured by eye movements. Additionally, we introduce new contextualized meaning maps in which scene patches were rated based upon how informative or recognizable they were in the context of the scene from which they derived. We found that both context-free and contextualized meaning explained significantly more of the overall variance in the spatial distribution of attention than image salience. Furthermore, meaning explained early attention to a significantly greater extent than image salience, contrary to predictions of the 'saliency first' hypothesis. Finally, both context-free and contextualized meaning predicted attention equivalently. These results support theories in which meaning plays a dominant role in attentional guidance during free viewing of real-world scenes.
Collapse
|
9
|
Henderson JM, Hayes TR, Peacock CE, Rehrig G. Meaning and Attentional Guidance in Scenes: A Review of the Meaning Map Approach. Vision (Basel) 2019; 3:E19. [PMID: 31735820 PMCID: PMC6802777 DOI: 10.3390/vision3020019] [Citation(s) in RCA: 26] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2019] [Revised: 04/24/2019] [Accepted: 05/07/2019] [Indexed: 11/16/2022] Open
Abstract
Perception of a complex visual scene requires that important regions be prioritized and attentionally selected for processing. What is the basis for this selection? Although much research has focused on image salience as an important factor guiding attention, relatively little work has focused on semantic salience. To address this imbalance, we have recently developed a new method for measuring, representing, and evaluating the role of meaning in scenes. In this method, the spatial distribution of semantic features in a scene is represented as a meaning map. Meaning maps are generated from crowd-sourced responses given by naïve subjects who rate the meaningfulness of a large number of scene patches drawn from each scene. Meaning maps are coded in the same format as traditional image saliency maps, and therefore both types of maps can be directly evaluated against each other and against maps of the spatial distribution of attention derived from viewers' eye fixations. In this review we describe our work focusing on comparing the influences of meaning and image salience on attentional guidance in real-world scenes across a variety of viewing tasks that we have investigated, including memorization, aesthetic judgment, scene description, and saliency search and judgment. Overall, we have found that both meaning and salience predict the spatial distribution of attention in a scene, but that when the correlation between meaning and salience is statistically controlled, only meaning uniquely accounts for variance in attention.
Collapse
Affiliation(s)
- John M. Henderson
- Center for Mind and Brain, 267 Cousteau Place, University of California, Davis, CA 95618, USA
- Department of Psychology, University of California, Davis, CA 95618, USA
| | - Taylor R. Hayes
- Center for Mind and Brain, 267 Cousteau Place, University of California, Davis, CA 95618, USA
| | - Candace E. Peacock
- Center for Mind and Brain, 267 Cousteau Place, University of California, Davis, CA 95618, USA
- Department of Psychology, University of California, Davis, CA 95618, USA
| | - Gwendolyn Rehrig
- Department of Psychology, University of California, Davis, CA 95618, USA
| |
Collapse
|
10
|
Abstract
Rich contextual and semantic information can be extracted from only a brief presentation of a natural scene. This is presumed to be activated quickly enough to guide initial eye movements into a scene. However, early, short-latency eye movements in natural scenes have been shown to be dependent on the salience distribution across the image (Anderson, Ort, Kruijne, Meeter, & Donk, 2015). In the present work, we manipulated the salience distribution across a natural scene by changing the global contrast. We showed participants a brief real or nonsense preview of the scene and examined the time-course of eye movement guidance. A real preview decreased the latency and increased the amplitude of initial saccades into the image, suggesting that the preview allowed observers to obtain additional contextual information that would otherwise not be available. However, the preview did not completely override the initial tendency for short-latency saccades to be guided by the underlying salience distribution of the image. We discuss these findings in the context of oculomotor selection based on the integration of contextual information and low-level features in a natural scene.
Collapse
|
11
|
Gong M, Xuan Y, Xu X, Fu X. The Effect of Consistency on Short-Term Memory for Scenes. Front Psychol 2017; 8:1712. [PMID: 29046654 PMCID: PMC5632670 DOI: 10.3389/fpsyg.2017.01712] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2017] [Accepted: 09/19/2017] [Indexed: 11/13/2022] Open
Abstract
Which is more detectable, the change of a consistent or an inconsistent object in a scene? This question has been debated for decades. We noted that the change of objects in scenes might simultaneously be accompanied with gist changes. In the present study we aimed to examine how the alteration of gist, as well as the consistency of the changed objects, modulated change detection. In Experiment 1, we manipulated the semantic content by either keeping or changing the consistency of the scene. Results showed that the changes of consistent and inconsistent scenes were equally detected. More importantly, the changes were more accurately detected when scene consistency changed than when the consistency remained unchanged, regardless of the consistency of the memory scenes. A phase-scrambled version of stimuli was adopted in Experiment 2 to decouple the possible confounding effect of low-level factors. The results of Experiment 2 demonstrated that the effect found in Experiment 1 was indeed due to the change of high-level semantic consistency rather than the change of low-level physical features. Together, the study suggests that the change of consistency plays an important role in scene short-term memory, which might be attributed to the sensitivity to the change of semantic content.
Collapse
Affiliation(s)
- Mingliang Gong
- State Key Laboratory of Brain and Cognitive Sciences, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.,Department of Psychology, Miami University, Oxford, OH, United States
| | - Yuming Xuan
- State Key Laboratory of Brain and Cognitive Sciences, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.,Department of Psychology, University of the Chinese Academy of Sciences, Beijing, China
| | - Xinwen Xu
- State Key Laboratory of Brain and Cognitive Sciences, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.,Department of Psychology, University of the Chinese Academy of Sciences, Beijing, China
| | - Xiaolan Fu
- State Key Laboratory of Brain and Cognitive Sciences, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.,Department of Psychology, University of the Chinese Academy of Sciences, Beijing, China
| |
Collapse
|
12
|
Abstract
Attention and eye movements provide a window into the selective processing of visual information. Evidence suggests that selection is influenced by various factors and is not always under the strategic control of the observer. The aims of this tutorial review are to give a brief introduction to eye movements and attention and to outline the conditions that help determine control. Evidence suggests that the ability to establish control depends on the complexity of the display as well as the point in time at which selection occurs. Stimulus-driven selection is more probable in simple displays than in complex natural scenes, but it critically depends on the timing of the response: Salience determines selection only when responses are triggered quickly following display presentation, and plays no role in longer-latency responses. The time course of selection is also important for the relationship between attention and eye movements. Specifically, attention and eye movements appear to act independently when oculomotor selection is quick, whereas attentional processes are able to influence oculomotor control when saccades are triggered only later in time. This relationship may also be modulated by whether the eye movement is controlled in a voluntary or an involuntary manner. To conclude, we present evidence that shows that visual control is limited in flexibility and that the mechanisms of selection are constrained by context and time. The outcome of visual selection changes with the situational context, and knowing the constraints of control is necessary to understanding when and how visual selection is truly controlled by the observer.
Collapse
|
13
|
Salient object changes influence overt attentional prioritization and object-based targeting in natural scenes. PLoS One 2017; 12:e0172132. [PMID: 28222190 PMCID: PMC5319693 DOI: 10.1371/journal.pone.0172132] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2016] [Accepted: 01/31/2017] [Indexed: 11/28/2022] Open
Abstract
A change to an object in natural scenes attracts attention when it occurs during a fixation. However, when a change occurs during a saccade, and is masked by saccadic suppression, it typically does not capture the gaze in a bottom-up manner. In the present work, we investigated how the type and direction of salient changes to objects affect the prioritization and targeting of objects in natural scenes. We asked observers to look around a scene in preparation for a later memory test. After a period of time, an object in the scene was increased or decreased in salience either during a fixation (with a transient signal) or during a saccade (without transient signal), or it was not changed at all. Changes that were made during a fixation attracted the eyes both when the change involved an increase and a decrease in salience. However, changes that were made during a saccade only captured the eyes when the change was an increase in salience, relative to the baseline no-change condition. These results suggest that the prioritization of object changes can be influenced by the underlying salience of the changed object. In addition, object changes that occurred with a transient signal (which is itself a salient signal) resulted in more central object targeting. Taken together, our results suggest that salient signals in a natural scene are an important component in both object prioritization and targeting in natural scene viewing, insofar as they align with object locations.
Collapse
|
14
|
Shen K, Bezgin G, Selvam R, McIntosh AR, Ryan JD. An Anatomical Interface between Memory and Oculomotor Systems. J Cogn Neurosci 2016; 28:1772-1783. [DOI: 10.1162/jocn_a_01007] [Citation(s) in RCA: 43] [Impact Index Per Article: 5.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/15/2022]
Abstract
Abstract
Visual behavior is guided by memories from prior experience and knowledge of the visual scene. The hippocampal system (HC), in particular, has been implicated in the guidance of saccades: Amnesic patients, following damage to the HC, exhibit selective deficits in their gaze patterns. However, the neural circuitry by which mnemonic representations influence the oculomotor system remains unknown. We used a data-driven, network-based approach on directed anatomical connectivity from the macaque brain to reveal an extensive set of polysnaptic pathways spanning the extrastriate, posterior parietal and prefrontal cortices that potentially mediate the exchange of information between the memory and visuo-oculomotor systems. We additionally show how the potential for directed information flow from the hippocampus to oculomotor control areas is exceptionally high. In particular, the dorsolateral pFC and FEF—regions known to be responsible for the cognitive control of saccades—are topologically well positioned to receive information from the hippocampus. Together with neuropsychological evidence of altered gaze patterns following damage to the hippocampus, our findings suggest that a reconsideration of hippocampal involvement in oculomotor guidance is needed.
Collapse
|
15
|
LaPointe MRP, Milliken B. Semantically incongruent objects attract eye gaze when viewing scenes for change. VISUAL COGNITION 2016. [DOI: 10.1080/13506285.2016.1185070] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
16
|
Arexis M, Maquestiaux F, Gaspelin N, Ruthruff E, Didierjean A. Attentional capture in driving displays. Br J Psychol 2016; 108:259-275. [PMID: 28369841 DOI: 10.1111/bjop.12197] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2015] [Revised: 02/22/2016] [Accepted: 03/07/2016] [Indexed: 11/27/2022]
Abstract
Drivers face frequent distraction on the roadways, but little is known about situations placing them at risk of misallocating visual attention. To investigate this issue, we asked participants to search for a red target embedded within simulated driving scenes (photographs taken from inside a car) in three experiments. Distraction was induced by presenting, via a GPS unit, red or green distractors positioned in an irrelevant location at which the target never appeared. If the salient distractor captures attention, visual search should be slower on distractor-present trials than distractor-absent trials. In Experiment 1, salient distractors yielded no such capture effect. In Experiment 2, we decreased the frequency of the salient distractor from 50% of trials to only 10% or 20% of trials. Capture effects were almost five times larger for the 10% occurrence group than for the 20% occurrence group. In Experiment 3, the amount of available central resources was manipulated by asking participants to either simultaneously monitor or ignore a stream of spoken digits. Capture effects were much larger for the dual-task group than for the single-task group. In summary, these findings identify risk factors for attentional capture in real-world driving scenes: distractor rarity and diversion of attention.
Collapse
Affiliation(s)
- Mahé Arexis
- Université de Franche-Comté, Besançon, France
| | | | | | - Eric Ruthruff
- University of New Mexico, Albuquerque, New Mexico, USA
| | - André Didierjean
- Université de Franche-Comté, Besançon, France.,Institut Universitaire de France, Paris, France
| |
Collapse
|
17
|
D’Mello SK. Giving Eyesight to the Blind: Towards Attention-Aware AIED. INTERNATIONAL JOURNAL OF ARTIFICIAL INTELLIGENCE IN EDUCATION 2016. [DOI: 10.1007/s40593-016-0104-1] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
18
|
The effects of saccade-contingent changes on oculomotor capture: salience is important even beyond the first oculomotor response. Atten Percept Psychophys 2014; 76:1803-14. [PMID: 24874259 DOI: 10.3758/s13414-014-0688-1] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Whenever a novel scene is presented, visual salience merely plays a transient role in oculomotor selection. Unique stimulus properties, such as a distinct and, thereby, salient color, affect the oculomotor response only when observers react relatively quickly. For slower responses, or for consecutive ones, salience-driven effects appear completely absent. To date, however, the circumstances that may reinstate the effects of salience over multiple eye movements are still unclear. Recent research shows that changes to a scene can attract gaze, even when these changes occur without a transient signal (i.e., during an eye movement). The aim of the present study was to investigate whether this capture is mediated through salience-driven or memory-guided processes. In three experiments, we examined how the nature of a change in salience that occurred during an eye movement affected consecutive saccades. The results demonstrate that the oculomotor system is exclusively susceptible to increases in salience from one fixation to the next, but only when these increases result in a uniquely high salience level. This suggests that even in the case of a saccade-contingent change, oculomotor selection behavior can be affected by salience-driven mechanisms, possibly to allow the automatic detection of uniquely distinct objects at any moment. The results and implications will be discussed in relation to current views on visual selection.
Collapse
|
19
|
Eye movements, visual search and scene memory, in an immersive virtual environment. PLoS One 2014; 9:e94362. [PMID: 24759905 PMCID: PMC3997357 DOI: 10.1371/journal.pone.0094362] [Citation(s) in RCA: 39] [Impact Index Per Article: 3.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2014] [Accepted: 03/13/2014] [Indexed: 12/02/2022] Open
Abstract
Visual memory has been demonstrated to play a role in both visual search and attentional prioritization in natural scenes. However, it has been studied predominantly in experimental paradigms using multiple two-dimensional images. Natural experience, however, entails prolonged immersion in a limited number of three-dimensional environments. The goal of the present experiment was to recreate circumstances comparable to natural visual experience in order to evaluate the role of scene memory in guiding eye movements in a natural environment. Subjects performed a continuous visual-search task within an immersive virtual-reality environment over three days. We found that, similar to two-dimensional contexts, viewers rapidly learn the location of objects in the environment over time, and use spatial memory to guide search. Incidental fixations did not provide obvious benefit to subsequent search, suggesting that semantic contextual cues may often be just as efficient, or that many incidentally fixated items are not held in memory in the absence of a specific task. On the third day of the experience in the environment, previous search items changed in color. These items were fixated upon with increased probability relative to control objects, suggesting that memory-guided prioritization (or Surprise) may be a robust mechanisms for attracting gaze to novel features of natural environments, in addition to task factors and simple spatial saliency.
Collapse
|
20
|
Ren Z, Gao S, Chia LT, Rajan D. Regularized feature reconstruction for spatio-temporal saliency detection. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2013; 22:3120-3132. [PMID: 23743773 DOI: 10.1109/tip.2013.2259837] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
Multimedia applications such as image or video retrieval, copy detection, and so forth can benefit from saliency detection, which is essentially a method to identify areas in images and videos that capture the attention of the human visual system. In this paper, we propose a new spatio-temporal saliency detection framework on the basis of regularized feature reconstruction. Specifically, for video saliency detection, both the temporal and spatial saliency detection are considered. For temporal saliency, we model the movement of the target patch as a reconstruction process using the patches in neighboring frames. A Laplacian smoothing term is introduced to model the coherent motion trajectories. With psychological findings that abrupt stimulus could cause a rapid and involuntary deployment of attention, our temporal model combines the reconstruction error, regularizer, and local trajectory contrast to measure the temporal saliency. For spatial saliency, a similar sparse reconstruction process is adopted to capture the regions with high center-surround contrast. Finally, the temporal saliency and spatial saliency are combined together to favor salient regions with high confidence for video saliency detection. We also apply the spatial saliency part of the spatio-temporal model to image saliency detection. Experimental results on a human fixation video dataset and an image saliency detection dataset show that our method achieves the best performance over several state-of-the-art approaches.
Collapse
Affiliation(s)
- Zhixiang Ren
- Centre for Multimedia and Network Technology, School of Computer Engineering, Nanyang Technological University, 639798, Singapore.
| | | | | | | |
Collapse
|
21
|
LaPointe MRP, Lupianez J, Milliken B. Context congruency effects in change detection: Opposing effects on detection and identification. VISUAL COGNITION 2013. [DOI: 10.1080/13506285.2013.787133] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
22
|
Brockmole JR, Davoli CC, Cronin DA. The Visual World in Sight and Mind. PSYCHOLOGY OF LEARNING AND MOTIVATION 2012. [DOI: 10.1016/b978-0-12-394293-7.00003-0] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
23
|
|
24
|
Matsukura M, Brockmole JR, Boot WR, Henderson JM. Oculomotor capture during real-world scene viewing depends on cognitive load. Vision Res 2011; 51:546-52. [PMID: 21310171 DOI: 10.1016/j.visres.2011.01.014] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2010] [Revised: 01/20/2011] [Accepted: 01/28/2011] [Indexed: 11/24/2022]
Abstract
It has been claimed that gaze control during scene viewing is largely governed by stimulus-driven, bottom-up selection mechanisms. Recent research, however, has strongly suggested that observers' top-down control plays a dominant role in attentional prioritization in scenes. A notable exception to this strong top-down control is oculomotor capture, where visual transients in a scene draw the eyes. One way to test whether oculomotor capture during scene viewing is independent of an observer's top-down goal setting is to reduce observers' cognitive resource availability. In the present study, we examined whether increasing observers' cognitive load influences the frequency and speed of oculomotor capture during scene viewing. In Experiment 1, we tested whether increasing observers' cognitive load modulates the degree of oculomotor capture by a new object suddenly appeared in a scene. Similarly, in Experiment 2, we tested whether increasing observers' cognitive load modulates the degree of oculomotor capture by an object's color change. In both experiments, the degree of oculomotor capture decreased as observers' cognitive resources were reduced. These results suggest that oculomotor capture during scene viewing is dependent on observers' top-down selection mechanisms.
Collapse
Affiliation(s)
- Michi Matsukura
- University of Iowa, Department of Psychology, 11 Seashore Hall E, Iowa City, IA 52242, USA.
| | | | | | | |
Collapse
|
25
|
DynAOI: a tool for matching eye-movement data with dynamic areas of interest in animations and movies. Behav Res Methods 2010; 42:179-87. [PMID: 20160298 DOI: 10.3758/brm.42.1.179] [Citation(s) in RCA: 35] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Analyzing gaze behavior with dynamic stimulus material is of growing importance in experimental psychology; however, there is still a lack of efficient analysis tools that are able to handle dynamically changing areas of interest. In this article, we present DynAOI, an open-source tool that allows for the definition of dynamic areas of interest. It works automatically with animations that are based on virtual three-dimensional models. When one is working with videos of real-world scenes, a three-dimensional model of the relevant content needs to be created first. The recorded eye-movement data are matched with the static and dynamic objects in the model underlying the video content, thus creating static and dynamic areas of interest. A validation study asking participants to track particular objects demonstrated that DynAOI is an efficient tool for handling dynamic areas of interest.
Collapse
|
26
|
Varakin DA, Loschky L. Object appearance and picture-specific viewpoint are not integrated in long-term memory. Q J Exp Psychol (Hove) 2009; 63:1181-200. [PMID: 19830629 DOI: 10.1080/17470210903254639] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
Abstract
Previous work has demonstrated that visual long-term memory (VLTM) stores detailed information about object appearance. The current experiments investigate whether object appearance information in VLTM is integrated within representations that contain picture-specific viewpoint information. In three experiments using both incidental and intentional encoding instructions, participants were unable to perform above chance on recognition tests that required recognizing the conjunction of object appearance and viewpoint information (Experiments 1a, 1b, 2, and 3). However, performance was better when object appearance information (Experiments 1a, 1b, and 2) or picture-specific viewpoint information (Experiment 3) alone was sufficient to succeed on the memory test. These results replicate previous work demonstrating good memory for object appearance and viewpoint. However the current results suggest that object appearance and viewpoint are not episodically integrated in VLTM.
Collapse
|
27
|
Rayner K. The 35th Sir Frederick Bartlett Lecture: Eye movements and attention in reading, scene perception, and visual search. Q J Exp Psychol (Hove) 2009; 62:1457-506. [PMID: 19449261 DOI: 10.1080/17470210902816461] [Citation(s) in RCA: 975] [Impact Index Per Article: 65.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
Abstract
Eye movements are now widely used to investigate cognitive processes during reading, scene perception, and visual search. In this article, research on the following topics is reviewed with respect to reading: (a) the perceptual span (or span of effective vision), (b) preview benefit, (c) eye movement control, and (d) models of eye movements. Related issues with respect to eye movements during scene perception and visual search are also reviewed. It is argued that research on eye movements during reading has been somewhat advanced over research on eye movements in scene perception and visual search and that some of the paradigms developed to study reading should be more widely adopted in the study of scene perception and visual search. Research dealing with “real-world” tasks and research utilizing the visual-world paradigm are also briefly discussed.
Collapse
|
28
|
Matsukura M, Brockmole JR, Henderson JM. Overt attentional prioritization of new objects and feature changes during real-world scene viewing. VISUAL COGNITION 2009. [DOI: 10.1080/13506280902868660] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
29
|
Abstract
The abrupt appearance of a new object captures attention, even when the object is task irrelevant. These findings suggest that abrupt onsets capture attention in a stimulus-driven manner and are not susceptible to top-down influences on attentional control. However, previous studies examining the ability of abrupt onsets to capture attention have used search displays that lacked significant complexity. Because attention is a limited capacity mechanism, it is possible that increasing the complexity, or perceptual load, of the search arrays may modulate capture by abrupt onsets. We used a flanker task to examine the effect of perceptual load on attentional capture by abruptly appearing objects. Subjects searched for a target letter through low-load (set size = 1) and high-load (set size = 6) displays. On each trial, irrelevant flankers also appeared, one as an onset and the other as an offset. Onset flankers affected search in low-load but not high-load displays. This modulation of attentional capture was not caused by generalized slowing when subjects searched through high-load displays; search for a single perceptually degraded target slowed response times but did not affect attentional capture. These findings demonstrate that attentional capture by an abrupt onset is attenuated when people search through high-load scenes.
Collapse
|
30
|
|
31
|
|