1
|
Fabian T. Exploring power-law behavior in human gaze shifts across tasks and populations. Cognition 2025; 257:106079. [PMID: 39904005 DOI: 10.1016/j.cognition.2025.106079] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2024] [Revised: 01/18/2025] [Accepted: 01/30/2025] [Indexed: 02/06/2025]
Abstract
Visual perception is an integral part of human cognition. Vision comprises sampling information and processing them. Tasks and stimuli influence human sampling behavior, while cognitive and neurological processing mechanisms remain unchanged. A question still controversial today is whether the components interact with each other. Some theories see the components of visual cognition as separate and their influence on gaze behavior as additive. Others see gaze behavior as an emergent structure of visual cognition that emerges through multiplicative interactions. One way to approach this problem is to examine the magnitude of gaze shifts. Demonstrating that gaze shifts show a constant behavior across tasks would argue for the existence of an independent component in human visual behavior. However, studies attempting to generally describe gaze shift magnitudes deliver contradictory results. In this work, we analyze data from numerous experiments to advance the debate on visual cognition by providing a more comprehensive view of visual behavior. The data show that the magnitude of eye movements, also called saccades, cannot be described by a consistent distribution across different experiments. However, we also propose a new way of measuring the magnitude of saccades: relative saccade lengths. We find that a saccade's length relative to the preceding saccade's length consistently follows a power-law distribution. We observe this distribution for all datasets we analyze, regardless of the task, stimulus, age, or native language of the participants. Our results indicate the existence of an independent component utilized by other cognitive processes without interacting with them. This suggests that a part of human visual cognition is based on an additive component that does not depend on stimulus features.
Collapse
Affiliation(s)
- Thomas Fabian
- Department of History and Social Sciences, Technical University Darmstadt, Darmstadt, Germany.
| |
Collapse
|
2
|
Korteland RJ, Kok E, Hulshof C, van Gog T. Teaching through their eyes: effects on optometry teachers' adaptivity and students' learning when teachers see students' gaze. ADVANCES IN HEALTH SCIENCES EDUCATION : THEORY AND PRACTICE 2024; 29:1735-1748. [PMID: 38598135 PMCID: PMC11549187 DOI: 10.1007/s10459-024-10325-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/06/2023] [Accepted: 03/17/2024] [Indexed: 04/11/2024]
Abstract
Adaptive teacher support fosters effective learning in one-to-one teaching sessions, which are a common way of learning complex visual tasks in the health sciences. Adaptive support is tailored to student needs, and this is difficult in complex visual tasks as visual problem-solving processes are covert and thus cannot be directly observed by the teacher. Eye-tracking apparatus can measure covert processes and make them visible in gaze displays: visualizations of where a student looks while executing a task. We investigate whether live dynamic gaze displays help teachers in being more adaptive to students' needs when teaching optical coherence tomography interpretation in one-to-one teaching sessions and whether this fosters learning. Forty-nine students and 10 teachers participated in a one-to-one teaching session in clinical optometry. In the control condition, teachers saw the learning task of the student and could discuss it with them, whereas in the gaze-display condition, teachers could additionally see where the student looked. After the 15-minute teaching session, a test was administered to examine achievement. Furthermore, students filled in the 'questionnaire on teacher support adaptivity', and teachers rated how adaptive their support was. Bayesian analyses provide some initial evidence that students did not experience support to be more adaptive in the gaze-display condition versus the control condition, nor were their post-test scores higher. Teachers rated their provided support as being more adaptive in the gaze-display versus the control condition. Further research could investigate if live dynamic gaze displays impact adaptive teaching when used over longer periods or with more teacher training.
Collapse
Affiliation(s)
- Robert-Jan Korteland
- Department of Education, Utrecht University, P.O. Box 80140, Utrecht, 3508 CS, Netherlands
| | - Ellen Kok
- Department of Education, Utrecht University, P.O. Box 80140, Utrecht, 3508 CS, Netherlands.
| | - Casper Hulshof
- Department of Education, Utrecht University, P.O. Box 80140, Utrecht, 3508 CS, Netherlands
| | - Tamara van Gog
- Department of Education, Utrecht University, P.O. Box 80140, Utrecht, 3508 CS, Netherlands
| |
Collapse
|
3
|
Quarmley M, Zelinsky G, Athar S, Yang Z, Drucker JH, Samaras D, Jarcho JM. Nonverbal behavioral patterns predict social rejection elicited aggression. Biol Psychol 2023; 183:108670. [PMID: 37652178 PMCID: PMC10591947 DOI: 10.1016/j.biopsycho.2023.108670] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2023] [Revised: 08/02/2023] [Accepted: 08/28/2023] [Indexed: 09/02/2023]
Abstract
Aggression elicited by social rejection is costly, prevalent, and often lethal. Attempts to predict rejection-elicited aggression using trait-based data have had little success. This may be because in-the-moment aggression is a complex process influenced by current states of attention, arousal, and affect which are poorly predicted by trait-level characteristics. In a study of young adults (N = 89; 18-25 years), machine learning tested the extent to which nonverbal behavioral indices of attention (eye gaze), arousal (pupillary reactivity), and affect (facial expressions) during a novel social interaction paradigm predicted subsequent aggression towards rejecting and accepting peers. Eye gaze and pupillary reactivity predicted aggressive behavior; predictions were more successful than measures of trait-based aggression and harsh parenting. These preliminary results suggest that nonverbal behavior may elucidate underlying mechanisms of in-the-moment aggression.
Collapse
Affiliation(s)
- M Quarmley
- Department of Psychology, Temple University, Philadelphia, PA, United States
| | - G Zelinsky
- Department of Psychology, Stony Brook University, Stony Brook, NY, United States
| | - S Athar
- Department of Computer Science, Stony Brook University, Stony Brook, NY, United States
| | - Z Yang
- Department of Computer Science, Stony Brook University, Stony Brook, NY, United States
| | | | - D Samaras
- Department of Computer Science, Stony Brook University, Stony Brook, NY, United States
| | - J M Jarcho
- Department of Psychology, Temple University, Philadelphia, PA, United States.
| |
Collapse
|
4
|
Wada Y. Consistency and stability of gaze behavior when reading manga. APPLIED COGNITIVE PSYCHOLOGY 2023. [DOI: 10.1002/acp.4059] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/03/2023]
|
5
|
Gautier J, El Haj M. Eyes don't lie: Eye movements differ during covert and overt autobiographical recall. Cognition 2023; 235:105416. [PMID: 36821995 DOI: 10.1016/j.cognition.2023.105416] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2022] [Revised: 02/14/2023] [Accepted: 02/15/2023] [Indexed: 02/24/2023]
Abstract
In everyday life, autobiographical memories are revisited silently (i.e., covert recall) or shared with others (i.e., overt recall), yet most research regarding eye movements and autobiographical recall has focused on overt recall. With that in mind, the aim of the current study was to evaluate eye movements during the retrieval of autobiographical memories (with a focus on emotion), recollected during covert and overt recall. Forty-three participants recalled personal memories out loud and silently, while wearing eye-tracking glasses, and rated these memories in terms of mental imagery and emotional intensity. Analyses showed fewer and longer fixations, fewer and shorter saccades, and fewer blinks during covert recall compared with overt recall. Participants perceived more mental images and had a more intense emotional experience during covert recall. These results are discussed considering cognitive load theories and the various functions of autobiographical recall. We theorize that fewer and longer fixations during covert recall may be due to more intense mental imagery. This study enriches the field of research on eye movements and autobiographical memory by addressing how we retrieve memories silently, a common activity of everyday life. More broadly, our results contribute to building objective tools to measure autobiographical memory, alongside already existing subjective scales.
Collapse
Affiliation(s)
- Joanna Gautier
- Nantes Université, Univ Angers, Laboratoire de Psychologie des Pays de la Loire (LPPL - EA 4638), Chemin de la Censive du Tertre, F44000 Nantes, France.
| | - Mohamad El Haj
- Nantes Université, Univ Angers, Laboratoire de Psychologie des Pays de la Loire (LPPL - EA 4638), Chemin de la Censive du Tertre, F44000 Nantes, France; CHU Nantes, Clinical Gerontology Department, Bd Jacques Monod, F44300, Nantes, France; Institut Universitaire de France, Paris, France
| |
Collapse
|
6
|
Kok EM, Jarodzka H, Sibbald M, van Gog T. Did You Get That? Predicting Learners' Comprehension of a Video Lecture from Visualizations of Their Gaze Data. Cogn Sci 2023; 47:e13247. [PMID: 36744751 PMCID: PMC10078589 DOI: 10.1111/cogs.13247] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2021] [Revised: 12/20/2022] [Accepted: 01/04/2023] [Indexed: 02/07/2023]
Abstract
In online lectures, unlike in face-to-face lectures, teachers lack access to (nonverbal) cues to check if their students are still "with them" and comprehend the lecture. The increasing availability of low-cost eye-trackers provides a promising solution. These devices measure unobtrusively where students look and can visualize these data to teachers. These visualizations might inform teachers about students' level of "with-me-ness" (i.e., do students look at the information that the teacher is currently talking about) and comprehension of the lecture, provided that (1) gaze measures of "with-me-ness" are related to comprehension, (2) people not trained in eye-tracking can predict students' comprehension from gaze visualizations, (3) we understand how different visualization techniques impact this prediction. We addressed these issues in two studies. In Study 1, 36 students watched a video lecture while being eye-tracked. The extent to which students looked at relevant information and the extent to which they looked at the same location as the teacher both correlated with students' comprehension (score on an open question) of the lecture. In Study 2, 50 participants watched visualizations of students' gaze (from Study 1), using six visualization techniques (dynamic and static versions of scanpaths, heatmaps, and focus maps) and were asked to predict students' posttest performance and to rate their ease of prediction. We found that people can use gaze visualizations to predict learners' comprehension above chance level, with minor differences between visualization techniques. Further research should investigate if teachers can act on the information provided by gaze visualizations and thereby improve students' learning.
Collapse
Affiliation(s)
- Ellen M Kok
- Department of Education, Utrecht University.,Department of Online Learning and Instruction, Open University of the Netherlands
| | - Halszka Jarodzka
- Department of Online Learning and Instruction, Open University of the Netherlands
| | - Matt Sibbald
- McMaster Education Research, Innovation and Theory (MERIT) Program, Faculty of Health Sciences, McMaster University
| | | |
Collapse
|
7
|
Two-dimensional and three-dimensional multiple object tracking learning performance in adolescent female soccer players: The role of flow experience reflected by heart rate variability. Physiol Behav 2023; 258:114009. [PMID: 36326537 DOI: 10.1016/j.physbeh.2022.114009] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2022] [Revised: 10/21/2022] [Accepted: 10/26/2022] [Indexed: 11/06/2022]
Abstract
Three-dimensional multiple object tracking (3D-MOT) has been used in various fields to mimic real-life tracking, especially in perceptual-cognitive skills training for soccer. Yet, the learning efficiency in 3D-MOT tasks has not been compared with 2D-MOT. Further, whether the advantage can be reflected by heart rate variability (HRV) based on the neurovisceral integration model should also be examined. Therefore, we used both 2D- and 3D-MOT in a brief adaptive task procedure for adolescent female soccer players with HRV measurement. A faster tracking speed threshold of participants was found in the 3D- compared to 2D-MOT, as well as average tracking speed in the last training period of 3D-MOT. Moreover, lower low frequency (LF) components of HRV in the 3D-MOT indicated a flow experience, demonstrating the provision of more attentional resources. Therefore, we observed that adolescent female soccer players demonstrated higher learning efficiency in 3D-MOT tasks in virtual reality (VR) through a higher flow experience. This study examined the learning efficiency between the two MOT tasks in the soccer domain using evidence from HRV and highlighted the utility and applicability of 3D-MOT application.
Collapse
|
8
|
Yang Z, Mondal S, Ahn S, Zelinsky G, Hoai M, Samaras D. Target-absent Human Attention. COMPUTER VISION - ECCV ... : ... EUROPEAN CONFERENCE ON COMPUTER VISION : PROCEEDINGS. EUROPEAN CONFERENCE ON COMPUTER VISION 2022; 13664:52-68. [PMID: 38144433 PMCID: PMC10745181 DOI: 10.1007/978-3-031-19772-7_4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/26/2023]
Abstract
The prediction of human gaze behavior is important for building human-computer interaction systems that can anticipate the user's attention. Computer vision models have been developed to predict the fixations made by people as they search for target objects. But what about when the target is not in the image? Equally important is to know how people search when they cannot find a target, and when they would stop searching. In this paper, we propose a data-driven computational model that addresses the search-termination problem and predicts the scanpath of search fixations made by people searching for targets that do not appear in images. We model visual search as an imitation learning problem and represent the internal knowledge that the viewer acquires through fixations using a novel state representation that we call Foveated Feature Maps (FFMs). FFMs integrate a simulated foveated retina into a pretrained ConvNet that produces an in-network feature pyramid, all with minimal computational overhead. Our method integrates FFMs as the state representation in inverse reinforcement learning. Experimentally, we improve the state of the art in predicting human target-absent search behavior on the COCO-Search18 dataset. Code is available at: https://github.com/cvlab-stonybrook/Target-absent-Human-Attention.
Collapse
Affiliation(s)
- Zhibo Yang
- Stony Brook University, Stony Brook, NY 11794, USA
| | | | - Seoyoung Ahn
- Stony Brook University, Stony Brook, NY 11794, USA
| | | | - Minh Hoai
- Stony Brook University, Stony Brook, NY 11794, USA
| | | |
Collapse
|
9
|
Chakraborty S, Samaras D, Zelinsky GJ. Weighting the factors affecting attention guidance during free viewing and visual search: The unexpected role of object recognition uncertainty. J Vis 2022; 22:13. [PMID: 35323870 PMCID: PMC8963662 DOI: 10.1167/jov.22.4.13] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2021] [Accepted: 02/18/2022] [Indexed: 11/24/2022] Open
Abstract
The factors determining how attention is allocated during visual tasks have been studied for decades, but few studies have attempted to model the weighting of several of these factors within and across tasks to better understand their relative contributions. Here we consider the roles of saliency, center bias, target features, and object recognition uncertainty in predicting the first nine changes in fixation made during free viewing and visual search tasks in the OSIE and COCO-Search18 datasets, respectively. We focus on the latter-most and least familiar of these factors by proposing a new method of quantifying uncertainty in an image, one based on object recognition. We hypothesize that the greater the number of object categories competing for an object proposal, the greater the uncertainty of how that object should be recognized and, hence, the greater the need for attention to resolve this uncertainty. As expected, we found that target features best predicted target-present search, with their dominance obscuring the use of other features. Unexpectedly, we found that target features were only weakly used during target-absent search. We also found that object recognition uncertainty outperformed an unsupervised saliency model in predicting free-viewing fixations, although saliency was slightly more predictive of search. We conclude that uncertainty in object recognition, a measure that is image computable and highly interpretable, is better than bottom-up saliency in predicting attention during free viewing.
Collapse
Affiliation(s)
| | - Dimitris Samaras
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
| | - Gregory J Zelinsky
- Department of Psychology, Stony Brook University, Stony Brook, NY, USA
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
| |
Collapse
|
10
|
Target specificity improves search, but how universal is the benefit? Atten Percept Psychophys 2020; 82:3878-3894. [DOI: 10.3758/s13414-020-02111-1] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
11
|
Inferring task performance and confidence from displays of eye movements. APPLIED COGNITIVE PSYCHOLOGY 2020. [DOI: 10.1002/acp.3721] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
12
|
Deep gaze pooling: Inferring and visually decoding search intents from human gaze fixations. Neurocomputing 2020. [DOI: 10.1016/j.neucom.2020.01.028] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022]
|
13
|
Changing perspectives on goal-directed attention control: The past, present, and future of modeling fixations during visual search. PSYCHOLOGY OF LEARNING AND MOTIVATION 2020. [DOI: 10.1016/bs.plm.2020.08.001] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/28/2022]
|
14
|
Yu CP, Liu H, Samaras D, Zelinsky GJ. Modelling attention control using a convolutional neural network designed after the ventral visual pathway. VISUAL COGNITION 2019. [DOI: 10.1080/13506285.2019.1661927] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
Affiliation(s)
- Chen-Ping Yu
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
- Department of Psychology, Harvard University, Cambridge, MA, USA
| | - Huidong Liu
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
| | - Dimitrios Samaras
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
| | - Gregory J. Zelinsky
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
- Department of Psychology, Stony Brook University, Stony Brook, NY, USA
| |
Collapse
|
15
|
Abstract
Computer classifiers have been successful at classifying various tasks using eye movement statistics. However, the question of human classification of task from eye movements has rarely been studied. Across two experiments, we examined whether humans could classify task based solely on the eye movements of other individuals. In Experiment 1, human classifiers were shown one of three sets of eye movements: Fixations, which were displayed as blue circles, with larger circles meaning longer fixation durations; Scanpaths, which were displayed as yellow arrows; and Videos, in which a neon green dot moved around the screen. There was an additional Scene manipulation in which eye movement properties were displayed either on the original scene where the task (Search, Memory, or Rating) was performed or on a black background in which no scene information was available. Experiment 2 used similar methods but only displayed Fixations and Videos with the same Scene manipulation. The results of both experiments showed successful classification of Search. Interestingly, Search was best classified in the absence of the original scene, particularly in the Fixation condition. Memory also was classified above chance with the strongest classification occurring with Videos in the presence of the scene. Additional analyses on the pattern of correct responses in these two conditions demonstrated which eye movement properties successful classifiers were using. These findings demonstrate conditions under which humans can extract information from eye movement characteristics in addition to providing insight into the relative success/failure of previous computer classifiers.
Collapse
|
16
|
Abstract
How people look at visual information reveals fundamental information about them; their interests and their states of mind. Previous studies showed that scanpath, i.e., the sequence of eye movements made by an observer exploring a visual stimulus, can be used to infer observer-related (e.g., task at hand) and stimuli-related (e.g., image semantic category) information. However, eye movements are complex signals and many of these studies rely on limited gaze descriptors and bespoke datasets. Here, we provide a turnkey method for scanpath modeling and classification. This method relies on variational hidden Markov models (HMMs) and discriminant analysis (DA). HMMs encapsulate the dynamic and individualistic dimensions of gaze behavior, allowing DA to capture systematic patterns diagnostic of a given class of observers and/or stimuli. We test our approach on two very different datasets. Firstly, we use fixations recorded while viewing 800 static natural scene images, and infer an observer-related characteristic: the task at hand. We achieve an average of 55.9% correct classification rate (chance = 33%). We show that correct classification rates positively correlate with the number of salient regions present in the stimuli. Secondly, we use eye positions recorded while viewing 15 conversational videos, and infer a stimulus-related characteristic: the presence or absence of original soundtrack. We achieve an average 81.2% correct classification rate (chance = 50%). HMMs allow to integrate bottom-up, top-down, and oculomotor influences into a single model of gaze behavior. This synergistic approach between behavior and machine learning will open new avenues for simple quantification of gazing behavior. We release SMAC with HMM, a Matlab toolbox freely available to the community under an open-source license agreement.
Collapse
Affiliation(s)
| | - Janet H Hsiao
- Department of Psychology, The University of Hong Kong, Pok Fu Lam, Hong Kong
| | - Antoni B Chan
- Department of Computer Science, City University of Hong Kong, Kowloon Tong, Hong Kong
| |
Collapse
|
17
|
How do targets, nontargets, and scene context influence real-world object detection? Atten Percept Psychophys 2017; 79:2021-2036. [PMID: 28660468 DOI: 10.3758/s13414-017-1359-9] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
18
|
Gallagher-Mitchell T, Simms V, Litchfield D. Learning from where 'eye' remotely look or point: Impact on number line estimation error in adults. Q J Exp Psychol (Hove) 2017; 71:1526-1534. [PMID: 28540753 DOI: 10.1080/17470218.2017.1335335] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
In this article, we present an investigation into the use of visual cues during number line estimation and their influence on cognitive processes for reducing number line estimation error. Participants completed a 0-1000 number line estimation task before and after a brief intervention in which they observed static-visual or dynamic-visual cues (control, anchor, gaze cursor, mouse cursor) and also made estimation marks to test effective number-target estimation. Results indicated that a significant pre-test to post-test reduction in estimation error was present for dynamic-visual cues of modelled eye-gaze and mouse cursor. However, there was no significant performance difference between pre- and post-test for the control or static anchor conditions. Findings are discussed in relation to the extent to which anchor points alone are meaningful in promoting successful segmentation of the number line and whether dynamic cues promote the utility of these locations in reducing error through attentional guidance.
Collapse
Affiliation(s)
- Thomas Gallagher-Mitchell
- 1 Department of Psychology, Edge Hill University, Ormskirk, UK.,2 Department of Psychology, Liverpool Hope University, Liverpool, UK
| | - Victoria Simms
- 3 School of Psychology, Ulster University, Coleraine, UK
| | | |
Collapse
|
19
|
van Wermeskerken M, Litchfield D, van Gog T. What Am I Looking at? Interpreting Dynamic and Static Gaze Displays. Cogn Sci 2017; 42:220-252. [PMID: 28295482 PMCID: PMC5811818 DOI: 10.1111/cogs.12484] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2015] [Revised: 11/28/2016] [Accepted: 01/11/2017] [Indexed: 11/30/2022]
Abstract
Displays of eye movements may convey information about cognitive processes but require interpretation. We investigated whether participants were able to interpret displays of their own or others' eye movements. In Experiments 1 and 2, participants observed an image under three different viewing instructions. Then they were shown static or dynamic gaze displays and had to judge whether it was their own or someone else's eye movements and what instruction was reflected. Participants were capable of recognizing the instruction reflected in their own and someone else's gaze display. Instruction recognition was better for dynamic displays, and only this condition yielded above chance performance in recognizing the display as one's own or another person's (Experiments 1 and 2). Experiment 3 revealed that order information in the gaze displays facilitated instruction recognition when transitions between fixated regions distinguish one viewing instruction from another. Implications of these findings are discussed.
Collapse
Affiliation(s)
- Margot van Wermeskerken
- Department of Education, Utrecht University.,Institute of Psychology, Erasmus University Rotterdam
| | | | - Tamara van Gog
- Department of Education, Utrecht University.,Institute of Psychology, Erasmus University Rotterdam
| |
Collapse
|
20
|
A Model of the Superior Colliculus Predicts Fixation Locations during Scene Viewing and Visual Search. J Neurosci 2016; 37:1453-1467. [PMID: 28039373 DOI: 10.1523/jneurosci.0825-16.2016] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2016] [Revised: 11/21/2016] [Accepted: 12/01/2016] [Indexed: 11/21/2022] Open
Abstract
Modern computational models of attention predict fixations using saliency maps and target maps, which prioritize locations for fixation based on feature contrast and target goals, respectively. But whereas many such models are biologically plausible, none have looked to the oculomotor system for design constraints or parameter specification. Conversely, although most models of saccade programming are tightly coupled to underlying neurophysiology, none have been tested using real-world stimuli and tasks. We combined the strengths of these two approaches in MASC, a model of attention in the superior colliculus (SC) that captures known neurophysiological constraints on saccade programming. We show that MASC predicted the fixation locations of humans freely viewing naturalistic scenes and performing exemplar and categorical search tasks, a breadth achieved by no other existing model. Moreover, it did this as well or better than its more specialized state-of-the-art competitors. MASC's predictive success stems from its inclusion of high-level but core principles of SC organization: an over-representation of foveal information, size-invariant population codes, cascaded population averaging over distorted visual and motor maps, and competition between motor point images for saccade programming, all of which cause further modulation of priority (attention) after projection of saliency and target maps to the SC. Only by incorporating these organizing brain principles into our models can we fully understand the transformation of complex visual information into the saccade programs underlying movements of overt attention. With MASC, a theoretical footing now exists to generate and test computationally explicit predictions of behavioral and neural responses in visually complex real-world contexts.SIGNIFICANCE STATEMENT The superior colliculus (SC) performs a visual-to-motor transformation vital to overt attention, but existing SC models cannot predict saccades to visually complex real-world stimuli. We introduce a brain-inspired SC model that outperforms state-of-the-art image-based competitors in predicting the sequences of fixations made by humans performing a range of everyday tasks (scene viewing and exemplar and categorical search), making clear the value of looking to the brain for model design. This work is significant in that it will drive new research by making computationally explicit predictions of SC neural population activity in response to naturalistic stimuli and tasks. It will also serve as a blueprint for the construction of other brain-inspired models, helping to usher in the next generation of truly intelligent autonomous systems.
Collapse
|
21
|
Martarelli CS, Chiquet S, Laeng B, Mast FW. Using space to represent categories: insights from gaze position. PSYCHOLOGICAL RESEARCH 2016; 81:721-729. [PMID: 27306547 DOI: 10.1007/s00426-016-0781-2] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2015] [Accepted: 06/04/2016] [Indexed: 11/30/2022]
Abstract
We investigated the boundaries among imagery, memory, and perception by measuring gaze during retrieved versus imagined visual information. Eye fixations during recall were bound to the location at which a specific stimulus was encoded. However, eye position information generalized to novel objects of the same category that had not been seen before. For example, encoding an image of a dog in a specific location enhanced the likelihood of looking at the same location during subsequent mental imagery of other mammals. The results suggest that eye movements can also be launched by abstract representations of categories and not exclusively by a single episode or a specific visual exemplar.
Collapse
Affiliation(s)
- Corinna S Martarelli
- Department of Psychology, University of Bern, Fabrikstrasse 8, 3012, Bern, Switzerland. .,Center for Cognition, Learning and Memory, University of Bern, Bern, Switzerland.
| | - Sandra Chiquet
- Department of Psychology, University of Bern, Fabrikstrasse 8, 3012, Bern, Switzerland.,Center for Cognition, Learning and Memory, University of Bern, Bern, Switzerland
| | - Bruno Laeng
- Department of Psychology, University of Oslo, Oslo, Norway
| | - Fred W Mast
- Department of Psychology, University of Bern, Fabrikstrasse 8, 3012, Bern, Switzerland.,Center for Cognition, Learning and Memory, University of Bern, Bern, Switzerland
| |
Collapse
|
22
|
Hout MC, Godwin HJ, Fitzsimmons G, Robbins A, Menneer T, Goldinger SD. Using multidimensional scaling to quantify similarity in visual search and beyond. Atten Percept Psychophys 2016; 78:3-20. [PMID: 26494381 PMCID: PMC5523409 DOI: 10.3758/s13414-015-1010-6] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Visual search is one of the most widely studied topics in vision science, both as an independent topic of interest, and as a tool for studying attention and visual cognition. A wide literature exists that seeks to understand how people find things under varying conditions of difficulty and complexity, and in situations ranging from the mundane (e.g., looking for one's keys) to those with significant societal importance (e.g., baggage or medical screening). A primary determinant of the ease and probability of success during search are the similarity relationships that exist in the search environment, such as the similarity between the background and the target, or the likeness of the non-targets to one another. A sense of similarity is often intuitive, but it is seldom quantified directly. This presents a problem in that similarity relationships are imprecisely specified, limiting the capacity of the researcher to examine adequately their influence. In this article, we present a novel approach to overcoming this problem that combines multi-dimensional scaling (MDS) analyses with behavioral and eye-tracking measurements. We propose a method whereby MDS can be repurposed to successfully quantify the similarity of experimental stimuli, thereby opening up theoretical questions in visual search and attention that cannot currently be addressed. These quantifications, in conjunction with behavioral and oculomotor measures, allow for critical observations about how similarity affects performance, information selection, and information processing. We provide a demonstration and tutorial of the approach, identify documented examples of its use, discuss how complementary computer vision methods could also be adopted, and close with a discussion of potential avenues for future application of this technique.
Collapse
|
23
|
Maxfield JT, Stalder WD, Zelinsky GJ. Effects of target typicality on categorical search. J Vis 2014; 14:1. [PMID: 25274990 PMCID: PMC4181372 DOI: 10.1167/14.12.1] [Citation(s) in RCA: 44] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2013] [Accepted: 07/25/2014] [Indexed: 11/24/2022] Open
Abstract
The role of target typicality in a categorical visual search task was investigated by cueing observers with a target name, followed by a five-item target present/absent search array in which the target images were rated in a pretest to be high, medium, or low in typicality with respect to the basic-level target cue. Contrary to previous work, we found that search guidance was better for high-typicality targets compared to low-typicality targets, as measured by both the proportion of immediate target fixations and the time to fixate the target. Consistent with previous work, we also found an effect of typicality on target verification times, the time between target fixation and the search judgment; as target typicality decreased, verification times increased. To model these typicality effects, we trained Support Vector Machine (SVM) classifiers on the target categories, and tested these on the corresponding specific targets used in the search task. This analysis revealed significant differences in classifier confidence between the high-, medium-, and low-typicality groups, paralleling the behavioral results. Collectively, these findings suggest that target typicality broadly affects both search guidance and verification, and that differences in typicality can be predicted by distance from an SVM classification boundary.
Collapse
Affiliation(s)
| | | | - Gregory J. Zelinsky
- Department of Psychology, Stony Brook University, Stony Brook, NY
- Department of Computer Science, Stony Brook University, Stony Brook, NY
| |
Collapse
|
24
|
Affiliation(s)
- Gregory J Zelinsky
- Departments of Psychology and Computer Science, Stony Brook University Stony Brook, NY, USA ; Center for Interdisciplinary Research (ZiF), University of Bielefeld Bielefeld, Germany
| |
Collapse
|