1
|
Chen Y, Wang J, Miao Z, Jiang Y, Wu X. The Target-Defining Attributes Can Determine the Effects of Attentional Control Settings in Singleton Search Mode. Behav Sci (Basel) 2025; 15:97. [PMID: 39851901 PMCID: PMC11762503 DOI: 10.3390/bs15010097] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2024] [Revised: 01/13/2025] [Accepted: 01/17/2025] [Indexed: 01/26/2025] Open
Abstract
The attentional control settings (ACSs) can help us efficiently select targets in complex real-world environments. Previous research has shown that category-specific ACS demands more attentional resources than feature-specific ACS. However, comparing natural or alphanumeric categories with color features does not distinguish the effects of processing hierarchy and target-defining properties. The present study employed a spatial cueing paradigm to better understand the effects of target-defining properties and search mode on attentional resources in visual search. The target was defined as a combination of shape feature (shape "X") and color category (green in different shades), which generated shape-specific ACS (sACS) and color-specific ACS (cACS). The degrees of shape matching (SM), color matching (CM), and spatial validity between the cue and target were manipulated. Search modes were manipulated by changing the homogeneity of distractors in either shape or color dimensions. Results show a main effect of CM across all four experiments, indicating that category can tune on attentional capture consistently. Importantly, the analysis between four experiments found different interactions across experiments, suggesting that the singleton search mode can reduce the effects of ACS and increase the interactions with other factors. In conclusion, this study suggests that the effects of ACS on attentional capture are determined by both target-defining properties and search mode, rather than processing hierarchy. The results indicate that attentional processes are highly dynamic and context-dependent, requiring a flexible allocation of resources to effectively prioritize relevant information.
Collapse
Affiliation(s)
- Ying Chen
- Department of Psychology, Tianjin University of Technology and Education, Tianjin 300222, China
| | - Junzhe Wang
- Faculty of Psychology, Tianjin Normal University, Tianjin 300387, China
| | - Zhiwei Miao
- Faculty of Psychology, Tianjin Normal University, Tianjin 300387, China
| | - Yunpeng Jiang
- Faculty of Psychology, Tianjin Normal University, Tianjin 300387, China
| | - Xia Wu
- Faculty of Psychology, Tianjin Normal University, Tianjin 300387, China
| |
Collapse
|
2
|
Robbins A, Evdokimov A. Distractor similarity and category variability effects in search. Atten Percept Psychophys 2024; 86:2231-2250. [PMID: 38982007 PMCID: PMC11480196 DOI: 10.3758/s13414-024-02924-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 06/14/2024] [Indexed: 07/11/2024]
Abstract
Categorical search involves looking for objects based on category information from long-term memory. Previous research has shown that search efficiency in categorical search is influenced by target/distractor similarity and category variability (i.e., heterogeneity). However, the interaction between these factors and their impact on different subprocesses of search remains unclear. This study examined the effects of target/distractor similarity and category variability on processes of categorical search. Using multidimensional scaling, we manipulated target/distractor similarity and measured category variability for target categories that participants searched for. Eye-tracking data were collected to examine attentional guidance and target verification. The results demonstrated that the effect of category variability on response times (RTs) was dependent on the level of target/distractor similarity. Specifically, when distractors were highly similar to target categories, there was a negative relation between RTs and variability, with low variability categories producing longer RTs than higher variability categories. Surprisingly, this trend was only present in the eye-tracking measures of target verification but not attentional guidance. Our results suggest that searchers more effectively guide attention to low-variability categories compared to high-variability categories, regardless of the degree of similarity between targets and distractors. However, low category variability interferes with target match decisions when distractors are highly similar to the category, thus the advantage that low category variability provides to searchers is not equal across processes of search.
Collapse
Affiliation(s)
- Arryn Robbins
- Department of Psychology, University of Richmond, 114 UR Drive, Rm 113, Richmond, VA, 27303, USA.
| | - Anatolii Evdokimov
- Department of Psychology, University of Richmond, 114 UR Drive, Rm 113, Richmond, VA, 27303, USA
| |
Collapse
|
3
|
Petilli MA, Rodio FM, Günther F, Marelli M. Visual search and real-image similarity: An empirical assessment through the lens of deep learning. Psychon Bull Rev 2024:10.3758/s13423-024-02583-4. [PMID: 39327401 DOI: 10.3758/s13423-024-02583-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/03/2024] [Indexed: 09/28/2024]
Abstract
The ability to predict how efficiently a person finds an object in the environment is a crucial goal of attention research. Central to this issue are the similarity principles initially proposed by Duncan and Humphreys, which outline how the similarity between target and distractor objects (TD) and between distractor objects themselves (DD) affect search efficiency. However, the search principles lack direct quantitative support from an ecological perspective, being a summary approximation of a wide range of lab-based results poorly generalisable to real-world scenarios. This study exploits deep convolutional neural networks to predict human search efficiency from computational estimates of similarity between objects populating, potentially, any visual scene. Our results provide ecological evidence supporting the similarity principles: search performance continuously varies across tasks and conditions and improves with decreasing TD similarity and increasing DD similarity. Furthermore, our results reveal a crucial dissociation: TD and DD similarities mainly operate at two distinct layers of the network: DD similarity at the intermediate layers of coarse object features and TD similarity at the final layers of complex features used for classification. This suggests that these different similarities exert their major effects at two distinct perceptual levels and demonstrates our methodology's potential to offer insights into the depth of visual processing on which the search relies. By combining computational techniques with visual search principles, this approach aligns with modern trends in other research areas and fulfils longstanding demands for more ecologically valid research in the field of visual search.
Collapse
Affiliation(s)
- Marco A Petilli
- Department of Psychology, University of Milano-Bicocca, Milano, Italy.
| | - Francesca M Rodio
- Institute for Advanced Studies, IUSS, Pavia, Italy
- Department of Brain and Behavioral Sciences, University of Pavia, Pavia, Italy
| | - Fritz Günther
- Department of Psychology, Humboldt University at Berlin, Berlin, Germany
| | - Marco Marelli
- Department of Psychology, University of Milano-Bicocca, Milano, Italy
- NeuroMI, Milan Center for Neuroscience, Milan, Italy
| |
Collapse
|
4
|
Walper D, Bendixen A, Grimm S, Schubö A, Einhäuser W. Attention deployment in natural scenes: Higher-order scene statistics rather than semantics modulate the N2pc component. J Vis 2024; 24:7. [PMID: 38848099 PMCID: PMC11166226 DOI: 10.1167/jov.24.6.7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Accepted: 04/19/2024] [Indexed: 06/13/2024] Open
Abstract
Which properties of a natural scene affect visual search? We consider the alternative hypotheses that low-level statistics, higher-level statistics, semantics, or layout affect search difficulty in natural scenes. Across three experiments (n = 20 each), we used four different backgrounds that preserve distinct scene properties: (a) natural scenes (all experiments); (b) 1/f noise (pink noise, which preserves only low-level statistics and was used in Experiments 1 and 2); (c) textures that preserve low-level and higher-level statistics but not semantics or layout (Experiments 2 and 3); and (d) inverted (upside-down) scenes that preserve statistics and semantics but not layout (Experiment 2). We included "split scenes" that contained different backgrounds left and right of the midline (Experiment 1, natural/noise; Experiment 3, natural/texture). Participants searched for a Gabor patch that occurred at one of six locations (all experiments). Reaction times were faster for targets on noise and slower on inverted images, compared to natural scenes and textures. The N2pc component of the event-related potential, a marker of attentional selection, had a shorter latency and a higher amplitude for targets in noise than for all other backgrounds. The background contralateral to the target had an effect similar to that on the target side: noise led to faster reactions and shorter N2pc latencies than natural scenes, although we observed no difference in N2pc amplitude. There were no interactions between the target side and the non-target side. Together, this shows that-at least when searching simple targets without own semantic content-natural scenes are more effective distractors than noise and that this results from higher-order statistics rather than from semantics or layout.
Collapse
Affiliation(s)
- Daniel Walper
- Physics of Cognition Group, Chemnitz University of Technology, Chemnitz, Germany
| | - Alexandra Bendixen
- Cognitive Systems Lab, Chemnitz University of Technology, Chemnitz, Germany
- https://www.tu-chemnitz.de/physik/SFKS/index.html.en
| | - Sabine Grimm
- Physics of Cognition Group, Chemnitz University of Technology, Chemnitz, Germany
- Cognitive Systems Lab, Chemnitz University of Technology, Chemnitz, Germany
| | - Anna Schubö
- Cognitive Neuroscience of Perception & Action, Philipps University Marburg, Marburg, Germany
- https://www.uni-marburg.de/en/fb04/team-schuboe
| | - Wolfgang Einhäuser
- Physics of Cognition Group, Chemnitz University of Technology, Chemnitz, Germany
- https://www.tu-chemnitz.de/physik/PHKP/index.html.en
| |
Collapse
|
5
|
Alexander RG, Venkatakrishnan A, Chanovas J, Ferguson S, Macknik SL, Martinez-Conde S. Why did Rubens add a parrot to Titian's The Fall of Man? A pictorial manipulation of joint attention. J Vis 2024; 24:1. [PMID: 38558160 PMCID: PMC10996941 DOI: 10.1167/jov.24.4.1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2023] [Accepted: 01/19/2024] [Indexed: 04/04/2024] Open
Abstract
Almost 400 years ago, Rubens copied Titian's The Fall of Man, albeit with important changes. Rubens altered Titian's original composition in numerous ways, including by changing the gaze directions of the depicted characters and adding a striking red parrot to the painting. Here, we quantify the impact of Rubens's choices on the viewer's gaze behavior. We displayed digital copies of Rubens's and Titian's artworks-as well as a version of Rubens's painting with the parrot digitally removed-on a computer screen while recording the eye movements produced by observers during free visual exploration of each image. To assess the effects of Rubens's changes to Titian's composition, we directly compared multiple gaze parameters across the different images. We found that participants gazed at Eve's face more frequently in Rubens's painting than in Titian's. In addition, gaze positions were more tightly focused for the former than for the latter, consistent with different allocations of viewer interest. We also investigated how gaze fixation on Eve's face affected the perceptual visibility of the parrot in Rubens's composition and how the parrot's presence versus its absence impacted gaze dynamics. Taken together, our results demonstrate that Rubens's critical deviations from Titian's painting have powerful effects on viewers' oculomotor behavior.
Collapse
Affiliation(s)
- Robert G Alexander
- Department of Psychology & Counseling, New York Institute of Technology, New York, NY, USA
| | - Ashwin Venkatakrishnan
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Jordi Chanovas
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
- Graduate Program in Neural and Behavioral Science, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Sophie Ferguson
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Stephen L Macknik
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Susana Martinez-Conde
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| |
Collapse
|
6
|
Zhou Z, Geng JJ. Learned associations serve as target proxies during difficult but not easy visual search. Cognition 2024; 242:105648. [PMID: 37897882 DOI: 10.1016/j.cognition.2023.105648] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2023] [Revised: 10/03/2023] [Accepted: 10/12/2023] [Indexed: 10/30/2023]
Abstract
The target template contains information in memory that is used to guide attention during visual search and is typically thought of as containing features of the actual target object. However, when targets are hard to find, it is advantageous to use other information in the visual environment that is predictive of the target's location to help guide attention. The purpose of these studies was to test if newly learned associations between face and scene category images lead observers to use scene information as a proxy for the face target. Our results showed that scene information was used as a proxy for the target to guide attention but only when the target face was difficult to discriminate from the distractor face; when the faces were easy to distinguish, attention was no longer guided by the scene unless the scene was presented earlier. The results suggest that attention is flexibly guided by both target features as well as features of objects that are predictive of the target location. The degree to which each contributes to guiding attention depends on the efficiency with which that information can be used to decode the location of the target in the current moment. The results contribute to the view that attentional guidance is highly flexible in its use of information to rapidly locate the target.
Collapse
Affiliation(s)
- Zhiheng Zhou
- Center for Mind and Brain, University of California, 267 Cousteau Place, Davis, CA 95618, USA.
| | - Joy J Geng
- Center for Mind and Brain, University of California, 267 Cousteau Place, Davis, CA 95618, USA; Department of Psychology, University of California, One Shields Ave, Davis, CA 95616, USA.
| |
Collapse
|
7
|
Miao Z, Wang J, Wang Y, Jiang Y, Chen Y, Wu X. The time course of category-based attentional template pre-activation depends on the category framework. Neuropsychologia 2023; 189:108667. [PMID: 37619937 DOI: 10.1016/j.neuropsychologia.2023.108667] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2023] [Revised: 08/20/2023] [Accepted: 08/21/2023] [Indexed: 08/26/2023]
Abstract
When searching for a target defined by a set of objects, attention can be directed toward task-relevant objects by creating a category-based attentional template (CAT). Previous studies have found that CAT can be activated before the onset of the target. However, the time course of CAT pre-activation and whether the category framework (prototypical or semantic) can modulate it remain unclear. To explore the time course of CAT pre-activation, we employed a rapid serial probe presentation paradigm (RSPP) with event-related potentials (ERPs). To investigate the effect of the category framework on the time course of CAT pre-activation, the target category was defined as the prototypical category (Experiment 1) or the semantic category (Experiment 2). The results showed that the prototype-based CAT was pre-activated 300 ms prior to the target, whereas the semantics-based CAT was pre-activated 1500 ms before the onset of the target. The difference in the time course of pre-activation between the two CAT types indicates that the category framework can modulate the time course of CAT pre-activation. Additionally, during the attentional selection phase, an overall comparison of the target revealed that a larger N2pc was elicited by the prototype-based CAT than by the semantics-based CAT, suggesting that the prototype-based CAT could capture more attention than the semantics-based CAT. The findings on the difference between the two CAT frameworks in the preparatory and attentional selection phases provide more evidence for categorical information in visual search and extend our understanding of the mechanism of categorical attentional selection.
Collapse
Affiliation(s)
- Zhiwei Miao
- Faculty of Psychology, Tianjin Normal University, Tianjin, China; Department of Psychology, Soochow University, Suzhou, Jiangsu, 215000, China
| | - Junzhe Wang
- Faculty of Psychology, Tianjin Normal University, Tianjin, China
| | - Yun Wang
- Faculty of Psychology, Tianjin Normal University, Tianjin, China
| | - Yunpeng Jiang
- Faculty of Psychology, Tianjin Normal University, Tianjin, China; Key Research Base of Humanities and Social Sciences of the Ministry of Education, Academy of Psychology and Behavior, Tianjin, China; Tianjin Social Science Laboratory of Students' Mental Development and Learning, Tianjin, China
| | - Ying Chen
- School of Vocational Education, Tianjin University of Technology and Education, Tianjin, China
| | - Xia Wu
- Faculty of Psychology, Tianjin Normal University, Tianjin, China; Key Research Base of Humanities and Social Sciences of the Ministry of Education, Academy of Psychology and Behavior, Tianjin, China; Tianjin Social Science Laboratory of Students' Mental Development and Learning, Tianjin, China.
| |
Collapse
|
8
|
Sakata C, Ueda Y, Moriguchi Y. Visual memory of a co-actor's target during joint search. PSYCHOLOGICAL RESEARCH 2023; 87:2068-2085. [PMID: 36976364 PMCID: PMC10043510 DOI: 10.1007/s00426-023-01819-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2022] [Accepted: 03/17/2023] [Indexed: 03/29/2023]
Abstract
Studies on joint action show that when two actors turn-takingly attend to each other's target that appears one at a time, a partner's target is accumulated in memory. However, in the real world, actors may not be certain that they attend to the same object because multiple objects often appear simultaneously. In this study, we asked participant pairs to search for different targets in parallel from multiple objects and investigated the memory of a partner's target. We employed the contextual cueing paradigm, in which repetitive search forms associative memory between a target and a configuration of distractors that facilitates search. During the learning phase, exemplars of three target categories (i.e., bird, shoe, and tricycle) were presented among unique objects, and participant pairs searched for them. In Experiment 1, it was followed by a memory test about target exemplars. Consequently, the partner's target was better recognized than the target that nobody searched for. In Experiments 2a and 2b, the memory test was replaced with the transfer phase, where one individual from the pair searched for the category that nobody had searched for while the other individual searched for the category the partner had searched for in the learning phase. The transfer phase did not show search facilitation underpinned by associative memory between the partner's target and distractors. These results suggest that when participant pairs search for different targets in parallel, they accumulate the partner's target in memory but may not form its associative memory with the distractors that facilitates its search.
Collapse
Affiliation(s)
- Chifumi Sakata
- Graduate School of Letters, Kyoto University, Yoshida Hon-Machi, Sakyo-Ku, Kyoto, 606-8501, Japan.
| | - Yoshiyuki Ueda
- Institute for the Future of Human Society, Kyoto University, 46 Yoshida Shimoadachi-Cho, Sakyo-Ku, Kyoto, 606-8501, Japan
| | - Yusuke Moriguchi
- Graduate School of Letters, Kyoto University, Yoshida Hon-Machi, Sakyo-Ku, Kyoto, 606-8501, Japan
| |
Collapse
|
9
|
Clement A, Grégoire L, Anderson BA. Generalisation of value-based attentional priority is category-specific. Q J Exp Psychol (Hove) 2023; 76:2401-2409. [PMID: 36453711 PMCID: PMC10319404 DOI: 10.1177/17470218221144318] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/16/2023]
Abstract
A large body of research suggests that previously reward-associated stimuli can capture attention. Recent evidence also suggests that value-driven attentional biases can occur for a particular category of objects. However, it is unclear how broadly these category-level attentional biases can generalise. In the present study, we examined whether value-driven attentional biases can generalise to new exemplars of a category or semantically related categories using a modified version of the value-driven attentional capture paradigm. In an initial training phase, participants searched for two categories of objects and were rewarded for correctly fixating members of one target category. In a subsequent test phase, participants searched for two new categories of objects. A new exemplar of one of the previous target categories or a member of a semantically related category could appear as a critical distractor in this phase. Participants were more likely to initially fixate the critical distractor and fixated the distractor longer when it was a new exemplar of the previously rewarded category. However, similar findings were not observed for members of semantically related categories. Together, these findings suggest that the generalisation of value-based attentional priority is category-specific.
Collapse
Affiliation(s)
- Andrew Clement
- Department of Psychological and Brain Sciences, Texas A&M University, College Station, TX, USA
| | - Laurent Grégoire
- Department of Psychological and Brain Sciences, Texas A&M University, College Station, TX, USA
| | - Brian A Anderson
- Department of Psychological and Brain Sciences, Texas A&M University, College Station, TX, USA
| |
Collapse
|
10
|
Yu X, Zhou Z, Becker SI, Boettcher SEP, Geng JJ. Good-enough attentional guidance. Trends Cogn Sci 2023; 27:391-403. [PMID: 36841692 DOI: 10.1016/j.tics.2023.01.007] [Citation(s) in RCA: 15] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2022] [Revised: 01/24/2023] [Accepted: 01/25/2023] [Indexed: 02/27/2023]
Abstract
Theories of attention posit that attentional guidance operates on information held in a target template within memory. The template is often thought to contain veridical target features, akin to a photograph, and to guide attention to objects that match the exact target features. However, recent evidence suggests that attentional guidance is highly flexible and often guided by non-veridical features, a subset of features, or only associated features. We integrate these findings and propose that attentional guidance maximizes search efficiency based on a 'good-enough' principle to rapidly localize candidate target objects. Candidates are then serially interrogated to make target-match decisions using more precise information. We suggest that good-enough guidance optimizes the speed-accuracy-effort trade-offs inherent in each stage of visual search.
Collapse
Affiliation(s)
- Xinger Yu
- Center for Mind and Brain, University of California Davis, Davis, CA, USA; Department of Psychology, University of California Davis, Davis, CA, USA
| | - Zhiheng Zhou
- Center for Mind and Brain, University of California Davis, Davis, CA, USA
| | - Stefanie I Becker
- School of Psychology, University of Queensland, Brisbane, QLD, Australia
| | | | - Joy J Geng
- Center for Mind and Brain, University of California Davis, Davis, CA, USA; Department of Psychology, University of California Davis, Davis, CA, USA.
| |
Collapse
|
11
|
How does searching for faces among similar-looking distractors affect distractor memory? Mem Cognit 2023:10.3758/s13421-023-01405-7. [PMID: 36849759 DOI: 10.3758/s13421-023-01405-7] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 02/09/2023] [Indexed: 03/01/2023]
Abstract
Prior research has shown that searching for multiple targets in a visual search task enhances distractor memory in a subsequent recognition test. Three non-mutually exclusive accounts have been offered to explain this phenomenon. The mental comparison hypothesis states that searching for multiple targets requires participants to make more mental comparisons between the targets and the distractors, which enhances distractor memory. The attention allocation hypothesis states that participants allocate more attention to distractors because a multiple-target search cue leads them to expect a more difficult search. Finally, the partial match hypothesis states that searching for multiple targets increases the amount of featural overlap between targets and distractors, which necessitates greater attention in order to reject each distractor. In two experiments, we examined these hypotheses by manipulating visual working memory (VWM) load and target-distractor similarity of AI-generated faces in a visual search (i.e., RSVP) task. Distractor similarity was manipulated using a multidimensional scaling model constructed from facial landmarks and other metadata of each face. In both experiments, distractors from multiple-target searches were recognized better than distractors from single-target searches. Experiment 2 additionally revealed that increased target-distractor similarity during search improved distractor recognition memory, consistent with the partial match hypothesis.
Collapse
|
12
|
DeBolt MC, Mitsven SG, Pomaranski KI, Cantrell LM, Luck SJ, Oakes LM. A new perspective on the role of physical salience in visual search: Graded effect of salience on infants' attention. Dev Psychol 2023; 59:326-343. [PMID: 36355689 PMCID: PMC9905344 DOI: 10.1037/dev0001460] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022]
Abstract
We tested 6- and 8-month-old White and non-White infants (N = 53 total, 28 girls) from Northern California in a visual search task to determine whether a unique item in an otherwise homogeneous display (a singleton) attracts attention because it is a unique singleton and "pops out" in a categorical manner, or whether attention instead varies in a graded manner on the basis of quantitative differences in physical salience. Infants viewed arrays of four or six items; one item was a singleton and the other items were identical distractors (e.g., a single cookie and three identical toy cars). At both ages, infants looked to the singletons first more often, were faster to look at singletons, and looked longer at singletons. However, when a computational model was used to quantify the relative salience of the singleton in each display-which varied widely among the different singleton-distractor combinations-we found a strong, graded effect of physical salience on attention and no evidence that singleton status per se influenced attention. In addition, consistent with other research on attention in infancy, the effect of salience was stronger for 6-month-old infants than for 8-month-old infants. Taken together, these results show that attention-getting and attention-holding in infancy vary continuously with quantitative variations in physical salience rather than depending in a categorical manner on whether an item is unique. (PsycInfo Database Record (c) 2023 APA, all rights reserved).
Collapse
Affiliation(s)
- Michaela C. DeBolt
- Department of Psychology, University of California, Davis
- Center for Mind and Brain, University of California, Davis
| | | | - Katherine I. Pomaranski
- Department of Psychology, University of California, Davis
- Center for Mind and Brain, University of California, Davis
| | - Lisa M. Cantrell
- Department of Child Development, California State University, Sacramento
| | - Steven J. Luck
- Department of Psychology, University of California, Davis
- Center for Mind and Brain, University of California, Davis
| | - Lisa M. Oakes
- Department of Psychology, University of California, Davis
- Center for Mind and Brain, University of California, Davis
| |
Collapse
|
13
|
Adamo SH, Roque N, Barufaldi B, Schmidt J, Mello-Thoms C, Lago M. Assessing satisfaction of search in virtual mammograms for experienced and novice searchers. J Med Imaging (Bellingham) 2023; 10:S11917. [PMID: 37485309 PMCID: PMC10359808 DOI: 10.1117/1.jmi.10.s1.s11917] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2022] [Revised: 06/08/2023] [Accepted: 06/20/2023] [Indexed: 07/25/2023] Open
Abstract
Purpose Satisfaction of search (SOS) is a phenomenon where searchers are more likely to miss a lesion/target after detecting a first lesion/target. Here, we investigated SOS for masses and calcifications in virtual mammograms with experienced and novice searchers to determine the extent to which: (1) SOS affects breast lesion detection, (2) similarity between lesions impacts detection, and (3) experience impacts SOS rates. Approach The open virtual clinical trials framework was used to simulate the breast anatomy of patients, and up to two simulated masses and/or single-calcifications were inserted into the breast models. Experienced searchers (residents, fellows, and radiologists with breast imaging experience) and novice searchers (undergraduates who had no breast imaging experience) were instructed to search for up to two lesions (masses and calcifications) per image. Results 2 × 2 mixed factors analysis of variances (ANOVAs) were run with: (1) single versus second lesion hit rates, (2) similar versus dissimilar second-lesion hit rates, and (3) similar versus dissimilar second-lesion response times as within-subject factors and experience as the between subject's factor. The ANOVAs demonstrated that: (1) experienced and novice searchers made a significant amount of SOS errors, (2) similarity had little impact on experienced searchers, but novice searchers were more likely to miss a dissimilar second lesion compared to when it was similar to a detected first lesion, (3) experienced and novice searchers were faster at finding similar compared to dissimilar second lesions. Conclusions We demonstrated that SOS is a significant cause of lesion misses in virtual mammograms and that reader experience impacts detection rates for similar compared to dissimilar abnormalities. These results suggest that experience may impact strategy and/or recognition with theoretical implications for determining why SOS occurs.
Collapse
Affiliation(s)
| | - Nelson Roque
- University of Central Florida, Orlando, Florida, United States
| | - Bruno Barufaldi
- University of Pennsylvania, Philadelphia, Pennsylvania, United States
| | - Joseph Schmidt
- University of Central Florida, Orlando, Florida, United States
| | | | - Miguel Lago
- U.S. Food and Drug Administration, Silver Spring, Maryland, United States
| |
Collapse
|
14
|
Yang Z, Mondal S, Ahn S, Zelinsky G, Hoai M, Samaras D. Target-absent Human Attention. COMPUTER VISION - ECCV ... : ... EUROPEAN CONFERENCE ON COMPUTER VISION : PROCEEDINGS. EUROPEAN CONFERENCE ON COMPUTER VISION 2022; 13664:52-68. [PMID: 38144433 PMCID: PMC10745181 DOI: 10.1007/978-3-031-19772-7_4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/26/2023]
Abstract
The prediction of human gaze behavior is important for building human-computer interaction systems that can anticipate the user's attention. Computer vision models have been developed to predict the fixations made by people as they search for target objects. But what about when the target is not in the image? Equally important is to know how people search when they cannot find a target, and when they would stop searching. In this paper, we propose a data-driven computational model that addresses the search-termination problem and predicts the scanpath of search fixations made by people searching for targets that do not appear in images. We model visual search as an imitation learning problem and represent the internal knowledge that the viewer acquires through fixations using a novel state representation that we call Foveated Feature Maps (FFMs). FFMs integrate a simulated foveated retina into a pretrained ConvNet that produces an in-network feature pyramid, all with minimal computational overhead. Our method integrates FFMs as the state representation in inverse reinforcement learning. Experimentally, we improve the state of the art in predicting human target-absent search behavior on the COCO-Search18 dataset. Code is available at: https://github.com/cvlab-stonybrook/Target-absent-Human-Attention.
Collapse
Affiliation(s)
- Zhibo Yang
- Stony Brook University, Stony Brook, NY 11794, USA
| | | | - Seoyoung Ahn
- Stony Brook University, Stony Brook, NY 11794, USA
| | | | - Minh Hoai
- Stony Brook University, Stony Brook, NY 11794, USA
| | | |
Collapse
|
15
|
Yeh LC, Peelen MV. The time course of categorical and perceptual similarity effects in visual search. J Exp Psychol Hum Percept Perform 2022; 48:1069-1082. [PMID: 35951407 PMCID: PMC7616436 DOI: 10.1037/xhp0001034] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
During visual search for objects (e.g., an apple), the surrounding distractor objects may share perceptual (tennis ball), categorical (banana), or both (peach) properties with the target. Previous studies showed that the perceptual similarity between target and distractor objects influences visual search. Here, we tested whether categorical target-distractor similarity also influences visual search, and how this influence depends on perceptual similarity. By orthogonally manipulating categorical and perceptual target-distractor similarity, we could investigate how and when the two similarities interactively affect search performance and neural correlates of spatial attention (N2pc) using electroencephalography (EEG). Behavioral results showed that categorical target-distractor similarity interacted with perceptual target-distractor similarity, such that the effect of categorical similarity was strongest when target and distractor objects were perceptually similar. EEG results showed that perceptual similarity influenced the early part of the N2pc (200-250 ms after stimulus onset), while categorical similarity influenced the later part (250-300 ms). Mirroring the behavioral results, categorical similarity interacted with perceptual similarity during this later time window, with categorical effects only observed for perceptually similar target-distractor pairs. Together, these results provide evidence for hierarchical processing in visual search: categorical properties influence spatial attention only when perceptual properties are insufficient to guide attention to the target. (PsycInfo Database Record (c) 2022 APA, all rights reserved).
Collapse
|
16
|
Abassi Abu Rukab S, Khayat N, Hochstein S. High-level visual search in children with autism. J Vis 2022; 22:6. [PMID: 35994261 PMCID: PMC9419456 DOI: 10.1167/jov.22.9.6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2021] [Accepted: 06/12/2022] [Indexed: 11/24/2022] Open
Abstract
Visual search has been classified as easy feature search, with rapid target detection and little set size dependence, versus slower difficult search with focused attention, with set size-dependent speed. Reverse hierarchy theory attributes these classes to rapid high cortical-level vision at a glance versus low-level vision with scrutiny, attributing easy search to high-level representations. Accordingly, faces "pop out" of heterogeneous object photographs. Individuals with autism have difficulties recognizing faces, and we now asked if this disability disturbs their search for faces. We compare search times and set size slopes for children with autism spectrum disorders (ASDs) and those with neurotypical development (NT) when searching for faces. Human face targets were found rapidly, with shallow set size slopes. The between-group difference between slopes (18.8 vs. 11.3 ms/item) is significant, suggesting that faces may not "pop out" as easily, but in our view does not warrant classifying ASD face search as categorically different from that of NT children. We also tested search for different target categories, dog and lion faces, and nonface basic categories, cars and houses. The ASD group was generally a bit slower than the NT group, and their slopes were somewhat steeper. Nevertheless, the overall dependencies on target category were similar: human face search fastest, nonface categories slowest, and dog and lion faces in between. We conclude that autism may spare vision at a glance, including face detection, despite its reported effects on face recognition, which may require vision with scrutiny. This dichotomy is consistent with the two perceptual modes suggested by reverse hierarchy theory.
Collapse
Affiliation(s)
- Safa'a Abassi Abu Rukab
- ELSC Edmond & Lily Safra Center for Brain Research and Silberman Institute for Life Sciences, Hebrew University, Jerusalem, Israel
| | - Noam Khayat
- ELSC Edmond & Lily Safra Center for Brain Research and Silberman Institute for Life Sciences, Hebrew University, Jerusalem, Israel
| | - Shaul Hochstein
- ELSC Edmond & Lily Safra Center for Brain Research and Silberman Institute for Life Sciences, Hebrew University, Jerusalem, Israel
| |
Collapse
|
17
|
Phelps AM, Alexander RG, Schmidt J. Negative cues minimize visual search specificity effects. Vision Res 2022; 196:108030. [PMID: 35313163 PMCID: PMC9090971 DOI: 10.1016/j.visres.2022.108030] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2021] [Revised: 03/02/2022] [Accepted: 03/08/2022] [Indexed: 11/28/2022]
Abstract
Prior target knowledge (i.e., positive cues) improves visual search performance. However, there is considerable debate about whether distractor knowledge (i.e., negative cues) can guide search. Some studies suggest the active suppression of negatively cued search items, while others suggest the initial capture of attention by negatively cued items. Prior work has used pictorial or specific text cues but has not explicitly compared them. We build on that work by comparing positive and negative cues presented pictorially and as categorical text labels using photorealistic objects and eye movement measures. Search displays contained a target (cued on positive trials), a lure from the target category (cued on negative trials), and four categorically-unrelated distractors. Search performance with positive cues resulted in stronger attentional guidance and faster object recognition for pictorial relative to categorical cues (i.e., a pictorial advantage, suggesting specific visual details afforded by pictorial cues improved search). However, in most search performance metrics, negative cues mitigate the pictorial advantage. Given that the negatively cued items captured attention, generated target guidance but mitigated the pictorial advantage, these results are partly consistent with both existing theories. Specific visual details provided in positive cues produce a large pictorial advantage in all measures, whereas specific visual details in negative cues only produce a small pictorial advantage for object recognition but not for attentional guidance. This asymmetry in the pictorial advantage suggests that the down-weighting of specific negatively cued visual features is less efficient than the up-weighting of specific positively cued visual features.
Collapse
Affiliation(s)
- Ashley M Phelps
- Department of Psychology, University of Central Florida, Orlando, FL, USA
| | - Robert G Alexander
- Departments of Ophthalmology, Neurology, and Physiology & Pharmacology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Joseph Schmidt
- Department of Psychology, University of Central Florida, Orlando, FL, USA.
| |
Collapse
|
18
|
Cimminella F, D'Innocenzo G, Sala SD, Iavarone A, Musella C, Coco MI. Preserved Extra-Foveal Processing of Object Semantics in Alzheimer's Disease. J Geriatr Psychiatry Neurol 2022; 35:418-433. [PMID: 34044661 DOI: 10.1177/08919887211016056] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Alzheimer's disease (AD) patients underperform on a range of tasks requiring semantic processing, but it is unclear whether this impairment is due to a generalised loss of semantic knowledge or to issues in accessing and selecting such information from memory. The objective of this eye-tracking visual search study was to determine whether semantic expectancy mechanisms known to support object recognition in healthy adults are preserved in AD patients. Furthermore, as AD patients are often reported to be impaired in accessing information in extra-foveal vision, we investigated whether that was also the case in our study. Twenty AD patients and 20 age-matched controls searched for a target object among an array of distractors presented extra-foveally. The distractors were either semantically related or unrelated to the target (e.g., a car in an array with other vehicles or kitchen items). Results showed that semantically related objects were detected with more difficulty than semantically unrelated objects by both groups, but more markedly by the AD group. Participants looked earlier and for longer at the critical objects when these were semantically unrelated to the distractors. Our findings show that AD patients can process the semantics of objects and access it in extra-foveal vision. This suggests that their impairments in semantic processing may reflect difficulties in accessing semantic information rather than a generalised loss of semantic memory.
Collapse
Affiliation(s)
- Francesco Cimminella
- Human Cognitive Neuroscience, Psychology, University of Edinburgh, Edinburgh, United Kingdom.,Laboratory of Experimental Psychology, Suor Orsola Benincasa University, Naples, Italy
| | | | - Sergio Della Sala
- Human Cognitive Neuroscience, Psychology, University of Edinburgh, Edinburgh, United Kingdom
| | | | - Caterina Musella
- Associazione Italiana Malattia d'Alzheimer (AIMA sezione Campania), Naples, Italy
| | - Moreno I Coco
- Faculdade de Psicologia, Universidade de Lisboa, Lisbon, Portugal.,School of Psychology, The University of East London, London, United Kingdom
| |
Collapse
|
19
|
Kershner AM, Hollingworth A. Real-world object categories and scene contexts conjointly structure statistical learning for the guidance of visual search. Atten Percept Psychophys 2022; 84:1304-1316. [PMID: 35426031 PMCID: PMC9010067 DOI: 10.3758/s13414-022-02475-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 03/12/2022] [Indexed: 12/04/2022]
Abstract
We examined how object categories and scene contexts act in conjunction to structure the acquisition and use of statistical regularities to guide visual search. In an exposure session, participants viewed five object exemplars in each of two colors in each of 42 real-world categories. Objects were presented individually against scene context backgrounds. Exemplars within a category were presented with different contexts as a function of color (e.g., the five red staplers were presented with a classroom scene, and the five blue staplers with an office scene). Participants then completed a visual search task, in which they searched for novel exemplars matching a category label cue among arrays of eight objects superimposed over a scene background. In the context-match condition, the color of the target exemplar was consistent with the color associated with that combination of category and scene context from the exposure phase (e.g., a red stapler in a classroom scene). In the context-mismatch condition, the color of the target was not consistent with that association (e.g., a red stapler in an office scene). In two experiments, search response time was reliably lower in the context-match than in the context-mismatch condition, demonstrating that the learning of category-specific color regularities was itself structured by scene context. The results indicate that categorical templates retrieved from long-term memory are biased toward the properties of recent exemplars and that this learning is organized in a scene-specific manner.
Collapse
Affiliation(s)
- Ariel M Kershner
- Department of Psychological and Brain Sciences, The University of Iowa, Iowa City, IA, 52242, USA.
| | - Andrew Hollingworth
- Department of Psychological and Brain Sciences, The University of Iowa, Iowa City, IA, 52242, USA
| |
Collapse
|
20
|
The detail is in the difficulty: Challenging search facilitates rich incidental object encoding. Mem Cognit 2021; 48:1214-1233. [PMID: 32562249 DOI: 10.3758/s13421-020-01051-3] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
When searching for objects in the environment, observers necessarily encounter other, nontarget, objects. Despite their irrelevance for search, observers often incidentally encode the details of these objects, an effect that is exaggerated as the search task becomes more challenging. Although it is well established that searchers create incidental memories for targets, less is known about the fidelity with which nontargets are remembered. Do observers store richly detailed representations of nontargets, or are these memories characterized by gist-level detail, containing only the information necessary to reject the item as a nontarget? We addressed this question across two experiments in which observers completed multiple-target (one to four potential targets) searches, followed by surprise alternative forced-choice (AFC) recognition tests for all encountered objects. To assess the detail of incidentally stored memories, we used similarity rankings derived from multidimensional scaling to manipulate the perceptual similarity across objects in 4-AFC (Experiment 1a) and 16-AFC (Experiments 1b and 2) tests. Replicating prior work, observers recognized more nontarget objects encountered during challenging, relative to easier, searches. More importantly, AFC results revealed that observers stored more than gist-level detail: When search objects were not recognized, observers systematically chose lures with higher perceptual similarity, reflecting partial encoding of the search object's perceptual features. Further, similarity effects increased with search difficulty, revealing that incidental memories for visual search objects are sharpened when the search task requires greater attentional processing.
Collapse
|
21
|
Dreneva A, Shvarts A, Chumachenko D, Krichevets A. Extrafoveal Processing in Categorical Search for Geometric Shapes: General Tendencies and Individual Variations. Cogn Sci 2021; 45:e13025. [PMID: 34379345 PMCID: PMC8459262 DOI: 10.1111/cogs.13025] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2020] [Revised: 06/10/2021] [Accepted: 06/27/2021] [Indexed: 11/29/2022]
Abstract
The paper addresses the capabilities and limitations of extrafoveal processing during a categorical visual search. Previous research has established that a target could be identified from the very first or without any saccade, suggesting that extrafoveal perception is necessarily involved. However, the limits in complexity defining the processed information are still not clear. We performed four experiments with a gradual increase of stimuli complexity to determine the role of extrafoveal processing in searching for the categorically defined geometric shape. The series of experiments demonstrated a significant role of extrafoveal processing while searching for simple two-dimensional shapes and its gradual decrease in a condition with more complicated three-dimensional shapes. The factors of objects' spatial orientation and distractor homogeneity significantly influenced both reaction time and the number of saccades required to identify a categorically defined target. An analysis of the individual p-value distributions revealed pronounced individual differences in using extrafoveal analysis and allowed examination of the performance of each particular participant. The condition with the forced prohibition of eye movements enabled us to investigate the efficacy of covert attention in the condition with complicated shapes. Our results indicate that both foveal and extrafoveal processing are simultaneously involved during a categorical search, and the specificity of their interaction is determined by the spatial orientation of objects, type of distractors, the prohibition to use overt attention, and individual characteristics of the participants.
Collapse
Affiliation(s)
- Anna Dreneva
- Faculty of PsychologyLomonosov Moscow State University
| | - Anna Shvarts
- Freudenthal InstituteFaculty of ScienceUtrecht University
| | | | | |
Collapse
|
22
|
Alexander RG, Yazdanie F, Waite S, Chaudhry ZA, Kolla S, Macknik SL, Martinez-Conde S. Visual Illusions in Radiology: Untrue Perceptions in Medical Images and Their Implications for Diagnostic Accuracy. Front Neurosci 2021; 15:629469. [PMID: 34177444 PMCID: PMC8226024 DOI: 10.3389/fnins.2021.629469] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2020] [Accepted: 04/19/2021] [Indexed: 11/13/2022] Open
Abstract
Errors in radiologic interpretation are largely the result of failures of perception. This remains true despite the increasing use of computer-aided detection and diagnosis. We surveyed the literature on visual illusions during the viewing of radiologic images. Misperception of anatomical structures is a potential cause of error that can lead to patient harm if disease is seen when none is present. However, visual illusions can also help enhance the ability of radiologists to detect and characterize abnormalities. Indeed, radiologists have learned to exploit certain perceptual biases in diagnostic findings and as training tools. We propose that further detailed study of radiologic illusions would help clarify the mechanisms underlying radiologic performance and provide additional heuristics to improve radiologist training and reduce medical error.
Collapse
Affiliation(s)
- Robert G Alexander
- Department of Ophthalmology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Neurology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Physiology and Pharmacology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Fahd Yazdanie
- Department of Radiology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Stephen Waite
- Department of Radiology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Zeshan A Chaudhry
- Department of Radiology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Srinivas Kolla
- Department of Radiology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Stephen L Macknik
- Department of Ophthalmology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Neurology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Physiology and Pharmacology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Susana Martinez-Conde
- Department of Ophthalmology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Neurology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Physiology and Pharmacology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| |
Collapse
|
23
|
Lavelle M, Alonso D, Luria R, Drew T. Visual working memory load plays limited, to no role in encoding distractor objects during visual search. VISUAL COGNITION 2021. [DOI: 10.1080/13506285.2021.1914256] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
Affiliation(s)
- Mark Lavelle
- Department of Psychology, University of Utah, Salt Lake City, UT, USA
| | - David Alonso
- Department of Psychology, University of Utah, Salt Lake City, UT, USA
| | - Roy Luria
- The School of Psychological Sciences, Tel Aviv University, Tel Aviv, Israel
- Sagol School of Neuroscience, Tel Aviv University, Tel Aviv, Israel
| | - Trafton Drew
- Department of Psychology, University of Utah, Salt Lake City, UT, USA
| |
Collapse
|
24
|
Schienle A, Potthoff J, Schönthaler E, Schlintl C. Disgust-Related Memory Bias in Children and Adults. EVOLUTIONARY PSYCHOLOGY 2021; 19:1474704921996585. [PMID: 33902359 PMCID: PMC10303556 DOI: 10.1177/1474704921996585] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2021] [Accepted: 02/02/2021] [Indexed: 11/17/2022] Open
Abstract
Studies with adults found a memory bias for disgust, such that memory for disgusting stimuli was enhanced compared to neutral and frightening stimuli. We investigated whether this bias is more pronounced in females and whether it is already present in children. Moreover, we analyzed whether the visual exploration of disgust stimuli during encoding is associated with memory retrieval. In a first recognition experiment with intentional learning, 50 adults (mean age; M = 23 years) and 52 children (M = 11 years) were presented with disgusting, frightening, and neutral pictures. Both children and adults showed a better recognition performance for disgusting images compared to the other image categories. Males and females did not differ in their memory performance. In a second free recall experiment with eye-tracking, 50 adults (M = 22 years) viewed images from the categories disgust, fear, and neutral. Disgusting and neutral images were matched for color, complexity, brightness, and contrast. The participants, who were not instructed to remember the stimuli, showed a disgust memory bias as well as shorter fixation durations and longer scan paths for disgusting images compared to neutral images. This "hyperscanning pattern" correlated with the number of correctly recalled disgust images. In conclusion, we found a disgust-related memory bias in both children and adults regardless of sex and independently of the memorization method used (recognition/free recall; intentional/incidental).
Collapse
Affiliation(s)
- Anne Schienle
- Clinical Psychology, University of Graz,
BioTechMed Graz, Austria
| | - Jonas Potthoff
- Clinical Psychology, University of Graz,
BioTechMed Graz, Austria
| | | | - Carina Schlintl
- Clinical Psychology, University of Graz,
BioTechMed Graz, Austria
| |
Collapse
|
25
|
Laxton V, Crundall D, Guest D, Howard CJ. Visual search for drowning swimmers: Investigating the impact of lifeguarding experience. APPLIED COGNITIVE PSYCHOLOGY 2020. [DOI: 10.1002/acp.3756] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Affiliation(s)
- Victoria Laxton
- Department of Psychology Nottingham Trent University Nottingham UK
| | - David Crundall
- Department of Psychology Nottingham Trent University Nottingham UK
| | - Duncan Guest
- Department of Psychology Nottingham Trent University Nottingham UK
| | | |
Collapse
|
26
|
Alexander RG, Waite S, Macknik SL, Martinez-Conde S. What do radiologists look for? Advances and limitations of perceptual learning in radiologic search. J Vis 2020; 20:17. [PMID: 33057623 PMCID: PMC7571277 DOI: 10.1167/jov.20.10.17] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/21/2019] [Accepted: 09/14/2020] [Indexed: 12/31/2022] Open
Abstract
Supported by guidance from training during residency programs, radiologists learn clinically relevant visual features by viewing thousands of medical images. Yet the precise visual features that expert radiologists use in their clinical practice remain unknown. Identifying such features would allow the development of perceptual learning training methods targeted to the optimization of radiology training and the reduction of medical error. Here we review attempts to bridge current gaps in understanding with a focus on computational saliency models that characterize and predict gaze behavior in radiologists. There have been great strides toward the accurate prediction of relevant medical information within images, thereby facilitating the development of novel computer-aided detection and diagnostic tools. In some cases, computational models have achieved equivalent sensitivity to that of radiologists, suggesting that we may be close to identifying the underlying visual representations that radiologists use. However, because the relevant bottom-up features vary across task context and imaging modalities, it will also be necessary to identify relevant top-down factors before perceptual expertise in radiology can be fully understood. Progress along these dimensions will improve the tools available for educating new generations of radiologists, and aid in the detection of medically relevant information, ultimately improving patient health.
Collapse
Affiliation(s)
- Robert G Alexander
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Stephen Waite
- Department of Radiology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Stephen L Macknik
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Susana Martinez-Conde
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| |
Collapse
|
27
|
Robbins A, Hout MC. Typicality guides attention during categorical search, but not universally so. Q J Exp Psychol (Hove) 2020; 73:1977-1999. [PMID: 32519925 DOI: 10.1177/1747021820936472] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
The degree to which an item is rated as being a typical member of its category influences an observer's ability to find that item during word-cued search. However, there are conflicting accounts as to whether or not typicality affects attentional guidance to categorical items, or whether it affects some other aspect of the search process. In this study, we employed word-cued search and eye tracking to disentangle typicality effects on attentional guidance and target verification across differing category cue specificities (i.e., superordinate or basic-level cues), while also varying the degree of similarity between targets and non-targets. We found that typicality influenced attentional guidance when searchers were cued at the superordinate level (e.g., clothing). When cues were provided at the basic level (e.g., pants), typicality did not influence attentional guidance, and only affected target verification when there was featural similarity between targets and non-targets. When a searcher uses a target template comprising features cued at the basic level, therefore, target/non-target similarity produces interference that affects attentional guidance, but we did not find evidence that it also affects target verification.
Collapse
Affiliation(s)
- Arryn Robbins
- Department of Psychological Sciences, Carthage College, Kenosha, WI, USA.,Department of Psychology, New Mexico State University, Las Cruces, NM, USA
| | - Michael C Hout
- Department of Psychology, New Mexico State University, Las Cruces, NM, USA
| |
Collapse
|
28
|
Cimminella F, Sala SD, Coco MI. Extra-foveal Processing of Object Semantics Guides Early Overt Attention During Visual Search. Atten Percept Psychophys 2020; 82:655-670. [PMID: 31792893 PMCID: PMC7246246 DOI: 10.3758/s13414-019-01906-1] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/05/2022]
Abstract
Eye-tracking studies using arrays of objects have demonstrated that some high-level processing of object semantics can occur in extra-foveal vision, but its role on the allocation of early overt attention is still unclear. This eye-tracking visual search study contributes novel findings by examining the role of object-to-object semantic relatedness and visual saliency on search responses and eye-movement behaviour across arrays of increasing size (3, 5, 7). Our data show that a critical object was looked at earlier and for longer when it was semantically unrelated than related to the other objects in the display, both when it was the search target (target-present trials) and when it was a target's semantically related competitor (target-absent trials). Semantic relatedness effects manifested already during the very first fixation after array onset, were consistently found for increasing set sizes, and were independent of low-level visual saliency, which did not play any role. We conclude that object semantics can be extracted early in extra-foveal vision and capture overt attention from the very first fixation. These findings pose a challenge to models of visual attention which assume that overt attention is guided by the visual appearance of stimuli, rather than by their semantics.
Collapse
Affiliation(s)
- Francesco Cimminella
- Human Cognitive Neuroscience, Psychology, University of Edinburgh, Edinburgh, UK.
- Laboratory of Experimental Psychology, Suor Orsola Benincasa University, Naples, Italy.
| | - Sergio Della Sala
- Human Cognitive Neuroscience, Psychology, University of Edinburgh, Edinburgh, UK
| | - Moreno I Coco
- Human Cognitive Neuroscience, Psychology, University of Edinburgh, Edinburgh, UK.
- School of Psychology, The University of East London, London, UK.
- Faculdade de Psicologia, Universidade de Lisboa, Lisbon, Portugal.
| |
Collapse
|
29
|
Changing perspectives on goal-directed attention control: The past, present, and future of modeling fixations during visual search. PSYCHOLOGY OF LEARNING AND MOTIVATION 2020. [DOI: 10.1016/bs.plm.2020.08.001] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/28/2022]
|
30
|
Geng JJ, Witkowski P. Template-to-distractor distinctiveness regulates visual search efficiency. Curr Opin Psychol 2019; 29:119-125. [PMID: 30743200 PMCID: PMC6625942 DOI: 10.1016/j.copsyc.2019.01.003] [Citation(s) in RCA: 44] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2018] [Revised: 12/13/2018] [Accepted: 01/04/2019] [Indexed: 11/18/2022]
Abstract
All models of attention include the concept of an attentional template (or a target or search template). The template is conceptualized as target information held in memory that is used for prioritizing sensory processing and determining if an object matches the target. It is frequently assumed that the template contains a veridical copy of the target. However, we review recent evidence showing that the template encodes a version of the target that is adapted to the current context (e.g. distractors, task, etc.); information held within the template may include only a subset of target features, real world knowledge, pre-existing perceptual biases, or even be a distorted version of the veridical target. We argue that the template contents are customized in order to maximize the ability to prioritize information that distinguishes targets from distractors. We refer to this as template-to-distractor distinctiveness and hypothesize that it contributes to visual search efficiency by exaggerating target-to-distractor dissimilarity.
Collapse
Affiliation(s)
- Joy J Geng
- Center for Mind and Brain, University of California Davis, Davis, CA, 95616, United States; Department of Psychology, University of California Davis, Davis, CA, 95616, United States.
| | - Phillip Witkowski
- Center for Mind and Brain, University of California Davis, Davis, CA, 95616, United States; Department of Psychology, University of California Davis, Davis, CA, 95616, United States
| |
Collapse
|
31
|
Face search in CCTV surveillance. COGNITIVE RESEARCH-PRINCIPLES AND IMPLICATIONS 2019; 4:37. [PMID: 31549263 PMCID: PMC6757089 DOI: 10.1186/s41235-019-0193-0] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/09/2019] [Accepted: 08/14/2019] [Indexed: 11/10/2022]
Abstract
Background We present a series of experiments on visual search in a highly complex environment, security closed-circuit television (CCTV). Using real surveillance footage from a large city transport hub, we ask viewers to search for target individuals. Search targets are presented in a number of ways, using naturally occurring images including their passports and photo ID, social media and custody images/videos. Our aim is to establish general principles for search efficiency within this realistic context. Results Across four studies we find that providing multiple photos of the search target consistently improves performance. Three different photos of the target, taken at different times, give substantial performance improvements by comparison to a single target. By contrast, providing targets in moving videos or with biographical context does not lead to improvements in search accuracy. Conclusions We discuss the multiple-image advantage in relation to a growing understanding of the importance of within-person variability in face recognition. Electronic supplementary material The online version of this article (10.1186/s41235-019-0193-0) contains supplementary material, which is available to authorized users.
Collapse
|
32
|
Alexander RG, Nahvi RJ, Zelinsky GJ. Specifying the precision of guiding features for visual search. J Exp Psychol Hum Percept Perform 2019; 45:1248-1264. [PMID: 31219282 PMCID: PMC6706321 DOI: 10.1037/xhp0000668] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Visual search is the task of finding things with uncertain locations. Despite decades of research, the features that guide visual search remain poorly specified, especially in realistic contexts. This study tested the role of two features-shape and orientation-both in the presence and absence of hue information. We conducted five experiments to describe preview-target mismatch effects, decreases in performance caused by differences between the image of the target as it appears in the preview and as it appears in the actual search display. These mismatch effects provide direct measures of feature importance, with larger performance decrements expected for more important features. Contrary to previous conclusions, our data suggest that shape and orientation only guide visual search when color is not available. By varying the probability of mismatch in each feature dimension, we also show that these patterns of feature guidance do not change with the probability that the previewed feature will be invalid. We conclude that the target representations used to guide visual search are much less precise than previously believed, with participants encoding and using color and little else. (PsycINFO Database Record (c) 2019 APA, all rights reserved).
Collapse
|
33
|
Guevara Pinto JD, Papesh MH. Incidental memory following rapid object processing: The role of attention allocation strategies. J Exp Psychol Hum Percept Perform 2019; 45:1174-1190. [PMID: 31219283 PMCID: PMC7202240 DOI: 10.1037/xhp0000664] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
When observers search for multiple (rather than singular) targets, they are slower and less accurate, yet have better incidental memory for nontarget items encountered during the task (Hout & Goldinger, 2010). One explanation for this may be that observers titrate their attention allocation based on the expected difficulty suggested by search cues. Difficult search cues may implicitly encourage observers to narrow their attention, simultaneously enhancing distractor encoding and hindering peripheral processing. Across three experiments, we manipulated the difficulty of search cues preceding passive visual search for real-world objects, using a Rapid Serial Visual Presentation (RSVP) task to equate item exposure durations. In all experiments, incidental memory was enhanced for distractors encountered while participants monitored for difficult targets. Moreover, in key trials, peripheral shapes appeared at varying eccentricities off center, allowing us to infer the spread and precision of participants' attentional windows. Peripheral item detection and identification decreased when search cues were difficult, even when the peripheral items appeared before targets. These results were not an artifact of sustained vigilance in miss trials, but instead reflect top-down modulation of attention allocation based on task demands. Implications for individual differences are discussed. (PsycINFO Database Record (c) 2019 APA, all rights reserved).
Collapse
|
34
|
Waite S, Grigorian A, Alexander RG, Macknik SL, Carrasco M, Heeger DJ, Martinez-Conde S. Analysis of Perceptual Expertise in Radiology - Current Knowledge and a New Perspective. Front Hum Neurosci 2019; 13:213. [PMID: 31293407 PMCID: PMC6603246 DOI: 10.3389/fnhum.2019.00213] [Citation(s) in RCA: 63] [Impact Index Per Article: 10.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/18/2019] [Accepted: 06/07/2019] [Indexed: 12/14/2022] Open
Abstract
Radiologists rely principally on visual inspection to detect, describe, and classify findings in medical images. As most interpretive errors in radiology are perceptual in nature, understanding the path to radiologic expertise during image analysis is essential to educate future generations of radiologists. We review the perceptual tasks and challenges in radiologic diagnosis, discuss models of radiologic image perception, consider the application of perceptual learning methods in medical training, and suggest a new approach to understanding perceptional expertise. Specific principled enhancements to educational practices in radiology promise to deepen perceptual expertise among radiologists with the goal of improving training and reducing medical error.
Collapse
Affiliation(s)
- Stephen Waite
- Department of Radiology, SUNY Downstate Medical Center, Brooklyn, NY, United States
| | - Arkadij Grigorian
- Department of Radiology, SUNY Downstate Medical Center, Brooklyn, NY, United States
| | - Robert G. Alexander
- Department of Ophthalmology, SUNY Downstate Medical Center, Brooklyn, NY, United States
- Department of Neurology, SUNY Downstate Medical Center, Brooklyn, NY, United States
- Department of Physiology/Pharmacology, SUNY Downstate Medical Center, Brooklyn, NY, United States
| | - Stephen L. Macknik
- Department of Ophthalmology, SUNY Downstate Medical Center, Brooklyn, NY, United States
- Department of Neurology, SUNY Downstate Medical Center, Brooklyn, NY, United States
- Department of Physiology/Pharmacology, SUNY Downstate Medical Center, Brooklyn, NY, United States
| | - Marisa Carrasco
- Department of Psychology and Center for Neural Science, New York University, New York, NY, United States
| | - David J. Heeger
- Department of Psychology and Center for Neural Science, New York University, New York, NY, United States
| | - Susana Martinez-Conde
- Department of Ophthalmology, SUNY Downstate Medical Center, Brooklyn, NY, United States
- Department of Neurology, SUNY Downstate Medical Center, Brooklyn, NY, United States
- Department of Physiology/Pharmacology, SUNY Downstate Medical Center, Brooklyn, NY, United States
| |
Collapse
|
35
|
Baier D, Ansorge U. Contingent capture during search for alphanumerical characters: A case of feature-based capture or of conceptual category membership? Vision Res 2019; 160:43-51. [PMID: 31078664 DOI: 10.1016/j.visres.2019.02.016] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2018] [Revised: 01/15/2019] [Accepted: 02/19/2019] [Indexed: 11/26/2022]
Abstract
To distinguish if search for alphanumerical characters is based on features or on conceptual category membership, we conducted two experiments where we presented upright and inverted characters as cues in a contingent-capture protocol. Here, only cues matching the top-down search template (e.g., a letter cue when searching for target letters) capture attention and lead to validity effects: shorter search times and fewer errors for validly than invalidly cued targets. Top-down nonmatching cues (e.g., a number cue when searching for target letters) do not capture attention. To tell a feature-based explanation from one based on conceptual category membership, we used both upright (canonical) and inverted characters as cues. These cues share the same features, but inverted cues cannot be conceptually categorized as easily as upright cues. Thus, we expected no difference between upright and inverted cues when search is feature-based, whereas inverted cues would elicit no or at least considerably weaker validity effects if search relies on conceptual category membership. Altogether, the results of both experiments (with overlapping and with separate sets of characters for cues and targets) provide evidence for search based on feature representations, as among other things, significant validity effects were found with upright and inverted characters as cues. However, an influence of category membership was also evident, as validity effects of inverted characters were diminished.
Collapse
Affiliation(s)
- Diane Baier
- Department of Basic Psychological Research and Research Methods, Faculty of Psychology, University of Vienna, Austria.
| | - Ulrich Ansorge
- Department of Basic Psychological Research and Research Methods, Faculty of Psychology, University of Vienna, Austria
| |
Collapse
|
36
|
Hättenschwiler N, Merks S, Sterchi Y, Schwaninger A. Traditional Visual Search vs. X-Ray Image Inspection in Students and Professionals: Are the Same Visual-Cognitive Abilities Needed? Front Psychol 2019; 10:525. [PMID: 30984052 PMCID: PMC6437096 DOI: 10.3389/fpsyg.2019.00525] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2018] [Accepted: 02/22/2019] [Indexed: 11/13/2022] Open
Abstract
The act of looking for targets amongst an array of distractors is a cognitive task that has been studied extensively over many decades and has many real-world applications. Research shows that specific visual-cognitive abilities are needed to efficiently and effectively locate a target among distractors. It is, however, not always clear whether the results from traditional, simplified visual search tasks conducted by students will extrapolate to an applied inspection tasks in which professionals search for targets that are more complex, ambiguous, and less salient. More concretely, there are several potential challenges when interpreting traditional visual search results in terms of their implications for the X-ray image inspection task. In this study, we tested whether a theoretical intelligence model with known facets of visual-cognitive abilities (visual processing Gv, short-term memory Gsm, and processing speed Gs) can predict performance in both a traditional visual search task and an X-ray image inspection task in both students and professionals. Results showed that visual search ability as measured with a traditional visual search task is not comparable to an applied X-ray image inspection task. Even though both tasks require aspects of the same visual-cognitive abilities, the overlap between the tasks was small. We concluded that different aspects of visual-cognitive abilities predict performance on the measured tasks. Furthermore, although our tested populations were comparable in terms of performance predictors based on visual-cognitive abilities, professionals outperformed students on an applied X-ray image inspection task. Hence, inferences from our research questions have to be treated with caution, because the comparability of the two populations depends on the task.
Collapse
Affiliation(s)
- Nicole Hättenschwiler
- School of Applied Psychology, University of Applied Sciences and Arts Northwestern Switzerland, Olten, Switzerland
| | - Sarah Merks
- School of Applied Psychology, University of Applied Sciences and Arts Northwestern Switzerland, Olten, Switzerland
| | - Yanik Sterchi
- School of Applied Psychology, University of Applied Sciences and Arts Northwestern Switzerland, Olten, Switzerland
| | - Adrian Schwaninger
- School of Applied Psychology, University of Applied Sciences and Arts Northwestern Switzerland, Olten, Switzerland
| |
Collapse
|
37
|
Friedman GN, Johnson L, Williams ZM. Long-Term Visual Memory and Its Role in Learning Suppression. Front Psychol 2018; 9:1896. [PMID: 30369895 PMCID: PMC6194155 DOI: 10.3389/fpsyg.2018.01896] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2018] [Accepted: 09/18/2018] [Indexed: 11/13/2022] Open
Abstract
Long-term memory is a core aspect of human learning that permits a wide range of skills and behaviors often important for survival. While this core ability has been broadly observed for procedural and declarative memory, whether similar mechanisms subserve basic sensory or perceptual processes remains unclear. Here, we use a visual learning paradigm to show that training humans to search for common visual features in the environment leads to a persistent improvement in performance over consecutive days but, surprisingly, suppresses the subsequent ability to learn similar visual features. This suppression is reversed if the memory is prevented from consolidating, while still permitting the ability to learn multiple visual features simultaneously. These findings reveal a memory mechanism that may enable salient sensory patterns to persist in memory over prolonged durations, but which also functions to prevent false-positive detection by proactively suppressing new learning.
Collapse
Affiliation(s)
- Gabriel N Friedman
- Department of Neurosurgery, Harvard Medical School, Massachusetts General Hospital, Boston, MA, United States
| | - Lance Johnson
- Department of Neurobiology, Harvard University, Cambridge, MA, United States
| | - Ziv M Williams
- Department of Neurosurgery, Harvard Medical School, Massachusetts General Hospital, Boston, MA, United States.,Harvard-MIT Health Sciences and Technology, Boston, MA, United States.,Program in Neuroscience, Harvard Medical School, Harvard University, Boston, MA, United States
| |
Collapse
|
38
|
Qiu W, Gao X, Han B. Eye Fixation Assisted Video Saliency Detection via Total Variation-based Pairwise Interaction. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2018; 27:4724-4739. [PMID: 29993549 DOI: 10.1109/tip.2018.2843680] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
As human visual attention is naturally biased towards foreground objects in a scene, it can be used to extract salient objects in video clips. In this work, we proposed a weakly supervised learning based video saliency detection algorithm utilizing eye fixations information from multiple subjects. Our main idea is to extend eye fixations to saliency regions step by step. First, visual seeds are collected using multiple color space geodesic distance based seed region mapping with filtered and extended eye fixations. This operation helps raw fixation points spread to the most likely salient regions, namely, visual seed regions. Second, in order to seize the essential scene structure from video sequences, we introduce the total variance based pairwise interaction model to learn the potential pairwise relationship between foreground and background within a frame or across video frames. In this vein, visual seed regions eventually grow into salient regions. Compared with previous approaches the generated saliency maps has two most outstanding properties: integrity and purity, which are conductive to segment the foreground and significant to the follow-up tasks. Extensive quantitative and qualitative experiments on various video sequences demonstrate that the proposed method outperforms state-of-theart image and video saliency detection algorithms.
Collapse
|
39
|
Alexander RG, Zelinsky GJ. Occluded information is restored at preview but not during visual search. J Vis 2018; 18:4. [PMID: 30347091 PMCID: PMC6181188 DOI: 10.1167/18.11.4] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2018] [Accepted: 07/26/2018] [Indexed: 11/30/2022] Open
Abstract
Objects often appear with some amount of occlusion. We fill in missing information using local shape features even before attending to those objects-a process called amodal completion. Here we explore the possibility that knowledge about common realistic objects can be used to "restore" missing information even in cases where amodal completion is not expected. We systematically varied whether visual search targets were occluded or not, both at preview and in search displays. Button-press responses were longest when the preview was unoccluded and the target was occluded in the search display. This pattern is consistent with a target-verification process that uses the features visible at preview but does not restore missing information in the search display. However, visual search guidance was weakest whenever the target was occluded in the search display, regardless of whether it was occluded at preview. This pattern suggests that information missing during the preview was restored and used to guide search, thereby resulting in a feature mismatch and poor guidance. If this process were preattentive, as with amodal completion, we should have found roughly equivalent search guidance across all conditions because the target would always be unoccluded or restored, resulting in no mismatch. We conclude that realistic objects are restored behind occluders during search target preview, even in situations not prone to amodal completion, and this restoration does not occur preattentively during search.
Collapse
Affiliation(s)
| | - Gregory J Zelinsky
- Department of Psychology, Stony Brook University, Stony Brook, NY, USA
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
| |
Collapse
|
40
|
Doherty BR, Charman T, Johnson MH, Scerif G, Gliga T. Visual search and autism symptoms: What young children search for and co-occurring ADHD matter. Dev Sci 2018; 21:e12661. [PMID: 29726058 DOI: 10.1111/desc.12661] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2016] [Accepted: 01/09/2018] [Indexed: 11/08/2023]
Abstract
Superior visual search is one of the most common findings in the autism spectrum disorder (ASD) literature. Here, we ascertain how generalizable these findings are across task and participant characteristics, in light of recent replication failures. We tested 106 3-year-old children at familial risk for ASD, a sample that presents high ASD and ADHD symptoms, and 25 control participants, in three multi-target search conditions: easy exemplar search (look for cats amongst artefacts), difficult exemplar search (look for dogs amongst chairs/tables perceptually similar to dogs), and categorical search (look for animals amongst artefacts). Performance was related to dimensional measures of ASD and ADHD, in agreement with current research domain criteria (RDoC). We found that ASD symptom severity did not associate with enhanced performance in search, but did associate with poorer categorical search in particular, consistent with literature describing impairments in categorical knowledge in ASD. Furthermore, ASD and ADHD symptoms were both associated with more disorganized search paths across all conditions. Thus, ASD traits do not always convey an advantage in visual search; on the contrary, ASD traits may be associated with difficulties in search depending upon the nature of the stimuli (e.g., exemplar vs. categorical search) and the presence of co-occurring symptoms.
Collapse
Affiliation(s)
- Brianna R Doherty
- Department of Experimental Psychology, University of Oxford, Oxford, UK
| | - Tony Charman
- Institute of Psychiatry, Psychology& Neuroscience, King's College London, London, UK
| | - Mark H Johnson
- Centre for Brain and Cognitive Development, Birkbeck College, University of London, London, UK
| | - Gaia Scerif
- Department of Experimental Psychology, University of Oxford, Oxford, UK
| | - Teodora Gliga
- Centre for Brain and Cognitive Development, Birkbeck College, University of London, London, UK
| |
Collapse
|
41
|
Non-singleton colors are not attended faster than categories, but they are encoded faster: A combined approach of behavior, modeling and ERPs. Vision Res 2017; 140:106-119. [DOI: 10.1016/j.visres.2017.06.013] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2017] [Revised: 06/05/2017] [Accepted: 06/08/2017] [Indexed: 11/22/2022]
|
42
|
Wang Z, Buetti S, Lleras A. Predicting Search Performance in Heterogeneous Visual Search Scenes with Real-World Objects. COLLABRA-PSYCHOLOGY 2017. [DOI: 10.1525/collabra.53] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Previous work in our lab has demonstrated that efficient visual search with a fixed target has a reaction time by set size function that is best characterized by logarithmic curves. Further, the steepness of these logarithmic curves is determined by the similarity between target and distractor items (Buetti et al., 2016). A theoretical account of these findings was proposed, namely that a parallel, unlimited capacity, exhaustive processing architecture is underlying such data. Here, we conducted two experiments to expand these findings to a set of real-world stimuli, in both homogeneous and heterogeneous search displays. We used computational simulations of this architecture to identify a way to predict RT performance in heterogeneous search using parameters estimated from homogeneous search data. Further, by examining the systematic deviation from our predictions in the observed data, we found evidence that early visual processing for individual items is not independent. Instead, items in homogeneous displays seemed to facilitate each other’s processing by a multiplicative factor. These results challenge previous accounts of heterogeneity effects in visual search, and demonstrate the explanatory and predictive power of an approach that combines computational simulations and behavioral data to better understand performance in visual search.
Collapse
Affiliation(s)
- Zhiyuan Wang
- Department of Psychology, University of Illinois at Urbana-Champaign, Champaign, Illinois, US
| | - Simona Buetti
- Department of Psychology, University of Illinois at Urbana-Champaign, Champaign, Illinois, US
| | - Alejandro Lleras
- Department of Psychology, University of Illinois at Urbana-Champaign, Champaign, Illinois, US
| |
Collapse
|
43
|
A Model of the Superior Colliculus Predicts Fixation Locations during Scene Viewing and Visual Search. J Neurosci 2016; 37:1453-1467. [PMID: 28039373 DOI: 10.1523/jneurosci.0825-16.2016] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2016] [Revised: 11/21/2016] [Accepted: 12/01/2016] [Indexed: 11/21/2022] Open
Abstract
Modern computational models of attention predict fixations using saliency maps and target maps, which prioritize locations for fixation based on feature contrast and target goals, respectively. But whereas many such models are biologically plausible, none have looked to the oculomotor system for design constraints or parameter specification. Conversely, although most models of saccade programming are tightly coupled to underlying neurophysiology, none have been tested using real-world stimuli and tasks. We combined the strengths of these two approaches in MASC, a model of attention in the superior colliculus (SC) that captures known neurophysiological constraints on saccade programming. We show that MASC predicted the fixation locations of humans freely viewing naturalistic scenes and performing exemplar and categorical search tasks, a breadth achieved by no other existing model. Moreover, it did this as well or better than its more specialized state-of-the-art competitors. MASC's predictive success stems from its inclusion of high-level but core principles of SC organization: an over-representation of foveal information, size-invariant population codes, cascaded population averaging over distorted visual and motor maps, and competition between motor point images for saccade programming, all of which cause further modulation of priority (attention) after projection of saliency and target maps to the SC. Only by incorporating these organizing brain principles into our models can we fully understand the transformation of complex visual information into the saccade programs underlying movements of overt attention. With MASC, a theoretical footing now exists to generate and test computationally explicit predictions of behavioral and neural responses in visually complex real-world contexts.SIGNIFICANCE STATEMENT The superior colliculus (SC) performs a visual-to-motor transformation vital to overt attention, but existing SC models cannot predict saccades to visually complex real-world stimuli. We introduce a brain-inspired SC model that outperforms state-of-the-art image-based competitors in predicting the sequences of fixations made by humans performing a range of everyday tasks (scene viewing and exemplar and categorical search), making clear the value of looking to the brain for model design. This work is significant in that it will drive new research by making computationally explicit predictions of SC neural population activity in response to naturalistic stimuli and tasks. It will also serve as a blueprint for the construction of other brain-inspired models, helping to usher in the next generation of truly intelligent autonomous systems.
Collapse
|
44
|
Idiosyncratic Patterns of Representational Similarity in Prefrontal Cortex Predict Attentional Performance. J Neurosci 2016; 37:1257-1268. [PMID: 28028199 DOI: 10.1523/jneurosci.1407-16.2016] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2016] [Revised: 11/08/2016] [Accepted: 12/09/2016] [Indexed: 11/21/2022] Open
Abstract
The efficiency of finding an object in a crowded environment depends largely on the similarity of nontargets to the search target. Models of attention theorize that the similarity is determined by representations stored within an "attentional template" held in working memory. However, the degree to which the contents of the attentional template are individually unique and where those idiosyncratic representations are encoded in the brain are unknown. We investigated this problem using representational similarity analysis of human fMRI data to measure the common and idiosyncratic representations of famous face morphs during an identity categorization task; data from the categorization task were then used to predict performance on a separate identity search task. We hypothesized that the idiosyncratic categorical representations of the continuous face morphs would predict their distractability when searching for each target identity. The results identified that patterns of activation in the lateral prefrontal cortex (LPFC) as well as in face-selective areas in the ventral temporal cortex were highly correlated with the patterns of behavioral categorization of face morphs and search performance that were common across subjects. However, the individually unique components of the categorization behavior were reliably decoded only in right LPFC. Moreover, the neural pattern in right LPFC successfully predicted idiosyncratic variability in search performance, such that reaction times were longer when distractors had a higher probability of being categorized as the target identity. These results suggest that the prefrontal cortex encodes individually unique components of categorical representations that are also present in attentional templates for target search. SIGNIFICANCE STATEMENT Everyone's perception of the world is uniquely shaped by personal experiences and preferences. Using functional MRI, we show that individual differences in the categorization of face morphs between two identities could be decoded from the prefrontal cortex and the ventral temporal cortex. Moreover, the individually unique representations in prefrontal cortex predicted idiosyncratic variability in attentional performance when looking for each identity in the "crowd" of another morphed face in a separate search task. Our results reveal that the representation of task-related information in prefrontal cortex is individually unique and preserved across categorization and search performance. This demonstrates the possibility of predicting individual behaviors across tasks with patterns of brain activity.
Collapse
|
45
|
Horstmann G, Herwig A, Becker SI. Distractor Dwelling, Skipping, and Revisiting Determine Target Absent Performance in Difficult Visual Search. Front Psychol 2016; 7:1152. [PMID: 27574510 PMCID: PMC4983613 DOI: 10.3389/fpsyg.2016.01152] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2016] [Accepted: 07/19/2016] [Indexed: 11/13/2022] Open
Abstract
Some targets in visual search are more difficult to find than others. In particular, a target that is similar to the distractors is more difficult to find than a target that is dissimilar to the distractors. Efficiency differences between easy and difficult searches are manifest not only in target-present trials but also in target-absent trials. In fact, even physically identical displays are searched through with different efficiency depending on the searched-for target. Here, we monitored eye movements in search for a target similar to the distractors (difficult search) versus a target dissimilar to the distractors (easy search). We aimed to examine three hypotheses concerning the causes of differential search efficiencies in target-absent trials: (a) distractor dwelling (b) distractor skipping, and (c) distractor revisiting. Reaction times increased with target similarity which is consistent with existing theories and replicates earlier results. Eye movement data indicated guidance in target trials, even though search was very slow. Dwelling, skipping, and revisiting contributed to low search efficiency in difficult search, with dwelling being the strongest factor. It is argued that differences in dwell time account for a large amount of total search time differences.
Collapse
Affiliation(s)
- Gernot Horstmann
- Department of Psychology, Bielefeld UniversityBielefeld, Germany; Cognitive Interaction Technology - Excellence Center, Bielefeld UniversityBielefeld, Germany; Center for Interdisciplinary Research, Bielefeld UniversityBielefeld, Germany
| | - Arvid Herwig
- Department of Psychology, Bielefeld UniversityBielefeld, Germany; Cognitive Interaction Technology - Excellence Center, Bielefeld UniversityBielefeld, Germany; Center for Interdisciplinary Research, Bielefeld UniversityBielefeld, Germany
| | | |
Collapse
|
46
|
Wu X, Liu X, Fu S. Feature- and category-specific attentional control settings are differently affected by attentional engagement in contingent attentional capture. Biol Psychol 2016; 118:8-16. [PMID: 27151739 DOI: 10.1016/j.biopsycho.2016.04.065] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2015] [Revised: 04/19/2016] [Accepted: 04/29/2016] [Indexed: 10/21/2022]
|
47
|
Hout MC, Godwin HJ, Fitzsimmons G, Robbins A, Menneer T, Goldinger SD. Using multidimensional scaling to quantify similarity in visual search and beyond. Atten Percept Psychophys 2016; 78:3-20. [PMID: 26494381 PMCID: PMC5523409 DOI: 10.3758/s13414-015-1010-6] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Visual search is one of the most widely studied topics in vision science, both as an independent topic of interest, and as a tool for studying attention and visual cognition. A wide literature exists that seeks to understand how people find things under varying conditions of difficulty and complexity, and in situations ranging from the mundane (e.g., looking for one's keys) to those with significant societal importance (e.g., baggage or medical screening). A primary determinant of the ease and probability of success during search are the similarity relationships that exist in the search environment, such as the similarity between the background and the target, or the likeness of the non-targets to one another. A sense of similarity is often intuitive, but it is seldom quantified directly. This presents a problem in that similarity relationships are imprecisely specified, limiting the capacity of the researcher to examine adequately their influence. In this article, we present a novel approach to overcoming this problem that combines multi-dimensional scaling (MDS) analyses with behavioral and eye-tracking measurements. We propose a method whereby MDS can be repurposed to successfully quantify the similarity of experimental stimuli, thereby opening up theoretical questions in visual search and attention that cannot currently be addressed. These quantifications, in conjunction with behavioral and oculomotor measures, allow for critical observations about how similarity affects performance, information selection, and information processing. We provide a demonstration and tutorial of the approach, identify documented examples of its use, discuss how complementary computer vision methods could also be adopted, and close with a discussion of potential avenues for future application of this technique.
Collapse
|
48
|
Hout MC, Goldinger SD. Target templates: the precision of mental representations affects attentional guidance and decision-making in visual search. Atten Percept Psychophys 2015; 77:128-49. [PMID: 25214306 PMCID: PMC4286498 DOI: 10.3758/s13414-014-0764-6] [Citation(s) in RCA: 98] [Impact Index Per Article: 9.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
When people look for things in the environment, they use target templates-mental representations of the objects they are attempting to locate-to guide attention and to assess incoming visual input as potential targets. However, unlike laboratory participants, searchers in the real world rarely have perfect knowledge regarding the potential appearance of targets. In seven experiments, we examined how the precision of target templates affects the ability to conduct visual search. Specifically, we degraded template precision in two ways: 1) by contaminating searchers' templates with inaccurate features, and 2) by introducing extraneous features to the template that were unhelpful. We recorded eye movements to allow inferences regarding the relative extents to which attentional guidance and decision-making are hindered by template imprecision. Our findings support a dual-function theory of the target template and highlight the importance of examining template precision in visual search.
Collapse
Affiliation(s)
- Michael C Hout
- Department of Psychology, New Mexico State University, P.O. Box 30001 / MSC 3452, Las Cruces, NM, 88003, USA,
| | | |
Collapse
|
49
|
Maxfield JT, Stalder WD, Zelinsky GJ. Effects of target typicality on categorical search. J Vis 2014; 14:1. [PMID: 25274990 PMCID: PMC4181372 DOI: 10.1167/14.12.1] [Citation(s) in RCA: 44] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2013] [Accepted: 07/25/2014] [Indexed: 11/24/2022] Open
Abstract
The role of target typicality in a categorical visual search task was investigated by cueing observers with a target name, followed by a five-item target present/absent search array in which the target images were rated in a pretest to be high, medium, or low in typicality with respect to the basic-level target cue. Contrary to previous work, we found that search guidance was better for high-typicality targets compared to low-typicality targets, as measured by both the proportion of immediate target fixations and the time to fixate the target. Consistent with previous work, we also found an effect of typicality on target verification times, the time between target fixation and the search judgment; as target typicality decreased, verification times increased. To model these typicality effects, we trained Support Vector Machine (SVM) classifiers on the target categories, and tested these on the corresponding specific targets used in the search task. This analysis revealed significant differences in classifier confidence between the high-, medium-, and low-typicality groups, paralleling the behavioral results. Collectively, these findings suggest that target typicality broadly affects both search guidance and verification, and that differences in typicality can be predicted by distance from an SVM classification boundary.
Collapse
Affiliation(s)
| | | | - Gregory J. Zelinsky
- Department of Psychology, Stony Brook University, Stony Brook, NY
- Department of Computer Science, Stony Brook University, Stony Brook, NY
| |
Collapse
|
50
|
Alexander RG, Schmidt J, Zelinsky GJ. Are summary statistics enough? Evidence for the importance of shape in guiding visual search. VISUAL COGNITION 2014; 22:595-609. [PMID: 26180505 DOI: 10.1080/13506285.2014.890989] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Abstract
Peripheral vision outside the focus of attention may rely on summary statistics. We used a gaze-contingent paradigm to directly test this assumption by asking whether search performance differed between targets and statistically-matched visualizations of the same targets. Four-object search displays included one statistically-matched object that was replaced by an unaltered version of the object during the first eye movement. Targets were designated by previews, which were never altered. Two types of statistically-matched objects were tested: One that maintained global shape and one that did not. Differences in guidance were found between targets and statistically-matched objects when shape was not preserved, suggesting that they were not informationally equivalent. Responses were also slower after target fixation when shape was not preserved, suggesting an extrafoveal processing of the target that again used shape information. We conclude that summary statistics must include some global shape information to approximate the peripheral information used during search.
Collapse
Affiliation(s)
| | - Joseph Schmidt
- Department of Psychology, Stony Brook University, USA ; Institute for Mind and Brain, University of South Carolina, USA
| | - Gregory J Zelinsky
- Department of Psychology, Stony Brook University, USA ; Department of Computer Science, Stony Brook University, USA
| |
Collapse
|