1
|
Skog E, Meese TS, Sargent IMJ, Ormerod A, Schofield AJ. Classification images for aerial images capture visual expertise for binocular disparity and a prior for lighting from above. J Vis 2024; 24:11. [PMID: 38607637 PMCID: PMC11019598 DOI: 10.1167/jov.24.4.11] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2023] [Accepted: 02/06/2024] [Indexed: 04/13/2024] Open
Abstract
Using a novel approach to classification images (CIs), we investigated the visual expertise of surveyors for luminance and binocular disparity cues simultaneously after screening for stereoacuity. Stereoscopic aerial images of hedges and ditches were classified in 10,000 trials by six trained remote sensing surveyors and six novices. Images were heavily masked with luminance and disparity noise simultaneously. Hedge and ditch images had reversed disparity on around half the trials meaning hedges became ditch-like and vice versa. The hedge and ditch images were also flipped vertically on around half the trials, changing the direction of the light source and completing a 2 × 2 × 2 stimulus design. CIs were generated by accumulating the noise textures associated with "hedge" and "ditch" classifications, respectively, and subtracting one from the other. Typical CIs had a central peak with one or two negative side-lobes. We found clear differences in the amplitudes and shapes of perceptual templates across groups and noise-type, with experts prioritizing binocular disparity and using this more effectively. Contrariwise, novices used luminance cues more than experts meaning that task motivation alone could not explain group differences. Asymmetries in the luminance CIs revealed individual differences for lighting interpretation, with experts less prone to assume lighting from above, consistent with their training on aerial images of UK scenes lit by a southerly sun. Our results show that (i) dual noise in images can be used to produce simultaneous CI pairs, (ii) expertise for disparity cues does not depend on stereoacuity, (iii) CIs reveal the visual strategies developed by experts, (iv) top-down perceptual biases can be overcome with long-term learning effects, and (v) CIs have practical potential for directing visual training.
Collapse
Affiliation(s)
- Emil Skog
- School of Psychology, College of Health and Life Sciences, Aston University, Birmingham, B4 7ET, UK
- Aston Laboratory for Immersive Virtual Environments, College of Health and Life Sciences, Aston University, Birmingham, B4 7ET, UK
- Department of Health, Learning and Technology, Luleå University of Technology, Luleå, Sweden
| | - Timothy S Meese
- Aston Laboratory for Immersive Virtual Environments, College of Health and Life Sciences, Aston University, Birmingham, B4 7ET, UK
- https://research.aston.ac.uk/en/persons/tim-s-meese
| | - Isabel M J Sargent
- Ordnance Survey, Adanac Drive, Southampton, SO16 0AS, UK
- Electronics and Computer Science, University of Southampton, University Road, Southampton, SO17 1BJ, UK
- http://www.os.uk/
| | - Andrew Ormerod
- Ordnance Survey, Adanac Drive, Southampton, SO16 0AS, UK
- http://www.os.uk/
| | - Andrew J Schofield
- School of Psychology, College of Health and Life Sciences, Aston University, Birmingham, B4 7ET, UK
- Aston Laboratory for Immersive Virtual Environments, College of Health and Life Sciences, Aston University, Birmingham, B4 7ET, UK
- https://research.aston.ac.uk/en/persons/andrew-schofield
| |
Collapse
|
2
|
Jiang C, Chen Z, Wolfe JM. Toward viewing behavior for aerial scene categorization. Cogn Res Princ Implic 2024; 9:17. [PMID: 38530617 PMCID: PMC10965882 DOI: 10.1186/s41235-024-00541-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2023] [Accepted: 03/07/2024] [Indexed: 03/28/2024] Open
Abstract
Previous work has demonstrated similarities and differences between aerial and terrestrial image viewing. Aerial scene categorization, a pivotal visual processing task for gathering geoinformation, heavily depends on rotation-invariant information. Aerial image-centered research has revealed effects of low-level features on performance of various aerial image interpretation tasks. However, there are fewer studies of viewing behavior for aerial scene categorization and of higher-level factors that might influence that categorization. In this paper, experienced subjects' eye movements were recorded while they were asked to categorize aerial scenes. A typical viewing center bias was observed. Eye movement patterns varied among categories. We explored the relationship of nine image statistics to observers' eye movements. Results showed that if the images were less homogeneous, and/or if they contained fewer or no salient diagnostic objects, viewing behavior became more exploratory. Higher- and object-level image statistics were predictive at both the image and scene category levels. Scanpaths were generally organized and small differences in scanpath randomness could be roughly captured by critical object saliency. Participants tended to fixate on critical objects. Image statistics included in this study showed rotational invariance. The results supported our hypothesis that the availability of diagnostic objects strongly influences eye movements in this task. In addition, this study provides supporting evidence for Loschky et al.'s (Journal of Vision, 15(6), 11, 2015) speculation that aerial scenes are categorized on the basis of image parts and individual objects. The findings were discussed in relation to theories of scene perception and their implications for automation development.
Collapse
Affiliation(s)
- Chenxi Jiang
- School of Remote Sensing and Information Engineering, Wuhan University, Wuhan, Hubei, China
| | - Zhenzhong Chen
- School of Remote Sensing and Information Engineering, Wuhan University, Wuhan, Hubei, China.
- Hubei Luojia Laboratory, Wuhan, Hubei, China.
| | - Jeremy M Wolfe
- Harvard Medical School, Boston, MA, USA
- Brigham & Women's Hospital, Boston, MA, USA
| |
Collapse
|
3
|
Li C, Ge S, Wang R. Similarities and Differences in the Outsiders and Insiders' Visual Preferences on Sacred Landscape. Front Psychol 2022; 13:743933. [PMID: 35874406 PMCID: PMC9301073 DOI: 10.3389/fpsyg.2022.743933] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2021] [Accepted: 06/20/2022] [Indexed: 11/13/2022] Open
Abstract
Previous studies have reported religious and non-religious people as having different psychological experiences when visiting sacred landscapes; however, the visual consensus and differences between diverse groups visiting them have rarely been considered. This study used subjective preference evaluation and experimental eye tracking to assess the visual preferences of different groups regarding sacred landscapes. Overall, 48 photos of the Han Chinese Buddhist temples were selected as stimulus materials, including the categories of squares, architecture, waterscapes, and plants. In all, 90 participants were classified into two groups of outsiders and insiders to view the photos. The consensus and differences in their visual preferences and eye movement metrics were evaluated. The results showed that the two groups were more inclined toward the visual preference of religious architectures than the natural landscape that people usually prefer. Another noteworthy discovery revealed the significant differences between the outsiders and the insiders in viewing and evaluating sacred landscapes; the immersion effect explains this result. Specifically, the group with a higher interaction with the environment had greater visual experiences, easier visual information coding, and larger visual exploration range. In addition, this study revealed familiarity with the religious background facilitated achieving a higher consistency between the landscape preference scores and the eye movement metrics. These findings expand the theory of religious environment perception and provided important insights for subsequent research on sacred landscape planning and management.
Collapse
|
4
|
Rhodes RE, Cowley HP, Huang JG, Gray-Roncal W, Wester BA, Drenkow N. Benchmarking Human Performance for Visual Search of Aerial Images. Front Psychol 2021; 12:733021. [PMID: 34970183 PMCID: PMC8713551 DOI: 10.3389/fpsyg.2021.733021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2021] [Accepted: 11/08/2021] [Indexed: 12/05/2022] Open
Abstract
Aerial images are frequently used in geospatial analysis to inform responses to crises and disasters but can pose unique challenges for visual search when they contain low resolution, degraded information about color, and small object sizes. Aerial image analysis is often performed by humans, but machine learning approaches are being developed to complement manual analysis. To date, however, relatively little work has explored how humans perform visual search on these tasks, and understanding this could ultimately help enable human-machine teaming. We designed a set of studies to understand what features of an aerial image make visual search difficult for humans and what strategies humans use when performing these tasks. Across two experiments, we tested human performance on a counting task with a series of aerial images and examined the influence of features such as target size, location, color, clarity, and number of targets on accuracy and search strategies. Both experiments presented trials consisting of an aerial satellite image; participants were asked to find all instances of a search template in the image. Target size was consistently a significant predictor of performance, influencing not only accuracy of selections but the order in which participants selected target instances in the trial. Experiment 2 demonstrated that the clarity of the target instance and the match between the color of the search template and the color of the target instance also predicted accuracy. Furthermore, color also predicted the order of selecting instances in the trial. These experiments establish not only a benchmark of typical human performance on visual search of aerial images but also identify several features that can influence the task difficulty level for humans. These results have implications for understanding human visual search on real-world tasks and when humans may benefit from automated approaches.
Collapse
Affiliation(s)
- Rebecca E. Rhodes
- Johns Hopkins University Applied Physics Laboratory, Laurel, MD, United States
| | | | | | | | | | - Nathan Drenkow
- Johns Hopkins University Applied Physics Laboratory, Laurel, MD, United States
| |
Collapse
|