1
|
Turnbull G, Lego S, Kennedy BL, Alexi J, Li YR, Engel MM, Mann G, Bayliss DM, Farrell S, Bell J. Sizing up the crowd: Assessing spatial integration difficulties in body size judgements across eating disorder symptomatology. Front Psychol 2023; 13:1003250. [PMID: 36687820 PMCID: PMC9853910 DOI: 10.3389/fpsyg.2022.1003250] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2022] [Accepted: 11/30/2022] [Indexed: 01/07/2023] Open
Abstract
Introduction Body size judgements are frequently biased, or inaccurate, and these errors are further exaggerated for individuals with eating disorders. Within the eating disorder literature, it has been suggested that exaggerated errors in body size judgements are due to difficulties with integration. Across two experiments, we developed a novel integration task, named the Ebbinghaus Illusion for Bodies in Virtual Reality (VR), to assess whether nearby bodies influence the perceived size of a single body. VR was used to simulate the appearance of a small crowd around a central target body. Method and Results In Experiment 1 (N = 412), participants were required to judge the size of a central female target within a crowd. Experiment 1 revealed an Ebbinghaus Illusion, in which a central female appeared larger when surrounded by small distractors, but comparatively smaller when surrounded by large distractors. In other words, the findings of Experiment 1 demonstrate that surrounding crowd information is integrated when judging an individual's body size; a novel measure of spatial integration (i.e., an Ebbinghaus Illusion for Bodies in VR). In Experiment 2 (N = 96), female participants were selected based on high (n = 43) and low (n = 53) eating disorder symptomatology. We examined whether the magnitude of this illusion would differ amongst those with elevated versus low eating disorder symptomatology, in accordance with weak central coherence theory, with the high symptomatology group displaying less spatial integration relative to the low group. The results of Experiment 2 similarly found an Ebbinghaus Illusion for Bodies in VR. However, illusion magnitude did not vary across high and low symptomatology groups. Discussion Overall, these findings demonstrate that surrounding crowd information is integrated when judging individual body size; however, those with elevated eating disorder symptomatology did not show any integration deficit on this broader measure of spatial integration.
Collapse
Affiliation(s)
- Georgia Turnbull
- School of Psychological Science, University of Western Australia, Crawley, WA, Australia,*Correspondence: Georgia Turnbull,
| | - Sophia Lego
- School of Psychological Science, University of Western Australia, Crawley, WA, Australia
| | - Briana L. Kennedy
- School of Psychological Science, University of Western Australia, Crawley, WA, Australia
| | - Joanna Alexi
- School of Psychological Science, University of Western Australia, Crawley, WA, Australia
| | - Yanqi R. Li
- School of Psychological Science, University of Western Australia, Crawley, WA, Australia
| | - Manja M. Engel
- Faculty of Social and Behavioural Sciences, Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, Netherlands
| | - Georgina Mann
- School of Psychological Science, University of Western Australia, Crawley, WA, Australia
| | - Donna M. Bayliss
- School of Psychological Science, University of Western Australia, Crawley, WA, Australia
| | - Simon Farrell
- School of Psychological Science, University of Western Australia, Crawley, WA, Australia
| | - Jason Bell
- School of Psychological Science, University of Western Australia, Crawley, WA, Australia
| |
Collapse
|
2
|
Turnbull G, Alexi J, Mann G, Li Y, Engel M, Bayliss DM, Farrell S, Bell J. EXPRESS: The Influence of Three-Dimensional Cues on Body Size Judgements. Q J Exp Psychol (Hove) 2022; 75:2318-2331. [PMID: 35034530 DOI: 10.1177/17470218221076850] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Research has shown that body size judgements are frequently biased, or inaccurate. Critically, judgement biases are further exaggerated for individuals with eating disorders, a finding that has been attributed to difficulties integrating body features into a perceptual whole. However, current understanding of which body features are integrated when judging body size is lacking. In this study, we examine whether individuals integrate three-dimensional (3D) cues to body volume when making body size judgements. Computer-generated body stimuli were presented in a 3D Virtual Reality (VR) environment. Participants (N = 412) were randomly assigned to one of two conditions: in one condition the to-be-judged body was displayed binocularly (containing 3D cues to body volume), in the other, bodies were presented monocularly (2D cues only). Across 150 trials, participants were required to make a body size judgement of a target female body from a third-person point of view using an unmarked visual analogue scale (VAS). It was found that 3D cues significantly influenced body size judgements. Namely, thin 3D bodies were judged smaller, and overweight 3D bodies were judged larger, than their 2D counterpart. Furthermore, to reconcile these effects, we present evidence that the two perceptual biases, regression to the mean and serial dependence, were reduced by the additional 3D feature information. Our findings increase our understanding of how body size is perceptually encoded and creates testable predictions for clinical populations exhibiting integration difficulties.
Collapse
Affiliation(s)
- Georgia Turnbull
- School of Psychological Science, University of Western Australia (M304), 35 Stirling Highway, Crawley WA 6009, Australia 2720
| | - Joanna Alexi
- School of Psychological Science, University of Western Australia (M304), 35 Stirling Highway, Crawley WA 6009, Australia 2720
| | - Georgina Mann
- School of Psychological Science, University of Western Australia (M304), 35 Stirling Highway, Crawley WA 6009, Australia 2720
| | - Yanqi Li
- School of Psychological Science, University of Western Australia (M304), 35 Stirling Highway, Crawley WA 6009, Australia 2720
| | - Manja Engel
- Utrecht University, Faculty of Social and Behavioural Sciences, Experimental Psychology/Helmholtz Institute. Heidelberglaan 1, 3584 CS Utrecht, Netherlands 8125
| | - Donna M Bayliss
- School of Psychological Science, University of Western Australia (M304), 35 Stirling Highway, Crawley WA 6009, Australia 2720
| | - Simon Farrell
- School of Psychological Science, University of Western Australia (M304), 35 Stirling Highway, Crawley WA 6009, Australia 2720
| | - Jason Bell
- School of Psychological Science, University of Western Australia (M304), 35 Stirling Highway, Crawley WA 6009, Australia 2720
| |
Collapse
|
3
|
Liu H, Laeng B, Czajkowski NO. Does stereopsis improve face identification? A study using a virtual reality display with integrated eye-tracking and pupillometry. Acta Psychol (Amst) 2020; 210:103142. [PMID: 32836112 DOI: 10.1016/j.actpsy.2020.103142] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2020] [Revised: 06/02/2020] [Accepted: 07/15/2020] [Indexed: 10/23/2022] Open
Abstract
Stereopsis is a powerful depth cue for humans, which may also contribute to object recognition. In particular, we surmise that face identification would benefit from the availability of stereoscopic depth cues, since facial perception may be based on three-dimensional (3D) representations. In this study, a virtual reality (VR) headset with integrated eye-tracking was used to present stereoscopic images of faces. As a monoscopic contrast condition, identical images of faces were displayed to the two eyes. We monitored the participants' gaze behavior and pupil diameters while they performed a sample-to-match face identification task. We found that accuracy was superior in the stereoscopic condition compared to the monoscopic condition for frontal and intermediate views, but not profiles. Moreover, pupillary diameters were smaller when identifying stereoscopically seen faces than when viewing them without stereometric cues, which we interpret as lower processing load for the former than the latter conditions. The analysis of gaze showed that participants tended to focus on regions of the face rich in volumetric information, more so in the stereoscopic condition than the monoscopic condition. Together, these findings suggest that a 3D representation of faces may be the natural format used by the visual system when assessing face identity. Stereoscopic information, by providing depth information, assists the construction of robust facial representations in memory.
Collapse
|
4
|
Quantifying the effect of viewpoint changes on sensitivity to face identity. Vision Res 2019; 165:1-12. [DOI: 10.1016/j.visres.2019.09.006] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2019] [Revised: 08/28/2019] [Accepted: 09/16/2019] [Indexed: 11/20/2022]
|
5
|
Arguin M, Marleau I, Aubin M, Zahabi S, Leek EC. A surface-based code contributes to visual shape perception. J Vis 2019; 19:6. [PMID: 31509602 DOI: 10.1167/19.11.6] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Considerable uncertainty remains regarding the types of features human vision uses for shape representation. Visual-search experiments are reported which assessed the hypothesis of a surface-based (i.e., edge-bounded polygons) code for shape representation in human vision. The results indicate slower search rates and/or longer response times when the target shape shares its constituent surfaces with distractors (conjunction condition) than when the target surfaces are unique in the display (nonconjunction condition). This demonstration is made using test conditions that strictly control any potential artifact pertaining to target-distractor similarity. The surface-based code suggested by this surface-conjunction effect is strictly 2-D, since the effect occurs even when the surfaces are shared between the target and distractors in the 2-D image but not in their 3-D instantiation. Congruently, this latter finding is unaltered by manipulations of the richness of the depth information offered by the stimuli. It is proposed that human vision uses a 2-D surface-based code for shape representation which, considering other key findings in the field, probably coexists with an alternative representation mode based on a type of structural description that can integrate information pertaining to the 3-D aspect of shapes.
Collapse
Affiliation(s)
- Martin Arguin
- Centre de Recherche en Neuropsychologie Expérimentale et Cognition, Département de psychologie, Université de Montréal, Montréal, Canada
| | - Ian Marleau
- CISSS de la Montérégie-Ouest, Installation Longueuil, Longueuil, Canada
| | - Mercédès Aubin
- CÉGEP de Jonquière, Département des Sciences humaines, Jonquière, Canada
| | - Sacha Zahabi
- Centre de Recherche en Neuropsychologie Expérimentale et Cognition, Département de psychologie, Université de Montréal, Montréal, Canada
| | - E Charles Leek
- School of Psychology, Institute of Life and Human Sciences, University of Liverpool, Liverpool, UK
| |
Collapse
|
6
|
Sama MA, Nestor A, Cant JS. Independence of viewpoint and identity in face ensemble processing. J Vis 2019; 19:2. [DOI: 10.1167/19.5.2] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Affiliation(s)
- Marco A. Sama
- Department of Psychology, University of Toronto Scarborough, Toronto, Canada
| | - Adrian Nestor
- Department of Psychology, University of Toronto Scarborough, Toronto, Canada
| | - Jonathan S. Cant
- Department of Psychology, University of Toronto Scarborough, Toronto, Canada
| |
Collapse
|
7
|
View specific generalisation effects in face recognition: Front and yaw comparison views are better than pitch. PLoS One 2018; 13:e0209927. [PMID: 30592761 PMCID: PMC6310264 DOI: 10.1371/journal.pone.0209927] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2018] [Accepted: 12/13/2018] [Indexed: 11/26/2022] Open
Abstract
It can be difficult to recognise new instances of an unfamiliar face. Recognition errors in this particular situation appear to be viewpoint dependent with error rates increasing with the angular distance between the face views. Studies using front views for comparison have shown that recognising faces rotated in yaw can be difficult and that recognition of faces rotated in pitch is more challenging still. Here we investigate the extent to which viewpoint dependent face recognition depends on the comparison view. Participants were assigned to one of four different comparison view groups: front, ¾ yaw (right), ¾ pitch-up (above) or ¾ pitch-down (below). On each trial, participants matched their particular comparison view to a range of yaw or pitch rotated test views. Results showed that groups with a front or ¾ yaw comparison view had superior overall performance and more successful generalisation to a broader range of both pitch and yaw test views compared to groups with pitch-up or pitch-down comparison views, both of which had a very restricted generalisation range. Regression analyses revealed the importance of image similarity between views for generalisation, with a lesser role for 3D face depth. These findings are consistent with a view interpolation solution to view generalisation of face recognition, with front and ¾ yaw views being most informative.
Collapse
|
8
|
Oliver ZJ, Cristino F, Roberts MV, Pegna AJ, Leek EC. Stereo viewing modulates three-dimensional shape processing during object recognition: A high-density ERP study. J Exp Psychol Hum Percept Perform 2018; 44:518-534. [PMID: 29022728 PMCID: PMC5896504 DOI: 10.1037/xhp0000444] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2016] [Revised: 03/29/2017] [Accepted: 04/10/2017] [Indexed: 11/17/2022]
Abstract
The role of stereo disparity in the recognition of 3-dimensional (3D) object shape remains an unresolved issue for theoretical models of the human visual system. We examined this issue using high-density (128 channel) recordings of event-related potentials (ERPs). A recognition memory task was used in which observers were trained to recognize a subset of complex, multipart, 3D novel objects under conditions of either (bi-) monocular or stereo viewing. In a subsequent test phase they discriminated previously trained targets from untrained distractor objects that shared either local parts, 3D spatial configuration, or neither dimension, across both previously seen and novel viewpoints. The behavioral data showed a stereo advantage for target recognition at untrained viewpoints. ERPs showed early differential amplitude modulations to shape similarity defined by local part structure and global 3D spatial configuration. This occurred initially during an N1 component around 145-190 ms poststimulus onset, and then subsequently during an N2/P3 component around 260-385 ms poststimulus onset. For mono viewing, amplitude modulation during the N1 was greatest between targets and distracters with different local parts for trained views only. For stereo viewing, amplitude modulation during the N2/P3 was greatest between targets and distracters with different global 3D spatial configurations and generalized across trained and untrained views. The results show that image classification is modulated by stereo information about the local part, and global 3D spatial configuration of object shape. The findings challenge current theoretical models that do not attribute functional significance to stereo input during the computation of 3D object shape. (PsycINFO Database Record
Collapse
|
9
|
Pegna AJ, Darque A, Roberts MV, Leek EC. Effects of stereoscopic disparity on early ERP components during classification of three-dimensional objects. Q J Exp Psychol (Hove) 2018; 71:1419-1430. [PMID: 28524772 DOI: 10.1080/17470218.2017.1333129] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
This study investigates the effects of stereo disparity on the perception of three-dimensional (3D) object shape. We tested the hypothesis that stereo input modulates the brain activity related to perceptual analyses of 3D shape configuration during image classification. High-density (256-channel) electroencephalogram (EEG) was used to record the temporal dynamics of visual shape processing under conditions of two-dimensional (2D) and 3D visual presentation. On each trial, observers made image classification judgements ('Same'/'Different') to two briefly presented, multi-part, novel objects. On different-object trials, stimuli could either share volumetric parts but not the global 3D shape configuration and have different parts but the same global 3D shape configuration or differ on both aspects. Analyses using mass univariate contrasts showed that the earliest sensitivity to 2D versus 3D viewing appeared as a negative deflection over posterior locations on the N1 component between 160 and 220 ms post-stimulus onset. Subsequently, event-related potential (ERP) modulations during the N2 time window between 240 and 370 ms were linked to image classification. N2 activity reflected two distinct components - an early N2 (240-290 ms) and a late N2 (290-370 ms) - that showed different patterns of responses to 2D and 3D input and differential sensitivity to 3D object structure. The results revealed that stereo input modulates the neural correlates of 3D object shape. We suggest that this reflects differential perceptual processing of object shape under conditions of stereo or mono input. These findings challenge current theories that attribute no functional role for stereo input during 3D shape perception.
Collapse
Affiliation(s)
- Alan J Pegna
- 1 Faculty of Psychology and Educational Sciences, University of Geneva, Geneva, Switzerland.,2 School of Psychology, University of Queensland, St Lucia, QLD, Australia
| | - Alexandra Darque
- 1 Faculty of Psychology and Educational Sciences, University of Geneva, Geneva, Switzerland
| | - Mark V Roberts
- 3 Wolfson Centre for Clinical and Cognitive Neuroscience, School of Psychology, Bangor University, Bangor, UK
| | - E Charles Leek
- 3 Wolfson Centre for Clinical and Cognitive Neuroscience, School of Psychology, Bangor University, Bangor, UK.,4 Laboratoire de Psychologie et NeuroCognition (LPNC), Université Grenoble Alpes, Saint-Martin-d'Hères, France
| |
Collapse
|
10
|
Schneider TM, Carbon CC. Taking the Perfect Selfie: Investigating the Impact of Perspective on the Perception of Higher Cognitive Variables. Front Psychol 2017. [PMID: 28649219 PMCID: PMC5465279 DOI: 10.3389/fpsyg.2017.00971] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/04/2023] Open
Abstract
Taking selfies is now becoming a standard human habit. However, as a social phenomenon, research is still in the fledgling stage and the scientific framework is sparse. Selfies allow us to share social information with others in a compact format. Furthermore, we are able to control important photographic and compositional aspects, such as perspective, which have a strong impact on the assessment of a face (e.g., demonstrated by the height-weight illusion, effects of gaze direction, faceism-index). In Study 1, we focused on the impact of perspective (left/right hemiface, above/below vs. frontal presentation) on higher cognitive variables and let 172 participants rate the perceived attractiveness, helpfulness, sympathy, dominance, distinctiveness, and intelligence, plus important information on health issues (e.g., body weight), on the basis of 14 3D faces. We could show that lateral snapshots yielded higher ratings for attractiveness compared to the classical frontal view. However, this effect was more pronounced for left hemifaces and especially female faces. Compared to the frontal condition, 30° right hemifaces were rated as more helpful, but only for female faces while faces viewed from above were perceived as significant less helpful. Direct comparison between left vs. right hemifaces revealed no effect. Relating to sympathy, we only found a significant effect for 30° right male hemifaces, but only in comparison to the frontal condition. Furthermore, female 30° right hemifaces were perceived as more intelligent. Relating to body weight, we replicated the so-called “height-weight illusion.” Other variables remained unaffected. In Study 2, we investigated the impact of a typical selfie-style condition by presenting the respective faces from a lateral (left/right) and tilted (lower/higher) vantage point. Most importantly, depending on what persons wish to express with a selfie, a systematic change of perspective can strongly optimize their message; e.g., increasing their attractiveness by shooting from above left, and in contrast, decreasing their expressed helpfulness by shooting from below. We could further extent past findings relating to the height-weight illusion and showed that an additional rotation of the camera positively affected the perception of body weight (lower body weight). We discuss potential explanations for perspective-related effects, especially gender-related ones.
Collapse
Affiliation(s)
- Tobias M Schneider
- Department of General Psychology and Methodology, University of BambergBamberg, Germany.,Bamberg Graduate School of Affective and Cognitive Sciences, University of BambergBamberg, Germany.,Research Group EPÆG (Ergonomics, Psychological Æsthetics, Gestalt)Bamberg, Germany
| | - Claus-Christian Carbon
- Department of General Psychology and Methodology, University of BambergBamberg, Germany.,Bamberg Graduate School of Affective and Cognitive Sciences, University of BambergBamberg, Germany.,Research Group EPÆG (Ergonomics, Psychological Æsthetics, Gestalt)Bamberg, Germany
| |
Collapse
|
11
|
Wang L, Chen W, Li H. Use of 3D faces facilitates facial expression recognition in children. Sci Rep 2017; 7:45464. [PMID: 28368008 PMCID: PMC5377359 DOI: 10.1038/srep45464] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2016] [Accepted: 02/28/2017] [Indexed: 11/17/2022] Open
Abstract
This study assessed whether presenting 3D face stimuli could facilitate children’s facial expression recognition. Seventy-one children aged between 3 and 6 participated in the study. Their task was to judge whether a face presented in each trial showed a happy or fearful expression. Half of the face stimuli were shown with 3D representations, whereas the other half of the images were shown as 2D pictures. We compared expression recognition under these conditions. The results showed that the use of 3D faces improved the speed of facial expression recognition in both boys and girls. Moreover, 3D faces improved boys’ recognition accuracy for fearful expressions. Since fear is the most difficult facial expression for children to recognize, the facilitation effect of 3D faces has important practical implications for children with difficulties in facial expression recognition. The potential benefits of 3D representation for other expressions also have implications for developing more realistic assessments of children’s expression recognition.
Collapse
Affiliation(s)
- Lamei Wang
- Joint International Research Laboratory of Child Development and Health, College of Psychology and Sociology, Shenzhen University, Shenzhen, China.,Institute of Psychology, Chinese Academy of Sciences, Beijing, China
| | - Wenfeng Chen
- Institute of Psychology, Chinese Academy of Sciences, Beijing, China
| | - Hong Li
- Joint International Research Laboratory of Child Development and Health, College of Psychology and Sociology, Shenzhen University, Shenzhen, China
| |
Collapse
|
12
|
Hakala J, Kätsyri J, Häkkinen J. Stereoscopy Amplifies Emotions Elicited by Facial Expressions. Iperception 2016; 6:2041669515615071. [PMID: 27551358 PMCID: PMC4975116 DOI: 10.1177/2041669515615071] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
Mediated facial expressions do not elicit emotions as strongly as real-life facial expressions, possibly due to the low fidelity of pictorial presentations in typical mediation technologies. In the present study, we investigated the extent to which stereoscopy amplifies emotions elicited by images of neutral, angry, and happy facial expressions. The emotional self-reports of positive and negative valence (which were evaluated separately) and arousal of 40 participants were recorded. The magnitude of perceived depth in the stereoscopic images was manipulated by varying the camera base at 15, 40, 65, 90, and 115 mm. The analyses controlled for participants' gender, gender match, emotional empathy, and trait alexithymia. The results indicated that stereoscopy significantly amplified the negative valence and arousal elicited by angry expressions at the most natural (65 mm) camera base, whereas stereoscopy amplified the positive valence elicited by happy expressions in both the narrowed and most natural (15-65 mm) base conditions. Overall, the results indicate that stereoscopy amplifies the emotions elicited by mediated emotional facial expressions when the depth geometry is close to natural. The findings highlight the sensitivity of the visual system to depth and its effect on emotions.
Collapse
Affiliation(s)
- Jussi Hakala
- Department of Computer Science, Aalto University, Espoo, Finland
| | - Jari Kätsyri
- Department of Computer Science, Aalto University, Espoo, Finland
| | - Jukka Häkkinen
- Institute of Behavioural Sciences, University of Helsinki, Finland
| |
Collapse
|
13
|
Papenmeier F, Schwan S. If you watch it move, you'll recognize it in 3D: Transfer of depth cues between encoding and retrieval. Acta Psychol (Amst) 2016; 164:90-5. [PMID: 26765253 DOI: 10.1016/j.actpsy.2015.12.010] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2015] [Revised: 12/07/2015] [Accepted: 12/20/2015] [Indexed: 11/19/2022] Open
Abstract
Viewing objects with stereoscopic displays provides additional depth cues through binocular disparity supporting object recognition. So far, it was unknown whether this results from the representation of specific stereoscopic information in memory or a more general representation of an object's depth structure. Therefore, we investigated whether continuous object rotation acting as depth cue during encoding results in a memory representation that can subsequently be accessed by stereoscopic information during retrieval. In Experiment 1, we found such transfer effects from continuous object rotation during encoding to stereoscopic presentations during retrieval. In Experiments 2a and 2b, we found that the continuity of object rotation is important because only continuous rotation and/or stereoscopic depth but not multiple static snapshots presented without stereoscopic information caused the extraction of an object's depth structure into memory. We conclude that an object's depth structure and not specific depth cues are represented in memory.
Collapse
|
14
|
Functional architecture for disparity in macaque inferior temporal cortex and its relationship to the architecture for faces, color, scenes, and visual field. J Neurosci 2015; 35:6952-68. [PMID: 25926470 DOI: 10.1523/jneurosci.5079-14.2015] [Citation(s) in RCA: 38] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
Binocular disparity is a powerful depth cue for object perception. The computations for object vision culminate in inferior temporal cortex (IT), but the functional organization for disparity in IT is unknown. Here we addressed this question by measuring fMRI responses in alert monkeys to stimuli that appeared in front of (near), behind (far), or at the fixation plane. We discovered three regions that showed preferential responses for near and far stimuli, relative to zero-disparity stimuli at the fixation plane. These "near/far" disparity-biased regions were located within dorsal IT, as predicted by microelectrode studies, and on the posterior inferotemporal gyrus. In a second analysis, we instead compared responses to near stimuli with responses to far stimuli and discovered a separate network of "near" disparity-biased regions that extended along the crest of the superior temporal sulcus. We also measured in the same animals fMRI responses to faces, scenes, color, and checkerboard annuli at different visual field eccentricities. Disparity-biased regions defined in either analysis did not show a color bias, suggesting that disparity and color contribute to different computations within IT. Scene-biased regions responded preferentially to near and far stimuli (compared with stimuli without disparity) and had a peripheral visual field bias, whereas face patches had a marked near bias and a central visual field bias. These results support the idea that IT is organized by a coarse eccentricity map, and show that disparity likely contributes to computations associated with both central (face processing) and peripheral (scene processing) visual field biases, but likely does not contribute much to computations within IT that are implicated in processing color.
Collapse
|
15
|
Caziot B, Backus BT. Stereoscopic Offset Makes Objects Easier to Recognize. PLoS One 2015; 10:e0129101. [PMID: 26079788 PMCID: PMC4469586 DOI: 10.1371/journal.pone.0129101] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2015] [Accepted: 05/05/2015] [Indexed: 11/21/2022] Open
Abstract
Binocular vision is obviously useful for depth perception, but it might also enhance other components of visual processing, such as image segmentation. We used naturalistic images to determine whether giving an object a stereoscopic offset of 15-120 arcmin of crossed disparity relative to its background would make the object easier to recognize in briefly presented (33-133 ms), temporally masked displays. Disparity had a beneficial effect across a wide range of disparities and display durations. Most of this benefit occurred whether or not the stereoscopic contour agreed with the object’s luminance contour. We attribute this benefit to an orienting of spatial attention that selected the object and its local background for enhanced 2D pattern processing. At longer display durations, contour agreement provided an additional benefit, and a separate experiment using random-dot stimuli confirmed that stereoscopic contours plausibly contributed to recognition at the longer display durations in our experiment. We conclude that in real-world situations binocular vision confers an advantage not only for depth perception, but also for recognizing objects from their luminance patterns and bounding contours.
Collapse
Affiliation(s)
- Baptiste Caziot
- Graduate Center for Vision Research, SUNY College of Optometry, 33 W. 42 St., New York, New York, 10036, United States of America
- SUNY Eye Institute, 33 W. 42 St., New York, New York, 10036, United States of America
- * E-mail:
| | - Benjamin T. Backus
- Graduate Center for Vision Research, SUNY College of Optometry, 33 W. 42 St., New York, New York, 10036, United States of America
- SUNY Eye Institute, 33 W. 42 St., New York, New York, 10036, United States of America
| |
Collapse
|
16
|
Cristino F, Davitt L, Hayward WG, Leek EC. Stereo disparity facilitates view generalization during shape recognition for solid multipart objects. Q J Exp Psychol (Hove) 2015; 68:2419-36. [PMID: 25679983 DOI: 10.1080/17470218.2015.1017512] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
Abstract
Current theories of object recognition in human vision make different predictions about whether the recognition of complex, multipart objects should be influenced by shape information about surface depth orientation and curvature derived from stereo disparity. We examined this issue in five experiments using a recognition memory paradigm in which observers (N = 134) memorized and then discriminated sets of 3D novel objects at trained and untrained viewpoints under either mono or stereo viewing conditions. In order to explore the conditions under which stereo-defined shape information contributes to object recognition we systematically varied the difficulty of view generalization by increasing the angular disparity between trained and untrained views. In one series of experiments, objects were presented from either previously trained views or untrained views rotated (15°, 30°, or 60°) along the same plane. In separate experiments we examined whether view generalization effects interacted with the vertical or horizontal plane of object rotation across 40° viewpoint changes. The results showed robust viewpoint-dependent performance costs: Observers were more efficient in recognizing learned objects from trained than from untrained views, and recognition was worse for extrapolated than for interpolated untrained views. We also found that performance was enhanced by stereo viewing but only at larger angular disparities between trained and untrained views. These findings show that object recognition is not based solely on 2D image information but that it can be facilitated by shape information derived from stereo disparity.
Collapse
Affiliation(s)
| | - Lina Davitt
- a School of Psychology , Bangor University , Bangor , UK
| | - William G Hayward
- b School of Psychology , University of Auckland , Auckland , New Zealand
| | - E Charles Leek
- c Wolfson Centre for Clinical and Cognitive Neuroscience, School of Psychology , Bangor University , Bangor , UK
| |
Collapse
|
17
|
Apps MAJ, Tsakiris M. Predictive codes of familiarity and context during the perceptual learning of facial identities. Nat Commun 2014; 4:2698. [PMID: 24220539 DOI: 10.1038/ncomms3698] [Citation(s) in RCA: 27] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2012] [Accepted: 10/01/2013] [Indexed: 11/09/2022] Open
Abstract
Face recognition is a key component of successful social behaviour. However, the computational processes that underpin perceptual learning and recognition as faces transition from unfamiliar to familiar are poorly understood. In predictive coding, learning occurs through prediction errors that update stimulus familiarity, but recognition is a function of both stimulus and contextual familiarity. Here we show that behavioural responses on a two-option face recognition task can be predicted by the level of contextual and facial familiarity in a computational model derived from predictive-coding principles. Using fMRI, we show that activity in the superior temporal sulcus varies with the contextual familiarity in the model, whereas activity in the fusiform face area covaries with the prediction error parameter that updated facial familiarity. Our results characterize the key computations underpinning the perceptual learning of faces, highlighting that the functional properties of face-processing areas conform to the principles of predictive coding.
Collapse
Affiliation(s)
- Matthew A J Apps
- 1] Nuffield Department of Clinical Neuroscience, University of Oxford, Oxford OX3 9DU, UK [2] Department of Experimental Psychology, University of Oxford, Oxford OX1 3UD, UK [3] Laboratory of Action and Body, Department of Psychology, University of London, Royal Holloway, Egham, Surrey TW20 0EX, UK
| | | |
Collapse
|
18
|
Valsecchi M, Caziot B, Backus BT, Gegenfurtner KR. The role of binocular disparity in rapid scene and pattern recognition. Iperception 2013; 4:122-36. [PMID: 23755357 PMCID: PMC3677332 DOI: 10.1068/i0587] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2013] [Revised: 03/22/2013] [Indexed: 11/03/2022] Open
Abstract
We investigated the contribution of binocular disparity to the rapid recognition of scenes and simpler spatial patterns using a paradigm combining backward masked stimulus presentation and short-term match-to-sample recognition. First, we showed that binocular disparity did not contribute significantly to the recognition of briefly presented natural and artificial scenes, even when the availability of monocular cues was reduced. Subsequently, using dense random dot stereograms as stimuli, we showed that observers were in principle able to extract spatial patterns defined only by disparity under brief, masked presentations. Comparing our results with the predictions from a cue-summation model, we showed that combining disparity with luminance did not per se disrupt the processing of disparity. Our results suggest that the rapid recognition of scenes is mediated mostly by a monocular comparison of the images, although we can rely on stereo in fast pattern recognition.
Collapse
Affiliation(s)
- Matteo Valsecchi
- Abteilung Allgemeine Psychologie, Justus-Liebig-Universität, Otto-Behaghel-Str. 10F, D-35394 Giessen, Germany; e-mail:
| | | | | | | |
Collapse
|
19
|
Bell J, Kanji J, Kingdom FA. Discrimination of rotated-in-depth curves is facilitated by stereoscopic cues, but curvature is not tuned for stereoscopic rotation-in-depth. Vision Res 2013. [DOI: 10.1016/j.visres.2012.11.003] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
20
|
On the contribution of binocular disparity to the long-term memory for natural scenes. PLoS One 2012; 7:e49947. [PMID: 23166799 PMCID: PMC3499513 DOI: 10.1371/journal.pone.0049947] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2012] [Accepted: 10/19/2012] [Indexed: 11/19/2022] Open
Abstract
Binocular disparity is a fundamental dimension defining the input we receive from the visual world, along with luminance and chromaticity. In a memory task involving images of natural scenes we investigate whether binocular disparity enhances long-term visual memory. We found that forest images studied in the presence of disparity for relatively long times (7s) were remembered better as compared to 2D presentation. This enhancement was not evident for other categories of pictures, such as images containing cars and houses, which are mostly identified by the presence of distinctive artifacts rather than by their spatial layout. Evidence from a further experiment indicates that observers do not retain a trace of stereo presentation in long-term memory.
Collapse
|
21
|
Abstract
The appearance of faces can be strongly affected by the characteristics of faces viewed previously. These perceptual after-effects reflect processes of sensory adaptation that are found throughout the visual system, but which have been considered only relatively recently in the context of higher level perceptual judgements. In this review, we explore the consequences of adaptation for human face perception, and the implications of adaptation for understanding the neural-coding schemes underlying the visual representation of faces. The properties of face after-effects suggest that they, in part, reflect response changes at high and possibly face-specific levels of visual processing. Yet, the form of the after-effects and the norm-based codes that they point to show many parallels with the adaptations and functional organization that are thought to underlie the encoding of perceptual attributes like colour. The nature and basis for human colour vision have been studied extensively, and we draw on ideas and principles that have been developed to account for norms and normalization in colour vision to consider potential similarities and differences in the representation and adaptation of faces.
Collapse
Affiliation(s)
- Michael A Webster
- Department of Psychology, University of Nevada, , Reno, NV 89557, USA.
| | | |
Collapse
|
22
|
Seeing an unfamiliar face in rotational motion does not aid identity discrimination across viewpoints. Vision Res 2010; 50:854-9. [DOI: 10.1016/j.visres.2010.02.013] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2010] [Accepted: 02/18/2010] [Indexed: 11/19/2022]
|
23
|
Abstract
We tested recognition of familiar objects in two different conditions: mono, where stimuli were displayed as flat, 2-D images, and stereo, where objects were displayed with stereoscopic depth information. In three experiments, participants performed a sequential matching task, where an object was rotated by up to 180 degrees between presentations. When the 180 degrees rotation resulted in large changes in depth for object components, recognition performance in the mono condition showed better performance at 180 degrees rotations than at smaller rotations, but stereo presentations showed a monotonic increase in response time with rotation. However, 180 degrees rotations that did not result in much depth variation showed similar patterns of results for mono and stereo conditions. These results suggest that in some circumstances, the lack of explicit 3-D information in 2-D images may influence the recognition of familiar objects when they are depicted on flat computer monitors.
Collapse
|
24
|
Farivar R. Dorsal–ventral integration in object recognition. ACTA ACUST UNITED AC 2009; 61:144-53. [DOI: 10.1016/j.brainresrev.2009.05.006] [Citation(s) in RCA: 48] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2008] [Revised: 04/29/2009] [Accepted: 05/23/2009] [Indexed: 10/20/2022]
|
25
|
Bell J, Kingdom FA. Global contour shapes are coded differently from their local components. Vision Res 2009; 49:1702-10. [DOI: 10.1016/j.visres.2009.04.012] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2008] [Revised: 03/25/2009] [Accepted: 04/09/2009] [Indexed: 10/20/2022]
|
26
|
Burton AM, Bindemann M. The role of view in human face detection. Vision Res 2009; 49:2026-36. [DOI: 10.1016/j.visres.2009.05.012] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2009] [Revised: 05/08/2009] [Accepted: 05/15/2009] [Indexed: 10/20/2022]
|
27
|
Bell J, Dickinson JE, Badcock DR. Radial frequency adaptation suggests polar-based coding of local shape cues. Vision Res 2008; 48:2293-301. [PMID: 18675841 DOI: 10.1016/j.visres.2008.07.003] [Citation(s) in RCA: 14] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2007] [Revised: 06/01/2008] [Accepted: 07/10/2008] [Indexed: 10/21/2022]
Abstract
The study of shape processing in the human visual system has frequently employed radial frequency (RF) patterns as conveniently manipulable stimuli. This study uses an adaptation paradigm to investigate how local shape information is sampled in the processing of RF contour shapes. Experiment 1 measured thresholds for detecting a fixed mean radius RF contour following adaptation to RF patterns which, in separate conditions, varied in mean radius and radial frequency. Results reveal that, adaptation is strongly tuned for RF over a range of pattern radii, but is not tuned for the number of cycles of radial modulation per visual degree of contour length; a characteristic that changes with both radius and radial frequency. Experiment 2 manipulated the polar angle separation on the fronto-parallel plane between curvature features on a fixed RF by foreshortening the pattern appearance (consistent with a rotation in depth) and shows that RF shape processing is tuned for fronto-parallel separation angles between curvature features. Results were near identical when a stereo rotation cue was added to the perspective modified RF. In the second part of Experiment 2 we showed that RF shape adaptation is also tuned for the polar angular extent of the curvature represented by the lobe at that angle. Collectively, our results indicate that the polar angle at which local curvature features appear, in addition to the angular extent of the curvature feature at that location, are both critical parameters for coding specific RF shapes.
Collapse
Affiliation(s)
- Jason Bell
- School of Psychology, The University of Western Australia, 35 Stirling Highway, Crawley 6009, WA, Australia.
| | | | | |
Collapse
|