1
|
Devi P, Solomon JA, Tyler CW, Dave TV, Kaliki S, Bharadwaj SR. Comparison of Depth-Related Visuomotor Task Performance in Uniocular Individuals and in Binocular Controls With and Without Temporary Monocular Occlusion. Invest Ophthalmol Vis Sci 2024; 65:32. [PMID: 39028979 PMCID: PMC11262539 DOI: 10.1167/iovs.65.8.32] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2024] [Accepted: 06/24/2024] [Indexed: 07/21/2024] Open
Abstract
Purpose Do one-eyed (uniocular) humans use monocular depth cues differently from those with intact binocularity to perform depth-related visuomotor tasks that emulate complex activities of daily living? If so, does performance depend on the participant's age, duration of uniocularity and head movements? Methods Forty-five uniocular cases (age range 6-37 years; 2.4 months-31.0 years of uniocularity) and 46 age-similar binocular controls performed a task that required them to pass a hoop around an electrified wire convoluted in depth multiple times, while avoiding contact as indicated by auditory feedback. The task was performed with and without head restraint, in random order. The error rate and speed were calculated from the frequency of contact between the hoop and wire and the total task duration (adjusting for error time), respectively, all determined from video recordings of the task. Head movements were analyzed from the videos using face-tracking software. Results Error rate decreased with age (P < 0.001) until the late teen years while speed revealed no such trend. Across all ages, the error rate increased and speed decreased in the absence of binocularity (P < 0.001). There was no additional error reduction with duration of uniocularity (P = 0.16). Head movements provided no advantage to task performance, despite generating parallax disparities comparable to binocular viewing. Conclusions Performance in a dynamic, depth-related visuomotor task is reduced in the absence of binocular viewing, independent of age-related performance level. This study finds no evidence for a prolonged experience with monocular depth cues being advantageous for such tasks over transient loss of binocularity.
Collapse
Affiliation(s)
- Preetirupa Devi
- Centre for Applied Vision Research, City, University of London, London, United Kingdom
- Brien Holden Institute of Optometry and Vision Sciences, L V Prasad Eye Institute, Hyderabad, India
- Brien Holden Eye Research Centre, L V Prasad Eye Institute, Hyderabad, India
| | - Joshua A. Solomon
- Centre for Applied Vision Research, City, University of London, London, United Kingdom
| | - Christopher W. Tyler
- Centre for Applied Vision Research, City, University of London, London, United Kingdom
| | - Tarjani V. Dave
- Hariram Motumal Nasta and Renu Hariram Nasta Ophthalmic Plastic Surgery Service, L V Prasad Eye Institute, Hyderabad, India
| | - Swathi Kaliki
- Operation Eyesight Universal Institute for Eye Cancer, L V Prasad Eye Institute, Hyderabad, India
| | - Shrikant R. Bharadwaj
- Brien Holden Institute of Optometry and Vision Sciences, L V Prasad Eye Institute, Hyderabad, India
- Brien Holden Eye Research Centre, L V Prasad Eye Institute, Hyderabad, India
| |
Collapse
|
2
|
Linton P, Morgan MJ, Read JCA, Vishwanath D, Creem-Regehr SH, Domini F. New Approaches to 3D Vision. Philos Trans R Soc Lond B Biol Sci 2023; 378:20210443. [PMID: 36511413 PMCID: PMC9745878 DOI: 10.1098/rstb.2021.0443] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2022] [Accepted: 10/25/2022] [Indexed: 12/15/2022] Open
Abstract
New approaches to 3D vision are enabling new advances in artificial intelligence and autonomous vehicles, a better understanding of how animals navigate the 3D world, and new insights into human perception in virtual and augmented reality. Whilst traditional approaches to 3D vision in computer vision (SLAM: simultaneous localization and mapping), animal navigation (cognitive maps), and human vision (optimal cue integration) start from the assumption that the aim of 3D vision is to provide an accurate 3D model of the world, the new approaches to 3D vision explored in this issue challenge this assumption. Instead, they investigate the possibility that computer vision, animal navigation, and human vision can rely on partial or distorted models or no model at all. This issue also highlights the implications for artificial intelligence, autonomous vehicles, human perception in virtual and augmented reality, and the treatment of visual disorders, all of which are explored by individual articles. This article is part of a discussion meeting issue 'New approaches to 3D vision'.
Collapse
Affiliation(s)
- Paul Linton
- Presidential Scholars in Society and Neuroscience, Center for Science and Society, Columbia University, New York, NY 10027, USA
- Italian Academy for Advanced Studies in America, Columbia University, New York, NY 10027, USA
- Visual Inference Lab, Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY 10027, USA
| | - Michael J. Morgan
- Department of Optometry and Visual Sciences, City, University of London, Northampton Square, London EC1V 0HB, UK
| | - Jenny C. A. Read
- Biosciences Institute, Newcastle University, Newcastle upon Tyne, Tyne & Wear NE2 4HH, UK
| | - Dhanraj Vishwanath
- School of Psychology and Neuroscience, University of St Andrews, St Andrews, Fife KY16 9JP, UK
| | | | - Fulvio Domini
- Department of Cognitive, Linguistic, and Psychological Sciences, Brown University, Providence, RI 02912-9067, USA
| |
Collapse
|
3
|
Abstract
Shape is an interesting property of objects because it is used in ordinary discourse in ways that seem to have little connection to how it is typically defined in mathematics. The present article describes how the concept of shape can be grounded within Euclidean and non-Euclidean geometry and also to human perception. It considers the formal methods that have been proposed for measuring the differences among shapes and how the performance of those methods compares with shape difference thresholds of human observers. It discusses how different types of shape change can be perceptually categorized. It also evaluates the specific data structures that have been used to represent shape in models of both human and machine vision, and it reviews the psychophysical evidence about the extent to which those models are consistent with human perception. Based on this review of the literature, we argue that shape is not one thing but rather a collection of many object attributes, some of which are more perceptually salient than others. Because the relative importance of these attributes can be context dependent, there is no obvious single definition of shape that is universally applicable in all situations.
Collapse
Affiliation(s)
- James T Todd
- Department of Psychology, The Ohio State University, Columbus, OH, USA
| | | |
Collapse
|
4
|
Scarfe P. Experimentally disambiguating models of sensory cue integration. J Vis 2022; 22:5. [PMID: 35019955 PMCID: PMC8762719 DOI: 10.1167/jov.22.1.5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Sensory cue integration is one of the primary areas in which a normative mathematical framework has been used to define the “optimal” way in which to make decisions based upon ambiguous sensory information and compare these predictions to behavior. The conclusion from such studies is that sensory cues are integrated in a statistically optimal fashion. However, numerous alternative computational frameworks exist by which sensory cues could be integrated, many of which could be described as “optimal” based on different criteria. Existing studies rarely assess the evidence relative to different candidate models, resulting in an inability to conclude that sensory cues are integrated according to the experimenter's preferred framework. The aims of the present paper are to summarize and highlight the implicit assumptions rarely acknowledged in testing models of sensory cue integration, as well as to introduce an unbiased and principled method by which to determine, for a given experimental design, the probability with which a population of observers behaving in accordance with one model of sensory integration can be distinguished from the predictions of a set of alternative models.
Collapse
Affiliation(s)
- Peter Scarfe
- Vision and Haptics Laboratory, School of Psychology and Clinical Language Sciences, University of Reading, Reading, UK.,
| |
Collapse
|
5
|
Cue vetoing in depth estimation: Physical and virtual stimuli. Vision Res 2021; 188:51-64. [PMID: 34289419 DOI: 10.1016/j.visres.2021.07.003] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2021] [Revised: 06/17/2021] [Accepted: 07/05/2021] [Indexed: 11/21/2022]
Abstract
Motion parallax and binocular disparity contribute to the perceived depth of three-dimensional (3D) objects. However, depth is often misperceived, even when both cues are available. This may be due in part to conflicts with unmodelled cues endemic to computerized displays. Here we evaluated the impact of display-based cue conflicts on depth cue integration by comparing perceived depth for physical and virtual objects. Truncated square pyramids were rendered using Blender and 3D printed. We assessed perceived depth using a discrimination task with motion parallax, binocular disparity, and their combination. Physical stimuli were presented with precise control over position and lighting. Virtual stimuli were viewed using a head-mounted display. To generate motion parallax, observers made lateral head movements using a chin rest on a motion platform. Observers indicated if the width of the front face appeared greater or less than the distance between this surface and the base. We found that accuracy was similar for virtual and physical pyramids. All estimates were more precise when depth was defined by binocular disparity than motion parallax. Our probabilistic model shows that a linear combination model does not adequately describe performance in either physical or virtual conditions. While there was inter-observer variability in weights, performance in all conditions was best predicted by a veto model that excludes the less reliable depth cue, in this case motion parallax.
Collapse
|
6
|
Campagnoli C, Domini F, Taylor JA. Taking aim at the perceptual side of motor learning: exploring how explicit and implicit learning encode perceptual error information through depth vision. J Neurophysiol 2021; 126:413-426. [PMID: 34161173 DOI: 10.1152/jn.00153.2021] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Motor learning in visuomotor adaptation tasks results from both explicit and implicit processes, each responding differently to an error signal. Although the motor output side of these processes has been extensively studied, the visual input side is relatively unknown. We investigated if and how depth perception affects the computation of error information by explicit and implicit motor learning. Two groups of participants made reaching movements to bring a virtual cursor to a target in the frontoparallel plane. The Delayed group was allowed to reaim and their feedback was delayed to emphasize explicit learning, whereas the camped group received task-irrelevant clamped cursor feedback and continued to aim straight at the target to emphasize implicit adaptation. Both groups played this game in a highly detailed virtual environment (depth condition), leveraging a cover task of playing darts in a virtual tavern, and in an empty environment (no-depth condition). The delayed group showed an increase in error sensitivity under depth relative to no-depth. In contrast, the clamped group adapted to the same degree under both conditions. The movement kinematics of the delayed participants also changed under the depth condition, consistent with the target appearing more distant, unlike the Clamped group. A comparison of the delayed behavioral data with a perceptual task from the same individuals showed that the greater reaiming in the depth condition was consistent with an increase in the scaling of the error distance and size. These findings suggest that explicit and implicit learning processes may rely on different sources of perceptual information.NEW & NOTEWORTHY We leveraged a classic sensorimotor adaptation task to perform a first systematic assessment of the role of perceptual cues in the estimation of an error signal in the 3-D space during motor learning. We crossed two conditions presenting different amounts of depth information, with two manipulations emphasizing explicit and implicit learning processes. Explicit learning responded to the visual conditions, consistent with perceptual reports, whereas implicit learning appeared to be independent of them.
Collapse
Affiliation(s)
- Carlo Campagnoli
- Department of Psychology, Princeton University, Princeton, New Jersey
| | - Fulvio Domini
- Department of Cognitive, Linguistic and Psychological Sciences, Brown University, Providence, Rhode Island
| | - Jordan A Taylor
- Department of Psychology, Princeton University, Princeton, New Jersey
| |
Collapse
|
7
|
Scarfe P, Glennerster A. Combining cues to judge distance and direction in an immersive virtual reality environment. J Vis 2021; 21:10. [PMID: 33900366 PMCID: PMC8083085 DOI: 10.1167/jov.21.4.10] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2019] [Accepted: 01/31/2021] [Indexed: 11/24/2022] Open
Abstract
When we move, the visual direction of objects in the environment can change substantially. Compared with our understanding of depth perception, the problem the visual system faces in computing this change is relatively poorly understood. Here, we tested the extent to which participants' judgments of visual direction could be predicted by standard cue combination rules. Participants were tested in virtual reality using a head-mounted display. In a simulated room, they judged the position of an object at one location, before walking to another location in the room and judging, in a second interval, whether an object was at the expected visual direction of the first. By manipulating the scale of the room across intervals, which was subjectively invisible to observers, we put two classes of cue into conflict, one that depends only on visual information and one that uses proprioceptive information to scale any reconstruction of the scene. We find that the sensitivity to changes in one class of cue while keeping the other constant provides a good prediction of performance when both cues vary, consistent with the standard cue combination framework. Nevertheless, by comparing judgments of visual direction with those of distance, we show that judgments of visual direction and distance are mutually inconsistent. We discuss why there is no need for any contradiction between these two conclusions.
Collapse
|
8
|
Abstract
With the increase in popularity of consumer virtual reality headsets, for research and other applications, it is important to understand the accuracy of 3D perception in VR. We investigated the perceptual accuracy of near-field virtual distances using a size and shape constancy task, in two commercially available devices. Participants wore either the HTC Vive or the Oculus Rift and adjusted the size of a virtual stimulus to match the geometric qualities (size and depth) of a physical stimulus they were able to refer to haptically. The judgments participants made allowed for an indirect measure of their perception of the egocentric, virtual distance to the stimuli. The data show under-constancy and are consistent with research from carefully calibrated psychophysical techniques. There was no difference in the degree of constancy found in the two headsets. We conclude that consumer virtual reality headsets provide a sufficiently high degree of accuracy in distance perception, to allow them to be used confidently in future experimental vision science, and other research applications in psychology.
Collapse
|
9
|
Muryy A, Siddharth N, Nardelli N, Glennerster A, Torr PHS. Lessons from reinforcement learning for biological representations of space. Vision Res 2020; 174:79-93. [PMID: 32683096 DOI: 10.1016/j.visres.2020.05.009] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2019] [Revised: 04/26/2020] [Accepted: 05/26/2020] [Indexed: 10/23/2022]
Abstract
Neuroscientists postulate 3D representations in the brain in a variety of different coordinate frames (e.g. 'head-centred', 'hand-centred' and 'world-based'). Recent advances in reinforcement learning demonstrate a quite different approach that may provide a more promising model for biological representations underlying spatial perception and navigation. In this paper, we focus on reinforcement learning methods that reward an agent for arriving at a target image without any attempt to build up a 3D 'map'. We test the ability of this type of representation to support geometrically consistent spatial tasks such as interpolating between learned locations using decoding of feature vectors. We introduce a hand-crafted representation that has, by design, a high degree of geometric consistency and demonstrate that, in this case, information about the persistence of features as the camera translates (e.g. distant features persist) can improve performance on the geometric tasks. These examples avoid Cartesian (in this case, 2D) representations of space. Non-Cartesian, learned representations provide an important stimulus in neuroscience to the search for alternatives to a 'cognitive map'.
Collapse
Affiliation(s)
- Alex Muryy
- School of Psychology and Clinical Language Sciences, University of Reading, UK
| | - N Siddharth
- Department of Engineering Science, University of Oxford, UK
| | | | - Andrew Glennerster
- School of Psychology and Clinical Language Sciences, University of Reading, UK.
| | | |
Collapse
|
10
|
Nityananda V, Read JCA. Stereopsis in animals: evolution, function and mechanisms. ACTA ACUST UNITED AC 2018; 220:2502-2512. [PMID: 28724702 PMCID: PMC5536890 DOI: 10.1242/jeb.143883] [Citation(s) in RCA: 51] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022]
Abstract
Stereopsis is the computation of depth information from views acquired simultaneously from different points in space. For many years, stereopsis was thought to be confined to primates and other mammals with front-facing eyes. However, stereopsis has now been demonstrated in many other animals, including lateral-eyed prey mammals, birds, amphibians and invertebrates. The diversity of animals known to have stereo vision allows us to begin to investigate ideas about its evolution and the underlying selective pressures in different animals. It also further prompts the question of whether all animals have evolved essentially the same algorithms to implement stereopsis. If so, this must be the best way to do stereo vision, and should be implemented by engineers in machine stereopsis. Conversely, if animals have evolved a range of stereo algorithms in response to different pressures, that could inspire novel forms of machine stereopsis appropriate for distinct environments, tasks or constraints. As a first step towards addressing these ideas, we here review our current knowledge of stereo vision in animals, with a view towards outlining common principles about the evolution, function and mechanisms of stereo vision across the animal kingdom. We conclude by outlining avenues for future work, including research into possible new mechanisms of stereo vision, with implications for machine vision and the role of stereopsis in the evolution of camouflage. Summary: Stereopsis has evolved independently in different animals. We review the various functions it serves and the variety of mechanisms that could underlie stereopsis in different species.
Collapse
Affiliation(s)
- Vivek Nityananda
- Wissenschaftskolleg zu Berlin, Institute for Advanced Study, Wallotstraße 19, Berlin 14193, Germany .,Newcastle University, Institute of Neuroscience, Henry Wellcome Building, Framlington Place, Newcastle Upon Tyne NE2 4HH, UK
| | - Jenny C A Read
- Newcastle University, Institute of Neuroscience, Henry Wellcome Building, Framlington Place, Newcastle Upon Tyne NE2 4HH, UK
| |
Collapse
|
11
|
Qian J, Petrov Y. A depth illusion supports the model of General Object Constancy: Size and depth constancies related by a same distance-scaling factor. Vision Res 2016; 129:77-86. [PMID: 27810350 DOI: 10.1016/j.visres.2016.09.015] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2016] [Revised: 09/16/2016] [Accepted: 09/16/2016] [Indexed: 11/28/2022]
Abstract
Perceptual constancy refers to the ability to stabilize the representation of an object even though the retinal image of the object undergoes variations. In previous studies, we proposed a General Object Constancy (GOC) hypothesis to demonstrate a common stabilization mechanism for perception of an object's features, such as size, contrast and depth, as the perceived distance varies. In the present study, we report another depth illusion supporting the GOC model. The stimuli comprised pairs of disks moving in a pattern of radial optic flow. Each pair consisted of a white disk positioned upper left to a dark disk, creating a percept of the white disk casting a shadow. As the pairs contracted towards the center of the screen in accordance with motion away from the observer, the two disks in each pair appeared to increase in contrast and separate farther away from each other both in the fronto-parallel plane (angular separation illusion) and in depth (depth separation illusion). While the contrast illusion and the angular separation illusion, which is a variant of the size illusion, replicated our previous findings, the illusion of depth separation revealed a depth constancy phenomenon. We further confirmed that the size and depth perception were related, e.g., the depth separation and the angular separation illusions were highly correlated across observers. Whereas the illusory increase in the angular separation between a disk and its 'shadow' could not be canceled by modulation of depth, decreasing the angular separation could offset the illusory increase in depth separation. The results can be explained by the GOC hypothesis: the visual system uses the same scaling factor to account for contrast, size (angular separation), and depth variations with distance; additionally, the perceived size of the object is used to scale its depth and contrast signals in order to achieve constancy.
Collapse
Affiliation(s)
- Jiehui Qian
- Department of Psychology, Sun Yat-Sen University, Guangzhou, China.
| | | |
Collapse
|
12
|
Bozzacchi C, Volcic R, Domini F. Grasping in absence of feedback: systematic biases endure extensive training. Exp Brain Res 2015; 234:255-65. [PMID: 26449965 DOI: 10.1007/s00221-015-4456-9] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2015] [Accepted: 09/25/2015] [Indexed: 11/26/2022]
Abstract
Reach-to-grasp movements performed without visual and haptic feedback of the hand are subject to systematic inaccuracies. Grasps directed at an object specified by binocular information usually end at the wrong distance with an incorrect final grip aperture. More specifically, moving the target object away from the observer leads to increasingly larger undershoots and smaller grip apertures. These systematic biases suggest that the visuomotor mapping is based on inaccurate estimates of an object's egocentric distance and 3D structure that compress the visual space. Here we ask whether the appropriate visuomotor mapping can be learned through an extensive exposure to trials where haptic and visual feedback of the hand is provided. By intermixing feedback trials with test trials without feedback, we aimed at maximizing the likelihood that the motor execution of test trials is positively influenced by that of preceding feedback trials. We found that the intermittent presence of feedback trials both (1) largely reduced the positioning error of the hand with respect to the object and (2) affected the shaping of the hand before the final grasp, leading to an overall more accurate performance. While this demonstrates an effective transfer of information from feedback trials to test trials, the remaining biases indicate that a compression of visual space is still taking place. The correct visuomotor mapping, therefore, could not be learned. We speculate that an accurate reconstruction of the scene at movement onset may not actually be needed. Instead, the online monitoring of the hand position relative to the object and the final contact with the object are sufficient for a successful execution of a grasp.
Collapse
Affiliation(s)
- Chiara Bozzacchi
- Center for Neuroscience and Cognitive Systems@UniTn, Istituto Italiano di Tecnologia (IIT), Corso Bettini, 31, 38068, Rovereto, TN, Italy.
| | - Robert Volcic
- Center for Neuroscience and Cognitive Systems@UniTn, Istituto Italiano di Tecnologia (IIT), Corso Bettini, 31, 38068, Rovereto, TN, Italy
| | - Fulvio Domini
- Center for Neuroscience and Cognitive Systems@UniTn, Istituto Italiano di Tecnologia (IIT), Corso Bettini, 31, 38068, Rovereto, TN, Italy
- Department of Cognitive, Linguistic and Psychological Sciences, Brown University, Providence, RI, 02912, USA
| |
Collapse
|
13
|
Semi-Autonomous Domestic Service Robots: Evaluation of a User Interface for Remote Manipulation and Navigation With Focus on Effects of Stereoscopic Display. Int J Soc Robot 2014. [DOI: 10.1007/s12369-014-0266-7] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
14
|
Nawrot M, Ratzlaff M, Leonard Z, Stroyan K. Modeling depth from motion parallax with the motion/pursuit ratio. Front Psychol 2014; 5:1103. [PMID: 25339926 PMCID: PMC4186274 DOI: 10.3389/fpsyg.2014.01103] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2014] [Accepted: 09/11/2014] [Indexed: 11/13/2022] Open
Abstract
The perception of unambiguous scaled depth from motion parallax relies on both retinal image motion and an extra-retinal pursuit eye movement signal. The motion/pursuit ratio represents a dynamic geometric model linking these two proximal cues to the ratio of depth to viewing distance. An important step in understanding the visual mechanisms serving the perception of depth from motion parallax is to determine the relationship between these stimulus parameters and empirically determined perceived depth magnitude. Observers compared perceived depth magnitude of dynamic motion parallax stimuli to static binocular disparity comparison stimuli at three different viewing distances, in both head-moving and head-stationary conditions. A stereo-viewing system provided ocular separation for stereo stimuli and monocular viewing of parallax stimuli. For each motion parallax stimulus, a point of subjective equality (PSE) was estimated for the amount of binocular disparity that generates the equivalent magnitude of perceived depth from motion parallax. Similar to previous results, perceived depth from motion parallax had significant foreshortening. Head-moving conditions produced even greater foreshortening due to the differences in the compensatory eye movement signal. An empirical version of the motion/pursuit law, termed the empirical motion/pursuit ratio, which models perceived depth magnitude from these stimulus parameters, is proposed.
Collapse
Affiliation(s)
- Mark Nawrot
- Department of Psychology, Center for Visual and Cognitive Neuroscience, North Dakota State University Fargo, ND, USA
| | - Michael Ratzlaff
- Department of Psychology, Center for Visual and Cognitive Neuroscience, North Dakota State University Fargo, ND, USA
| | - Zachary Leonard
- Department of Psychology, Center for Visual and Cognitive Neuroscience, North Dakota State University Fargo, ND, USA
| | - Keith Stroyan
- Math Department, University of Iowa Iowa City, IA, USA
| |
Collapse
|
15
|
Chen KB, Kimmel RA, Bartholomew A, Ponto K, Gleicher ML, Radwin RG. Manually locating physical and virtual reality objects. HUMAN FACTORS 2014; 56:1163-1176. [PMID: 25277024 DOI: 10.1177/0018720814523067] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
OBJECTIVE In this study, we compared how users locate physical and equivalent three-dimensional images of virtual objects in a cave automatic virtual environment (CAVE) using the hand to examine how human performance (accuracy, time, and approach) is affected by object size, location, and distance. BACKGROUND Virtual reality (VR) offers the promise to flexibly simulate arbitrary environments for studying human performance. Previously, VR researchers primarily considered differences between virtual and physical distance estimation rather than reaching for close-up objects. METHOD Fourteen participants completed manual targeting tasks that involved reaching for corners on equivalent physical and virtual boxes of three different sizes. Predicted errors were calculated from a geometric model based on user interpupillary distance, eye location, distance from the eyes to the projector screen, and object. RESULTS Users were 1.64 times less accurate (p < .001) and spent 1.49 times more time (p = .01) targeting virtual versus physical box corners using the hands. Predicted virtual targeting errors were on average 1.53 times (p < .05) greater than the observed errors for farther virtual targets but not significantly different for close-up virtual targets. CONCLUSION Target size, location, and distance, in addition to binocular disparity, affected virtual object targeting inaccuracy. Observed virtual box inaccuracy was less than predicted for farther locations, suggesting possible influence of cues other than binocular vision. APPLICATION Human physical interaction with objects in VR for simulation, training, and prototyping involving reaching and manually handling virtual objects in a CAVE are more accurate than predicted when locating farther objects.
Collapse
|
16
|
Vienne C, Blondé L, Mamassian P. Perceived rigidity in motion-in-depth increases with contour perspective. Perception 2014; 43:481-98. [PMID: 25154283 DOI: 10.1068/p7642] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Abstract
When observers are asked to match the depth of an object according to its height, they often report systematic errors depending on viewing distance. Systematic biases can also arise while vergence distances are induced by binocular disparities. Observers of stereoscopic images tend to overestimate the depth of objects displayed in front of the screen, while the depth of objects displayed behind the screen plane is underestimated. This phenomenon creates a serious problem in that veridicality in depth perception appears distorted when one attempts to render the metrics of a captured 3-D world. These distortions could also subsist with structure-from-motion information and during motion-in-depth. Observers judged the circularity of transparent rotating cylinders that were either static or moving in depth. Crossed results show that participants could precisely retrieve the best modulation between presented depth and width. As this effect could be amplified with stimuli containing stronger perspective cues (ie contour perspective), participants judged the rigidity of spinning cubes, moving along the line of sight, which were either edges-defined or defined by randomly textured surfaces (dots). The results showed that, although depth constancy was not improved by contour perspective, perceived rigidity was increased by perspective when the best scaling estimate was displayed. This finding suggests that appropriate binocular disparity information in combination to monocular signal is necessary for stereoscopic depth perception.
Collapse
|
17
|
Scarfe P, Hibbard PB. Reverse correlation reveals how observers sample visual information when estimating three-dimensional shape. Vision Res 2013; 86:115-27. [PMID: 23665429 DOI: 10.1016/j.visres.2013.04.016] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2012] [Revised: 03/14/2013] [Accepted: 04/24/2013] [Indexed: 11/18/2022]
Abstract
Human observers exhibit large systematic distance-dependent biases when estimating the three-dimensional (3D) shape of objects defined by binocular image disparities. This has led some to question the utility of disparity as a cue to 3D shape and whether accurate estimation of 3D shape is at all possible. Others have argued that accurate perception is possible, but only with large continuous perspective transformations of an object. Using a stimulus that is known to elicit large distance-dependent perceptual bias (random dot stereograms of elliptical cylinders) we show that contrary to these findings the simple adoption of a more naturalistic viewing angle completely eliminates this bias. Using behavioural psychophysics, coupled with a novel surface-based reverse correlation methodology, we show that it is binocular edge and contour information that allows for accurate and precise perception and that observers actively exploit and sample this information when it is available.
Collapse
Affiliation(s)
- Peter Scarfe
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB1 5QJ, United Kingdom.
| | | |
Collapse
|
18
|
Hibbard PB, Goutcher R, O'Kane LM, Scarfe P. Misperception of aspect ratio in binocularly viewed surfaces. Vision Res 2012; 70:34-43. [PMID: 22925917 DOI: 10.1016/j.visres.2012.08.003] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2012] [Revised: 08/03/2012] [Accepted: 08/07/2012] [Indexed: 10/28/2022]
Abstract
The horizontal-vertical illusion, in which the vertical dimension is overestimated relative to the horizontal direction, has been explained in terms of the statistical relationship between the lengths of lines in the world, and the lengths of their projections onto the retina (Howe & Purves, 2002). The current study shows that this illusion affects the apparent aspect ratio of shapes, and investigates how it interacts with binocular cues to surface slant. One way in which statistical information could give rise to the horizontal-vertical illusion would be through prior assumptions about the distribution of slant. This prior would then be expected to interact with retinal cues to slant. We determined the aspect ratio of stereoscopically viewed ellipses that appeared circular. We show that observers' judgements of aspect ratio were affected by surface slant, but that the largest image vertical:horizontal aspect ratio that was considered to be a surface with a circular profile was always found for surfaces close to fronto-parallel. This is not consistent with a Bayesian model in which the horizontal-vertical illusion arises from a non-uniform prior probability distribution for slant. Rather, we suggest that assumptions about the slant of surfaces affect apparent aspect ratio in a manner that is more heuristic, and partially dissociated from apparent slant.
Collapse
Affiliation(s)
- Paul B Hibbard
- School of Psychology, University of St. Andrews, St. Mary's Quad, St. Andrews, Fife, UK.
| | | | | | | |
Collapse
|
19
|
Stroyan K, Nawrot M. Visual depth from motion parallax and eye pursuit. J Math Biol 2012; 64:1157-88. [PMID: 21695531 PMCID: PMC3348271 DOI: 10.1007/s00285-011-0445-1] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2011] [Revised: 05/26/2011] [Indexed: 10/18/2022]
Abstract
A translating observer viewing a rigid environment experiences "motion parallax", the relative movement upon the observer's retina of variously positioned objects in the scene. This retinal movement of images provides a cue to the relative depth of objects in the environment, however retinal motion alone cannot mathematically determine relative depth of the objects. Visual perception of depth from lateral observer translation uses both retinal image motion and eye movement. In Nawrot and Stroyan (Vision Res 49:1969-1978, 2009) we showed mathematically that the ratio of the rate of retinal motion over the rate of smooth eye pursuit mathematically determines depth relative to the fixation point in central vision. We also reported on psychophysical experiments indicating that this ratio is the important quantity for perception. Here we analyze the motion/pursuit cue for the more general, and more complicated, case when objects are distributed across the horizontal viewing plane beyond central vision. We show how the mathematical motion/pursuit cue varies with different points across the plane and with time as an observer translates. If the time varying retinal motion and smooth eye pursuit are the only signals used for this visual process, it is important to know what is mathematically possible to derive about depth and structure. Our analysis shows that the motion/pursuit ratio determines an excellent description of depth and structure in these broader stimulus conditions, provides a detailed quantitative hypothesis of these visual processes for the perception of depth and structure from motion parallax, and provides a computational foundation to analyze the dynamic geometry of future experiments.
Collapse
Affiliation(s)
- Keith Stroyan
- Mathematics Department, University of Iowa, Iowa City, IA, 52242, USA.
| | | |
Collapse
|
20
|
Sherman A, Papathomas TV, Jain A, Keane BP. The Role of Stereopsis, Motion Parallax, Perspective and Angle Polarity in Perceiving 3-D Shape. ACTA ACUST UNITED AC 2012; 25:263-85. [DOI: 10.1163/187847511x576802] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
21
|
Domini F, Shah R, Caudek C. Do we perceive a flattened world on the monitor screen? Acta Psychol (Amst) 2011; 138:359-66. [PMID: 21986481 DOI: 10.1016/j.actpsy.2011.07.007] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2011] [Revised: 07/27/2011] [Accepted: 07/29/2011] [Indexed: 11/28/2022] Open
Abstract
The current model of three-dimensional perception hypothesizes that the brain integrates the depth cues in a statistically optimal fashion through a weighted linear combination with weights proportional to the reliabilities obtained for each cue in isolation (Landy, Maloney, Johnston, & Young, 1995). Even though many investigations support such theoretical framework, some recent empirical findings are at odds with this view (e.g., Domini, Caudek, & Tassinari, 2006). Failures of linear cue integration have been attributed to cue-conflict and to unmodelled cues to flatness present in computer-generated displays. We describe two cue-combination experiments designed to test the integration of stereo and motion cues, in the presence of consistent or conflicting blur and accommodation information (i.e., when flatness cues are either absent, with physical stimuli, or present, with computer-generated displays). In both conditions, we replicated the results of Domini et al. (2006): The amount of perceived depth increased as more cues were available, also producing an over-estimation of depth in some conditions. These results can be explained by the Intrinsic Constraint model, but not by linear cue combination.
Collapse
Affiliation(s)
- Fulvio Domini
- Department of Cognitive, Linguistic & Psychological Sciences, Brown University, Providence, RI 02912, USA.
| | | | | |
Collapse
|
22
|
Gillam B, Palmisano SA, Govan DG. Depth interval estimates from motion parallax and binocular disparity beyond interaction space. Perception 2011; 40:39-49. [PMID: 21513183 DOI: 10.1068/p6868] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
Abstract
Static and dynamic observers provided binocular and monocular estimates of the depths between real objects lying well beyond interaction space. On each trial, pairs of LEDs were presented inside a dark railway tunnel. The nearest LED was always 40 m from the observer, with the depth separation between LED pairs ranging from 0 up to 248 m. Dynamic binocular viewing was found to produce the greatest (ie most veridical) estimates of depth magnitude, followed next by static binocular viewing, and then by dynamic monocular viewing. (No significant depth was seen with static monocular viewing.) We found evidence that both binocular and monocular dynamic estimates of depth were scaled for the observation distance when the ground plane and walls of the tunnel were visible up to the nearest LED. We conclude that both motion parallax and stereopsis provide useful long-distance depth information and that motion-parallax information can enhance the degree of stereoscopic depth seen.
Collapse
Affiliation(s)
- Barbara Gillam
- School of Psychology, University of New South Wales, Sydney, NSW 2052, Australia.
| | | | | |
Collapse
|
23
|
Integration of disparity and velocity information for haptic and perceptual judgments of object depth. Acta Psychol (Amst) 2011; 136:300-10. [PMID: 21237442 DOI: 10.1016/j.actpsy.2010.12.003] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/16/2010] [Revised: 12/09/2010] [Accepted: 12/10/2010] [Indexed: 11/23/2022] Open
Abstract
Do reach-to-grasp (prehension) movements require a metric representation of three-dimensional (3D) layouts and objects? We propose a model relying only on direct sensory information to account for the planning and execution of prehension movements in the absence of haptic feedback and when the hand is not visible. In the present investigation, we isolate relative motion and binocular disparity information from other depth cues and we study their efficacy for reach-to-grasp movements and visual judgments. We show that (i) the amplitude of the grasp increases when relative motion is added to binocular disparity information, even if depth from disparity information is already veridical, and (ii) similar distortions of derived depth are found for haptic tasks and perceptual judgments. With a quantitative test, we demonstrate that our results are consistent with the Intrinsic Constraint model and do not require 3D metric inferences (Domini, Caudek, & Tassinari, 2006). By contrast, the linear cue integration model (Landy, Maloney, Johnston, & Young, 1995) cannot explain the present results, even if the flatness cues are taken into account.
Collapse
|
24
|
Lee YL, Bingham GP. Large perspective changes yield perception of metric shape that allows accurate feedforward reaches-to-grasp and it persists after the optic flow has stopped! Exp Brain Res 2010; 204:559-73. [PMID: 20563715 DOI: 10.1007/s00221-010-2323-2] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2009] [Accepted: 06/02/2010] [Indexed: 11/25/2022]
Abstract
Lee et al. (Percept Psychophys 70:1032-1046, 2008a) investigated whether visual perception of metric shape could be calibrated when used to guide feedforward reaches-to-grasp. It could not. Seated participants viewed target objects (elliptical cylinders) in normal lighting using stereo vision and free head movements that allowed small (approximately 10 degrees) perspective changes. The authors concluded that poor perception of metric shape was the reason reaches-to-grasp should be visually guided online. However, Bingham and Lind (Percept Psychophys 70:524-540, 2008) showed that large perspective changes (> or =45 degrees) yield good perception of metric shape. So, now we repeated the Lee et al.'s study with the addition of information from large perspective changes. The results were accurate feedforward reaches-to-grasp reflecting accurate perception of both metric shape and metric size. Large perspective changes occur when one locomotes into a workspace in which reaches-to-grasp are subsequently performed. Does the resulting perception of metric shape persist after the large perspective changes have ceased? Experiments 2 and 3 tested reaches-to-grasp with delays (Exp. 2, 5-s delay; Exp. 3, approximately 16-s delay) and multiple objects to be grasped after a single viewing. Perception of metric shape and metric size persisted yielding accurate reaches-to-grasp. We advocate the study of nested actions using a dynamic approach to perception/action.
Collapse
|
25
|
The accuracy of metric judgements: Perception of surface normal. Vision Res 2010; 50:1140-57. [DOI: 10.1016/j.visres.2010.03.006] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2007] [Revised: 03/11/2010] [Accepted: 03/11/2010] [Indexed: 11/17/2022]
|
26
|
Di Luca M, Domini F, Caudek C. Inconsistency of perceived 3D shape. Vision Res 2010; 50:1519-31. [PMID: 20470815 DOI: 10.1016/j.visres.2010.05.006] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/05/2010] [Revised: 05/05/2010] [Accepted: 05/05/2010] [Indexed: 11/16/2022]
Abstract
Internal consistency of local depth, slant, and curvature judgments was studied by asking participants to match two 3D surfaces rendered by different mixtures of 3D cues (velocity, texture, and shading). We found that perceptual judgments were not consistent with each other, with cue-specific distortions. Adding multiple cues did not eliminate the inconsistencies of the judgments. These results can be predicted by the Intrinsic Constraint (IC) model according to which the perceptual metric local estimates are a monotonically increasing function of the Signal-to-Noise Ratio of the optimal combination of direct information of 3D shape (Domini, Caudek, & Tassinari, 2006).
Collapse
Affiliation(s)
- M Di Luca
- Max Planck Institute for Biological Cybernetics, Tuebingen, Germany
| | | | | |
Collapse
|
27
|
Domini F, Caudek C. Matching perceived depth from disparity and from velocity: Modeling and psychophysics. Acta Psychol (Amst) 2010; 133:81-9. [PMID: 19963200 DOI: 10.1016/j.actpsy.2009.10.003] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2009] [Revised: 10/06/2009] [Accepted: 10/08/2009] [Indexed: 10/20/2022] Open
Abstract
We asked observers to match in depth a disparity-only stimulus with a velocity-only stimulus. The observers' responses revealed systematic biases: the two stimuli appeared to be matched in depth when they were produced by the projection of different distal depth extents. We discuss two alternative models of depth recovery that could account for these results. (1) Depth matches could be obtained by scaling the image signals by constants not specified by optical information, and (2) depth matches could be obtained by equating the stimuli in terms of their signal-to-noise ratios (see Domini & Caudek, 2009). We show that the systematic failures of shape constancy revealed by observers' judgments are well accounted for by the hypothesis that the apparent depth of a stimulus is determined by the magnitude of the retinal signals relative to the uncertainty (i.e., internal noise) arising from the measurement of those signals.
Collapse
|
28
|
Nawrot M, Stroyan K. The motion/pursuit law for visual depth perception from motion parallax. Vision Res 2009; 49:1969-78. [PMID: 19463848 PMCID: PMC2735858 DOI: 10.1016/j.visres.2009.05.008] [Citation(s) in RCA: 30] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2008] [Revised: 04/23/2009] [Accepted: 05/04/2009] [Indexed: 11/21/2022]
Abstract
One of vision's most important functions is specification of the layout of objects in the 3D world. While the static optical geometry of retinal disparity explains the perception of depth from binocular stereopsis, we propose a new formula to link the pertinent dynamic geometry to the computation of depth from motion parallax. Mathematically, the ratio of retinal image motion (motion) and smooth pursuit of the eye (pursuit) provides the necessary information for the computation of relative depth from motion parallax. We show that this could have been obtained with the approaches of Nakayama and Loomis [Nakayama, K., & Loomis, J. M. (1974). Optical velocity patterns, velocity-sensitive neurons, and space perception: A hypothesis. Perception, 3, 63-80] or Longuet-Higgens and Prazdny [Longuet-Higgens, H. C., & Prazdny, K. (1980). The interpretation of a moving retinal image. Proceedings of the Royal Society of London Series B, 208, 385-397] by adding pursuit to their treatments. Results of a psychophysical experiment show that changes in the motion/pursuit ratio have a much better relationship to changes in the perception of depth from motion parallax than do changes in motion or pursuit alone. The theoretical framework provided by the motion/pursuit law provides the quantitative foundation necessary to study this fundamental visual depth perception ability.
Collapse
Affiliation(s)
- Mark Nawrot
- Center for Visual Neuroscience, Department of Psychology, North Dakota State University, Fargo, ND 58104, USA.
| | | |
Collapse
|
29
|
Tassinari H, Domini F. The intrinsic constraint model for stereo-motion integration. Perception 2008; 37:79-95. [PMID: 18399249 DOI: 10.1068/p5501] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Abstract
How the visual system integrates the information provided by several depth cues is central for vision research. Here, we present a model for how the human visual system combines disparity and velocity information. The model provides a depth interpretation to a subspace defined by the covariation of the two signals. We show that human performance is consistent with the predictions of the model, and compare them with those of another theoretical approach, the modified weak-fusion model. We discuss the validity of each approach as a model for human perception of 3-D shape from multiple cues to depth.
Collapse
Affiliation(s)
- Hadley Tassinari
- Department of Cognitive and Linguistic Sciences, Brown University, Box 1978, Providence, RI 02912, USA
| | | |
Collapse
|
30
|
Fernandez JM, Farell B. A neural model for the integration of stereopsis and motion parallax in structure-from-motion. Neurocomputing 2008; 71:1629-1641. [PMID: 19255615 DOI: 10.1016/j.neucom.2007.04.006] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
We introduce a model for the computation of structure-from-motion based on the physiology of visual cortical areas MT and MST. The model assumes that the perception of depth from motion is related to the firing of a subset of MT neurons tuned to both velocity and disparity. The model's MT neurons are connected to each other laterally to form modulatory receptive-field surrounds that are gated by feedback connections from area MST. This allows the building up of a depth map from motion in area MT, even in absence of disparity in the input. Depth maps from motion and from stereo are combined by a weighted average at a final stage. The model's predictions for the interaction between motion and stereo cues agree with previous psychophysical data, both when the cues are consistent with each other or when they are contradictory. In particular, the model shows nonlinearities as a result of early interactions between motion and stereo before their depth maps are averaged. The two cues interact in a way that represents an alternative to the "modified weak fusion" model of depth-cue combination.
Collapse
|
31
|
YAKUSHIJIN REIKO. Independence of sampling of motion parallax and binocular disparity in discrimination of depth. JAPANESE PSYCHOLOGICAL RESEARCH 2007. [DOI: 10.1111/j.1468-5884.2007.00351.x] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
32
|
Melmoth DR, Storoni M, Todd G, Finlay AL, Grant S. Dissociation between vergence and binocular disparity cues in the control of prehension. Exp Brain Res 2007; 183:283-98. [PMID: 17665181 DOI: 10.1007/s00221-007-1041-x] [Citation(s) in RCA: 43] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2006] [Accepted: 06/20/2007] [Indexed: 11/27/2022]
Abstract
Binocular vision provides important advantages for controlling reach-to-grasp movements. We examined the possible source(s) of these advantages by comparing prehension proficiency under four different binocular viewing conditions, created by randomly placing a neutral lens (control), an eight dioptre prism (Base In or Base Out) or a low-power (2.00-3.75 dioptre) Plus lens over the eye opposite the moving limb. The Base In versus Base Out prisms were intended to selectively alter vergence-specified distance (VSD) information, such that the targets appeared beyond or closer than their actual physical position, respectively. The Plus lens was individually tailored to reduce each subject's disparity sensitivity (to 400-800 arc s), while minimizing effects on distance estimation. In pre-testing, subjects pointed (without visual feedback) to mid-line targets at different distances, and produced the systematic directional errors expected of uncorrected movements programmed under each of the perturbed conditions. For the prehension tasks, subjects reached and precision grasped (with visual feedback available) cylindrical objects (two sizes and three locations), either following a 3 s preview in which to plan their actions or immediately after the object became visible. Viewing condition markedly affected performance, but the planning time allowed did not. Participants made the most errors suggesting premature collision with the object (shortest 'braking' times after peak deceleration; fastest velocity and widest grip at initial contact) under Base In prism viewing, consistent with over-reaching movements programmed to transport the hand beyond the actual target due to its 'further' VSD. Conversely, they produced the longest terminal reaches and grip closure times, with multiple corrections just before and after object contact, under the Plus lens (reduced disparity) condition. Base Out prism performance was intermediate between these two, with significant increases in additional forward movements during the transport end-phase, indicative of initial under-reaching in response to the target's 'nearer' VSD. Our findings suggest dissociations between the role of vergence and binocular disparity in natural prehension movements, with vergence contributing mainly to reach planning and high-grade disparity cues providing particular advantages for grasp-point selection during grip programming and application.
Collapse
Affiliation(s)
- Dean R Melmoth
- Department of Optometry and Visual Science, The Henry Wellcome Laboratories for Visual Sciences, City University, Northampton Square, London, EC1V 0HB, UK
| | | | | | | | | |
Collapse
|
33
|
Rushton SK, Duke PA. The use of direction and distance information in the perception of approach trajectory. Vision Res 2007; 47:899-912. [PMID: 17321562 DOI: 10.1016/j.visres.2006.11.019] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/03/2005] [Revised: 10/11/2006] [Accepted: 11/12/2006] [Indexed: 10/23/2022]
Abstract
A pair of projectiles travelling on parallel trajectories produce differing patterns of retinal motion when they originate at different distances. For an observer to recognise that the two trajectories are parallel she must "factor out" the effect of distance on retinal motion. The observer faces a similar problem when physically parallel trajectories originate at different lateral positions; here direction must be "factored out". We report the results of a series of experiments designed to determine if observers can do this. The observers' task was to judge whether the direction of travel of an approaching sphere (test trajectory) was to the left or right of parallel to a previously shown trajectory (reference trajectory). In the first set of experiments the reference and test trajectories started from different lateral positions. In the final experiment they started from different distances. From the pattern of judgements we determined a set of perceptually parallel trajectories. Perceptually parallel trajectories deviated significantly from physically parallel. We conclude that under circumstances comparable to those encountered when catching a ball in flight, observers do not have access to accurate estimates of trajectory direction.
Collapse
Affiliation(s)
- Simon K Rushton
- Centre for Vision Research and Department of Psychology, York University, 4700 Keele Street, North York, ON, Canada M3J 1P3.
| | | |
Collapse
|
34
|
Domini F, Caudek C, Tassinari H. Stereo and motion information are not independently processed by the visual system. Vision Res 2006; 46:1707-23. [PMID: 16412492 DOI: 10.1016/j.visres.2005.11.018] [Citation(s) in RCA: 42] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2004] [Revised: 11/10/2005] [Accepted: 11/15/2005] [Indexed: 11/16/2022]
Abstract
Many visual tasks are carried out by using multiple sources of sensory information to estimate environmental properties. In this paper, we present a model for how the visual system combines disparity and velocity information. We propose that, in a first stage of processing, the best possible estimate of the affine structure is obtained by computing a composite score from the disparity and velocity signals. In a second stage, a maximum likelihood Euclidean interpretation is assigned to the recovered affine structure. In two experiments, we show that human performance is consistent with the predictions of our model. The present results are also discussed in the framework of another theoretical approach of the depth cue combination process termed Modified Weak Fusion.
Collapse
Affiliation(s)
- Fulvio Domini
- Department of Cognitive and Linguistic Sciences, Brown University, USA.
| | | | | |
Collapse
|
35
|
Glennerster A, Tcheang L, Gilson SJ, Fitzgibbon AW, Parker AJ. Humans ignore motion and stereo cues in favor of a fictional stable world. Curr Biol 2006; 16:428-32. [PMID: 16488879 PMCID: PMC2833396 DOI: 10.1016/j.cub.2006.01.019] [Citation(s) in RCA: 45] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2005] [Revised: 01/06/2006] [Accepted: 01/09/2006] [Indexed: 11/17/2022]
Abstract
As we move through the world, our eyes acquire a sequence of images. The information from this sequence is sufficient to determine the structure of a three-dimensional scene, up to a scale factor determined by the distance that the eyes have moved. Previous evidence shows that the human visual system accounts for the distance the observer has walked and the separation of the eyes when judging the scale, shape, and distance of objects. However, in an immersive virtual-reality environment, observers failed to notice when a scene expanded or contracted, despite having consistent information about scale from both distance walked and binocular vision. This failure led to large errors in judging the size of objects. The pattern of errors cannot be explained by assuming a visual reconstruction of the scene with an incorrect estimate of interocular separation or distance walked. Instead, it is consistent with a Bayesian model of cue integration in which the efficacy of motion and disparity cues is greater at near viewing distances. Our results imply that observers are more willing to adjust their estimate of interocular separation or distance walked than to accept that the scene has changed in size.
Collapse
Affiliation(s)
- Andrew Glennerster
- Department of Physiology, Anatomy and Genetics, Sherrington Building, University of Oxford, Parks Road, Oxford OX1 3PT, United Kingdom.
| | | | | | | | | |
Collapse
|
36
|
Scarfe P, Hibbard PB. Disparity-defined objects moving in depth do not elicit three-dimensional shape constancy. Vision Res 2005; 46:1599-610. [PMID: 16364392 DOI: 10.1016/j.visres.2005.11.002] [Citation(s) in RCA: 17] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2005] [Revised: 11/01/2005] [Accepted: 11/03/2005] [Indexed: 11/27/2022]
Abstract
Observers generally fail to recover three-dimensional shape accurately from binocular disparity. Typically, depth is overestimated at near distances and underestimated at far distances [Johnston, E. B. (1991). Systematic distortions of shape from stereopsis. Vision Research, 31, 1351-1360]. A simple prediction from this is that disparity-defined objects should appear to expand in depth when moving towards the observer, and compress in depth when moving away. However, additional information is provided when an object moves from which 3D Euclidean shape can be recovered, be this through the addition of structure from motion information [Richards, W. (1985). Structure from stereo and motion. Journal of the Optical Society of America A, 2, 343-349], or the use of non-generic strategies [Todd, J. T., & Norman, J. F. (2003). The visual perception of 3-D shape from multiple cues: Are observers capable of perceiving metric structure? Perception and Psychophysics, 65, 31-47]. Here, we investigated shape constancy for objects moving in depth. We found that to be perceived as constant in shape, objects needed to contract in depth when moving toward the observer, and expand in depth when moving away, countering the effects of incorrect distance scaling (Johnston, 1991). This is a striking example of the failure of shape constancy, but one that is predicted if observers neither accurately estimate object distance in order to recover Euclidean shape, nor are able to base their responses on a simpler processing strategy.
Collapse
Affiliation(s)
- P Scarfe
- School of Psychology, University of St. Andrews, St. Andrews, Fife KY16 9JP, UK.
| | | |
Collapse
|
37
|
Melmoth DR, Grant S. Advantages of binocular vision for the control of reaching and grasping. Exp Brain Res 2005; 171:371-88. [PMID: 16323004 DOI: 10.1007/s00221-005-0273-x] [Citation(s) in RCA: 105] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2005] [Accepted: 10/20/2005] [Indexed: 10/25/2022]
Abstract
Theoretical considerations suggest that binocular information should provide advantages, compared to monocular viewing, for the planning and execution of natural reaching and grasping actions, but empirical support for this is quite equivocal. We have examined these predictions on a simple prehension task in which normal subjects reached, grasped and lifted isolated cylindrical household objects (two sizes, four locations) in a well-lit environment, using binocular vision or with one eye occluded. Various kinematic measures reflecting the programming and on-line control of the movements were quantified, in combination with analyses of different types of error occurring in the velocity, spatial path and grip aperture profiles of each trial. There was little consistent effect of viewing condition on the early phase of the reach, up to and including the peak deceleration, but all other aspects of performance were superior under binocular control. Subjects adopted a cautious approach when binocular information was unavailable: they extended the end phase of the reach and pre-shaped their hand with a wider grip aperture further away from the object. Despite these precautions, initial grip application was poorly coordinated with target contact and was inaccurately scaled to the objects' dimensions, with the subsequent post-contact phase of the grasp significantly more prolonged, error-prone and variable compared to binocular performance. These effects were obtained in two separate experiments in which the participants' performed the task under randomized or more predictable (blocked) viewing conditions. Our data suggest that binocular vision offers particular advantages for controlling the terminal reach and the grasp. We argue that these benefits derive from binocular disparity processing linked to changes in relative hand-target distance, and that this depth information is independently used to regulate the progress of the approaching hand and to guide the digits to the (pre-selected) contact points on the object, thereby ensuring that the grip is securely applied.
Collapse
Affiliation(s)
- Dean R Melmoth
- Department of Optometry and Visual Science, The Henry Wellcome Laboratories for Visual Sciences, City University, Northampton Square, EC1V 0HB, London, UK
| | | |
Collapse
|
38
|
Tcheang L, Gilson SJ, Glennerster A. Systematic distortions of perceptual stability investigated using immersive virtual reality. Vision Res 2005; 45:2177-89. [PMID: 15845248 PMCID: PMC2833395 DOI: 10.1016/j.visres.2005.02.006] [Citation(s) in RCA: 27] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2004] [Revised: 02/02/2005] [Accepted: 02/02/2005] [Indexed: 11/28/2022]
Abstract
Using an immersive virtual reality system, we measured the ability of observers to detect the rotation of an object when its movement was yoked to the observer's own translation. Most subjects had a large bias such that a static object appeared to rotate away from them as they moved. Thresholds for detecting target rotation were similar to those for an equivalent speed discrimination task carried out by static observers, suggesting that visual discrimination is the predominant limiting factor in detecting target rotation. Adding a stable visual reference frame almost eliminated the bias. Varying the viewing distance of the target had little effect, consistent with observers underestimating distance walked. However, accuracy of walking to a briefly presented visual target was high and not consistent with an underestimation of distance walked. We discuss implications for theories of a task-independent representation of visual space.
Collapse
Affiliation(s)
- Lili Tcheang
- University Laboratory of Physiology, Parks Road, Oxford, OX1 3PT
| | - Stuart J. Gilson
- University Laboratory of Physiology, Parks Road, Oxford, OX1 3PT
| | | |
Collapse
|
39
|
Norman JF, Clayton AM, Shular CF, Thompson SR. Aging and the perception of depth and 3-D shape from motion parallax. Psychol Aging 2005; 19:506-14. [PMID: 15383000 DOI: 10.1037/0882-7974.19.3.506] [Citation(s) in RCA: 42] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
The ability of younger and older observers to perceive 3-D shape and depth from motion parallax was investigated. In Experiment 1, the observers discriminated among differently curved 3-dimensional (3-D) surfaces in the presence of noise. In Experiment 2, the surfaces' shape was held constant and the amount of front-to-back depth was varied; the observers estimated the amount of depth they perceived. The effects of age were strongly task dependent. The younger observers' performance in Experiment 1 was almost 60% higher than that of the older observers. In contrast, no age effect was obtained in Experiment 2. Older observers can effectively perceive variations in depth from patterns of motion parallax, but their ability to discriminate 3-D shape is significantly compromised.
Collapse
Affiliation(s)
- J Farley Norman
- Department of Psychology, Western Kentucky University, Bowling Green, KY 42101, USA.
| | | | | | | |
Collapse
|
40
|
Muller C, Lankheet MJM, Van De Grind WA. Binocular correlation does not improve coherence detection for fronto-parallel motion. Vision Res 2004; 44:1961-9. [PMID: 15145689 DOI: 10.1016/j.visres.2004.03.013] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2002] [Revised: 11/04/2003] [Indexed: 11/16/2022]
Abstract
We studied the low-level interactions between motion coherence detection and binocular correlation detection. It is well-established that e.g. depth information from motion parallax and from binocular disparities is effectively integrated. The question we aimed to answer is whether such interactions also exist at the very first correlation level that both mechanisms might have in common. First we quantitatively compared motion coherence detection and binocular correlation detection using similar stimuli (random pixels arrays, RPAs) and the same noise masking paradigm (luminance signal to noise ratio, LSNR). This showed that human observers are much more sensitive to motion than to binocular correlation. Adding noise therefore has a much stronger effect on binocular correlation than on motion detection. Next we manipulated the shape of the stimulus aperture to equalize LSNR thresholds for motion and binocular correlation. Motion sensitivity could be progressively reduced by shortening the length of the motion path, while keeping the aperture area constant. Changing the shape of the aperture did not affect binocular correlation sensitivity. A 'balanced' stimulus, one with equal strengths of motion and binocular correlation signals was then used to study the mutual interactions. In accordance with previous results, motion was found to greatly facilitate binocular correlation. Binocular correlation, however did not facilitate motion detection. We conclude that interactions are asymmetrical; fronto-parallel motion is primarily detected monocularly and this information can then be used to facilitate binocular correlation, but binocular correlation cannot improve motion sensitivity.
Collapse
Affiliation(s)
- Chris Muller
- Functional Neurobiology and Helmholtz Research School, Utrecht University, Padualaan 8, 3584 CH Utrecht, The Netherlands
| | | | | |
Collapse
|
41
|
Champion RA, Brenner E, Mamassian P, Simmons DR. No evidence for sequential effects of the interaction of stereo and motion cues in judgements of perceived shape. Vision Res 2003; 44:483-7. [PMID: 14680774 DOI: 10.1016/j.visres.2003.10.003] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
Abstract
The interaction of the depth cues of binocular disparity and motion parallax could potentially be used by the visual system to recover an estimate of the viewing distance. The present study investigated whether an interaction of stereo and motion has effects that persist over time to influence the perception of shape from stereo when the motion information is removed. Static stereoscopic ellipsoids were presented following the presentation of rotating stereoscopic ellipsoids, which were located either at the same or a different viewing distance. It was predicted that shape judgements for static stimuli would be better after presentation of a rotating stimulus at the same viewing distance, than after presentation of one at a different viewing distance. No such difference was found. It was concluded that an interaction between stereo and motion depth cues does not influence the perception of subsequently presented static objects.
Collapse
Affiliation(s)
- Rebecca A Champion
- Department of Psychology, University of Glasgow, 58 Hillhead Street, Glasgow G12 8QB, Scotland, UK
| | | | | | | |
Collapse
|
42
|
Watt SJ, Bradshaw MF. The visual control of reaching and grasping: binocular disparity and motion parallax. J Exp Psychol Hum Percept Perform 2003; 29:404-15. [PMID: 12760624 DOI: 10.1037/0096-1523.29.2.404] [Citation(s) in RCA: 51] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
The primary visual sources of depth and size information are binocular cues and motion parallax. Here, the authors determine the efficacy of these cues to control prehension by presenting them in isolation from other visual cues. When only binocular cues were available, reaches showed normal scaling of the transport and grasp components with object distance and size. However, when only motion parallax was available, only the transport component scaled reliably. No additional increase in scaling was found when both cues were available simultaneously. Therefore, although equivalent information is available from binocular and motion parallax information, the latter may be of relatively limited use for the control of the grasp. Binocular disparity appears selectively important for the control of the grasp.
Collapse
Affiliation(s)
- Simon J Watt
- Department of Psychology, University of Surrey, Guildford, England.
| | | |
Collapse
|
43
|
Todd JT, Norman JF. The visual perception of 3-D shape from multiple cues: are observers capable of perceiving metric structure? PERCEPTION & PSYCHOPHYSICS 2003; 65:31-47. [PMID: 12699307 DOI: 10.3758/bf03194781] [Citation(s) in RCA: 98] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Three experiments are reported in which observers judged the three-dimensional (3-D) structures of virtual or real objects defined by various combinations of texture, motion, and binocular disparity under a wide variety of conditions. The tasks employed in these studies involved adjusting the depth of an object to match its width, adjusting the planes of a dihedral angle so that they appeared orthogonal, and adjusting the shape of an object so that it appeared to match another at a different viewing distance. The results obtained on all of these tasks revealed large constant errors and large individual differences among observers. There were also systematic failures of constancy over changes in viewing distance, orientation, or response task. When considered in conjunction with other, similar reports in the literature, these findings provide strong evidence that human observers do not have accurate perceptions of 3-D metric structure.
Collapse
Affiliation(s)
- James T Todd
- Department of Psychology, Ohio State University, Columbus, Ohio 43210, USA.
| | | |
Collapse
|
44
|
Barrett DJ, Bradshaw MF, Rose D, Everatt J, Simpson PJ. Reflexive shifts of covert attention operate in an egocentric coordinate frame. Perception 2002; 30:1083-91. [PMID: 11694084 DOI: 10.1068/p3165] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
Abstract
Covert shifts of attention have been shown to improve detection and discrimination thresholds for a range of visual stimuli. Although there is some evidence to suggest that the allocation of attention to a particular region of interest occurs in a retinotopic frame of reference, the importance of an allocentric, or object-based, framework has gained widespread empirical support. The current experiment investigates the nature of the spatial representation in which covert shifts of attention occur in response to a reflexive prime. Primes and targets were presented in four conditions designed to vary systematically the validity of the spatial relationship between the prime and target in egocentric or allocentric coordinate frameworks. A significant advantage, in terms of reaction time and correct identification, was found for targets located in positions previously primed in an egocentric (but not allocentric) framework whereas there was no advantage for locations primed in an allocentric (but not egocentric) framework. These results suggest that the allocation of covert spatial attention within an egocentric framework may be more important than previously thought.
Collapse
Affiliation(s)
- D J Barrett
- Department of Psychology, University of Surrey, Guildford, UK.
| | | | | | | | | |
Collapse
|