1
|
DiRisio GF, Ra Y, Qiu Y, Anzai A, DeAngelis GC. Neurons in Primate Area MSTd Signal Eye Movement Direction Inferred from Dynamic Perspective Cues in Optic Flow. J Neurosci 2023; 43:1888-1904. [PMID: 36725323 PMCID: PMC10027048 DOI: 10.1523/jneurosci.1885-22.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2022] [Revised: 01/18/2023] [Accepted: 01/24/2023] [Indexed: 02/03/2023] Open
Abstract
Smooth eye movements are common during natural viewing; we frequently rotate our eyes to track moving objects or to maintain fixation on an object during self-movement. Reliable information about smooth eye movements is crucial to various neural computations, such as estimating heading from optic flow or judging depth from motion parallax. While it is well established that extraretinal signals (e.g., efference copies of motor commands) carry critical information about eye velocity, the rotational optic flow field produced by eye rotations also carries valuable information. Although previous work has shown that dynamic perspective cues in optic flow can be used in computations that require estimates of eye velocity, it has remained unclear where and how the brain processes these visual cues and how they are integrated with extraretinal signals regarding eye rotation. We examined how neurons in the dorsal region of the medial superior temporal area (MSTd) of two male rhesus monkeys represent the direction of smooth pursuit eye movements based on both visual cues (dynamic perspective) and extraretinal signals. We find that most MSTd neurons have matched preferences for the direction of eye rotation based on visual and extraretinal signals. Moreover, neural responses to combinations of these signals are well predicted by a weighted linear summation model. These findings demonstrate a neural substrate for representing the velocity of smooth eye movements based on rotational optic flow and establish area MSTd as a key node for integrating visual and extraretinal signals into a more generalized representation of smooth eye movements.SIGNIFICANCE STATEMENT We frequently rotate our eyes to smoothly track objects of interest during self-motion. Information about eye velocity is crucial for a variety of computations performed by the brain, including depth perception and heading perception. Traditionally, information about eye rotation has been thought to arise mainly from extraretinal signals, such as efference copies of motor commands. Previous work shows that eye velocity can also be inferred from rotational optic flow that accompanies smooth eye movements, but the neural origins of these visual signals about eye rotation have remained unknown. We demonstrate that macaque neurons signal the direction of smooth eye rotation based on visual signals, and that they integrate both visual and extraretinal signals regarding eye rotation in a congruent fashion.
Collapse
Affiliation(s)
- Grace F DiRisio
- Department of Brain and Cognitive Sciences, Center for Visual Science, University of Rochester, Rochester, New York 14627
- Department of Neurobiology, University of Chicago, Chicago, Illinois 60637
| | - Yongsoo Ra
- Department of Brain and Cognitive Sciences, Center for Visual Science, University of Rochester, Rochester, New York 14627
- Department of Neurobiology, Harvard Medical School, Boston, Massachusetts 02115
| | - Yinghui Qiu
- Department of Brain and Cognitive Sciences, Center for Visual Science, University of Rochester, Rochester, New York 14627
- College of Veterinary Medicine, Cornell University, Ithaca, New York 14853-6401
| | - Akiyuki Anzai
- Department of Brain and Cognitive Sciences, Center for Visual Science, University of Rochester, Rochester, New York 14627
| | - Gregory C DeAngelis
- Department of Brain and Cognitive Sciences, Center for Visual Science, University of Rochester, Rochester, New York 14627
| |
Collapse
|
2
|
French RL, DeAngelis GC. Scene-relative object motion biases depth percepts. Sci Rep 2022; 12:18480. [PMID: 36323845 PMCID: PMC9630409 DOI: 10.1038/s41598-022-23219-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2022] [Accepted: 10/27/2022] [Indexed: 11/07/2022] Open
Abstract
An important function of the visual system is to represent 3D scene structure from a sequence of 2D images projected onto the retinae. During observer translation, the relative image motion of stationary objects at different distances (motion parallax) provides potent depth information. However, if an object moves relative to the scene, this complicates the computation of depth from motion parallax since there will be an additional component of image motion related to scene-relative object motion. To correctly compute depth from motion parallax, only the component of image motion caused by self-motion should be used by the brain. Previous experimental and theoretical work on perception of depth from motion parallax has assumed that objects are stationary in the world. Thus, it is unknown whether perceived depth based on motion parallax is biased by object motion relative to the scene. Naïve human subjects viewed a virtual 3D scene consisting of a ground plane and stationary background objects, while lateral self-motion was simulated by optic flow. A target object could be either stationary or moving laterally at different velocities, and subjects were asked to judge the depth of the object relative to the plane of fixation. Subjects showed a far bias when object and observer moved in the same direction, and a near bias when object and observer moved in opposite directions. This pattern of biases is expected if subjects confound image motion due to self-motion with that due to scene-relative object motion. These biases were large when the object was viewed monocularly, and were greatly reduced, but not eliminated, when binocular disparity cues were provided. Our findings establish that scene-relative object motion can confound perceptual judgements of depth during self-motion.
Collapse
Affiliation(s)
- Ranran L. French
- grid.16416.340000 0004 1936 9174Department of Brain and Cognitive Sciences, Center for Visual Science, University of Rochester, Rochester, USA
| | - Gregory C. DeAngelis
- grid.16416.340000 0004 1936 9174Department of Brain and Cognitive Sciences, Center for Visual Science, University of Rochester, Rochester, USA
| |
Collapse
|
3
|
Abstract
Motion perception is a critical function of the visual system. In a three-dimensional environment, multiple sensory cues carry information about an object's motion trajectory. Previous work has quantified the contribution of binocular motion cues, such as interocular velocity differences and changing disparities over time, as well as monocular motion cues, such as size and density changes. However, even when these cues are presented in concert, observers will systematically misreport the direction of motion-in-depth. Although in the majority of laboratory experiments head position is held fixed using a chin or head rest, an observer's head position is subject to involuntary small movements under real-world viewing conditions. Here, we considered the potential impact of such “head jitter” on motion-in-depth perception. We presented visual stimuli in a head-mounted virtual reality device that facilitated low latency head tracking and asked observers to judge 3D object motion. We found performance improved when we updated the visual display consistent with the small changes in head position. When we disrupted or delayed head movement–contingent updating of the visual display, the proportion of motion-in-depth misreports again increased, reflected in both a reduction in sensitivity and an increase in bias. Our findings identify a critical function of head jitter in visual motion perception, which has been obscured in most (head-fixed and non-head jitter contingent) laboratory experiments.
Collapse
Affiliation(s)
- Jacqueline M Fulvio
- Department of Psychology, University of Wisconsin - Madison, Madison, WI, USA.,
| | - Huiyuan Miao
- Department of Psychology, Vanderbilt University, Nashville, TN, USA.,
| | - Bas Rokers
- Psychology, New York University - Abu Dhabi, Abu Dhabi, UAE.,
| |
Collapse
|
4
|
Rogers B. Optic Flow: Perceiving and Acting in a 3-D World. Iperception 2021; 12:2041669520987257. [PMID: 33613957 PMCID: PMC7869175 DOI: 10.1177/2041669520987257] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2020] [Accepted: 12/18/2020] [Indexed: 11/15/2022] Open
Abstract
In 1979, James Gibson completed his third and final book "The Ecological Approach to Visual Perception". That book can be seen as the synthesis of the many radical ideas he proposed over the previous 30 years - the concept of information and its sufficiency, the necessary link between perception and action, the need to see perception in relation to an animal's particular ecological niche and the meanings (affordances) offered by the visual world. One of the fundamental concepts that lies beyond all of Gibson's thinking is that of optic flow: the constantly changing patterns of light that reach our eyes and the information it provides. My purpose in writing this paper has been to evaluate the legacy of Gibson's conceptual ideas and to consider how his ideas have influenced and changed the way we study perception.
Collapse
|
5
|
Abstract
Distance is important: From an ecological perspective, knowledge about the distance to either prey or predator is vital. However, the distance of an unknown sound source is particularly difficult to assess, especially in anechoic environments. In vision, changes in perspective resulting from observer motion produce a reliable, consistent, and unambiguous impression of depth known as motion parallax. Here we demonstrate with formal psychophysics that humans can exploit auditory motion parallax, i.e., the change in the dynamic binaural cues elicited by self-motion, to assess the relative depths of two sound sources. Our data show that sensitivity to relative depth is best when subjects move actively; performance deteriorates when subjects are moved by a motion platform or when the sound sources themselves move. This is true even though the dynamic binaural cues elicited by these three types of motion are identical. Our data demonstrate a perceptual strategy to segregate intermittent sound sources in depth and highlight the tight interaction between self-motion and binaural processing that allows assessment of the spatial layout of complex acoustic scenes.
Collapse
|
6
|
Holmin J, Nawrot M. Aging does not affect integration times for the perception of depth from motion parallax. Vision Res 2017; 140:81-88. [PMID: 28859970 PMCID: PMC5905337 DOI: 10.1016/j.visres.2017.05.016] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2016] [Revised: 05/27/2017] [Accepted: 05/29/2017] [Indexed: 11/28/2022]
Abstract
To successfully navigate throughout the world, observers must rapidly recover depth information. One depth cue that is especially important for a moving observer is motion parallax. To perceive unambiguous depth from motion parallax, the visual system must integrate information from two different proximal signals, retinal image motion and a pursuit eye movement. Previous research has shown that aging affects both of these necessary components for motion parallax depth perception, but no research has yet investigated how aging affects the mechanism for integrating motion and pursuit information to recover depth from motion parallax. The goal of the current experiment was to assess the integration time required by older adults to process depth information. In four psychophysical conditions, younger and older observers made motion and depth judgments about stationary or translating random-dot stimuli. Stimulus presentations in all four psychophysical conditions were followed by a high-contrast pattern mask, and minimum stimulus presentation durations (stimulus-to-mask onset asynchrony, or SOA) were measured. These SOAs reflect the minimum neural processing time required to make motion and motion parallax depth judgments. Pursuit latency was also measured. The results revealed that, after accounting for age-related delays in motion processing and pursuit onset, older and younger adults required similar temporal intervals to combine retinal image motion with an internal pursuit signal for the perception of depth. These results suggest that the mechanism for motion and pursuit integration is not affected by age.
Collapse
Affiliation(s)
- Jessica Holmin
- Center for Visual and Cognitive Neuroscience, Department of Psychology, North Dakota State University, P.O. Box 6050, Fargo, ND 58108-6050, USA
| | - Mark Nawrot
- Center for Visual and Cognitive Neuroscience, Department of Psychology, North Dakota State University, P.O. Box 6050, Fargo, ND 58108-6050, USA.
| |
Collapse
|
7
|
Kim HR, Angelaki DE, DeAngelis GC. The neural basis of depth perception from motion parallax. Philos Trans R Soc Lond B Biol Sci 2017; 371:rstb.2015.0256. [PMID: 27269599 DOI: 10.1098/rstb.2015.0256] [Citation(s) in RCA: 18] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 01/06/2016] [Indexed: 11/12/2022] Open
Abstract
In addition to depth cues afforded by binocular vision, the brain processes relative motion signals to perceive depth. When an observer translates relative to their visual environment, the relative motion of objects at different distances (motion parallax) provides a powerful cue to three-dimensional scene structure. Although perception of depth based on motion parallax has been studied extensively in humans, relatively little is known regarding the neural basis of this visual capability. We review recent advances in elucidating the neural mechanisms for representing depth-sign (near versus far) from motion parallax. We examine a potential neural substrate in the middle temporal visual area for depth perception based on motion parallax, and we explore the nature of the signals that provide critical inputs for disambiguating depth-sign.This article is part of the themed issue 'Vision in our three-dimensional world'.
Collapse
Affiliation(s)
- HyungGoo R Kim
- Department of Brain and Cognitive Sciences, Center for Visual Science, University of Rochester, NY 14627, USA
| | - Dora E Angelaki
- Department of Neuroscience, Baylor College of Medicine, Houston, TX 77030, USA Department of Electrical and Computer Engineering, Rice University, Houston, TX 77005, USA
| | - Gregory C DeAngelis
- Department of Brain and Cognitive Sciences, Center for Visual Science, University of Rochester, NY 14627, USA
| |
Collapse
|
8
|
Gain Modulation as a Mechanism for Coding Depth from Motion Parallax in Macaque Area MT. J Neurosci 2017; 37:8180-8197. [PMID: 28739582 DOI: 10.1523/jneurosci.0393-17.2017] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2017] [Revised: 06/30/2017] [Accepted: 07/20/2017] [Indexed: 11/21/2022] Open
Abstract
Observer translation produces differential image motion between objects that are located at different distances from the observer's point of fixation [motion parallax (MP)]. However, MP can be ambiguous with respect to depth sign (near vs far), and this ambiguity can be resolved by combining retinal image motion with signals regarding eye movement relative to the scene. We have previously demonstrated that both extra-retinal and visual signals related to smooth eye movements can modulate the responses of neurons in area MT of macaque monkeys, and that these modulations generate neural selectivity for depth sign. However, the neural mechanisms that govern this selectivity have remained unclear. In this study, we analyze responses of MT neurons as a function of both retinal velocity and direction of eye movement, and we show that smooth eye movements modulate MT responses in a systematic, temporally precise, and directionally specific manner to generate depth-sign selectivity. We demonstrate that depth-sign selectivity is primarily generated by multiplicative modulations of the response gain of MT neurons. Through simulations, we further demonstrate that depth can be estimated reasonably well by a linear decoding of a population of MT neurons with response gains that depend on eye velocity. Together, our findings provide the first mechanistic description of how visual cortical neurons signal depth from MP.SIGNIFICANCE STATEMENT Motion parallax is a monocular cue to depth that commonly arises during observer translation. To compute from motion parallax whether an object appears nearer or farther than the point of fixation requires combining retinal image motion with signals related to eye rotation, but the neurobiological mechanisms have remained unclear. This study provides the first mechanistic account of how this interaction takes place in the responses of cortical neurons. Specifically, we show that smooth eye movements modulate the gain of responses of neurons in area MT in a directionally specific manner to generate selectivity for depth sign from motion parallax. We also show, through simulations, that depth could be estimated from a population of such gain-modulated neurons.
Collapse
|
9
|
Affiliation(s)
- Jessica Holmin
- Department of Psychology, North Dakota State University, Fargo, North Dakota, USA
| | - Shanda Lauer
- Department of Psychology, North Dakota State University, Fargo, North Dakota, USA
| | - Mark Nawrot
- Department of Psychology, North Dakota State University, Fargo, North Dakota, USA
| |
Collapse
|
10
|
Dolins FL, Schweller K, Milne S. Technology advancing the study of animal cognition: using virtual reality to present virtually simulated environments to investigate nonhuman primate spatial cognition. Curr Zool 2017; 63:97-108. [PMID: 29491967 PMCID: PMC5804157 DOI: 10.1093/cz/zow121] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2016] [Revised: 05/27/2016] [Accepted: 12/22/2016] [Indexed: 11/29/2022] Open
Abstract
Virtual simulated environments provide multiple ways of testing cognitive function and evaluating problem solving with humans (e.g., Woollett et al. 2009). The use of such interactive technology has increasingly become an essential part of modern life (e.g., autonomously driving vehicles, global positioning systems (GPS), and touchscreen computers; Chinn and Fairlie 2007; Brown 2011). While many nonhuman animals have their own forms of "technology", such as chimpanzees who create and use tools, in captive animal environments the opportunity to actively participate with interactive technology is not often made available. Exceptions can be found in some state-of-the-art zoos and laboratory facilities (e.g., Mallavarapu and Kuhar 2005). When interactive technology is available, captive animals often selectively choose to engage with it. This enhances the animal's sense of control over their immediate surroundings (e.g., Clay et al. 2011; Ackerman 2012). Such self-efficacy may help to fulfill basic requirements in a species' daily activities using problem solving that can involve foraging and other goal-oriented behaviors. It also assists in fulfilling the strong underlying motivation for contrafreeloading and exploration expressed behaviorally by many species in captivity (Young 1999). Moreover, being able to present nonhuman primates virtual reality environments under experimental conditions provides the opportunity to gain insight into their navigational abilities and spatial cognition. It allows for insight into the generation and application of internal mental representations of landmarks and environments under multiple conditions (e.g., small- and large-scale space) and subsequent spatial behavior. This paper reviews methods using virtual reality developed to investigate the spatial cognitive abilities of nonhuman primates, and great apes in particular, in comparison with that of humans of multiple age groups. We make recommendations about training, best practices, and also pitfalls to avoid.
Collapse
Affiliation(s)
- Francine L. Dolins
- Department of Behavioral Sciences, University of Michigan-Dearborn, Dearborn, MI, 48128, USA
| | | | - Scott Milne
- Impact Madagascar, Nanisana, Antananarivo, 101, Madagascar, Africa
| |
Collapse
|
11
|
Ratzlaff M, Nawrot M. A Pursuit Theory Account for the Perception of Common Motion in Motion Parallax. Perception 2016; 45:991-1007. [PMID: 27060180 PMCID: PMC4990516 DOI: 10.1177/0301006616643679] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2023]
Abstract
The visual system uses an extraretinal pursuit eye movement signal to disambiguate the perception of depth from motion parallax. Visual motion in the same direction as the pursuit is perceived nearer in depth while visual motion in the opposite direction as pursuit is perceived farther in depth. This explanation of depth sign applies to either an allocentric frame of reference centered on the fixation point or an egocentric frame of reference centered on the observer. A related problem is that of depth order when two stimuli have a common direction of motion. The first psychophysical study determined whether perception of egocentric depth order is adequately explained by a model employing an allocentric framework, especially when the motion parallax stimuli have common rather than divergent motion. A second study determined whether a reversal in perceived depth order, produced by a reduction in pursuit velocity, is also explained by this model employing this allocentric framework. The results show than an allocentric model can explain both the egocentric perception of depth order with common motion and the perceptual depth order reversal created by a reduction in pursuit velocity. We conclude that an egocentric model is not the only explanation for perceived depth order in these common motion conditions.
Collapse
Affiliation(s)
- Michael Ratzlaff
- Center for Visual and Cognitive Neuroscience, Department of Psychology, North Dakota State University, Fargo, ND, USA
| | - Mark Nawrot
- Center for Visual and Cognitive Neuroscience, Department of Psychology, North Dakota State University, Fargo, ND, USA
| |
Collapse
|
12
|
Abstract
There are many similarities between binocular disparity and motion parallax as sources of information about the structure and layout of 3-D objects and surfaces. The former can be thought of as a transformation that maps one eye's image onto the other while the latter is a transformation that maps the changes in one eye's image over time. There are many empirical similarities in the ways we use the two sources of information but there are also significant differences. A consideration of those differences leads to the conclusion that, rather than seeing motion parallax as a close analogue of binocular stereopsis, motion parallax is better thought of as a special case of the kinetic depth effect in which the depth order of the depicted 3-D object or surface can be disambiguated by vertical perspective information.
Collapse
Affiliation(s)
- Brian Rogers
- Department of Experimental Psychology, University of Oxford, UK
| |
Collapse
|
13
|
Rogers B. The Effectiveness of Vertical Perspective and Pursuit Eye Movements for Disambiguating Motion Parallax Transformations. Perception 2016; 45:1279-1303. [PMID: 27343187 DOI: 10.1177/0301006616655815] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
In the kinetic depth effect, the direction of the perceived depth and the direction of apparent rotation of a 3-D structure are linked, and typically ambiguous, whereas depth from motion parallax during both observer- and object-movement is stable and unambiguous. Rogers and Rogers demonstrated that the vertical perspective transformations play an important role in disambiguating the direction of the perceived depth in parallax-defined surfaces but more recently Nawrot et al. have proposed that pursuit eye movements provide the crucial disambiguating information. Theoretical considerations suggest that pursuit eye movements could not, in principle, provide the necessary information because 3-D objects as surfaces may rotate during observer- or object-movement. The empirical evidence presented here shows that vertical perspective transformations are sufficient for the unambiguous perception of parallax depth whereas pursuit eye movements are not necessary and may not even be sufficient.
Collapse
Affiliation(s)
- Brian Rogers
- Department of Experimental Psychology, University of Oxford, UK
| |
Collapse
|
14
|
Abstract
Successful navigation in the world requires effective visuospatial processing. Unfortunately, older adults have many visuospatial deficits, which can have severe real-world consequences. Although some of these age effects are well documented, some others, such as the perception of depth from motion parallax, are poorly understood. Depth perception from motion parallax requires intact retinal image motion and pursuit eye movement processing. Decades of research have shown that both motion processing and pursuit eye movements are affected by age; it follows that older adults may also be less sensitive to depth from motion parallax. The goals of the present study were to characterize motion parallax depth thresholds in older adults, and to explain older adults' sensitivity to depth from motion parallax in terms of motion and pursuit deficits. Younger and older adults' motion thresholds and pursuit accuracy were measured. Observers' depth thresholds across several different stimulus conditions were measured, as well. Older adults had higher motion thresholds and less accurate pursuit than younger adults. They were also less sensitive to depth from motion parallax at slow and moderate pursuit speeds. Although older adults had higher motion thresholds than younger adults, they used the available motion signals optimally, and age differences in motion processing could not account for the older adults' increased depth thresholds. Rather, these age effects can be explained by changes in older adults' pursuit signals.
Collapse
|
15
|
Holmin J, Nawrot M. Motion parallax thresholds for unambiguous depth perception. Vision Res 2015; 115:40-7. [PMID: 26232612 DOI: 10.1016/j.visres.2015.07.002] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2014] [Revised: 06/29/2015] [Accepted: 07/23/2015] [Indexed: 10/23/2022]
Abstract
The perception of unambiguous depth from motion parallax arises from the neural integration of retinal image motion and extra-retinal eye movement signals. It is only recently that these parameters have been articulated in the form of the motion/pursuit ratio. In the current study, we explored the lower limits of the parameter space in which observers could accurately perform near/far relative depth-sign discriminations for a translating random-dot stimulus. Stationary observers pursued a translating random dot stimulus containing relative image motion. Their task was to indicate the location of the peak in an approximate square-wave stimulus. We measured thresholds for depth from motion parallax, quantified as motion/pursuit ratios, as well as lower motion thresholds and pursuit accuracy. Depth thresholds were relatively stable at pursuit velocities 5-20 deg/s, and increased at lower and higher velocities. The pattern of results indicates that minimum motion/pursuit ratios are limited by motion and pursuit signals, both independently and in combination with each other. At low and high pursuit velocities, depth thresholds were limited by inaccurate pursuit signals. At moderate pursuit velocities, depth thresholds were limited by motion signals.
Collapse
Affiliation(s)
- Jessica Holmin
- Center for Visual and Cognitive Neuroscience, Department of Psychology, North Dakota State University, P.O. Box 6050, Fargo, ND 58108-6050, United States
| | - Mark Nawrot
- Center for Visual and Cognitive Neuroscience, Department of Psychology, North Dakota State University, P.O. Box 6050, Fargo, ND 58108-6050, United States.
| |
Collapse
|
16
|
Abstract
As an observer translates, objects lying at different distances from the observer have differential image motion on the retina (motion parallax). It is well established psychophysically that humans perceive depth rather precisely from motion parallax and that extraretinal signals may be used to correctly perceive the sign of depth (near vs far) when binocular and pictorial depth cues are absent or weak. However, the neural basis for this capacity remains poorly understood. We have shown previously that neurons in the macaque middle temporal (MT) area combine retinal image motion with smooth eye movement command signals to signal depth sign from motion parallax. However, those studies were performed in animals that were required simply to track a visual target, thus precluding direct comparisons between neural activity and behavior. Here, we examine the activity of MT neurons in rhesus monkeys that were trained to discriminate depth sign based on motion parallax, in the absence of binocular disparity and pictorial depth cues. We find that the most sensitive MT neurons approach behavioral sensitivity, whereas the average neuron is twofold to threefold less sensitive than the animal. We also find that MT responses are predictive of perceptual decisions (independent of the visual stimulus), consistent with a role for MT in providing sensory signals for this behavior. Our findings suggest that, in addition to its established roles in processing stereoscopic depth, area MT is well suited to contribute to perception of depth based on motion parallax.
Collapse
|
17
|
Kim HR, Angelaki DE, DeAngelis GC. A novel role for visual perspective cues in the neural computation of depth. Nat Neurosci 2014; 18:129-37. [PMID: 25436667 PMCID: PMC4281299 DOI: 10.1038/nn.3889] [Citation(s) in RCA: 31] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2014] [Accepted: 11/02/2014] [Indexed: 11/10/2022]
Abstract
As we explore a scene, our eye movements add global patterns of motion to the retinal image, complicating visual motion produced by self-motion or moving objects. Conventionally, it has been assumed that extra-retinal signals, such as efference copy of smooth pursuit commands, are required to compensate for the visual consequences of eye rotations. We consider an alternative possibility: namely, that the visual system can infer eye rotations from global patterns of image motion. We visually simulated combinations of eye translation and rotation, including perspective distortions that change dynamically over time. We demonstrate that incorporating these “dynamic perspective” cues allows the visual system to generate selectivity for depth sign from motion parallax in macaque area MT, a computation that was previously thought to require extra-retinal signals regarding eye velocity. Our findings suggest novel neural mechanisms that analyze global patterns of visual motion to perform computations that require knowledge of eye rotations.
Collapse
Affiliation(s)
- HyungGoo R Kim
- Department of Brain and Cognitive Sciences, Center for Visual Science, University of Rochester, Rochester, New York, USA
| | - Dora E Angelaki
- Department of Neuroscience, Baylor College of Medicine, Houston, Texas, USA
| | - Gregory C DeAngelis
- Department of Brain and Cognitive Sciences, Center for Visual Science, University of Rochester, Rochester, New York, USA
| |
Collapse
|
18
|
Nawrot M, Ratzlaff M, Leonard Z, Stroyan K. Modeling depth from motion parallax with the motion/pursuit ratio. Front Psychol 2014; 5:1103. [PMID: 25339926 PMCID: PMC4186274 DOI: 10.3389/fpsyg.2014.01103] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2014] [Accepted: 09/11/2014] [Indexed: 11/13/2022] Open
Abstract
The perception of unambiguous scaled depth from motion parallax relies on both retinal image motion and an extra-retinal pursuit eye movement signal. The motion/pursuit ratio represents a dynamic geometric model linking these two proximal cues to the ratio of depth to viewing distance. An important step in understanding the visual mechanisms serving the perception of depth from motion parallax is to determine the relationship between these stimulus parameters and empirically determined perceived depth magnitude. Observers compared perceived depth magnitude of dynamic motion parallax stimuli to static binocular disparity comparison stimuli at three different viewing distances, in both head-moving and head-stationary conditions. A stereo-viewing system provided ocular separation for stereo stimuli and monocular viewing of parallax stimuli. For each motion parallax stimulus, a point of subjective equality (PSE) was estimated for the amount of binocular disparity that generates the equivalent magnitude of perceived depth from motion parallax. Similar to previous results, perceived depth from motion parallax had significant foreshortening. Head-moving conditions produced even greater foreshortening due to the differences in the compensatory eye movement signal. An empirical version of the motion/pursuit law, termed the empirical motion/pursuit ratio, which models perceived depth magnitude from these stimulus parameters, is proposed.
Collapse
Affiliation(s)
- Mark Nawrot
- Department of Psychology, Center for Visual and Cognitive Neuroscience, North Dakota State University Fargo, ND, USA
| | - Michael Ratzlaff
- Department of Psychology, Center for Visual and Cognitive Neuroscience, North Dakota State University Fargo, ND, USA
| | - Zachary Leonard
- Department of Psychology, Center for Visual and Cognitive Neuroscience, North Dakota State University Fargo, ND, USA
| | - Keith Stroyan
- Math Department, University of Iowa Iowa City, IA, USA
| |
Collapse
|
19
|
Fesi JD, Thomas AL, Gilmore RO. Cortical responses to optic flow and motion contrast across patterns and speeds. Vision Res 2014; 100:56-71. [PMID: 24751405 DOI: 10.1016/j.visres.2014.04.004] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2013] [Revised: 03/05/2014] [Accepted: 04/09/2014] [Indexed: 11/26/2022]
Abstract
Motion provides animals with fast and robust cues for navigation and object detection. In the first case, stereotyped patterns of optic flow inform a moving observer about the direction and speed of its own movement. In the case of object detection, regional differences in motion allow for the segmentation of figures from their background, even in the absence of color or shading cues. Previous research has investigated human electrophysiological responses to global motion across speeds, but only focused upon one type of optic flow pattern. Here, we compared steady-state visual evoked potential (SSVEP) responses across patterns and speeds, both for optic flow and for motion-defined figure patterns, to assess the extent to which the processes are pattern-general or pattern-specific. For optic flow, pattern and speed effects on response amplitudes varied substantially across channels, suggesting pattern-specific processing at slow speeds and pattern-general activity at fast speeds. Responses for coherence- and direction-defined figures were comparatively more uniform, with similar response profiles and spatial distributions. Self- and object-motion patterns activate some of the same circuits, but these data suggest differential sensitivity: not only across the two classes of motion, but also across the patterns within each class, and across speeds. Thus, the results demonstrate that cortical processing of global motion is complex and activates a distributed network.
Collapse
Affiliation(s)
- Jeremy D Fesi
- Department of Ophthalmology, McGill University, 687 Pine Avenue West, Montreal, QC H3A 1A1, Canada.
| | - Amanda L Thomas
- Department of Psychology, The Pennsylvania State University, 114 Moore Building, University Park, PA 16802, United States
| | - Rick O Gilmore
- Department of Psychology, The Pennsylvania State University, 114 Moore Building, University Park, PA 16802, United States; Social, Life, & Engineering Sciences Imaging Center, The Pennsylvania State University, University Park, PA 16802, United States
| |
Collapse
|
20
|
George JM, Johnson JI, Nawrot M. In pursuit of perspective: does vertical perspective disambiguate depth from motion parallax? Perception 2014; 42:631-41. [PMID: 24422245 DOI: 10.1068/p7250] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
Abstract
Motion parallax provides a dynamic, unambiguous, monocular visual depth cue. However, the lateral image motion in computer-generated motion parallax displays is depth-sign ambiguous. While mounting evidence indicates that the visual system uses an extra-retinal signal from the pursuit system to disambiguate depth, vertical perspective is a potential confound because it co-varies with the stimulus translation that produces the pursuit signal. Here the role of an extra-retinal pursuit signal and the role of vertical perspective in disambiguating depth from motion parallax were investigated. Through the careful isolation of each cue, the results indicate that observers have excellent depth discrimination with an extra-retinal pursuit cue alone, but have poor discrimination with vertical perspective alone. The conclusion is that vertical perspective does not play a role in the disambiguation of depth in small computer-generated motion parallax displays.
Collapse
Affiliation(s)
- Jonathon M George
- Center for Visual and Cognitive Neuroscience, Department of Psychology, North Dakota State University, Fargo, ND 58108, USA
| | - Joshua I Johnson
- Center for Visual and Cognitive Neuroscience, Department of Psychology, North Dakota State University, Fargo, ND 58108, USA
| | - Mark Nawrot
- Center for Visual and Cognitive Neuroscience, Department of Psychology, North Dakota State University, Fargo, ND 58108, USA
| |
Collapse
|
21
|
Nawrot E, Nawrot M. The role of eye movements in depth from motion parallax during infancy. J Vis 2013; 13:13.14.15. [PMID: 24353309 DOI: 10.1167/13.14.15] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Motion parallax is a motion-based, monocular depth cue that uses an object's relative motion and velocity as a cue to relative depth. In adults, and in monkeys, a smooth pursuit eye movement signal is used to disambiguate the depth-sign provided by these relative motion cues. The current study investigates infants' perception of depth from motion parallax and the development of two oculomotor functions, smooth pursuit and the ocular following response (OFR) eye movements. Infants 8 to 20 weeks of age were presented with three tasks in a single session: depth from motion parallax, smooth pursuit tracking, and OFR to translation. The development of smooth pursuit was significantly related to age, as was sensitivity to motion parallax. OFR eye movements also corresponded to both age and smooth pursuit gain, with groups of infants demonstrating asymmetric function in both types of eye movements. These results suggest that the development of the eye movement system may play a crucial role in the sensitivity to depth from motion parallax in infancy. Moreover, describing the development of these oculomotor functions in relation to depth perception may aid in the understanding of certain visual dysfunctions.
Collapse
Affiliation(s)
- Elizabeth Nawrot
- Department of Psychology, Minnesota State University Moorhead, Moorhead, MN, USA
| | | |
Collapse
|
22
|
Joint representation of depth from motion parallax and binocular disparity cues in macaque area MT. J Neurosci 2013; 33:14061-74, 14074a. [PMID: 23986242 DOI: 10.1523/jneurosci.0251-13.2013] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
Perception of depth is based on a variety of cues, with binocular disparity and motion parallax generally providing more precise depth information than pictorial cues. Much is known about how neurons in visual cortex represent depth from binocular disparity or motion parallax, but little is known about the joint neural representation of these depth cues. We recently described neurons in the middle temporal (MT) area that signal depth sign (near vs far) from motion parallax; here, we examine whether and how these neurons also signal depth from binocular disparity. We find that most MT neurons in rhesus monkeys (Macaca Mulatta) are selective for depth sign based on both disparity and motion parallax cues. However, the depth-sign preferences (near or far) are not always aligned: 56% of MT neurons have matched depth-sign preferences ("congruent" cells) whereas the remaining 44% of neurons prefer near depth from motion parallax and far depth from disparity, or vice versa ("opposite" cells). For congruent cells, depth-sign selectivity increases when disparity cues are added to motion parallax, but this enhancement does not occur for opposite cells. This suggests that congruent cells might contribute to perceptual integration of depth cues. We also found that neurons are clustered in MT according to their depth tuning based on motion parallax, similar to the known clustering of MT neurons for binocular disparity. Together, these findings suggest that area MT is involved in constructing a representation of 3D scene structure that takes advantage of multiple depth cues available to mobile observers.
Collapse
|
23
|
Stroyan K, Nawrot M. Visual depth from motion parallax and eye pursuit. J Math Biol 2012; 64:1157-88. [PMID: 21695531 PMCID: PMC3348271 DOI: 10.1007/s00285-011-0445-1] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2011] [Revised: 05/26/2011] [Indexed: 10/18/2022]
Abstract
A translating observer viewing a rigid environment experiences "motion parallax", the relative movement upon the observer's retina of variously positioned objects in the scene. This retinal movement of images provides a cue to the relative depth of objects in the environment, however retinal motion alone cannot mathematically determine relative depth of the objects. Visual perception of depth from lateral observer translation uses both retinal image motion and eye movement. In Nawrot and Stroyan (Vision Res 49:1969-1978, 2009) we showed mathematically that the ratio of the rate of retinal motion over the rate of smooth eye pursuit mathematically determines depth relative to the fixation point in central vision. We also reported on psychophysical experiments indicating that this ratio is the important quantity for perception. Here we analyze the motion/pursuit cue for the more general, and more complicated, case when objects are distributed across the horizontal viewing plane beyond central vision. We show how the mathematical motion/pursuit cue varies with different points across the plane and with time as an observer translates. If the time varying retinal motion and smooth eye pursuit are the only signals used for this visual process, it is important to know what is mathematically possible to derive about depth and structure. Our analysis shows that the motion/pursuit ratio determines an excellent description of depth and structure in these broader stimulus conditions, provides a detailed quantitative hypothesis of these visual processes for the perception of depth and structure from motion parallax, and provides a computational foundation to analyze the dynamic geometry of future experiments.
Collapse
Affiliation(s)
- Keith Stroyan
- Mathematics Department, University of Iowa, Iowa City, IA, 52242, USA.
| | | |
Collapse
|
24
|
Nawrot M, Stroyan K. Integration time for the perception of depth from motion parallax. Vision Res 2012; 59:64-71. [PMID: 22406543 DOI: 10.1016/j.visres.2012.02.007] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2011] [Revised: 01/26/2012] [Accepted: 02/21/2012] [Indexed: 10/28/2022]
Abstract
The perception of depth from relative motion is believed to be a slow process that "builds-up" over a period of observation. However, in the case of motion parallax, the potential accuracy of the depth estimate suffers as the observer translates during the viewing period. Our recent quantitative model for the perception of depth from motion parallax proposes that relative object depth (d) can be determined from retinal image motion (dθ/dt), pursuit eye movement (dα/dt), and fixation distance (f) by the formula: d/f≈dθ/dα. Given the model's dynamics, it is important to know the integration time required by the visual system to recover dα and dθ, and then estimate d. Knowing the minimum integration time reveals the incumbent error in this process. A depth-phase discrimination task was used to determine the time necessary to perceive depth-sign from motion parallax. Observers remained stationary and viewed a briefly translating random-dot motion parallax stimulus. Stimulus duration varied between trials. Fixation on the translating stimulus was monitored and enforced with an eye-tracker. The study found that relative depth discrimination can be performed with presentations as brief as 16.6 ms, with only two stimulus frames providing both retinal image motion and the stimulus window motion for pursuit (mean range=16.6-33.2 ms). This was found for conditions in which, prior to stimulus presentation, the eye was engaged in ongoing pursuit or the eye was stationary. A large high-contrast masking stimulus disrupted depth-discrimination for stimulus presentations less than 70-75 ms in both pursuit and stationary conditions. This interval might be linked to ocular-following response eye-movement latencies. We conclude that neural mechanisms serving depth from motion parallax generate a depth estimate much more quickly than previously believed. We propose that additional sluggishness might be due to the visual system's attempt to determine the maximum dθ/dα ratio for a selection of points on a complicated stimulus.
Collapse
Affiliation(s)
- Mark Nawrot
- Center for Visual Neuroscience, Department of Psychology, North Dakota State University, Fargo, ND 58108, USA.
| | | |
Collapse
|
25
|
Hanes DA. Mathematical requirements of visual–vestibular integration. J Math Biol 2011; 65:1245-66. [DOI: 10.1007/s00285-011-0494-5] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2011] [Revised: 11/16/2011] [Indexed: 10/15/2022]
|
26
|
Dokka K, MacNeilage PR, DeAngelis GC, Angelaki DE. Estimating distance during self-motion: a role for visual-vestibular interactions. J Vis 2011; 11:11.13.2. [PMID: 22045777 DOI: 10.1167/11.13.2] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
A fundamental challenge for the visual system is to extract the 3D spatial structure of the environment. When an observer translates without moving the eyes, the retinal speed of a stationary object is related to its distance by a scale factor that depends on the velocity of the observer's self-motion. Here, we aim to test whether the brain uses vestibular cues to self-motion to estimate distance to stationary surfaces in the environment. This relationship was systematically probed using a two-alternative forced-choice task in which distance perceived from monocular image motion during passive body translation was compared to distance perceived from binocular disparity while subjects were stationary. We show that perceived distance from motion depended on both observer velocity and retinal speed. For a given head speed, slower retinal speeds led to the perception of farther distances. Likewise, for a given retinal speed, slower head speeds led to the perception of nearer distances. However, these relationships were weak in some subjects and absent in others, and distance estimated from self-motion and retinal image motion was substantially compressed relative to distance estimated from binocular disparity. Overall, our findings suggest that the combination of retinal image motion and vestibular signals related to head velocity can provide a rudimentary capacity for distance estimation.
Collapse
Affiliation(s)
- Kalpana Dokka
- Department of Anatomy and Neurobiology, Washington University in St. Louis, USA
| | | | | | | |
Collapse
|