1
|
Gerharz L, Brenner E, Billino J, Voudouris D. Age effects on predictive eye movements for action. J Vis 2024; 24:8. [PMID: 38856982 PMCID: PMC11166221 DOI: 10.1167/jov.24.6.8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2024] [Accepted: 04/22/2024] [Indexed: 06/11/2024] Open
Abstract
When interacting with the environment, humans typically shift their gaze to where information is to be found that is useful for the upcoming action. With increasing age, people become slower both in processing sensory information and in performing their movements. One way to compensate for this slowing down could be to rely more on predictive strategies. To examine whether we could find evidence for this, we asked younger (19-29 years) and older (55-72 years) healthy adults to perform a reaching task wherein they hit a visual target that appeared at one of two possible locations. In separate blocks of trials, the target could appear always at the same location (predictable), mainly at one of the locations (biased), or at either location randomly (unpredictable). As one might expect, saccades toward predictable targets had shorter latencies than those toward less predictable targets, irrespective of age. Older adults took longer to initiate saccades toward the target location than younger adults, even when the likely target location could be deduced. Thus we found no evidence of them relying more on predictive gaze. Moreover, both younger and older participants performed more saccades when the target location was less predictable, but again no age-related differences were found. Thus we found no tendency for older adults to rely more on prediction.
Collapse
Affiliation(s)
- Leonard Gerharz
- Experimental Psychology, Justus Liebig University Giessen, Giessen, Germany
- https://orcid.org/0009-0006-0487-2609
| | - Eli Brenner
- Department of Human Movement Science, Vrije Universiteit Amsterdam, Amsterdam, The Netherlands
| | - Jutta Billino
- Experimental Psychology, Justus Liebig University Giessen, Giessen, Germany
| | - Dimitris Voudouris
- Experimental Psychology, Justus Liebig University Giessen, Giessen, Germany
| |
Collapse
|
2
|
Heins F, Lappe M. Oculomotor behavior can be adjusted on the basis of artificial feedback signals indicating externally caused errors. PLoS One 2024; 19:e0302872. [PMID: 38768134 PMCID: PMC11104623 DOI: 10.1371/journal.pone.0302872] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2023] [Accepted: 04/15/2024] [Indexed: 05/22/2024] Open
Abstract
Whether a saccade is accurate and has reached the target cannot be evaluated during its execution, but relies on post-saccadic feedback. If the eye has missed the target object, a secondary corrective saccade has to be made to align the fovea with the target. If a systematic post-saccadic error occurs, adaptive changes to the oculomotor behavior are made, such as shortening or lengthening the saccade amplitude. Systematic post-saccadic errors are typically attributed internally to erroneous motor commands. The corresponding adaptive changes to the motor command reduce the error and the need for secondary corrective saccades, and, in doing so, restore accuracy and efficiency. However, adaptive changes to the oculomotor behavior also occur if a change in saccade amplitude is beneficial for task performance, or if it is rewarded. Oculomotor learning thus is more complex than reducing a post-saccadic position error. In the current study, we used a novel oculomotor learning paradigm and investigated whether human participants are able to adapt their oculomotor behavior to improve task performance even when they attribute the error externally. The task was to indicate the intended target object among several objects to a simulated human-machine interface by making eye movements. The participants were informed that the system itself could make errors. The decoding process depended on a distorted landing point of the saccade, resulting in decoding errors. Two different types of visual feedback were added to the post-saccadic scene and we compared how participants used the different feedback types to adjust their oculomotor behavior to avoid errors. We found that task performance improved over time, regardless of the type of feedback. Thus, error feedback from the simulated human-machine interface was used for post-saccadic error evaluation. This indicates that 1) artificial visual feedback signals and 2) externally caused errors might drive adaptive changes to oculomotor behavior.
Collapse
Affiliation(s)
- Frauke Heins
- Institute for Psychology and Otto-Creutzfeldt Center for Cognitive and Behavioral Neuroscience, University of Münster, Münster, Germany
| | - Markus Lappe
- Institute for Psychology and Otto-Creutzfeldt Center for Cognitive and Behavioral Neuroscience, University of Münster, Münster, Germany
| |
Collapse
|
3
|
Brenner E, Smeets JBJ. Continuous use of visual information about the position of the moving hand. Exp Brain Res 2023:10.1007/s00221-023-06658-x. [PMID: 37386195 PMCID: PMC10386923 DOI: 10.1007/s00221-023-06658-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2023] [Accepted: 06/22/2023] [Indexed: 07/01/2023]
Abstract
People generally look at a target when they want to reach for it. Doing so presumably helps them continuously update their judgments about the target's position and motion. But not looking at their hand does not prevent people from updating judgments about its position on the basis of visual information, because people do respond to experimental perturbations of visual information about the position of their hand. Here, we study such responses by adding jitter to the movement of a cursor that follows participants' fingers. We analyse the response to the jitter in a way that reveals how the vigour of the response depends on the moment during the movement at which the change in cursor position occurs. We compare the change in vigour to that for equivalent jitter in the position of the target. We find that participants respond to jitter in the position of a cursor in much the same way as they respond to jitter in the target's position. The responses are more vigorous late in the movement, when adjustments need to be made within less time, but similarly so for the cursor as for the target. The responses are weaker for the cursor, presumably because of the jitter-free kinaesthetic information about the position of the finger.
Collapse
Affiliation(s)
- Eli Brenner
- Department of Human Movement Sciences, Vrije Universiteit Amsterdam, Van der Boechorststraat 7, 1081 BT, Amsterdam, The Netherlands.
| | - Jeroen B J Smeets
- Department of Human Movement Sciences, Vrije Universiteit Amsterdam, Van der Boechorststraat 7, 1081 BT, Amsterdam, The Netherlands
| |
Collapse
|
4
|
Whitwell RL, Garach MA, Goodale MA, Sperandio I. Looking at the Ebbinghaus illusion: differences in neurocomputational requirements, not gaze-mediated attention, explain a classic perception-action dissociation. Philos Trans R Soc Lond B Biol Sci 2023; 378:20210459. [PMID: 36511402 PMCID: PMC9745866 DOI: 10.1098/rstb.2021.0459] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2022] [Accepted: 08/11/2022] [Indexed: 12/15/2022] Open
Abstract
Perceiving and grasping an object present an animal with different sets of computational problems. The solution in primates entails the specialization of separate neural networks for visual processing with different object representations. This explains why the Ebbinghaus illusion minimally affects the grasping hand's in-flight aperture, which normally scales with target size, even though the size of the target disc remains misperceived. An attractive alternative account, however, posits that grasps are refractory to the illusion because participants fixate on the target and fail to attend to the surrounding context. To test this account, we tracked both limb and gaze while participants made forced-choice judgments of relative disc size in the Ebbinghaus illusion or did so in combination with grasping or manually estimating the size of one of the discs. We replicated the classic dissociation: grasp aperture was refractory to the measured illusory effect on perceived size, while judgments and manual estimates of disc size were not. Importantly, the number of display-wide saccades per second and the percentage of total fixation time or fixations directed at the selected disc failed to explain the dissociation. Our findings support the contention that object perception and goal-directed action rely on distinct visual representations. This article is part of a discussion meeting issue 'New approaches to 3D vision'.
Collapse
Affiliation(s)
- Robert L. Whitwell
- Department of Psychology, University of Western Ontario, London, Ontario, Canada N6A 5C2
| | - Mehul A. Garach
- Division of Orthopaedics, St Michael's Hospital, University of Toronto, Toronto, Ontario, Canada M5B 1W8
| | - Melvyn A. Goodale
- Department of Psychology, University of Western Ontario, London, Ontario, Canada N6A 5C2
| | - Irene Sperandio
- Department of Psychology and Cognitive Science, University of Trento, Rovereto (TN) 38068, Italy
| |
Collapse
|
5
|
Voudouris D, Fiehler K. The role of grasping demands on tactile suppression. Hum Mov Sci 2022; 83:102957. [DOI: 10.1016/j.humov.2022.102957] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2021] [Revised: 03/27/2022] [Accepted: 04/30/2022] [Indexed: 11/15/2022]
|
6
|
Lombardi M, Maiettini E, De Tommaso D, Wykowska A, Natale L. Toward an Attentive Robotic Architecture: Learning-Based Mutual Gaze Estimation in Human–Robot Interaction. Front Robot AI 2022; 9:770165. [PMID: 35321344 PMCID: PMC8935014 DOI: 10.3389/frobt.2022.770165] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/03/2021] [Accepted: 02/10/2022] [Indexed: 11/23/2022] Open
Abstract
Social robotics is an emerging field that is expected to grow rapidly in the near future. In fact, it is increasingly more frequent to have robots that operate in close proximity with humans or even collaborate with them in joint tasks. In this context, the investigation of how to endow a humanoid robot with social behavioral skills typical of human–human interactions is still an open problem. Among the countless social cues needed to establish a natural social attunement, this article reports our research toward the implementation of a mechanism for estimating the gaze direction, focusing in particular on mutual gaze as a fundamental social cue in face-to-face interactions. We propose a learning-based framework to automatically detect eye contact events in online interactions with human partners. The proposed solution achieved high performance both in silico and in experimental scenarios. Our work is expected to be the first step toward an attentive architecture able to endorse scenarios in which the robots are perceived as social partners.
Collapse
Affiliation(s)
- Maria Lombardi
- Humanoid Sensing and Perception, Istituto Italiano di Tecnologia, Genova, Italy
- *Correspondence: Maria Lombardi,
| | - Elisa Maiettini
- Humanoid Sensing and Perception, Istituto Italiano di Tecnologia, Genova, Italy
| | - Davide De Tommaso
- Social Cognition in Human-Robot Interaction, Istituto Italiano di Tecnologia, Genova, Italy
| | - Agnieszka Wykowska
- Social Cognition in Human-Robot Interaction, Istituto Italiano di Tecnologia, Genova, Italy
| | - Lorenzo Natale
- Humanoid Sensing and Perception, Istituto Italiano di Tecnologia, Genova, Italy
| |
Collapse
|
7
|
Crowe EM, Smeets JBJ, Brenner E. The response to background motion: Characteristics of a movement stabilization mechanism. J Vis 2021; 21:3. [PMID: 34617956 PMCID: PMC8504189 DOI: 10.1167/jov.21.11.3] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
When making goal-directed movements toward a target, our hand deviates from its path in the direction of sudden background motion. We propose that this manual following response arises because ongoing movements are constantly guided toward the planned movement endpoint. Such guidance is needed to compensate for modest, unexpected self-motion. Our proposal is that the compensation for such self-motion does not involve a sophisticated analysis of the global optic flow. Instead, we propose that any motion in the vicinity of the planned endpoint is attributed to the endpoint's egocentric position having shifted in the direction of the motion. The ongoing movement is then stabilized relative to the shifted endpoint. In six experiments, we investigate what aspects of motion determine this shift of planned endpoint. We asked participants to intercept a moving target when it reached a certain area. During the target's motion, background structures briefly moved either leftward or rightward. Participants’ hands responded to background motion even when each background structure was only briefly visible or when the vast majority of background structures remained static. The response was not restricted to motion along the target's path but was most sensitive to motion close to where the target was to be hit, both in the visual field and in depth. In this way, a movement stabilization mechanism provides a comprehensive explanation of many aspects of the manual following response.
Collapse
Affiliation(s)
- Emily M Crowe
- Department of Human Movement Sciences, Institute of Brain and Behavior Amsterdam, Amsterdam Movement Sciences, Vrije Universiteit Amsterdam, The Netherlands.,
| | - Jeroen B J Smeets
- Department of Human Movement Sciences, Institute of Brain and Behavior Amsterdam, Amsterdam Movement Sciences, Vrije Universiteit Amsterdam, The Netherlands.,
| | - Eli Brenner
- Department of Human Movement Sciences, Institute of Brain and Behavior Amsterdam, Amsterdam Movement Sciences, Vrije Universiteit Amsterdam, The Netherlands.,
| |
Collapse
|
8
|
Sivakumar P, Quinlan DJ, Stubbs KM, Culham JC. Grasping performance depends upon the richness of hand feedback. Exp Brain Res 2021; 239:835-846. [PMID: 33403432 DOI: 10.1007/s00221-020-06025-0] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2020] [Accepted: 12/19/2020] [Indexed: 11/28/2022]
Abstract
Although visual feedback of the hand allows fast and accurate grasping actions, little is known about whether the nature of feedback of the hand affects performance. We investigated kinematics during precision grasping (with the index finger and thumb) when participants received different levels of hand feedback, with or without visual feedback of the target. Specifically, we compared performance when participants saw (1) no hand feedback; (2) only the two critical points on the index finger and thumb tips; (3) 21 points on all digit tips and hand joints; (4) 21 points connected by a "skeleton", or (5) full feedback of the hand wearing a glove. When less hand feedback was available, participants took longer to execute the movement because they allowed more time to slow the reach and close the hand. When target feedback was unavailable, participants took longer to plan the movement and reached with higher velocity. We were particularly interested in investigating maximum grip aperture (MGA), which can reflect the margin of error that participants allow to compensate for uncertainty. A trend suggested that MGA was smallest when ample feedback was available (skeleton and full hand feedback, regardless of target feedback) and when only essential information about hand and target was provided (2-point hand feedback + target feedback) but increased when non-essential points were included (21-point feedback). These results suggest that visual feedback of the hand affects grasping performance and that, while more feedback is usually beneficial, this is not necessarily always the case.
Collapse
Affiliation(s)
- Prajith Sivakumar
- Department of Biology, University of Western Ontario, London, Canada.,Brain and Mind Institute, University of Western Ontario, Western Interdisciplinary Research Building, London, ON, Canada
| | - Derek J Quinlan
- Brain and Mind Institute, University of Western Ontario, Western Interdisciplinary Research Building, London, ON, Canada.,BrainsCAN, University of Western Ontario, London, ON, Canada.,Department of Psychology, Huron University College, London, ON, Canada
| | - Kevin M Stubbs
- Brain and Mind Institute, University of Western Ontario, Western Interdisciplinary Research Building, London, ON, Canada.,BrainsCAN, University of Western Ontario, London, ON, Canada.,Department of Psychology, University of Western Ontario, London, ON, Canada
| | - Jody C Culham
- Brain and Mind Institute, University of Western Ontario, Western Interdisciplinary Research Building, London, ON, Canada. .,Department of Psychology, University of Western Ontario, London, ON, Canada.
| |
Collapse
|
9
|
Kuling IA, Laan L, Lopik EV, Smeets JBJ, Brenner E. Looking Precisely at Your Fingertip Requires Visual Guidance of Gaze. Perception 2020; 49:1252-1259. [PMID: 33086914 PMCID: PMC7675761 DOI: 10.1177/0301006620965133] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Abstract
People often look at objects that they are holding in their hands. It is therefore reasonable to expect them to be able to direct their gaze precisely with respect to their fingers. However, we know that people make reproducible idiosyncratic errors of up to a few centimetres when they try to align a visible cursor to their own finger hidden below a surface. To find out whether they also make idiosyncratic errors when they try to look at their finger, we asked participants to hold their finger in front of their head in the dark, and look at it. Participants made idiosyncratic errors of a similar magnitude to those previously found when matching a visual cursor to their hidden finger. This shows that proprioceptive position sense of finger and gaze are not aligned, suggesting that people rely on vision to guide their gaze to their own finger.
Collapse
Affiliation(s)
| | - Lotte Laan
- Vrije Universiteit Amsterdam, The Netherlands
| | | | | | - Eli Brenner
- Vrije Universiteit Amsterdam, The Netherlands
| |
Collapse
|
10
|
Goettker A, Fiehler K, Voudouris D. Somatosensory target information is used for reaching but not for saccadic eye movements. J Neurophysiol 2020; 124:1092-1102. [DOI: 10.1152/jn.00258.2020] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
A systematic investigation of contributions of different somatosensory modalities (proprioception, kinesthesia, tactile) for goal-directed movements is missing. Here we demonstrate that while eye movements are not affected by different types of somatosensory information, reach precision improves when two different types of information are available. Moreover, reach accuracy and gaze precision to unseen somatosensory targets improve when performing coordinated eye-hand movements, suggesting bidirectional contributions of efferent information in reach and eye movement control.
Collapse
Affiliation(s)
- Alexander Goettker
- Experimental Psychology, Justus Liebig University Giessen, Giessen, Germany
| | - Katja Fiehler
- Experimental Psychology, Justus Liebig University Giessen, Giessen, Germany
- Center for Mind, Brain and Behavior (CMBB), University of Marburg and Justus Liebig University, Giessen, Germany
| | - Dimitris Voudouris
- Experimental Psychology, Justus Liebig University Giessen, Giessen, Germany
| |
Collapse
|
11
|
Cesanek E, Domini F. Depth cue reweighting requires altered correlations with haptic feedback. J Vis 2019; 19:3. [PMID: 31826248 DOI: 10.1167/19.14.3] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Depth cue reweighting is a feedback-driven learning process that modifies the relative influences of different sources of three-dimensional shape information in perceptual judgments and or motor planning. In this study, we investigated the mechanism supporting reweighting of stereo and texture information by manipulating the haptic feedback obtained during a series of grasping movements. At the end of each grasp, the fingers closed down on a physical object that was consistent with one of the two cues, depending on the condition. Previous studies have shown that this style of visuomotor training leads to cue reweighting for perceptual judgments, but the time course has never been documented for a single training session, and many questions remain regarding the underlying mechanism, such as the pattern of feedback signals required to drive reweighting. We address these issues in two experiments, finding short-term changes in the motor response consistent with cue reweighting: the slope of the grip aperture with respect to the reliable cue increased, whereas the slope with respect to the unreliable cue decreased. Critically, Experiment 2 shows that slope changes do not occur when one of the cues is rendered with a constant bias; the grip aperture simply becomes uniformly larger or smaller. Our findings support a model of cue reweighting driven by altered correlations between haptic feedback and individual cues, rather than simple mismatches, which can be resolved by other mechanisms such as sensorimotor adaptation or cue recalibration.
Collapse
Affiliation(s)
- Evan Cesanek
- Department of Cognitive, Linguistic, & Psychological Sciences, Brown University, Providence, RI, USA
| | - Fulvio Domini
- Department of Cognitive, Linguistic, & Psychological Sciences, Brown University, Providence, RI, USA
| |
Collapse
|
12
|
Gregori V, Cognolato M, Saetta G, Atzori M, Gijsberts A. On the Visuomotor Behavior of Amputees and Able-Bodied People During Grasping. Front Bioeng Biotechnol 2019; 7:316. [PMID: 31799243 PMCID: PMC6874164 DOI: 10.3389/fbioe.2019.00316] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2019] [Accepted: 10/24/2019] [Indexed: 11/15/2022] Open
Abstract
Visual attention is often predictive for future actions in humans. In manipulation tasks, the eyes tend to fixate an object of interest even before the reach-to-grasp is initiated. Some recent studies have proposed to exploit this anticipatory gaze behavior to improve the control of dexterous upper limb prostheses. This requires a detailed understanding of visuomotor coordination to determine in which temporal window gaze may provide helpful information. In this paper, we verify and quantify the gaze and motor behavior of 14 transradial amputees who were asked to grasp and manipulate common household objects with their missing limb. For comparison, we also include data from 30 able-bodied subjects who executed the same protocol with their right arm. The dataset contains gaze, first person video, angular velocities of the head, and electromyography and accelerometry of the forearm. To analyze the large amount of video, we developed a procedure based on recent deep learning methods to automatically detect and segment all objects of interest. This allowed us to accurately determine the pixel distances between the gaze point, the target object, and the limb in each individual frame. Our analysis shows a clear coordination between the eyes and the limb in the reach-to-grasp phase, confirming that both intact and amputated subjects precede the grasp with their eyes by more than 500 ms. Furthermore, we note that the gaze behavior of amputees was remarkably similar to that of the able-bodied control group, despite their inability to physically manipulate the objects.
Collapse
Affiliation(s)
- Valentina Gregori
- Department of Computer, Control, and Management Engineering, University of Rome La Sapienza, Rome, Italy.,VANDAL Laboratory, Istituto Italiano di Tecnologia, Genoa, Italy
| | - Matteo Cognolato
- Information Systems Institute, University of Applied Sciences Western Switzerland (HES-SO Valais), Sierre, Switzerland.,Rehabilitation Engineering Laboratory, Department of Health Sciences and Technology, ETH Zurich, Zurich, Switzerland
| | - Gianluca Saetta
- Department of Neurology, University Hospital of Zurich, Zurich, Switzerland
| | - Manfredo Atzori
- Information Systems Institute, University of Applied Sciences Western Switzerland (HES-SO Valais), Sierre, Switzerland
| | | | - Arjan Gijsberts
- VANDAL Laboratory, Istituto Italiano di Tecnologia, Genoa, Italy
| |
Collapse
|
13
|
Grant S, Conway ML. Some binocular advantages for planning reach, but not grasp, components of prehension. Exp Brain Res 2019; 237:1239-1255. [PMID: 30850853 PMCID: PMC6557882 DOI: 10.1007/s00221-019-05503-4] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2018] [Accepted: 02/25/2019] [Indexed: 11/04/2022]
Abstract
Proficient (fast, accurate, precise) hand actions for reaching-to-grasp 3D objects are known to benefit significantly from the use of binocular vision compared to one eye alone. We examined whether these binocular advantages derive from increased reliability in encoding the goal object’s properties for feedforward planning of prehension movements or from enhanced feedback mediating their online control. Adult participants reached for, precision grasped and lifted cylindrical table-top objects (two sizes, 2 distances) using binocular vision or only their dominant/sighting eye or their non-dominant eye to program and fully execute their movements or using each of the three viewing conditions only to plan their reach-to-grasp during a 1 s preview, with vision occluded just before movement onset. Various kinematic measures of reaching and grasping proficiency, including corrective error rates, were quantified and compared by view, feedback and object type. Some significant benefits of binocular over monocular vision when they were just available for pre-movement planning were retained for the reach regardless of target distance, including higher peak velocities, straighter paths and shorter low velocity approach times, although these latter were contaminated by more velocity corrections and by poorer coordination with object contact. By contrast, virtually all binocular advantages for grasping, including improvements in peak grip aperture scaling, the accuracy and precision of digit placements at object contact and shorter grip application times preceding the lift, were eliminated with no feedback available, outcomes that were influenced by the object’s size. We argue that vergence cues can improve the reliability of binocular internal representations of object distance for the feedforward programming of hand transport, whereas the major benefits of binocular vision for enhancing grasping performance derive exclusively from its continuous presence online.
Collapse
Affiliation(s)
- Simon Grant
- Applied Vision Research Centre, City, University of London, Northampton Square, London, EC1V 0HB, UK.
| | - Miriam L Conway
- Applied Vision Research Centre, City, University of London, Northampton Square, London, EC1V 0HB, UK
| |
Collapse
|