1
|
McCormack GL, Kulowski KA. Image Size and the Range of Clear and Single Binocular Vision in 3D Displays. Optom Vis Sci 2021; 98:947-958. [PMID: 34460455 PMCID: PMC8409099 DOI: 10.1097/opx.0000000000001745] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022] Open
Abstract
SIGNIFICANCE The range of clear and single binocular vision differs between 3D displays and clinical prism vergences, but this difference is unexplained. This difference prevents clinicians from predicting the range of clear and single binocular vision in 3D-viewing patients. In this study, we tested a hypothesis for this difference. PURPOSE The purpose of this study was to determine whether changing fixation target size in 3D viewing significantly affects the vergence ranges and, if so, then to determine whether the target size effect is driven by fusional vergence gain changes, threshold of blur changes, or both. METHODS Twenty-one visually normal adults aged 18 to 28 years viewed 3D images at 40 cm in an electronic stereoscopic. The fixation target, a Maltese cross, moved in depth at 2∆/s by way of changing crossed or uncrossed disparity until blur and diplopia ensued. We used four target sizes: (1) small (width × height, 0.21° × 0.63°), (2) medium (1.43° × 4.3°), (3) large (3.6° × 10.8°), and (4) 3D (size changing congruently with disparity). The effect of target size on responses was tested by mixed ANOVAs. RESULT Mean convergence blurs and breaks increased with target size by 40% (P < .001) and 71% (P < .001), respectively, and in divergence by 33% (P = .03) and 30% (P = .04), respectively. The increases in break magnitude with target size implicate fusional vergence gain change in the size effect. Increasing target size raised the threshold of blur from 1.06 to 1.82 D in convergence and from 0.97 to 1.48 D in divergence (P = .008). CONCLUSIONS Growing fixation target size in 3D viewing increases fusional vergence gain and blur thresholds, which together increase the limits of clear and single binocular vision. Therefore, the clarity of a 3D image depends not only on its disparity but also on the size of the viewed image.
Collapse
|
2
|
The impact of retinal motion on stereoacuity for physical targets. Vision Res 2019; 161:43-51. [PMID: 31202606 DOI: 10.1016/j.visres.2019.06.003] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2018] [Revised: 06/04/2019] [Accepted: 06/05/2019] [Indexed: 11/22/2022]
Abstract
In a series of studies using physical targets, we examined the effect of lateral retinal motion on stereoscopic depth discrimination thresholds. We briefly presented thin vertical lines, along with a fixation marker, at speeds ranging from 0 to 16 deg·s-1. Previous investigations of the effect of retinal motion on stereoacuity consistently show that there is little impact of retinal motion up to 2 deg·s-1, however, thresholds appear to rise steeply at higher velocities (greater than 3 deg·s-1). These prior experiments used computerized displays to generate their stimuli. In contrast, with our physical targets we find that stereoacuity is stable up to 16 deg·s-1, even in the presence of appreciable smearing due to visual persistence. We show that this discrepancy cannot be explained by differences in viewing time, prevalence of motion smear or by high frequency flicker due to display updates. We conclude that under natural viewing conditions observers are able to make depth discrimination judgements using binocular disparity signals that are rapidly acquired at stimulus onset.
Collapse
|
3
|
McDonough KL, Hudson M, Bach P. Cues to intention bias action perception toward the most efficient trajectory. Sci Rep 2019; 9:6472. [PMID: 30996227 PMCID: PMC6470138 DOI: 10.1038/s41598-019-42204-y] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2018] [Accepted: 03/21/2019] [Indexed: 11/15/2022] Open
Abstract
Humans interpret others' behaviour as intentional and expect them to take the most energy-efficient path to achieve their goals. Recent studies show that these expectations of efficient action take the form of a prediction of an ideal "reference" trajectory, against which observed actions are evaluated, distorting their perceptual representation towards this expected path. Here we tested whether these predictions depend upon the implied intentionality of the stimulus. Participants saw videos of an actor reaching either efficiently (straight towards an object or arched over an obstacle) or inefficiently (straight towards obstacle or arched over empty space). The hand disappeared mid-trajectory and participants reported the last seen position on a touch-screen. As in prior research, judgments of inefficient actions were biased toward efficiency expectations (straight trajectories upwards to avoid obstacles, arched trajectories downward towards goals). In two further experimental groups, intentionality cues were removed by replacing the hand with a non-agentive ball (group 2), and by removing the action's biological motion profile (group 3). Removing these cues substantially reduced perceptual biases. Our results therefore confirm that the perception of others' actions is guided by expectations of efficient actions, which are triggered by the perception of semantic and motion cues to intentionality.
Collapse
Affiliation(s)
| | - Matthew Hudson
- University of Plymouth, School of Psychology, Plymouth, PL48AA, UK
- School of Business, National College of Ireland, Mayor Street, Dublin 1, Ireland
| | - Patric Bach
- University of Plymouth, School of Psychology, Plymouth, PL48AA, UK
| |
Collapse
|
4
|
Hudson M, McDonough KL, Edwards R, Bach P. Perceptual teleology: expectations of action efficiency bias social perception. Proc Biol Sci 2018; 285:rspb.2018.0638. [PMID: 30089623 PMCID: PMC6111183 DOI: 10.1098/rspb.2018.0638] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2018] [Accepted: 07/12/2018] [Indexed: 01/09/2023] Open
Abstract
Primates interpret conspecific behaviour as goal-directed and expect others to achieve goals by the most efficient means possible. While this teleological stance is prominent in evolutionary and developmental theories of social cognition, little is known about the underlying mechanisms. In predictive models of social cognition, a perceptual prediction of an ideal efficient trajectory would be generated from prior knowledge against which the observed action is evaluated, distorting the perception of unexpected inefficient actions. To test this, participants observed an actor reach for an object with a straight or arched trajectory on a touch screen. The actions were made efficient or inefficient by adding or removing an obstructing object. The action disappeared mid-trajectory and participants touched the last seen screen position of the hand. Judgements of inefficient actions were biased towards the efficient prediction (straight trajectories upward to avoid the obstruction, arched trajectories downward towards the target). These corrections increased when the obstruction's presence/absence was explicitly acknowledged, and when the efficient trajectory was explicitly predicted. Additional supplementary experiments demonstrated that these biases occur during ongoing visual perception and/or immediately after motion offset. The teleological stance is at least partly perceptual, providing an ideal reference trajectory against which actual behaviour is evaluated.
Collapse
Affiliation(s)
- Matthew Hudson
- School of Psychology, University of Plymouth, Plymouth, UK
| | | | - Rhys Edwards
- School of Psychology, University of Plymouth, Plymouth, UK
| | - Patric Bach
- School of Psychology, University of Plymouth, Plymouth, UK
| |
Collapse
|
5
|
Hammett ST, Georgeson MA, Barbieri-Hesse GS. Motion, Flash, and Flicker: A Unified Spatiotemporal Model of Perceived Edge Sharpening. Perception 2016; 32:1221-32. [PMID: 14700257 DOI: 10.1068/p3400] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
Abstract
Blurred edges appear sharper in motion than when they are stationary. We proposed a model of this motion sharpening that invokes a local, nonlinear contrast transducer function (Hammett et al, 1998 Vision Research38 2099–2108). Response saturation in the transducer compresses or ‘clips’ the input spatial waveform, rendering the edges as sharper. To explain the increasing distortion of drifting edges at higher speeds, the degree of nonlinearity must increase with speed or temporal frequency. A dynamic contrast gain control before the transducer can account for both the speed dependence and approximate contrast invariance of motION sharpening (Hammett et al, 2003 Vision Research in press). We show here that this model also predicts perceived sharpening of briefly flashed and flickering edges, and we show that the model can account fairly well for experimental data from all three modes of presentation (motion, flash, and flicker). At moderate durations and lower temporal frequencies the gain control attenuates the input signal, thus protecting it from later compression by the transducer. The gain control is somewhat sluggish, and so it suffers both a slow onset, and loss of power at high temporal frequencies. Consequently, brief presentations and high temporal frequencies of drift and flicker are less protected from distortion, and show greater perceptual sharpening.
Collapse
Affiliation(s)
- Stephen T Hammett
- Department of Psychology, Royal Holloway College, University of London, Egham, Surrey TW20 0EX, UK.
| | | | | |
Collapse
|
6
|
Öğmen H, Herzog MH. A New Conceptualization of Human Visual Sensory-Memory. Front Psychol 2016; 7:830. [PMID: 27375519 PMCID: PMC4899472 DOI: 10.3389/fpsyg.2016.00830] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2016] [Accepted: 05/18/2016] [Indexed: 11/16/2022] Open
Abstract
Memory is an essential component of cognition and disorders of memory have significant individual and societal costs. The Atkinson–Shiffrin “modal model” forms the foundation of our understanding of human memory. It consists of three stores: Sensory Memory (SM), whose visual component is called iconic memory, Short-Term Memory (STM; also called working memory, WM), and Long-Term Memory (LTM). Since its inception, shortcomings of all three components of the modal model have been identified. While the theories of STM and LTM underwent significant modifications to address these shortcomings, models of the iconic memory remained largely unchanged: A high capacity but rapidly decaying store whose contents are encoded in retinotopic coordinates, i.e., according to how the stimulus is projected on the retina. The fundamental shortcoming of iconic memory models is that, because contents are encoded in retinotopic coordinates, the iconic memory cannot hold any useful information under normal viewing conditions when objects or the subject are in motion. Hence, half-century after its formulation, it remains an unresolved problem whether and how the first stage of the modal model serves any useful function and how subsequent stages of the modal model receive inputs from the environment. Here, we propose a new conceptualization of human visual sensory memory by introducing an additional component whose reference-frame consists of motion-grouping based coordinates rather than retinotopic coordinates. We review data supporting this new model and discuss how it offers solutions to the paradoxes of the traditional model of sensory memory.
Collapse
Affiliation(s)
- Haluk Öğmen
- Department of Electrical and Computer Engineering, University of HoustonHouston, TX, USA; Center for Neuro-Engineering and Cognitive Science, University of HoustonHouston, TX, USA
| | - Michael H Herzog
- Laboratory of Psychophysics, Ecole Polytechnique Fédérale de Lausanne (EPFL) Lausanne, Switzerland
| |
Collapse
|
7
|
Retinotopy of visual masking and non-retinotopic perception during masking. Atten Percept Psychophys 2015; 77:1263-84. [DOI: 10.3758/s13414-015-0844-2] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
8
|
Oh T, Park J, Seshadrinathan K, Lee S, Bovik AC. No-reference sharpness assessment of camera-shaken images by analysis of spectral structure. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2014; 23:5428-5439. [PMID: 25350928 DOI: 10.1109/tip.2014.2364925] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
The tremendous explosion of image-, video-, and audio-enabled mobile devices, such as tablets and smart-phones in recent years, has led to an associated dramatic increase in the volume of captured and distributed multimedia content. In particular, the number of digital photographs being captured annually is approaching 100 billion in just the U.S. These pictures are increasingly being acquired by inexperienced, casual users under highly diverse conditions leading to a plethora of distortions, including blur induced by camera shake. In order to be able to automatically detect, correct, or cull images impaired by shake-induced blur, it is necessary to develop distortion models specific to and suitable for assessing the sharpness of camera-shaken images. Toward this goal, we have developed a no-reference framework for automatically predicting the perceptual quality of camera-shaken images based on their spectral statistics. Two kinds of features are defined that capture blur induced by camera shake. One is a directional feature, which measures the variation of the image spectrum across orientations. The second feature captures the shape, area, and orientation of the spectral contours of camera shaken images. We demonstrate the performance of an algorithm derived from these features on new and existing databases of images distorted by camera shake.
Collapse
|
9
|
Pavan A, Bellacosa Marotti R. Motion streaks do not influence the perceived position of stationary flashed objects. ScientificWorldJournal 2012; 2012:909547. [PMID: 22645464 PMCID: PMC3354717 DOI: 10.1100/2012/909547] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2011] [Accepted: 12/05/2011] [Indexed: 12/05/2022] Open
Abstract
In the present study, we investigated whether motion streaks, produced by fast moving dots Geisler 1999, distort the positional map of stationary flashed objects producing the well-known motion-induced position shift illusion (MIPS). The illusion relies on motion-processing mechanisms that induce local distortions in the positional map of the stimulus which is derived by shape-processing mechanisms. To measure the MIPS, two horizontally offset Gaussian blobs, placed above and below a central fixation point, were flashed over two fields of dots moving in opposite directions. Subjects judged the position of the top Gaussian blob relative to the bottom one. The results showed that neither fast (motion streaks) nor slow moving dots influenced the perceived spatial position of the stationary flashed objects, suggesting that background motion does not interact with the shape-processing mechanisms involved in MIPS.
Collapse
Affiliation(s)
- Andrea Pavan
- Cognitive Neuroscience Sector, International School for Advanced Studies-Via Bonomea 265, 34136 Trieste, Italy.
| | | |
Collapse
|
10
|
Herzog MH, Otto TU, Ogmen H. The fate of visible features of invisible elements. Front Psychol 2012; 3:119. [PMID: 22557985 PMCID: PMC3338119 DOI: 10.3389/fpsyg.2012.00119] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2012] [Accepted: 04/01/2012] [Indexed: 11/13/2022] Open
Abstract
To investigate the integration of features, we have developed a paradigm in which an element is rendered invisible by visual masking. Still, the features of the element are visible as part of other display elements presented at different locations and times (sequential metacontrast). In this sense, we can "transport" features non-retinotopically across space and time. The features of the invisible element integrate with features of other elements if and only if the elements belong to the same spatio-temporal group. The mechanisms of this kind of feature integration seem to be quite different from classical mechanisms proposed for feature binding. We propose that feature processing, binding, and integration occur concurrently during processes that group elements into wholes.
Collapse
Affiliation(s)
- Michael H Herzog
- Laboratory of Psychophysics, Ecole Polytechnique Fédérale de Lausanne (EPFL) Lausanne, Switzerland
| | | | | |
Collapse
|
11
|
|
12
|
Hoffman DM, Karasev VI, Banks MS. Temporal presentation protocols in stereoscopic displays: Flicker visibility, perceived motion, and perceived depth. JOURNAL OF THE SOCIETY FOR INFORMATION DISPLAY 2011; 19:271-297. [PMID: 21572544 PMCID: PMC3092720 DOI: 10.1889/jsid19.3.271] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/28/2023]
Abstract
Most stereoscopic displays rely on field-sequential presentation to present different images to the left and right eyes. With sequential presentation, images are delivered to each eye in alternation with dark intervals, and each eye receives its images in counter phase with the other eye. This type of presentation can exacerbate image artifacts including flicker, and the appearance of unsmooth motion. To address the flicker problem, some methods repeat images multiple times before updating to new ones. This greatly reduces flicker visibility, but makes motion appear less smooth. This paper describes an investigation of how different presentation methods affect the visibility of flicker, motion artifacts, and distortions in perceived depth. It begins with an examination of these methods in the spatio-temporal frequency domain. From this examination, it describes a series of predictions for how presentation rate, object speed, simultaneity of image delivery to the two eyes, and other properties ought to affect flicker, motion artifacts, and depth distortions, and reports a series of experiments that tested these predictions. The results confirmed essentially all of the predictions. The paper concludes with a summary and series of recommendations for the best approach to minimize these undesirable effects.
Collapse
|
13
|
The representation of moving 3-D objects in apparent motion perception. Atten Percept Psychophys 2009; 71:1294-304. [DOI: 10.3758/app.71.6.1294] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
14
|
Oğmen H. A theory of moving form perception: Synergy between masking, perceptual grouping, and motion computation in retinotopic and non-retinotopic representations. Adv Cogn Psychol 2008; 3:67-84. [PMID: 20517499 PMCID: PMC2864981 DOI: 10.2478/v10053-008-0015-2] [Citation(s) in RCA: 24] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2006] [Accepted: 11/21/2006] [Indexed: 11/20/2022] Open
Abstract
Because object and self-motion are ubiquitous in natural viewing conditions, understanding how the human visual system achieves a relatively clear perception for moving objects is a fundamental problem in visual perception. Several studies have shown that the visible persistence of a briefly presented stationary stimulus is approximately 120 ms under normal viewing conditions. Based on this duration of visible persistence, we would expect moving objects to appear highly blurred. However, in human vision, objects in motion typically appear relatively sharp and clear. We suggest that clarity of form in dynamic viewing is achieved by a synergy between masking, perceptual grouping, and motion computation across retinotopic and non-retinotopic representations. We also argue that dissociations observed in masking are essential to create and maintain this synergy.
Collapse
Affiliation(s)
- Haluk Oğmen
- Department of Electrical & Computer Engineering, Center for Neuro-Engineering & Cognitive Science, University of Houston, Houston, TX 77204-4005 USA
| |
Collapse
|
15
|
Scharnowski F, Hermens F, Kammer T, Oğmen H, Herzog MH. Feature fusion reveals slow and fast visual memories. J Cogn Neurosci 2007; 19:632-41. [PMID: 17381254 DOI: 10.1162/jocn.2007.19.4.632] [Citation(s) in RCA: 22] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Although the visual system can achieve a coarse classification of its inputs in a relatively short time, the synthesis of qualia-rich and detailed percepts can take substantially more time. If these prolonged computations were to take place in a retinotopic space, moving objects would generate extensive smear. However, under normal viewing conditions, moving objects appear relatively sharp and clear, suggesting that a substantial part of visual short-term memory takes place at a nonretinotopic locus. By using a retinotopic feature fusion and a nonretinotopic feature attribution paradigm, we provide evidence for a relatively fast retinotopic buffer and a substantially slower nonretinotopic memory. We present a simple model that can account for the dynamics of these complementary memory processes. Taken together, our results indicate that the visual system can accomplish temporal integration of information while avoiding smear by breaking off sensory memory into fast and slow components that are implemented in retinotopic and nonretinotopic loci, respectively.
Collapse
|
16
|
Chung STL, Patel SS, Bedell HE, Yilmaz O. Spatial and temporal properties of the illusory motion-induced position shift for drifting stimuli. Vision Res 2007; 47:231-43. [PMID: 17190608 PMCID: PMC2734886 DOI: 10.1016/j.visres.2006.10.008] [Citation(s) in RCA: 42] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2006] [Revised: 09/27/2006] [Accepted: 10/24/2006] [Indexed: 11/18/2022]
Abstract
The perceived position of a stationary Gaussian window of a Gabor target shifts in the direction of motion of the Gabor's carrier stimulus, implying the presence of interactions between the specialized visual areas that encode form, position, and motion. The purpose of this study was to examine the temporal and spatial properties of this illusory motion-induced position shift (MIPS). We measured the magnitude of the MIPS for a pair of horizontally separated (2 or 8deg) truncated-Gabor stimuli (carrier=1 or 4cpd sinusoidal grating, Gaussian envelope SD=18arc min, 50% contrast) or a pair of Gaussian-windowed random-texture patterns that drifted vertically in opposite directions. The magnitude of the MIPS was measured for drift speeds up to 16deg/s and for stimulus durations up to 453ms. The temporal properties of the MIPS depended on the drift speed. At low velocities, the magnitude of the MIPS increased monotonically with the stimulus duration. At higher velocities, the magnitude of the MIPS increased with duration initially, then decreased between approximately 45 and 75ms before rising to reach a steady-state value at longer durations. In general, the magnitude of the MIPS was larger when the truncated-Gabor or random-texture stimuli were more spatially separated, but was similar for the different types of carrier stimuli. Our results are consistent with a framework that suggests that perceived form is modulated dynamically during stimulus motion.
Collapse
Affiliation(s)
- Susana T L Chung
- College of Optometry, University of Houston, Houston, TX 77204, USA.
| | | | | | | |
Collapse
|
17
|
Tsui SY, Khuu SK, Hayes A. The perceived position shift of a pattern that contains internal motion is accompanied by a change in the pattern’s apparent size and shape. Vision Res 2007; 47:402-10. [PMID: 17184808 DOI: 10.1016/j.visres.2006.11.003] [Citation(s) in RCA: 12] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2006] [Revised: 11/06/2006] [Accepted: 11/07/2006] [Indexed: 11/17/2022]
Abstract
When the sinusoidal grating of a "Gabor pattern" is drifted, the apparent position of the pattern shifts in the direction of motion [De Valois, R. L., & De Valois, K. K. (1991). Vernier acuity with stationary moving Gabors. Vision Research, 31, 1619-1626]. We investigated the underlying cause of this illusion by determining whether the effect is a consequence of the internal motion shifting the perceived position of the whole pattern, or a consequence of a shift in the perceived location of the centroid (centre of mass) of the Gabor envelope. While each of these two possible distortions can account for a perceived positional offset, they give different predictions for the apparent size of the stimulus. A simple shift in perceived position results in no change in apparent size, while a centroid shift will likely result in either a decrease or an increase in the pattern's apparent size, depending on whether the trailing or leading edge of the Gabor stimulus is most affected by motion. We examined whether there is a change in the apparent size of Gabor patterns containing a range of grating motion speeds. We found that the perceived size of the pattern increased in the presence of motion as a function of speed, and is thus consistent with a centroid-shift explanation. We verified that this size change is a consequence of an increase in contrast at the leading edge, since the leading edge appears elongated relative to the trailing edge. We furthermore showed that the apparent-position shifts due to motion can be negated by displacing the centroid in the opposite direction to the motion.
Collapse
Affiliation(s)
- Sum Yin Tsui
- Department of Psychology, The University of Hong Kong, Pokfulam Road, Hong Kong, China.
| | | | | |
Collapse
|
18
|
Whitney D, Bressler DW. Spatially asymmetric response to moving patterns in the visual cortex: re-examining the local sign hypothesis. Vision Res 2006; 47:50-9. [PMID: 17049580 PMCID: PMC3890257 DOI: 10.1016/j.visres.2006.08.030] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2006] [Revised: 08/11/2006] [Accepted: 08/15/2006] [Indexed: 11/28/2022]
Abstract
One of the most fundamental functions of the visual system is to code the positions of objects. Most studies, especially those using fMRI, widely assume that the location of the peak retinotopic activity generated in the visual cortex by an object is the position assigned to that object-this is a simplified version of the local sign hypothesis. Here, we employed a novel technique to compare the pattern of responses to moving and stationary objects and found that the local sign hypothesis is false. By spatially correlating populations of voxel responses to different moving and stationary stimuli in different positions, we recovered the modulation transfer function for moving patterns. The results show that the pattern of responses to a moving object is best correlated with the response to a static object that is located behind the moving one. The pattern of responses across the visual cortex was able to distinguish object positions separated by about 0.25 deg visual angle, equivalent to approximately 0.25 mm cortical distance. We also found that the position assigned to a pattern is not simply dictated by the peak activity-the shape of the luminance envelope and the resulting shape of the population response, including the shape and skew in the response at the edges of the pattern, influences where the visual cortex assigns the object's position. Therefore, visually coded position is not conveyed by the peak but by the overall profile of activity.
Collapse
Affiliation(s)
- David Whitney
- The Center for Mind and Brain, The Department of Psychology, The University of California, Davis, CA 95616, USA.
| | | |
Collapse
|
19
|
Abstract
The relative visual position of a briefly flashed stimulus is systematically modified in the presence of motion signals. We investigated the two-dimensional distortion of the positional representation of a flash relative to a moving stimulus. Analysis of the spatial pattern of mislocalization revealed that the perceived position of a flash was not uniformly displaced, but instead shifted toward a single point of convergence that followed the moving object from behind at a fixed distance. Although the absolute magnitude of mislocalization increased with motion speed, the convergence point remained unaffected. The motion modified the perceived position of a flash, but had little influence on the perceived shape of a spatially extended flash stimulus. These results demonstrate that motion anisotropically distorts positional representation after the shapes of objects are represented. Furthermore, the results imply that the flash-lag effect may be considered a special case of two-dimensional anisotropic distortion.
Collapse
Affiliation(s)
- Katsumi Watanabe
- Research Center for Advanced Science and Technology, The University of Tokyo, Tokyo, Japan.
| | | |
Collapse
|
20
|
Whitney D, Goltz HC, Thomas CG, Gati JS, Menon RS, Goodale MA. Flexible retinotopy: motion-dependent position coding in the visual cortex. Science 2003; 302:878-81. [PMID: 14500849 PMCID: PMC3849414 DOI: 10.1126/science.1087839] [Citation(s) in RCA: 84] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
Abstract
Although the visual cortex is organized retinotopically, it is not clear whether the cortical representation of position necessarily reflects perceived position. Using functional magnetic resonance imaging (fMRI), we show that the retinotopic representation of a stationary object in the cortex was systematically shifted when visual motion was present in the scene. Whereas the object could appear shifted in the direction of the visual motion, the representation of the object in the visual cortex was always shifted in the opposite direction. The results show that the representation of position in the primary visual cortex, as revealed by fMRI, can be dissociated from perceived location.
Collapse
Affiliation(s)
- David Whitney
- Department of Psychology, University of Western Ontario, London, Ontario N6A 5C2, Canada.
| | | | | | | | | | | |
Collapse
|
21
|
Abstract
On the basis of the contrast sensitivity loss in amblyopia which mainly affects higher spatial frequencies, one would expect amblyopes to perceive sharp edges as blurred. We show that they perceive sharp edges as sharp and have veridical edge blur perception. Contrary to the currently accepted view, this suggests that the amblyopic visual system is not characterized by a blurred visual representation.
Collapse
Affiliation(s)
- Robert F Hess
- Department of Ophthalmology, McGill Vision Research Unit, McGill University, Que., Montreal, Canada.
| | | | | | | |
Collapse
|
22
|
Hammett ST, Georgeson MA, Bedingham S, Barbieri-Hesse GS. Motion sharpening and contrast: gain control precedes compressive non-linearity? Vision Res 2003; 43:1187-99. [PMID: 12705958 DOI: 10.1016/s0042-6989(03)00071-3] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
Abstract
Blurred edges appear sharper in motion than when they are stationary. We (Vision Research 38 (1998) 2108) have previously shown how such distortions in perceived edge blur may be accounted for by a model which assumes that luminance contrast is encoded by a local contrast transducer whose response becomes progressively more compressive as speed increases. If the form of the transducer is fixed (independent of contrast) for a given speed, then a strong prediction of the model is that motion sharpening should increase with increasing contrast. We measured the sharpening of periodic patterns over a large range of contrasts, blur widths and speeds. The results indicate that whilst sharpening increases with speed it is practically invariant with contrast. The contrast invariance of motion sharpening is not explained by an early, static compressive non-linearity alone. However, several alternative explanations are also inconsistent with these results. We show that if a dynamic contrast gain control precedes the static non-linear transducer then motion sharpening, its speed dependence, and its invariance with contrast, can be predicted with reasonable accuracy.
Collapse
Affiliation(s)
- Stephen T Hammett
- Department of Psychology, Royal Holloway College, University of London, Egham, Surrey TW20 0EX, UK.
| | | | | | | |
Collapse
|
23
|
Abstract
It is widely supposed that things tend to look blurred when they are moving fast. Previous work has shown that this is true for sharp edges but, paradoxically, blurred edges look sharper when they are moving than when stationary. This is 'motion sharpening'. We show that blurred edges also look up to 50% sharper when they are presented briefly (8-24 ms) than at longer durations (100-500 ms) without motion. This argues strongly against high-level models of sharpening based specifically on compensation for motion blur. It also argues against a recent, low-level, linear filter model that requires motion to produce sharpening. No linear filter model can explain our finding that sharpening was similar for sinusoidal and non-sinusoidal gratings, since linear filters can never distort sine waves. We also conclude that the idea of a 'default' assumption of sharpness is not supported by experimental evidence. A possible source of sharpening is a nonlinearity in the contrast response of early visual mechanisms to fast or transient temporal changes, perhaps based on the magnocellular (M-cell) pathway. Our finding that sharpening is not diminished at low contrast sets strong constraints on the nature of the nonlinearity.
Collapse
|
24
|
Abstract
Human observers are not normally conscious of blur from moving objects [Nature 284 (1980) 164]. Several recent reports have even shown that blurred images appear sharper when drifting than when stationary and have suggested different non-linear mechanisms to explain this phenomenon [Vision Res. 36 (1996) 2729; Vision Res. 38 (1998) 2099]. We demonstrate here that even though distortions of drifting narrow-band sine-wave gratings cannot be explained by linear mechanisms, these mechanisms may have an important role in sharpening of moving edges. We show first that the effective spatial filter for a moving object that is formed by a simple difference-of-Gaussians spatial filter and the typical biphasic temporal impulse response function can be approximated by a combination of Gaussian filters only. When this filter is applied to moving, Gaussian-blurred edges, regions of blurring and sharpening are found over the same ranges of blur widths and velocities where recent experimental findings have shown them to exist. In general, that means that the output of the filter shows blurring in response to small blur widths and sharpening in response to larger blur widths.
Collapse
Affiliation(s)
- A K Pääkkönen
- Department of Clinical Neuropysiology, Kuopio University and Kuopio University Hospital, PO Box 1777, FIN-70211, Kuopio, Finland.
| | | |
Collapse
|
25
|
Power GJ, Sturtz KE. Signal-to-noise ratio for temporal integrated drifting images: a model for perceived image sharpening. APPLIED OPTICS 2000; 39:6613-6620. [PMID: 18354675 DOI: 10.1364/ao.39.006613] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
A formulation of signal-to-noise ratio is constructed that uses temporal integrated images from image sequences. Given a blurred image that drifts horizontally at various speeds and at various linear blurs, we prove that this formulation of the signal-to-noise ratio consistently increases with an increase in speed. This increase is shown to model the trends in the human vision system by which drifting blurred images are perceived with increased sharpness. The existing widely used objective quality techniques fail to model the perceptual increase in sharpness. This new formulation, along with other objective quality measures, is tested on several blurred drifting image sequences. The new formulation reflects the theoretically predicted increase in perceived sharpness.
Collapse
Affiliation(s)
- G J Power
- Air Force Research Laboratory, Target Recognition Branch, AFRLySNAT, Building 620, 2241 Avionics Circle, Wright-Patterson Air Force Base, Ohio 45433-7321, USA.
| | | |
Collapse
|
26
|
Abstract
The apparent contrast of a center pattern depends on the contrast of its surround. To examine the suprathreshold perception of moving patterns, we measured the perceived contrast of a moving grating while the direction and speed of the surround patterns varied. Subjects matched the apparent contrast of a center patch embedded in surround patches to that of a patch with no surround pattern. Temporal frequency, Michelson contrast and movement direction of both center and surround patterns varied systematically. We found that: (1) contrast reduction is most prominent when the center and surround have the same velocity (velocity selectivity); (2) contrast enhancement occurs when the surround moves at a higher speed than the center, if the difference in temporal frequencies of center and surround exceeds 10-20, independent of the directional relationship between center and surround; (3) contrast reduction is stronger for higher surround contrasts with lower center contrasts; and (4) contrast enhancement is relatively unaffected by center and surround contrasts. We conclude that the contrast perception of moving patterns is influenced by directionally-selective mechanisms except at high temporal frequencies. Our results further suggest that there is not only the lateral inhibition often assumed to influence contrast gain control, but also an excitatory connection between motion encoding units.
Collapse
Affiliation(s)
- T Takeuchi
- NTT Communication Science Laboratories, Atsugi-shi, Morinosato-Wakamiya 3-1, 243-0198, Kanagawa, Japan.
| | | |
Collapse
|
27
|
Galvin SJ, O'Shea RP, Squire AM, Hailstone DS. Sharpness overconstancy: the roles of visibility and current context. Vision Res 1999; 39:2649-57. [PMID: 10492827 DOI: 10.1016/s0042-6989(98)00306-x] [Citation(s) in RCA: 15] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Abstract
In a previous study we found that blurred edges presented in peripheral vision look sharper than when they are looked at directly, a phenomenon we have called peripheral sharpness overconstancy (Galvin et al. (1997). Vision Research, 37, 2035-2039). In the current study we show that when visibility of the stimulus edges is compromised by very brief presentations, we can demonstrate sharpness overconstancy for static, foveal viewing. We also test whether the degree of sharpening is a function of the current visual context, but find no difference between the peripheral sharpness overconstancy (at 24 degrees eccentricity) of edges measured in a blurred context and that measured in a sharp context. We conclude that if the visual system does carry a template for sharp edges which contributes to edge appearance when visibility is poor, then that template is resistant to changes in context.
Collapse
Affiliation(s)
- S J Galvin
- Department of Psychology, University of Otago, Dunedin, New Zealand.
| | | | | | | |
Collapse
|
28
|
Hammett ST, Georgeson MA, Gorea A. Motion blur and motion sharpening: temporal smear and local contrast non-linearity. Vision Res 1998; 38:2099-108. [PMID: 9797970 DOI: 10.1016/s0042-6989(97)00430-6] [Citation(s) in RCA: 28] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
Abstract
Blurred images may appear sharper when drifting than when stationary. But, paradoxically, moving sharp edges may appear more blurred. To resolve this paradox, the perceived sharpness of drifting, blurred, square wave gratings was compared with that of their static analogues over a range of speeds, blurs and spatial frequencies. Both motion blur and motion sharpening occurred, depending upon the physical blur of the patterns. For large extents of blur (> 10 min arc) moving patterns always appeared sharper than their static analogues, but for small blurs (< 10 min arc) moving edges appeared more blurred than stationary ones. We present a quantitative model for the distortion of waveforms in motion based on two factors: (i) visual temporal integration that smears moving images, and (ii) a local contrast non-linearity that increasingly sharpens the effective profile of edges as speed and contrast increase. We suggest that a plausible account of the speed-dependent non-linearity is the differential recruitment of M and P cells at different speeds.
Collapse
Affiliation(s)
- S T Hammett
- Department of Psychology, University of Glasgow, UK.
| | | | | |
Collapse
|
29
|
Purushothaman G, Oğmen H, Chen S, Bedell HE. Motion deblurring in a neural network model of retino-cortical dynamics. Vision Res 1998; 38:1827-42. [PMID: 9797961 DOI: 10.1016/s0042-6989(97)00350-7] [Citation(s) in RCA: 32] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Abstract
Simulations of a neural network model of retino-cortical dynamics (Oğmen H, Neural Netw 6 (1993) 245-273) are presented. The temporal-step response of the model to a single dot (spatial impulse) consists of three post-retinal phases: reset, feed-forward dominant and feedback dominant. In response to a single moving dot, the model predicts the perception of extensive blur. This extensive blur is proposed to be due to the relative spatial and temporal offsets between transient and sustained signals conveyed from retina to post-retinal levels. In response to a pair of horizontally separated dots moving in the horizontal direction, the model predicts extensive blur for the trailing dot irrespective of dot-to-dot separation. For the leading dot, the model predicts a decrease in perceived blur for long exposure durations when dot-to-dot separations are small. The reduction of perceived blur at long exposure durations for small dot-to-dot separations is proposed to stem from the spatio-temporal overlap between the transient activity generated by the trailing dot and the sustained activity generated by the leading dot. The model also predicts that targets moving at higher speeds generate more blur even when blur is normalized with respect to speed. The mechanism in the model generating this effect is a slow inhibition within the sustained channel. These predictions are compared with recent psychophysical data (Chen S, Bedell HE, Oğmen H, Vis Res 35 (1995) 2315-2328) and are found to be in excellent agreement. The model is used to offer a coherent explanation for several controversial findings published in the literature. This computational study shows that a model without any motion-compensation mechanism can give a good account of motion deblurring phenomenon and supplements our recent experimental study which provided evidence against motion-compensation type models in explaining the motion deblurring phenomenon (Chen S, Bedell HE, Oğmen H, Vis Res 35 (1995) 2315-2328).
Collapse
Affiliation(s)
- G Purushothaman
- Department of Electrical and Computer Engineering, University of Houston, TX 77204-4793, USA
| | | | | | | |
Collapse
|
30
|
Abstract
The effect of motion sharpening upon blur discrimination thresholds was examined for a range of speeds and blur widths. Blur discrimination thresholds were measured for drifting edges whose blur was either physically or perceptually constant. Under conditions where edges were kept at a constant physical blur width, discrimination thresholds rose as a function of speed as previously reported. However, when the perceived blur of edges was held constant, discrimination performance was more-or-less constant for speeds up to at least 6.3 deg sec-1. The results indicate that the deterioration of blur discrimination performance with speed may be due to motion sharpening and not motion blur as has previously been suggested. The results are discussed in terms of a scheme whereby a non-linearity in motion processing serves to sharpen moving edges, whilst the finite integration time of the system tends to smear them.
Collapse
Affiliation(s)
- S T Hammett
- Laboratoire de Psychologie Expérimentale, Université René Descartes (Paris V), France.
| |
Collapse
|
31
|
Abstract
Although much has been learned about the spatial sampling and filtering properties of peripheral vision, little attention has been paid to the remarkably clear appearance of the peripheral visual field. To study the apparent sharpness of stimuli presented in the periphery, we presented Gaussian blurred horizontal edges at 8.3, 16.6, 24, 32, and 40 deg eccentricity. Observers adjusted the sharpness of a similar edge, viewed foveally, to match the appearance of the peripheral stimulus. All observers matched blurred peripheral stimuli with sharper foveal stimuli. We have called this effect "sharpness overconstancy". For field sizes of 4 deg, there was greater overconstancy at larger eccentricities. Scaling the field size of the peripheral stimuli by a cortical magnification factor produced sharpness overconstancy which was independent of eccentricity. In both cases, there was a slight sharpness underconstancy for peripherally presented edges blurred only slightly. We consider various explanations of peripheral sharpness overconstancy.
Collapse
Affiliation(s)
- S J Galvin
- Department of Psychology, University of Otago, Dunedin, New Zealand.
| | | | | | | |
Collapse
|