1
|
Ohtsu M, Kurata A, Hirai K, Tanaka M, Horiuchi T. Evaluating the Influence of ipRGCs on Color Discrimination. J Imaging 2022; 8:jimaging8060154. [PMID: 35735953 PMCID: PMC9225537 DOI: 10.3390/jimaging8060154] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2022] [Revised: 05/22/2022] [Accepted: 05/26/2022] [Indexed: 11/23/2022] Open
Abstract
To investigate the influence of intrinsically photosensitive retinal ganglion cells (ipRGCs) on color discrimination, it is necessary to create two metameric light stimuli (metameric ipRGC stimuli) with the same amount of cone and rod stimulation, but different amounts of ipRGC stimulation. However, since the spectral sensitivity functions of cones and rods overlap with those of ipRGCs in a wavelength band, it has been difficult to independently control the amount of stimulation of ipRGCs only. In this study, we first propose a method for calculating metameric ipRGC stimulation based on the orthogonal basis functions of human photoreceptor cells. Then, we clarify the controllable range of metameric ipRGC stimulation within a color gamut. Finally, to investigate the color discrimination by metameric ipRGC stimuli, we conduct subjective evaluation experiments on 24 chromaticity coordinates using a multispectral projector. The results reveal a correlation between differences in the amount of ipRGC stimulation and differences in color appearance, indicating that ipRGCs may influence color discrimination.
Collapse
Affiliation(s)
- Masaya Ohtsu
- Graduate School of Science and Engineering, Chiba University, Yayoi-cho 1-33, Inage-ku, Chiba 263-8522, Japan; (A.K.); (K.H.); (T.H.)
- Correspondence: ; Tel.: +81-43-290-3485
| | - Akihiro Kurata
- Graduate School of Science and Engineering, Chiba University, Yayoi-cho 1-33, Inage-ku, Chiba 263-8522, Japan; (A.K.); (K.H.); (T.H.)
| | - Keita Hirai
- Graduate School of Science and Engineering, Chiba University, Yayoi-cho 1-33, Inage-ku, Chiba 263-8522, Japan; (A.K.); (K.H.); (T.H.)
| | - Midori Tanaka
- Graduate School of Global and Transdisciplinary Studies, Chiba University, Yayoi-cho 1-33, Inage-ku, Chiba 263-8522, Japan;
| | - Takahiko Horiuchi
- Graduate School of Science and Engineering, Chiba University, Yayoi-cho 1-33, Inage-ku, Chiba 263-8522, Japan; (A.K.); (K.H.); (T.H.)
| |
Collapse
|
2
|
Chen S, Duinkharjav B, Sun X, Wei LY, Petrangeli S, Echevarria J, Silva C, Sun Q. Instant Reality: Gaze-Contingent Perceptual Optimization for 3D Virtual Reality Streaming. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2022; 28:2157-2167. [PMID: 35148266 DOI: 10.1109/tvcg.2022.3150522] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Media streaming, with an edge-cloud setting, has been adopted for a variety of applications such as entertainment, visualization, and design. Unlike video/audio streaming where the content is usually consumed passively, virtual reality applications require 3D assets stored on the edge to facilitate frequent edge-side interactions such as object manipulation and viewpoint movement. Compared to audio and video streaming, 3D asset streaming often requires larger data sizes and yet lower latency to ensure sufficient rendering quality, resolution, and latency for perceptual comfort. Thus, streaming 3D assets faces remarkably additional than streaming audios/videos, and existing solutions often suffer from long loading time or limited quality. To address this challenge, we propose a perceptually-optimized progressive 3D streaming method for spatial quality and temporal consistency in immersive interactions. On the cloud-side, our main idea is to estimate perceptual importance in 2D image space based on user gaze behaviors, including where they are looking and how their eyes move. The estimated importance is then mapped to 3D object space for scheduling the streaming priorities for edge-side rendering. Since this computational pipeline could be heavy, we also develop a simple neural network to accelerate the cloud-side scheduling process. We evaluate our method via subjective studies and objective analysis under varying network conditions (from 3G to 5G) and edge devices (HMD and traditional displays), and demonstrate better visual quality and temporal consistency than alternative solutions.
Collapse
|
3
|
Richard B, Shafto P. Sensitivity to the slope of the amplitude spectrum is dependent on the spectral slopes of recently viewed environments: A visual adaptation study in modified reality. Vision Res 2022; 197:108056. [PMID: 35489239 DOI: 10.1016/j.visres.2022.108056] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2021] [Revised: 04/06/2022] [Accepted: 04/13/2022] [Indexed: 11/29/2022]
Abstract
Scenes contain many statistical regularities that could benefit visual processing if accounted for by the visual system. One such statistic is the orientation-averaged slope (α) of the amplitude spectrum of natural scenes. Human observers show different discrimination sensitivity to α: sensitivity is highest for α values between 1.0 and 1.2 and decreases as α is steepened or shallowed. The range of α for peak discrimination sensitivity is concordant with the average α of natural scenes, which may indicate that visual mechanisms are optimized to process information at α values commonly encountered in the environment. Here we explore the association between peak discrimination sensitivity and the most viewed αs in natural environments. Specifically, we verified whether discrimination sensitivity depends on the recently viewed environments. Observers were immersed, using a Head-Mounted Display, in an environment that was either unaltered or had its average α steepened or shallowed by 0.4. Discrimination thresholds were affected by the average shift in α, but this effect was most prominent following adaptation to a shallowed environment. We modeled these data with a Bayesian observer and explored whether a change in the prior or a change in the likelihood best explained the psychophysical effects. Change in discrimination thresholds following adaptation could be explained by a shift in the central tendency of the prior concordant with the shift of the environment, in addition to a change in the likelihood. Our findings suggest that expectations on the occurrence of α that result from a lifetime of exposure remain plastic and able to accommodate for the statistical structure of recently viewed environments.
Collapse
Affiliation(s)
- Bruno Richard
- Department of Mathematics and Computer Science, Rutgers University - Newark, 101 Warren Street, Rm 216, Newark, NJ 07102, USA.
| | - Patrick Shafto
- Department of Mathematics and Computer Science, Rutgers University - Newark, 101 Warren Street, Rm 216, Newark, NJ 07102, USA; School of Mathematics, Institute for Advanced Study, Princeton, NJ, USA
| |
Collapse
|
4
|
Abbas Farishta R, Yang CL, Farivar R. Blur Representation in the Amblyopic Visual System Using Natural and Synthetic Images. Invest Ophthalmol Vis Sci 2022; 63:3. [PMID: 34982147 PMCID: PMC8742520 DOI: 10.1167/iovs.63.1.3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Purpose Amblyopia is diagnosed as a reduced acuity in an otherwise healthy eye, which indicates that the deficit is not happening in the eye, but in the brain. One suspected mechanism explaining these deficits is an elevated amount of intrinsic blur in the amblyopic visual system compared to healthy observers. This "internally produced blur" can be estimated by the "equivalent intrinsic blur method", which measures blur discrimination thresholds while systematically increasing the external blur in the physical stimulus. Surprisingly, amblyopes do not exhibit elevated intrinsic blur when measured with an edge stimulus. Given the fundamental ways in which they differ, synthetic stimuli, such as edges, are likely to generate contrasting blur perception compared to natural stimuli, such as pictures. Because our visual system is presumably tuned to process natural stimuli, testing artificial stimuli only could result in performances that are not ecologically valid. Methods We tested this hypothesis by measuring, for the first time, the perception of blur added to natural images in amblyopia and compared discrimination performance for natural images and synthetic edges in healthy and amblyopic groups. Results Our results demonstrate that patients with amblyopia exhibit higher levels of intrinsic blur than control subjects when tested on natural images. This difference was not observed when using edges. Conclusions Our results suggest that intrinsic blur is elevated in the visual system representing vision from the amblyopic eye and that distinct statistics of images can generate different blur perception.
Collapse
Affiliation(s)
- Reza Abbas Farishta
- McGill Vision Research, Department of Ophthalmology and Visual Sciences, McGill University, Montréal, Québec, Canada
| | - Charlene L Yang
- McGill Vision Research, Department of Ophthalmology and Visual Sciences, McGill University, Montréal, Québec, Canada
| | - Reza Farivar
- McGill Vision Research, Department of Ophthalmology and Visual Sciences, McGill University, Montréal, Québec, Canada
| |
Collapse
|
5
|
Abstract
Within the spectrum of a natural image, the amplitude of modulation decreases with spatial frequency. The speed of such an amplitude decrease, or the amplitude spectrum slope, of an image affects the perceived aesthetic value. Additionally, a human observer would consider a symmetric image more appealing than they would an asymmetric one. We investigated how these two factors jointly affect aesthetic preferences by manipulating both the amplitude spectrum slope and the symmetric level of images to assess their effects on aesthetic preference on a 6-point Likert scale. Our results showed that the preference ratings increased with the symmetry level but had an inverted U-shaped relation to amplitude spectrum slope. In addition, a strong interaction existed between symmetry level and amplitude spectrum slope on preference rating, in that symmetry can amplify the amplitude spectrum slope’s effects. A quadratic function of the spectrum slope can describe such effects. That is, preference is an inverted U-shaped function of spectrum slope whose intercept is determined by the number of symmetry axes. The modulation depth of the quadratic function manifests the interaction between the two factors.
Collapse
|
6
|
To MPS, Tolhurst DJ. V1-based modeling of discrimination between natural scenes within the luminance and isoluminant color planes. J Vis 2019; 19:9. [PMID: 30650432 DOI: 10.1167/19.1.9] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
We have been developing a computational visual difference predictor model that can predict how human observers rate the perceived magnitude of suprathreshold differences between pairs of full-color naturalistic scenes (To, Lovell, Troscianko, & Tolhurst, 2010). The model is based closely on V1 neurophysiology and has recently been updated to more realistically implement sequential application of nonlinear inhibitions (contrast normalization followed by surround suppression; To, Chirimuuta, & Tolhurst, 2017). The model is based originally on a reliable luminance model (Watson & Solomon, 1997) which we have extended to the red/green and blue/yellow opponent planes, assuming that the three planes (luminance, red/green, and blue/yellow) can be modeled similarly to each other with narrow-band oriented filters. This paper examines whether this may be a false assumption, by decomposing our original full-color stimulus images into monochromatic and isoluminant variants, which observers rate separately and which we model separately. The ratings for the original full-color scenes correlate better with the new ratings for the monochromatic variants than for the isoluminant ones, suggesting that luminance cues carry more weight in observers' ratings to full-color images. The ratings for the original full-color stimuli can be predicted from the new monochromatic and isoluminant rating data by combining them by Minkowski summation with power m = 2.71, consistent with other studies involving feature summation. The model performed well at predicting ratings for monochromatic stimuli, but was weaker for isoluminant stimuli, indicating that mirroring the monochromatic models is not sufficient to model the color planes. We discuss several alternative strategies to improve the color modeling.
Collapse
Affiliation(s)
- Michelle P S To
- Department of Psychology, Lancaster University, Lancaster, UK
| | | |
Collapse
|
7
|
Grootswagers T, Ritchie JB, Wardle SG, Heathcote A, Carlson TA. Asymmetric Compression of Representational Space for Object Animacy Categorization under Degraded Viewing Conditions. J Cogn Neurosci 2017; 29:1995-2010. [DOI: 10.1162/jocn_a_01177] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Abstract
Animacy is a robust organizing principle among object category representations in the human brain. Using multivariate pattern analysis methods, it has been shown that distance to the decision boundary of a classifier trained to discriminate neural activation patterns for animate and inanimate objects correlates with observer RTs for the same animacy categorization task [Ritchie, J. B., Tovar, D. A., & Carlson, T. A. Emerging object representations in the visual system predict reaction times for categorization. PLoS Computational Biology, 11, e1004316, 2015; Carlson, T. A., Ritchie, J. B., Kriegeskorte, N., Durvasula, S., & Ma, J. Reaction time for object categorization is predicted by representational distance. Journal of Cognitive Neuroscience, 26, 132–142, 2014]. Using MEG decoding, we tested if the same relationship holds when a stimulus manipulation (degradation) increases task difficulty, which we predicted would systematically decrease the distance of activation patterns from the decision boundary and increase RTs. In addition, we tested whether distance to the classifier boundary correlates with drift rates in the linear ballistic accumulator [Brown, S. D., & Heathcote, A. The simplest complete model of choice response time: Linear ballistic accumulation. Cognitive Psychology, 57, 153–178, 2008]. We found that distance to the classifier boundary correlated with RT, accuracy, and drift rates in an animacy categorization task. Split by animacy, the correlations between brain and behavior were sustained longer over the time course for animate than for inanimate stimuli. Interestingly, when examining the distance to the classifier boundary during the peak correlation between brain and behavior, we found that only degraded versions of animate, but not inanimate, objects had systematically shifted toward the classifier decision boundary as predicted. Our results support an asymmetry in the representation of animate and inanimate object categories in the human brain.
Collapse
Affiliation(s)
- Tijl Grootswagers
- Macquarie University, Australia
- ARC Centre of Excellence in Cognition and Its Disorders, Australia
- University of Sydney
| | | | - Susan G. Wardle
- Macquarie University, Australia
- ARC Centre of Excellence in Cognition and Its Disorders, Australia
| | | | - Thomas A. Carlson
- ARC Centre of Excellence in Cognition and Its Disorders, Australia
- University of Sydney
| |
Collapse
|
8
|
Spehar B, Walker N, Taylor RP. Taxonomy of Individual Variations in Aesthetic Responses to Fractal Patterns. Front Hum Neurosci 2016; 10:350. [PMID: 27458365 PMCID: PMC4937063 DOI: 10.3389/fnhum.2016.00350] [Citation(s) in RCA: 37] [Impact Index Per Article: 4.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2016] [Accepted: 06/28/2016] [Indexed: 11/13/2022] Open
Abstract
In two experiments, we investigate group and individual preferences in a range of different types of patterns with varying fractal-like scaling characteristics. In Experiment 1, we used 1/f filtered grayscale images as well as their thresholded (black and white) and edges only counterparts. Separate groups of observers viewed different types of images varying in slope of their amplitude spectra. Although with each image type, the groups exhibited the "universal" pattern of preference for intermediate amplitude spectrum slopes, we identified 4 distinct sub-groups in each case. Sub-group 1 exhibited a typical peak preference for intermediate amplitude spectrum slopes ("intermediate"; approx. 50%); sub-group 2 exhibited a linear increase in preference with increasing amplitude spectrum slope ("smooth"; approx. 20%), while sub-group 3 exhibited a linear decrease in preference as a function of the amplitude spectrum slope ("sharp"; approx. 20%). Sub-group 4 revealed no significant preference ("other"; approx. 10%). In Experiment 2, we extended the range of different image types and investigated preferences within the same observers. We replicate the results of our first experiment and show that individual participants exhibit stable patterns of preference across a wide range of image types. In both experiments, Q-mode factor analysis identified two principal factors that were able to explain more than 80% of interindividual variations in preference across all types of images, suggesting a highly similar dimensional structure of interindividual variations in preference for fractal-like scaling characteristics.
Collapse
Affiliation(s)
- Branka Spehar
- School of Psychology, UNSW Australia Sydney, NSW, Australia
| | | | | |
Collapse
|
9
|
Spehar B, Wong S, van de Klundert S, Lui J, Clifford CWG, Taylor RP. Beauty and the beholder: the role of visual sensitivity in visual preference. Front Hum Neurosci 2015; 9:514. [PMID: 26441611 PMCID: PMC4585069 DOI: 10.3389/fnhum.2015.00514] [Citation(s) in RCA: 37] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2015] [Accepted: 09/03/2015] [Indexed: 11/17/2022] Open
Abstract
For centuries, the essence of aesthetic experience has remained one of the most intriguing mysteries for philosophers, artists, art historians and scientists alike. Recently, views emphasizing the link between aesthetics, perception and brain function have become increasingly prevalent (Ramachandran and Hirstein, 1999; Zeki, 1999; Livingstone, 2002; Ishizu and Zeki, 2013). The link between art and the fractal-like structure of natural images has also been highlighted (Spehar et al., 2003; Graham and Field, 2007; Graham and Redies, 2010). Motivated by these claims and our previous findings that humans display a consistent preference across various images with fractal-like statistics, here we explore the possibility that observers’ preference for visual patterns might be related to their sensitivity for such patterns. We measure sensitivity to simple visual patterns (sine-wave gratings varying in spatial frequency and random textures with varying scaling exponent) and find that they are highly correlated with visual preferences exhibited by the same observers. Although we do not attempt to offer a comprehensive neural model of aesthetic experience, we demonstrate a strong relationship between visual sensitivity and preference for simple visual patterns. Broadly speaking, our results support assertions that there is a close relationship between aesthetic experience and the sensory coding of natural stimuli.
Collapse
Affiliation(s)
- Branka Spehar
- School of Psychology, UNSW Australia Sydney, NSW, Australia
| | - Solomon Wong
- School of Psychology, UNSW Australia Sydney, NSW, Australia
| | | | - Jessie Lui
- School of Psychology, UNSW Australia Sydney, NSW, Australia
| | | | | |
Collapse
|
10
|
Tajima S, Komine K. Saliency-based color accessibility. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2015; 24:1115-1126. [PMID: 25608304 DOI: 10.1109/tip.2015.2393056] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
Perception of color varies markedly between individuals because of differential expression of photopigments in retinal cones. However, it has been difficult to quantify the individual cognitive variation in colored scene and to predict its complex impacts on the behaviors. We developed a method for quantifying and visualizing information loss and gain resulting from individual differences in spectral sensitivity based on visual salience. We first modeled the visual salience for color-deficient observers, and found that the predicted losses and gains in local image salience derived from normal and color-blind models were correlated with the subjective judgment of image saliency in psychophysical experiments, i.e., saliency loss predicted reduced image preference in color-deficient observers. Moreover,saliency-guided image manipulations sufficiently compensated for individual differences in saliency. This visual saliency approach allows for quantification of information extracted from complex visual scenes and can be used as an image compensation to enhance visual accessibility by color-deficient individuals.
Collapse
|
11
|
Pearce B, Crichton S, Mackiewicz M, Finlayson GD, Hurlbert A. Chromatic illumination discrimination ability reveals that human colour constancy is optimised for blue daylight illuminations. PLoS One 2014; 9:e87989. [PMID: 24586299 PMCID: PMC3929610 DOI: 10.1371/journal.pone.0087989] [Citation(s) in RCA: 70] [Impact Index Per Article: 6.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2013] [Accepted: 01/02/2014] [Indexed: 11/19/2022] Open
Abstract
The phenomenon of colour constancy in human visual perception keeps surface colours constant, despite changes in their reflected light due to changing illumination. Although colour constancy has evolved under a constrained subset of illuminations, it is unknown whether its underlying mechanisms, thought to involve multiple components from retina to cortex, are optimised for particular environmental variations. Here we demonstrate a new method for investigating colour constancy using illumination matching in real scenes which, unlike previous methods using surface matching and simulated scenes, allows testing of multiple, real illuminations. We use real scenes consisting of solid familiar or unfamiliar objects against uniform or variegated backgrounds and compare discrimination performance for typical illuminations from the daylight chromaticity locus (approximately blue-yellow) and atypical spectra from an orthogonal locus (approximately red-green, at correlated colour temperature 6700 K), all produced in real time by a 10-channel LED illuminator. We find that discrimination of illumination changes is poorer along the daylight locus than the atypical locus, and is poorest particularly for bluer illumination changes, demonstrating conversely that surface colour constancy is best for blue daylight illuminations. Illumination discrimination is also enhanced, and therefore colour constancy diminished, for uniform backgrounds, irrespective of the object type. These results are not explained by statistical properties of the scene signal changes at the retinal level. We conclude that high-level mechanisms of colour constancy are biased for the blue daylight illuminations and variegated backgrounds to which the human visual system has typically been exposed.
Collapse
Affiliation(s)
- Bradley Pearce
- Institute of Neuroscience, Faculty of Medical Sciences, Newcastle University, England, United Kingdom
| | - Stuart Crichton
- Institute of Neuroscience, Faculty of Medical Sciences, Newcastle University, England, United Kingdom
| | - Michal Mackiewicz
- School of Computing Sciences, University of East Anglia, Norwich, England, United Kingdom
| | - Graham D. Finlayson
- School of Computing Sciences, University of East Anglia, Norwich, England, United Kingdom
| | - Anya Hurlbert
- Institute of Neuroscience, Faculty of Medical Sciences, Newcastle University, England, United Kingdom
| |
Collapse
|
12
|
How sensitive is the human visual system to the local statistics of natural images? PLoS Comput Biol 2013; 9:e1002873. [PMID: 23358106 PMCID: PMC3554546 DOI: 10.1371/journal.pcbi.1002873] [Citation(s) in RCA: 43] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2012] [Accepted: 11/21/2012] [Indexed: 11/19/2022] Open
Abstract
A key hypothesis in sensory system neuroscience is that sensory representations are adapted to the statistical regularities in sensory signals and thereby incorporate knowledge about the outside world. Supporting this hypothesis, several probabilistic models of local natural image regularities have been proposed that reproduce neural response properties. Although many such physiological links have been made, these models have not been linked directly to visual sensitivity. Previous psychophysical studies of sensitivity to natural image regularities focus on global perception of large images, but much less is known about sensitivity to local natural image regularities. We present a new paradigm for controlled psychophysical studies of local natural image regularities and compare how well such models capture perceptually relevant image content. To produce stimuli with precise statistics, we start with a set of patches cut from natural images and alter their content to generate a matched set whose joint statistics are equally likely under a probabilistic natural image model. The task is forced choice to discriminate natural patches from model patches. The results show that human observers can learn to discriminate the higher-order regularities in natural images from those of model samples after very few exposures and that no current model is perfect for patches as small as 5 by 5 pixels or larger. Discrimination performance was accurately predicted by model likelihood, an information theoretic measure of model efficacy, indicating that the visual system possesses a surprisingly detailed knowledge of natural image higher-order correlations, much more so than current image models. We also perform three cue identification experiments to interpret how model features correspond to perceptually relevant image features. Several aspects of primate visual physiology have been identified as adaptations to local regularities of natural images. However, much less work has measured visual sensitivity to local natural image regularities. Most previous work focuses on global perception of large images and shows that observers are more sensitive to visual information when image properties resemble those of natural images. In this work we measure human sensitivity to local natural image regularities using stimuli generated by patch-based probabilistic natural image models that have been related to primate visual physiology. We find that human observers can learn to discriminate the statistical regularities of natural image patches from those represented by current natural image models after very few exposures and that discriminability depends on the degree of regularities captured by the model. The quick learning we observed suggests that the human visual system is biased for processing natural images, even at very fine spatial scales, and that it has a surprisingly large knowledge of the regularities in natural images, at least in comparison to the state-of-the-art statistical models of natural images.
Collapse
|
13
|
Hansen BC, Hess RF. On the effectiveness of noise masks: naturalistic vs. un-naturalistic image statistics. Vision Res 2012; 60:101-13. [PMID: 22484251 DOI: 10.1016/j.visres.2012.03.017] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2011] [Revised: 01/10/2012] [Accepted: 03/23/2012] [Indexed: 11/28/2022]
Abstract
It has been argued that the human visual system is optimized for identification of broadband objects embedded in stimuli possessing orientation averaged power spectra fall-offs that obey the 1/f(β) relationship typically observed in natural scene imagery (i.e., β=2.0 on logarithmic axes). Here, we were interested in whether individual spatial channels leading to recognition are functionally optimized for narrowband targets when masked by noise possessing naturalistic image statistics (β=2.0). The current study therefore explores the impact of variable β noise masks on the identification of narrowband target stimuli ranging in spatial complexity, while simultaneously controlling for physical or perceived differences between the masks. The results show that β=2.0 noise masks produce the largest identification thresholds regardless of target complexity, and thus do not seem to yield functionally optimized channel processing. The differential masking effects are discussed in the context of contrast gain control.
Collapse
Affiliation(s)
- Bruce C Hansen
- Department of Psychology & Neuroscience Program, Colgate University, Hamilton, NY 13346, USA.
| | | |
Collapse
|
14
|
Watson AB, Ahumada AJ. Blur clarified: A review and synthesis of blur discrimination. J Vis 2011; 11:11.5.10. [DOI: 10.1167/11.5.10] [Citation(s) in RCA: 53] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Affiliation(s)
- Andrew B. Watson
- NASA Ames Research Center, Moffett Field, CA, USAhttp://vision.arc.nasa.gov/
| | - Albert J. Ahumada
- NASA Ames Research Center, Moffett Field, CA, USAhttp://vision.arc.nasa.gov/personnel/al/
| |
Collapse
|
15
|
Moca VV, Ţincaş I, Melloni L, Mureşan RC. Visual exploration and object recognition by lattice deformation. PLoS One 2011; 6:e22831. [PMID: 21818397 PMCID: PMC3144955 DOI: 10.1371/journal.pone.0022831] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2011] [Accepted: 07/01/2011] [Indexed: 11/19/2022] Open
Abstract
Mechanisms of explicit object recognition are often difficult to investigate and require stimuli with controlled features whose expression can be manipulated in a precise quantitative fashion. Here, we developed a novel method (called "Dots"), for generating visual stimuli, which is based on the progressive deformation of a regular lattice of dots, driven by local contour information from images of objects. By applying progressively larger deformation to the lattice, the latter conveys progressively more information about the target object. Stimuli generated with the presented method enable a precise control of object-related information content while preserving low-level image statistics, globally, and affecting them only little, locally. We show that such stimuli are useful for investigating object recognition under a naturalistic setting--free visual exploration--enabling a clear dissociation between object detection and explicit recognition. Using the introduced stimuli, we show that top-down modulation induced by previous exposure to target objects can greatly influence perceptual decisions, lowering perceptual thresholds not only for object recognition but also for object detection (visual hysteresis). Visual hysteresis is target-specific, its expression and magnitude depending on the identity of individual objects. Relying on the particular features of dot stimuli and on eye-tracking measurements, we further demonstrate that top-down processes guide visual exploration, controlling how visual information is integrated by successive fixations. Prior knowledge about objects can guide saccades/fixations to sample locations that are supposed to be highly informative, even when the actual information is missing from those locations in the stimulus. The duration of individual fixations is modulated by the novelty and difficulty of the stimulus, likely reflecting cognitive demand.
Collapse
Affiliation(s)
- Vasile V. Moca
- Department of Experimental and Theoretical Neuroscience, Center for Cognitive and Neural Studies (Coneural), Romanian Institute of Science and Technology, Cluj-Napoca, Romania
| | - Ioana Ţincaş
- Department of Experimental and Theoretical Neuroscience, Center for Cognitive and Neural Studies (Coneural), Romanian Institute of Science and Technology, Cluj-Napoca, Romania
| | - Lucia Melloni
- Department of Neurophysiology, Max Planck Institute for Brain Research, Frankfurt am Main, Hessen, Germany
| | - Raul C. Mureşan
- Department of Experimental and Theoretical Neuroscience, Center for Cognitive and Neural Studies (Coneural), Romanian Institute of Science and Technology, Cluj-Napoca, Romania
- Department of Neurophysiology, Max Planck Institute for Brain Research, Frankfurt am Main, Hessen, Germany
- * E-mail:
| |
Collapse
|
16
|
Discrimination of natural scenes in central and peripheral vision. Vision Res 2011; 51:1686-98. [DOI: 10.1016/j.visres.2011.05.010] [Citation(s) in RCA: 25] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2010] [Revised: 05/12/2011] [Accepted: 05/20/2011] [Indexed: 11/20/2022]
|
17
|
From spatial frequency contrast to edge preponderance: the differential modulation of early visual evoked potentials by natural scene stimuli. Vis Neurosci 2011; 28:221-37. [DOI: 10.1017/s095252381100006x] [Citation(s) in RCA: 33] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
AbstractThe contrast response function of early visual evoked potentials elicited by sinusoidal gratings is known to exhibit characteristic potentials closely associated with the processes of parvocellular and magnocellular pathways. Specifically, the N1 component has been linked with parvocellular processes, while the P1 component has been linked with magnocellular processes. However, little is known regarding the response properties of the N1 and P1 components during the processing and encoding of complex (i.e., broadband) stimuli such as natural scenes. Here, we examine how established physical characteristics of natural scene imagery modulate the N1 and P1 components in humans by providing a systematic investigation of component modulation as visual stimuli are gradually built up from simple sinusoidal gratings to highly complex natural scene imagery. The results suggest that the relative dominance in signal output of the N1 and P1 components is dependent on spatial frequency (SF) luminance contrast for simple stimuli up to natural scene imagery possessing few edges. However, such a dependency shifts to a dominant N1 signal for natural scenes possessing abundant edge content and operates independently of SF luminance contrast.
Collapse
|
18
|
Rouse DM, Hemami SS, Pépion R, Le Callet P. Estimating the usefulness of distorted natural images using an image contour degradation measure. JOURNAL OF THE OPTICAL SOCIETY OF AMERICA. A, OPTICS, IMAGE SCIENCE, AND VISION 2011; 28:157-188. [PMID: 21293521 DOI: 10.1364/josaa.28.000157] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
Quality estimators aspire to quantify the perceptual resemblance, but not the usefulness, of a distorted image when compared to a reference natural image. However, humans can successfully accomplish tasks (e.g., object identification) using visibly distorted images that are not necessarily of high quality. A suite of novel subjective experiments reveals that quality does not accurately predict utility (i.e., usefulness). Thus, even accurate quality estimators cannot accurately estimate utility. In the absence of utility estimators, leading quality estimators are assessed as both quality and utility estimators and dismantled to understand those image characteristics that distinguish utility from quality. A newly proposed utility estimator demonstrates that a measure of contour degradation is sufficient to accurately estimate utility and is argued to be compatible with shape-based theories of object perception.
Collapse
Affiliation(s)
- David M Rouse
- Visual Communications Laboratory, School of Electrical and Computer Engineering, Cornell University, 356 Rhodes Hall, Ithaca, New York 14850, USA.
| | | | | | | |
Collapse
|
19
|
To MPS, Baddeley RJ, Troscianko T, Tolhurst DJ. A general rule for sensory cue summation: evidence from photographic, musical, phonetic and cross-modal stimuli. Proc Biol Sci 2010; 278:1365-72. [PMID: 20961902 DOI: 10.1098/rspb.2010.1888] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
The Euclidean and MAX metrics have been widely used to model cue summation psychophysically and computationally. Both rules happen to be special cases of a more general Minkowski summation rule , where m = 2 and ∞, respectively. In vision research, Minkowski summation with power m = 3-4 has been shown to be a superior model of how subthreshold components sum to give an overall detection threshold. Recently, we have previously reported that Minkowski summation with power m = 2.84 accurately models summation of suprathreshold visual cues in photographs. In four suprathreshold discrimination experiments, we confirm the previous findings with new visual stimuli and extend the applicability of this rule to cue combination in auditory stimuli (musical sequences and phonetic utterances, where m = 2.95 and 2.54, respectively) and cross-modal stimuli (m = 2.56). In all cases, Minkowski summation with power m = 2.5-3 outperforms the Euclidean and MAX operator models. We propose that this reflects the summation of neuronal responses that are not entirely independent but which show some correlation in their magnitudes. Our findings are consistent with electrophysiological research that demonstrates signal correlations (r = 0.1-0.2) between sensory neurons when these are presented with natural stimuli.
Collapse
Affiliation(s)
- M P S To
- Department of Physiology, Development and Neuroscience, University of Cambridge, Downing Street, Cambridge CB2 3EG, UK.
| | | | | | | |
Collapse
|
20
|
Morgan MJ. Features and the 'primal sketch'. Vision Res 2010; 51:738-53. [PMID: 20696182 DOI: 10.1016/j.visres.2010.08.002] [Citation(s) in RCA: 28] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2010] [Revised: 08/01/2010] [Accepted: 08/02/2010] [Indexed: 10/19/2022]
Abstract
This review is concerned primarily with psychophysical and physiological evidence relevant to the question of the existence of spatial features or spatial primitives in human vision. The review will be almost exclusively confined to features defined in the luminance domain. The emphasis will be on the experimental and computational methods that have been used for revealing features, rather than on a detailed comparison between different models of feature extraction. Color and texture fall largely outside the scope of the review, though the principles may be similar. Stereo matching and motion matching are also largely excluded because they are covered in other contributions to this volume, although both have addressed the question of the spatial primitives involved in matching. Similarities between different psychophysically-based model will be emphasized rather than minor differences. All the models considered in the review are based on the extraction of directional spatial derivatives of the luminance profile, typically the first and second, but in one case the third order, and all have some form of non-linearity, be it rectification or thresholding.
Collapse
Affiliation(s)
- Michael J Morgan
- Applied Vision Research Centre, Department of Optometry, City University, Northampton Square, London EC1V0HB, UK.
| |
Collapse
|
21
|
The role of higher order image statistics in masking scene gist recognition. Atten Percept Psychophys 2010; 72:427-44. [PMID: 20139457 DOI: 10.3758/app.72.2.427] [Citation(s) in RCA: 29] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
In the present article, we investigated whether higher order image statistics, which are known to be carried by the Fourier phase spectrum, are sufficient to affect scene gist recognition. In Experiment 1, we compared the scene gist masking strength of four masking image types that varied in their degrees of second- and higher order relationships: normal scene images, scene textures, phase-randomized scene images, and white noise. Masking effects were the largest for masking images that possessed significant higher order image statistics (scene images and scene textures) as compared with masking images that did not (phase-randomized scenes and white noise), with scene image masks yielding the largest masking effects. In a control study, we eliminated all differences in the second-order statistics of the masks, while maintaining differences in their higher order statistics by comparing masking by scene textures rather than by their phase-randomized versions, and showed that the former produced significantly stronger gist masking. Experiments 2 and 3 were designed to test whether conceptual masking could account for the differences in the strength of the scene texture and phase-randomized masks used in Experiment 1, and revealed that the recognizability of scene texture masks explained just 1% of their masking variance. Together, the results suggest that (1) masks containing the higher order statistical structure of scenes are more effective at masking scene gist processing than are masks lacking such structure, and (2) much of the disruption of scene gist recognition that one might be tempted to attribute to conceptual masking is due to spatial masking.
Collapse
|
22
|
Tajima S, Okada M. Discriminating natural image statistics from neuronal population codes. PLoS One 2010; 5:e9704. [PMID: 20360849 PMCID: PMC2845616 DOI: 10.1371/journal.pone.0009704] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2009] [Accepted: 02/25/2010] [Indexed: 11/18/2022] Open
Abstract
The power law provides an efficient description of amplitude spectra of natural scenes. Psychophysical studies have shown that the forms of the amplitude spectra are clearly related to human visual performance, indicating that the statistical parameters in natural scenes are represented in the nervous system. However, the underlying neuronal computation that accounts for the perception of the natural image statistics has not been thoroughly studied. We propose a theoretical framework for neuronal encoding and decoding of the image statistics, hypothesizing the elicited population activities of spatial-frequency selective neurons observed in the early visual cortex. The model predicts that frequency-tuned neurons have asymmetric tuning curves as functions of the amplitude spectra falloffs. To investigate the ability of this neural population to encode the statistical parameters of the input images, we analyze the Fisher information of the stochastic population code, relating it to the psychophysically measured human ability to discriminate natural image statistics. The nature of discrimination thresholds suggested by the computational model is consistent with experimental data from previous studies. Of particular interest, a reported qualitative disparity between performance in fovea and parafovea can be explained based on the distributional difference over preferred frequencies of neurons in the current model. The threshold shows a peak at a small falloff parameter when the neuronal preferred spatial frequencies are narrowly distributed, whereas the threshold peak vanishes for a neural population with a more broadly distributed frequency preference. These results demonstrate that the distributional property of neuronal stimulus preference can play a crucial role in linking microscopic neurophysiological phenomena and macroscopic human behaviors.
Collapse
Affiliation(s)
- Satohiro Tajima
- Department of Complexity Science and Engineering, The University of Tokyo, Kashiwa, Chiba, Japan.
| | | |
Collapse
|
23
|
Troscianko T, Benton CP, Lovell PG, Tolhurst DJ, Pizlo Z. Camouflage and visual perception. Philos Trans R Soc Lond B Biol Sci 2009; 364:449-61. [PMID: 18990671 DOI: 10.1098/rstb.2008.0218] [Citation(s) in RCA: 100] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
How does an animal conceal itself from visual detection by other animals? This review paper seeks to identify general principles that may apply in this broad area. It considers mechanisms of visual encoding, of grouping and object encoding, and of search. In most cases, the evidence base comes from studies of humans or species whose vision approximates to that of humans. The effort is hampered by a relatively sparse literature on visual function in natural environments and with complex foraging tasks. However, some general constraints emerge as being potentially powerful principles in understanding concealment--a 'constraint' here means a set of simplifying assumptions. Strategies that disrupt the unambiguous encoding of discontinuities of intensity (edges), and of other key visual attributes, such as motion, are key here. Similar strategies may also defeat grouping and object-encoding mechanisms. Finally, the paper considers how we may understand the processes of search for complex targets in complex scenes. The aim is to provide a number of pointers towards issues, which may be of assistance in understanding camouflage and concealment, particularly with reference to how visual systems can detect the shape of complex, concealed objects.
Collapse
Affiliation(s)
- Tom Troscianko
- Department of Experimental Psychology, University of Bristol, 12a Priory Road, Bristol BS8 1TU, UK.
| | | | | | | | | |
Collapse
|
24
|
To M, Lovell PG, Troscianko T, Tolhurst DJ. Summation of perceptual cues in natural visual scenes. Proc Biol Sci 2008; 275:2299-308. [PMID: 18628119 PMCID: PMC2495046 DOI: 10.1098/rspb.2008.0692] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
Natural visual scenes are rich in information, and any neural system analysing them must piece together the many messages from large arrays of diverse feature detectors. It is known how threshold detection of compound visual stimuli (sinusoidal gratings) is determined by their components' thresholds. We investigate whether similar combination rules apply to the perception of the complex and suprathreshold visual elements in naturalistic visual images. Observers gave magnitude estimations (ratings) of the perceived differences between pairs of images made from photographs of natural scenes. Images in some pairs differed along one stimulus dimension such as object colour, location, size or blur. But, for other image pairs, there were composite differences along two dimensions (e.g. both colour and object-location might change). We examined whether the ratings for such composite pairs could be predicted from the two ratings for the respective pairs in which only one stimulus dimension had changed. We found a pooling relationship similar to that proposed for simple stimuli: Minkowski summation with exponent 2.84 yielded the best predictive power (r=0.96), an exponent similar to that generally reported for compound grating detection. This suggests that theories based on detecting simple stimuli can encompass visual processing of complex, suprathreshold stimuli.
Collapse
Affiliation(s)
- M To
- Department of Physiology, Development and Neuroscience, University of Cambridge, Downing Street, Cambridge CB2 3EG, UK.
| | | | | | | |
Collapse
|
25
|
Abstract
AbstractRecent work has revealed multiple pathways for cross-orientation suppression in cat and human vision. In particular, ipsiocular and interocular pathways appear to assert their influence before binocular summation in human but have different (1) spatial tuning, (2) temporal dependencies, and (3) adaptation after-effects. Here we use mask components that fall outside the excitatory passband of the detecting mechanism to investigate the rules for pooling multiple mask components within these pathways. We measured psychophysical contrast masking functions for vertical 1 cycle/deg sine-wave gratings in the presence of left or right oblique (±45 deg) 3 cycles/deg mask gratings with contrast C%, or a plaid made from their sum, where each component (i) had contrast 0.5Ci%. Masks and targets were presented to two eyes (binocular), one eye (monoptic), or different eyes (dichoptic). Binocular-masking functions superimposed when plotted against C, but in the monoptic and dichoptic conditions, the grating produced slightly more suppression than the plaid when Ci ≥ 16%. We tested contrast gain control models involving two types of contrast combination on the denominator: (1) spatial pooling of the mask after a local nonlinearity (to calculate either root mean square contrast or energy) and (2) “linear suppression” (Holmes & Meese, 2004, Journal of Vision4, 1080–1089), involving the linear sum of the mask component contrasts. Monoptic and dichoptic masking were typically better fit by the spatial pooling models, but binocular masking was not: it demanded strict linear summation of the Michelson contrast across mask orientation. Another scheme, in which suppressive pooling followed compressive contrast responses to the mask components (e.g., oriented cortical cells), was ruled out by all of our data. We conclude that the different processes that underlie monoptic and dichoptic masking use the same type of contrast pooling within their respective suppressive fields, but the effects do not sum to predict the binocular case.
Collapse
|
26
|
Baker DH, Meese TS, Hess RF. Contrast masking in strabismic amblyopia: attenuation, noise, interocular suppression and binocular summation. Vision Res 2008; 48:1625-40. [PMID: 18547600 DOI: 10.1016/j.visres.2008.04.017] [Citation(s) in RCA: 110] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2008] [Revised: 04/18/2008] [Accepted: 04/22/2008] [Indexed: 11/19/2022]
Abstract
To investigate amblyopic contrast vision at threshold and above we performed pedestal-masking (contrast discrimination) experiments with a group of eight strabismic amblyopes using horizontal sinusoidal gratings (mainly 3c/deg) in monocular, binocular and dichoptic configurations balanced across eye (i.e. five conditions). With some exceptions in some observers, the four main results were as follows. (1) For the monocular and dichoptic conditions, sensitivity was less in the amblyopic eye than in the good eye at all mask contrasts. (2) Binocular and monocular dipper functions superimposed in the good eye. (3) Monocular masking functions had a normal dipper shape in the good eye, but facilitation was diminished in the amblyopic eye. (4) A less consistent result was normal facilitation in dichoptic masking when testing the good eye, but a loss of this when testing the amblyopic eye. This pattern of amblyopic results was replicated in a normal observer by placing a neutral density filter in front of one eye. The two-stage model of binocular contrast gain control [Meese, T.S., Georgeson, M.A. & Baker, D.H. (2006). Binocular contrast vision at and above threshold. Journal of Vision 6, 1224-1243.] was 'lesioned' in several ways to assess the form of the amblyopic deficit. The most successful model involves attenuation of signal and an increase in noise in the amblyopic eye, and intact stages of interocular suppression and binocular summation. This implies a behavioural influence from monocular noise in the amblyopic visual system as well as in normal observers with an ND filter over one eye.
Collapse
Affiliation(s)
- Daniel H Baker
- School of Life and Health Sciences, Aston University, Birmingham, UK.
| | | | | |
Collapse
|
27
|
Foster DH, Amano K, Nascimento SMC. Color constancy in natural scenes explained by global image statistics. Vis Neurosci 2006; 23:341-9. [PMID: 16961965 PMCID: PMC1896061 DOI: 10.1017/s0952523806233455] [Citation(s) in RCA: 41] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2006] [Accepted: 03/09/2006] [Indexed: 11/07/2022]
Abstract
To what extent do observers' judgments of surface color with natural scenes depend on global image statistics? To address this question, a psychophysical experiment was performed in which images of natural scenes under two successive daylights were presented on a computer-controlled high-resolution color monitor. Observers reported whether there was a change in reflectance of a test surface in the scene. The scenes were obtained with a hyperspectral imaging system and included variously trees, shrubs, grasses, ferns, flowers, rocks, and buildings. Discrimination performance, quantified on a scale of 0 to 1 with a color-constancy index, varied from 0.69 to 0.97 over 21 scenes and two illuminant changes, from a correlated color temperature of 25,000 K to 6700 K and from 4000 K to 6700 K. The best account of these effects was provided by receptor-based rather than colorimetric properties of the images. Thus, in a linear regression, 43% of the variance in constancy index was explained by the log of the mean relative deviation in spatial cone-excitation ratios evaluated globally across the two images of a scene. A further 20% was explained by including the mean chroma of the first image and its difference from that of the second image and a further 7% by the mean difference in hue. Together, all four global color properties accounted for 70% of the variance and provided a good fit to the effects of scene and of illuminant change on color constancy, and, additionally, of changing test-surface position. By contrast, a spatial-frequency analysis of the images showed that the gradient of the luminance amplitude spectrum accounted for only 5% of the variance.
Collapse
Affiliation(s)
- David H Foster
- Sensing, Imaging, and Signal Processing Group, School of Electrical and Electronic Engineering, University of Manchester, Manchester, United Kingdom.
| | | | | |
Collapse
|
28
|
Foster DH, Amano K, Nascimento SMC, Foster MJ. Frequency of metamerism in natural scenes. JOURNAL OF THE OPTICAL SOCIETY OF AMERICA. A, OPTICS, IMAGE SCIENCE, AND VISION 2006; 23:2359-72. [PMID: 16985522 PMCID: PMC2040061 DOI: 10.1364/josaa.23.002359] [Citation(s) in RCA: 106] [Impact Index Per Article: 5.6] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
Estimates of the frequency of metameric surfaces, which appear the same to the eye under one illuminant but different under another, were obtained from 50 hyperspectral images of natural scenes. The degree of metamerism was specified with respect to a color-difference measure after allowing for full chromatic adaptation. The relative frequency of metameric pairs of surfaces, expressed as a proportion of all pairs of surfaces in a scene, was very low. Depending on the criterion degree of metamerism, it ranged from about 10(-6) to 10(-4) for the largest illuminant change tested, which was from a daylight of correlated color temperature 25,000 K to one of 4000 K. But, given pairs of surfaces that were indistinguishable under one of these illuminants, the conditional relative frequency of metamerism was much higher, from about 10(-2) to 10(-1), sufficiently large to affect visual inferences about material identity.
Collapse
Affiliation(s)
- David H Foster
- Sensing, Imaging, and Signal Processing Group, School of Electrical and Electronic Engineering, University of Manchester, UK
| | | | | | | |
Collapse
|
29
|
Bosworth RG, Bartlett MS, Dobkins KR. Image statistics of American Sign Language: comparison with faces and natural scenes. JOURNAL OF THE OPTICAL SOCIETY OF AMERICA. A, OPTICS, IMAGE SCIENCE, AND VISION 2006; 23:2085-96. [PMID: 16912735 DOI: 10.1364/josaa.23.002085] [Citation(s) in RCA: 16] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
Several lines of evidence suggest that the image statistics of the environment shape visual abilities. To date, the image statistics of natural scenes and faces have been well characterized using Fourier analysis. We employed Fourier analysis to characterize images of signs in American Sign Language (ASL). These images are highly relevant to signers who rely on ASL for communication, and thus the image statistics of ASL might influence signers' visual abilities. Fourier analysis was conducted on 105 static images of signs, and these images were compared with analyses of 100 natural scene images and 100 face images. We obtained two metrics from our Fourier analysis: mean amplitude and entropy of the amplitude across the image set (which is a measure from information theory) as a function of spatial frequency and orientation. The results of our analyses revealed interesting differences in image statistics across the three different image sets, setting up the possibility that ASL experience may alter visual perception in predictable ways. In addition, for all image sets, the mean amplitude results were markedly different from the entropy results, which raises the interesting question of which aspect of an image set (mean amplitude or entropy of the amplitude) is better able to account for known visual abilities.
Collapse
Affiliation(s)
- Rain G Bosworth
- Department of Psychology, University of California, San Diego, La Jolla, 92093, USA
| | | | | |
Collapse
|