51
|
Masquelier T. Relative spike time coding and STDP-based orientation selectivity in the early visual system in natural continuous and saccadic vision: a computational model. J Comput Neurosci 2011; 32:425-41. [PMID: 21938439 DOI: 10.1007/s10827-011-0361-9] [Citation(s) in RCA: 23] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2011] [Revised: 09/05/2011] [Accepted: 09/08/2011] [Indexed: 10/17/2022]
Abstract
We have built a phenomenological spiking model of the cat early visual system comprising the retina, the Lateral Geniculate Nucleus (LGN) and V1's layer 4, and established four main results (1) When exposed to videos that reproduce with high fidelity what a cat experiences under natural conditions, adjacent Retinal Ganglion Cells (RGCs) have spike-time correlations at a short timescale (~30 ms), despite neuronal noise and possible jitter accumulation. (2) In accordance with recent experimental findings, the LGN filters out some noise. It thus increases the spike reliability and temporal precision, the sparsity, and, importantly, further decreases down to ~15 ms adjacent cells' correlation timescale. (3) Downstream simple cells in V1's layer 4, if equipped with Spike Timing-Dependent Plasticity (STDP), may detect these fine-scale cross-correlations, and thus connect principally to ON- and OFF-centre cells with Receptive Fields (RF) aligned in the visual space, and thereby become orientation selective, in accordance with Hubel and Wiesel (Journal of Physiology 160:106-154, 1962) classic model. Up to this point we dealt with continuous vision, and there was no absolute time reference such as a stimulus onset, yet information was encoded and decoded in the relative spike times. (4) We then simulated saccades to a static image and benchmarked relative spike time coding and time-to-first spike coding w.r.t. to saccade landing in the context of orientation representation. In both the retina and the LGN, relative spike times are more precise, less affected by pre-landing history and global contrast than absolute ones, and lead to robust contrast invariant orientation representations in V1.
Collapse
Affiliation(s)
- Timothée Masquelier
- Unit for Brain and Cognition, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain.
| |
Collapse
|
52
|
Onat S, König P, Jancke D. Natural Scene Evoked Population Dynamics across Cat Primary Visual Cortex Captured with Voltage-Sensitive Dye Imaging. Cereb Cortex 2011; 21:2542-54. [DOI: 10.1093/cercor/bhr038] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/14/2022] Open
|
53
|
Cecchi GA, Rao AR, Xiao Y, Kaplan E. Statistics of natural scenes and cortical color processing. J Vis 2010; 10:21. [PMID: 20884516 DOI: 10.1167/10.11.21] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
We investigate the spatial correlations of orientation and color information in natural images. We find that the correlation of orientation information falls off rapidly with increasing distance, while color information is more highly correlated over longer distances. We show that orientation and color information are statistically independent in natural images and that the spatial correlation of jointly encoded orientation and color information decays faster than that of color alone. Our findings suggest that: (a) orientation and color information should be processed in separate channels and (b) the organization of cortical color and orientation selectivity at low spatial frequencies is a reflection of the cortical adaptation to the statistical structure of the visual world. These findings are in agreement with biological observations, as form and color are thought to be represented by different classes of neurons in the primary visual cortex, and the receptive fields of color-selective neurons are larger than those of orientation-selective neurons. The agreement between our findings and biological observations supports the ecological theory of perception.
Collapse
|
54
|
Weiller D, Märtin R, Dähne S, Engel AK, König P. Involving motor capabilities in the formation of sensory space representations. PLoS One 2010; 5:e10377. [PMID: 20442849 PMCID: PMC2860999 DOI: 10.1371/journal.pone.0010377] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/24/2009] [Accepted: 03/21/2010] [Indexed: 11/19/2022] Open
Abstract
A goal of sensory coding is to capture features of sensory input that are behaviorally relevant. Therefore, a generic principle of sensory coding should take into account the motor capabilities of an agent. Up to now, unsupervised learning of sensory representations with respect to generic coding principles has been limited to passively received sensory input. Here we propose an algorithm that reorganizes an agent's representation of sensory space by maximizing the predictability of sensory state transitions given a motor action. We applied the algorithm to the sensory spaces of a number of simple, simulated agents with different motor parameters, moving in two-dimensional mazes. We find that the optimization algorithm generates compact, isotropic representations of space, comparable to hippocampal place fields. As expected, the size and spatial distribution of these place fields-like representations adapt to the motor parameters of the agent as well as to its environment. The representations prove to be well suited as a basis for path planning and navigation. They not only possess a high degree of state-transition predictability, but also are temporally stable. We conclude that the coding principle of predictability is a promising candidate for understanding place field formation as the result of sensorimotor reorganization.
Collapse
Affiliation(s)
- Daniel Weiller
- Department of Neurobiopsychology, Institute of Cognitive Science, University of Osnabrück, Osnabrück, Germany.
| | | | | | | | | |
Collapse
|
55
|
Rapela J, Felsen G, Touryan J, Mendel JM, Grzywacz NM. ePPR: a new strategy for the characterization of sensory cells from input/output data. NETWORK (BRISTOL, ENGLAND) 2010; 21:35-90. [PMID: 20735338 DOI: 10.3109/0954898x.2010.488714] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2023]
Abstract
A central goal of systems neuroscience is to characterize the transformation of sensory input to spiking output in single neurons. This problem is complicated by the large dimensionality of the inputs. To cope with this problem, previous methods have estimated simplified versions of a generic linear-nonlinear (LN) model and required, in most cases, stimuli with constrained statistics. Here we develop the extended Projection Pursuit Regression (ePPR) algorithm that allows the estimation of all of the parameters, in space and time, of a generic LN model using arbitrary stimuli. We first prove that ePPR models can uniformly approximate, to an arbitrary degree of precision, any continuous function. To test this generality empirically, we use ePPR to recover the parameters of models of cortical cells that cannot be represented exactly with an ePPR model. Next we evaluate ePPR with physiological data from primary visual cortex, and show that it can characterize both simple and complex cells, from their responses to both natural and random stimuli. For both simulated and physiological data, we show that ePPR compares favorably to spike-triggered and information-theoretic techniques. To the best of our knowledge, this article contains the first demonstration of a method that allows the estimation of an LN model of visual cells, containing multiple spatio-temporal filters, from their responses to natural stimuli.
Collapse
Affiliation(s)
- Joaquín Rapela
- Department of Electrical Engineering, University of Southern California, Hedco Neuroscience Building, Los Angeles, CA 90089-2520, USA.
| | | | | | | | | |
Collapse
|
56
|
Bex PJ, Solomon SG, Dakin SC. Contrast sensitivity in natural scenes depends on edge as well as spatial frequency structure. J Vis 2009; 9:1.1-19. [PMID: 19810782 PMCID: PMC3612947 DOI: 10.1167/9.10.1] [Citation(s) in RCA: 79] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
The contrast sensitivity function is routinely measured in the laboratory with sine-wave gratings presented on homogenous gray backgrounds; natural images are instead composed of a broad range of spatial and temporal structures. In order to extend channel-based models of visual processing to more natural conditions, we examined how contrast sensitivity varies with the context in which it is measured. We report that contrast sensitivity is quite different under laboratory than natural viewing conditions: adaptation or masking with natural scenes attenuates contrast sensitivity at low spatial and temporal frequencies. Expressed another way, viewing stimuli presented on homogenous screens overcomes chronic adaptation to the natural environment and causes a sharp, unnatural increase in sensitivity to low spatial and temporal frequencies. Consequently, the standard contrast sensitivity function is a poor indicator of sensitivity to structure in natural scenes. The magnitude of masking by natural scenes is relatively independent of local contrast but depends strongly on the density of edges even though neither greatly affects the local amplitude spectrum. These results suggest that sensitivity to spatial structure in natural scenes depends on the distribution of local edges as well as the local amplitude spectrum.
Collapse
Affiliation(s)
- Peter J Bex
- Department of Ophthalmology, Schepens Eye Research Institute, Harvard Medical School, Boston, MA 02114, USA.
| | | | | |
Collapse
|
57
|
Berkes P, Turner RE, Sahani M. A structured model of video reproduces primary visual cortical organisation. PLoS Comput Biol 2009; 5:e1000495. [PMID: 19730679 PMCID: PMC2726939 DOI: 10.1371/journal.pcbi.1000495] [Citation(s) in RCA: 27] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2008] [Accepted: 07/31/2009] [Indexed: 11/19/2022] Open
Abstract
The visual system must learn to infer the presence of objects and features in the world from the images it encounters, and as such it must, either implicitly or explicitly, model the way these elements interact to create the image. Do the response properties of cells in the mammalian visual system reflect this constraint? To address this question, we constructed a probabilistic model in which the identity and attributes of simple visual elements were represented explicitly and learnt the parameters of this model from unparsed, natural video sequences. After learning, the behaviour and grouping of variables in the probabilistic model corresponded closely to functional and anatomical properties of simple and complex cells in the primary visual cortex (V1). In particular, feature identity variables were activated in a way that resembled the activity of complex cells, while feature attribute variables responded much like simple cells. Furthermore, the grouping of the attributes within the model closely parallelled the reported anatomical grouping of simple cells in cat V1. Thus, this generative model makes explicit an interpretation of complex and simple cells as elements in the segmentation of a visual scene into basic independent features, along with a parametrisation of their moment-by-moment appearances. We speculate that such a segmentation may form the initial stage of a hierarchical system that progressively separates the identity and appearance of more articulated visual elements, culminating in view-invariant object recognition.
Collapse
Affiliation(s)
- Pietro Berkes
- Gatsby Computational Neuroscience Unit, London, United Kingdom
| | | | - Maneesh Sahani
- Gatsby Computational Neuroscience Unit, London, United Kingdom
| |
Collapse
|
58
|
Uhlhaas PJ, Pipa G, Lima B, Melloni L, Neuenschwander S, Nikolić D, Singer W. Neural synchrony in cortical networks: history, concept and current status. Front Integr Neurosci 2009; 3:17. [PMID: 19668703 PMCID: PMC2723047 DOI: 10.3389/neuro.07.017.2009] [Citation(s) in RCA: 426] [Impact Index Per Article: 26.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2009] [Accepted: 07/11/2009] [Indexed: 12/02/2022] Open
Abstract
Following the discovery of context-dependent synchronization of oscillatory neuronal responses in the visual system, the role of neural synchrony in cortical networks has been expanded to provide a general mechanism for the coordination of distributed neural activity patterns. In the current paper, we present an update of the status of this hypothesis through summarizing recent results from our laboratory that suggest important new insights regarding the mechanisms, function and relevance of this phenomenon. In the first part, we present recent results derived from animal experiments and mathematical simulations that provide novel explanations and mechanisms for zero and nero-zero phase lag synchronization. In the second part, we shall discuss the role of neural synchrony for expectancy during perceptual organization and its role in conscious experience. This will be followed by evidence that indicates that in addition to supporting conscious cognition, neural synchrony is abnormal in major brain disorders, such as schizophrenia and autism spectrum disorders. We conclude this paper with suggestions for further research as well as with critical issues that need to be addressed in future studies.
Collapse
Affiliation(s)
- Peter J. Uhlhaas
- Department of Neurophysiology, Max Planck Institute for Brain ResearchFrankfurt am Main, Germany
- Laboratory for Neurophysiology and Neuroimaging, Department of Psychiatry, Johann Wolfgang Goethe UniversitätFrankfurt am Main, Germany
| | - Gordon Pipa
- Department of Neurophysiology, Max Planck Institute for Brain ResearchFrankfurt am Main, Germany
- Frankfurt Institute for Advanced Studies, Johann Wolfgang Goethe UniversitätFrankfurt am Main, Germany
| | - Bruss Lima
- Department of Neurophysiology, Max Planck Institute for Brain ResearchFrankfurt am Main, Germany
| | - Lucia Melloni
- Department of Neurophysiology, Max Planck Institute for Brain ResearchFrankfurt am Main, Germany
| | - Sergio Neuenschwander
- Department of Neurophysiology, Max Planck Institute for Brain ResearchFrankfurt am Main, Germany
| | - Danko Nikolić
- Department of Neurophysiology, Max Planck Institute for Brain ResearchFrankfurt am Main, Germany
- Frankfurt Institute for Advanced Studies, Johann Wolfgang Goethe UniversitätFrankfurt am Main, Germany
| | - Wolf Singer
- Department of Neurophysiology, Max Planck Institute for Brain ResearchFrankfurt am Main, Germany
- Frankfurt Institute for Advanced Studies, Johann Wolfgang Goethe UniversitätFrankfurt am Main, Germany
| |
Collapse
|
59
|
Einhäuser W, Moeller GU, Schumann F, Conradt J, Vockeroth J, Bartl K, Schneider E, König P. Eye-Head Coordination during Free Exploration in Human and Cat. Ann N Y Acad Sci 2009; 1164:353-66. [DOI: 10.1111/j.1749-6632.2008.03709.x] [Citation(s) in RCA: 20] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
60
|
Ohayon S, Harmening W, Wagner H, Rivlin E. Through a barn owl's eyes: interactions between scene content and visual attention. BIOLOGICAL CYBERNETICS 2008; 98:115-132. [PMID: 18066583 DOI: 10.1007/s00422-007-0199-4] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/31/2007] [Accepted: 10/22/2007] [Indexed: 05/25/2023]
Abstract
In this study we investigated visual attention properties of freely behaving barn owls, using a miniature wireless camera attached to their heads. The tubular eye structure of barn owls makes them ideal subjects for this research since it limits their eye movements. Video sequences recorded from the owl's point of view capture part of the visual scene as seen by the owl. Automated analysis of video sequences revealed that during an active search task, owls repeatedly and consistently direct their gaze in a way that brings objects of interest to a specific retinal location (retinal fixation area). Using a projective model that captures the geometry between the eye and the camera, we recovered the corresponding location in the recorded images (image fixation area). Recording in various types of environments (aviary, office, outdoors) revealed significant statistical differences of low level image properties at the image fixation area compared to values extracted at random image patches. These differences are in agreement with results obtained in primates in similar studies. To investigate the role of saliency and its contribution to drawing the owl's attention, we used a popular bottom-up computational model. Saliency values at the image fixation area were typically greater than at random patches, yet were only 20% out of the maximal saliency value, suggesting a top-down modulation of gaze control.
Collapse
Affiliation(s)
- Shay Ohayon
- Israel Institute of Technology (Technion), Haifa 32000, Israel.
| | | | | | | |
Collapse
|
61
|
Abstract
Statistically efficient processing schemes focus the resources of a signal processing system on the range of statistically probable signals. Relying on the statistical properties of retinal motion signals during ego-motion we propose a nonlinear processing scheme for retinal flow. It maximizes the mutual information between the visual input and its neural representation, and distributes the processing load uniformly over the neural resources. We derive predictions for the receptive fields of motion sensitive neurons in the velocity space. The properties of the receptive fields are tightly connected to their position in the visual field, and to their preferred retinal velocity. The velocity tuning properties show characteristics of properties of neurons in the motion processing pathway of the primate brain.
Collapse
Affiliation(s)
- Dirk Calow
- Department of Psychology, Westfalische Wilhelms University, Munster, Germany.
| | | |
Collapse
|
62
|
Calow D, Lappe M. Local statistics of retinal optic flow for self-motion through natural sceneries. NETWORK (BRISTOL, ENGLAND) 2007; 18:343-374. [PMID: 18360939 DOI: 10.1080/09548980701642277] [Citation(s) in RCA: 22] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
Image analysis in the visual system is well adapted to the statistics of natural scenes. Investigations of natural image statistics have so far mainly focused on static features. The present study is dedicated to the measurement and the analysis of the statistics of optic flow generated on the retina during locomotion through natural environments. Natural locomotion includes bouncing and swaying of the head and eye movement reflexes that stabilize gaze onto interesting objects in the scene while walking. We investigate the dependencies of the local statistics of optic flow on the depth structure of the natural environment and on the ego-motion parameters. To measure these dependencies we estimate the mutual information between correlated data sets. We analyze the results with respect to the variation of the dependencies over the visual field, since the visual motions in the optic flow vary depending on visual field position. We find that retinal flow direction and retinal speed show only minor statistical interdependencies. Retinal speed is statistically tightly connected to the depth structure of the scene. Retinal flow direction is statistically mostly driven by the relation between the direction of gaze and the direction of ego-motion. These dependencies differ at different visual field positions such that certain areas of the visual field provide more information about ego-motion and other areas provide more information about depth. The statistical properties of natural optic flow may be used to tune the performance of artificial vision systems based on human imitating behavior, and may be useful for analyzing properties of natural vision systems.
Collapse
Affiliation(s)
- Dirk Calow
- Department of Psychology, Westf- Wilhelms University, Fliednerstr. 21, 48149 Münster, Germany.
| | | |
Collapse
|
63
|
Abstract
The purpose of our nervous system is to allow us to successfully interact with our environment. This normative idea is formalized by decision theory that defines which choices would be most beneficial. We live in an uncertain world, and each decision may have many possible outcomes; choosing the best decision is thus complicated. Bayesian decision theory formalizes these problems in the presence of uncertainty and often provides compact models that predict observed behavior. With its elegant formalization of the problems faced by the nervous system, it promises to become a major inspiration for studies in neuroscience.
Collapse
Affiliation(s)
- Konrad Körding
- Department of Physical Medicine and Rehabilitation, Institute of Neuroscience, Northwestern University and Rehabilitation Institute of Chicago, Room O-922, 345 East Superior Street, Chicago, IL 60611, USA.
| |
Collapse
|
64
|
Einhäuser W, Schumann F, Bardins S, Bartl K, Böning G, Schneider E, König P. Human eye-head co-ordination in natural exploration. NETWORK (BRISTOL, ENGLAND) 2007; 18:267-297. [PMID: 17926195 DOI: 10.1080/09548980701671094] [Citation(s) in RCA: 53] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
During natural behavior humans continuously adjust their gaze by moving head and eyes, yielding rich dynamics of the retinal input. Sensory coding models, however, typically assume visual input as smooth or a sequence of static images interleaved by volitional gaze shifts. Are these assumptions valid during free exploration behavior in natural environments? We used an innovative technique to simultaneously record gaze and head movements in humans, who freely explored various environments (forest, train station, apartment). Most movements occur along the cardinal axes, and the predominance of vertical or horizontal movements depends on the environment. Eye and head movements co-occur more frequently than their individual statistics predicts under an independence assumption. The majority of co-occurring movements point in opposite directions, consistent with a gaze-stabilizing role of eye movements. Nevertheless, a substantial fraction of eye movements point in the same direction as co-occurring head movements. Even under the very most conservative assumptions, saccadic eye movements alone cannot account for these synergistic movements. Hence nonsaccadic eye movements that interact synergistically with head movements to adjust gaze cannot be neglected in natural visual input. Natural retinal input is continuously dynamic, and cannot be faithfully modeled as a mere sequence of static frames with interleaved large saccades.
Collapse
|
65
|
|
66
|
Abstract
Abstract
It is generally believed that the visual system is adapted to the statistics of the visual world. Measuring and understanding these statistics require precise knowledge of the structure of the signals reaching fovea during image scanning. However, despite the fact that eye movements cause retinal stimulation to change several times in a second, it is implicitly assumed that images are sampled uniformly during natural viewing. By analyzing the eye movements of three rhesus monkeys freely viewing natural scenes, we report here significant anisotropy in stimulus statistics at the center of gaze. We find that fixation on an image patch is more likely to be followed by a saccade to a nearby patch of similar orientation structure or by a saccade to a more distant patch of largely dissimilar orientation structure. Furthermore, we show that orientation-selective neurons in the primary visual cortex (V1) can take advantage of eye movement statistics to selectively improve their discrimination performance.
Collapse
Affiliation(s)
- Valentin Dragoi
- University of Texas-Houston Medical School, Houston, TX 77030, USA.
| | | |
Collapse
|
67
|
König P, Krüger N. Symbols as self-emergent entities in an optimization process of feature extraction and predictions. BIOLOGICAL CYBERNETICS 2006; 94:325-34. [PMID: 16496197 DOI: 10.1007/s00422-006-0050-3] [Citation(s) in RCA: 22] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/10/2005] [Accepted: 01/10/2006] [Indexed: 05/06/2023]
Abstract
In the mammalian cortex the early sensory processing can be characterized as feature extraction resulting in local and analogue low-level representations. As a direct consequence, these map directly to the environment, but interpretation under natural conditions is ambiguous. In contrast, high-level representations for cognitive processing, e.g. language, require symbolic representations characterized by expression and syntax. The representations are binary, structured and disambiguated. However, do these fundamental functional distinctions translate into a fundamental distinction of the respective brain areas and their anatomical and physiological properties? Here we argue that the distinction between early sensory processing and higher cognitive functions may not be based on structural differences of cortical areas; instead similar learning principles acting on input signals with different statistics give rise to the observed variations of function. Firstly, we give an account of present research describing neuronal properties at early stages of sensory systems as a consequence of an optimization process over the set of natural stimuli. Secondly, addressing a stage following early visual processing we suggest to extend the unsupervised learning scheme by including predictive processes. These contain the widely used objective of temporal coherence as a special case and are a powerful approach to resolve ambiguities. Furthermore, in combination with a prior on the bandwidth of information exchange between units it leads to a condensation of information. Thirdly, as a crucial step, not only are predictive units optimized, but the selectivity of the feature extractors are adapted to allow optimal predictability. Thus, over and beyond making useful predictions, we propose that the predictability of a stimulus be in itself a selection criterion for further processing. In a hierarchical system the combined optimization process leads to entities that represent condensed pieces of knowledge and that are not analogue anymore. Instead, these entities work as arguments in a framework of transformations that realize predictions. Thus, the criteria of predictability and condensation in an optimization of sensory representations relate directly to the two defining properties of symbols of expression and syntax. In this paper, we sketch an unsupervised learning process that gradually transforms analogue local representations into discrete binary representations by means of four hypotheses. We propose that in this optimization process acting in a hierarchical system, entities emerge at, higher levels that fulfil the criteria defining symbols, instantiating qualitatively different representations at similarly structured low and high levels.
Collapse
Affiliation(s)
- Peter König
- Department of Neurobiopsychology, Institute of Cognitive Science, University Osnabrück, Albrechtstr. 28, 49076, Osnabrück, Germany.
| | | |
Collapse
|
68
|
Stocker AA, Simoncelli EP. Noise characteristics and prior expectations in human visual speed perception. Nat Neurosci 2006; 9:578-85. [PMID: 16547513 DOI: 10.1038/nn1669] [Citation(s) in RCA: 399] [Impact Index Per Article: 21.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2005] [Accepted: 02/21/2006] [Indexed: 11/09/2022]
Abstract
Human visual speed perception is qualitatively consistent with a Bayesian observer that optimally combines noisy measurements with a prior preference for lower speeds. Quantitative validation of this model, however, is difficult because the precise noise characteristics and prior expectations are unknown. Here, we present an augmented observer model that accounts for the variability of subjective responses in a speed discrimination task. This allowed us to infer the shape of the prior probability as well as the internal noise characteristics directly from psychophysical data. For all subjects, we found that the fitted model provides an accurate description of the data across a wide range of stimulus parameters. The inferred prior distribution shows significantly heavier tails than a Gaussian, and the amplitude of the internal noise is approximately proportional to stimulus speed and depends inversely on stimulus contrast. The framework is general and should prove applicable to other experiments and perceptual modalities.
Collapse
Affiliation(s)
- Alan A Stocker
- Howard Hughes Medical Institute, Center for Neural Science and Courant Institute of Mathematical Sciences, New York University, 4 Washington Place Rm 809, New York, New York 10003, USA.
| | | |
Collapse
|
69
|
Zanker JM, Zeil J. Movement-induced motion signal distributions in outdoor scenes. NETWORK (BRISTOL, ENGLAND) 2005; 16:357-76. [PMID: 16611590 DOI: 10.1080/09548980500497758] [Citation(s) in RCA: 13] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/08/2023]
Abstract
The movement of an observer generates a characteristic field of velocity vectors on the retina (Gibson 1950). Because such optic flow-fields are useful for navigation, many theoretical, psychophysical and physiological studies have addressed the question how ego-motion parameters such as direction of heading can be estimated from optic flow. Little is known, however, about the structure of optic flow under natural conditions. To address this issue, we recorded sequences of panoramic images along accurately defined paths in a variety of outdoor locations and used these sequences as input to a two-dimensional array of correlation-based motion detectors (2DMD). We find that (a) motion signal distributions are sparse and noisy with respect to local motion directions; (b) motion signal distributions contain patches (motion streaks) which are systematically oriented along the principal flow-field directions; (c) motion signal distributions show a distinct, dorso-ventral topography, reflecting the distance anisotropy of terrestrial environments; (d) the spatiotemporal tuning of the local motion detector we used has little influence on the structure of motion signal distributions, at least for the range of conditions we tested; and (e) environmental motion is locally noisy throughout the visual field, with little spatial or temporal correlation; it can therefore be removed by temporal averaging and is largely over-ridden by image motion caused by observer movement. Our results suggest that spatial or temporal integration is important to retrieve reliable information on the local direction and size of motion vectors, because the structure of optic flow is clearly detectable in the temporal average of motion signal distributions. Ego-motion parameters can be reliably retrieved from such averaged distributions under a range of environmental conditions. These observations raise a number of questions about the role of specific environmental and computational constraints in the processing of natural optic flow.
Collapse
Affiliation(s)
- J M Zanker
- Department of Psychology, Royal Holloway, University of London, Egham, Surrey TW20 0EX, UK.
| | | |
Collapse
|
70
|
Carandini M, Demb JB, Mante V, Tolhurst DJ, Dan Y, Olshausen BA, Gallant JL, Rust NC. Do we know what the early visual system does? J Neurosci 2005; 25:10577-97. [PMID: 16291931 PMCID: PMC6725861 DOI: 10.1523/jneurosci.3726-05.2005] [Citation(s) in RCA: 318] [Impact Index Per Article: 15.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2005] [Revised: 10/10/2005] [Accepted: 10/11/2005] [Indexed: 11/21/2022] Open
Abstract
We can claim that we know what the visual system does once we can predict neural responses to arbitrary stimuli, including those seen in nature. In the early visual system, models based on one or more linear receptive fields hold promise to achieve this goal as long as the models include nonlinear mechanisms that control responsiveness, based on stimulus context and history, and take into account the nonlinearity of spike generation. These linear and nonlinear mechanisms might be the only essential determinants of the response, or alternatively, there may be additional fundamental determinants yet to be identified. Research is progressing with the goals of defining a single "standard model" for each stage of the visual pathway and testing the predictive power of these models on the responses to movies of natural scenes. These predictive models represent, at a given stage of the visual pathway, a compact description of visual computation. They would be an invaluable guide for understanding the underlying biophysical and anatomical mechanisms and relating neural responses to visual perception.
Collapse
Affiliation(s)
- Matteo Carandini
- Smith-Kettlewell Eye Research Institute, San Francisco, California 94115, USA.
| | | | | | | | | | | | | | | |
Collapse
|
71
|
Boeddeker N, Lindemann JP, Egelhaaf M, Zeil J. Responses of blowfly motion-sensitive neurons to reconstructed optic flow along outdoor flight paths. J Comp Physiol A Neuroethol Sens Neural Behav Physiol 2005; 191:1143-55. [PMID: 16133502 DOI: 10.1007/s00359-005-0038-9] [Citation(s) in RCA: 45] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2005] [Revised: 06/29/2005] [Accepted: 07/02/2005] [Indexed: 11/24/2022]
Abstract
The retinal image flow a blowfly experiences in its daily life on the wing is determined by both the structure of the environment and the animal's own movements. To understand the design of visual processing mechanisms, there is thus a need to analyse the performance of neurons under natural operating conditions. To this end, we recorded flight paths of flies outdoors and reconstructed what they had seen, by moving a panoramic camera along exactly the same paths. The reconstructed image sequences were later replayed on a fast, panoramic flight simulator to identified, motion sensitive neurons of the so-called horizontal system (HS) in the lobula plate of the blowfly, which are assumed to extract self-motion parameters from optic flow. We show that under real life conditions HS-cells not only encode information about self-rotation, but are also sensitive to translational optic flow and, thus, indirectly signal information about the depth structure of the environment. These properties do not require an elaboration of the known model of these neurons, because the natural optic flow sequences generate--at least qualitatively--the same depth-related response properties when used as input to a computational HS-cell model and to real neurons.
Collapse
Affiliation(s)
- N Boeddeker
- Lehrstuhl Neurobiologie, Universität Bielefeld, Postfach 10 01 31, 33501 Bielefeld, Germany.
| | | | | | | |
Collapse
|
72
|
Einhäuser W, Hipp J, Eggert J, Körner E, König P. Learning viewpoint invariant object representations using a temporal coherence principle. BIOLOGICAL CYBERNETICS 2005; 93:79-90. [PMID: 16021516 DOI: 10.1007/s00422-005-0585-8] [Citation(s) in RCA: 28] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/19/2004] [Accepted: 05/23/2005] [Indexed: 05/03/2023]
Abstract
Invariant object recognition is arguably one of the major challenges for contemporary machine vision systems. In contrast, the mammalian visual system performs this task virtually effortlessly. How can we exploit our knowledge on the biological system to improve artificial systems? Our understanding of the mammalian early visual system has been augmented by the discovery that general coding principles could explain many aspects of neuronal response properties. How can such schemes be transferred to system level performance? In the present study we train cells on a particular variant of the general principle of temporal coherence, the "stability" objective. These cells are trained on unlabeled real-world images without a teaching signal. We show that after training, the cells form a representation that is largely independent of the viewpoint from which the stimulus is looked at. This finding includes generalization to previously unseen viewpoints. The achieved representation is better suited for view-point invariant object classification than the cells' input patterns. This property to facilitate view-point invariant classification is maintained even if training and classification take place in the presence of an--also unlabeled--distractor object. In summary, here we show that unsupervised learning using a general coding principle facilitates the classification of real-world objects, that are not segmented from the background and undergo complex, non-isomorphic, transformations.
Collapse
Affiliation(s)
- Wolfgang Einhäuser
- Institute of Neuroinformatics, University & ETH Zürich, Zürich, Switzerland.
| | | | | | | | | |
Collapse
|
73
|
Suzuki M, Floreano D, Di Paolo EA. The contribution of active body movement to visual development in evolutionary robots. Neural Netw 2005; 18:656-65. [PMID: 16112555 DOI: 10.1016/j.neunet.2005.06.043] [Citation(s) in RCA: 16] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
Abstract
Inspired by the pioneering work by Held and Hein (1963) on the development of kitten visuo-motor systems, we explore the role of active body movement in the developmental process of the visual system by using robots. The receptive fields in an evolved mobile robot are developed during active or passive movement with a Hebbian learning rule. In accordance to experimental observations in kittens, we show that the receptive fields and behavior of the robot developed under active condition significantly differ from those developed under passive condition. A possible explanation of this difference is derived by correlating receptive field formation and behavioral performance in the two conditions.
Collapse
Affiliation(s)
- Mototaka Suzuki
- Laboratory of Intelligent Systems, Swiss Federal Institute of Technology (EPFL) CH-1015 Lausanne, Switzerland.
| | | | | |
Collapse
|
74
|
Abstract
Visual processing has been widely investigated with narrow band stimuli at low contrasts. We used a masking paradigm to examine how visual sensitivity under these conditions compares with the perception of the direction of heading in real scenes (i.e., with dynamic natural images at high contrasts). We first confirmed and extended previous studies showing biases in the amplitude distribution for spatial frequency, temporal frequency, speed and direction in dynamic natural movies. We then measured contrast thresholds for identification of the direction of motion for an observer traveling at various speeds. In spite of differences in contrast sensitivity and large non-uniformities in the amplitude content of the stimuli, contrast thresholds were relatively invariant of spatial frequency and completely invariant of temporal frequency, speed and direction. Our results suggest that visual processing normalises responses to supra-threshold structure at different spatial and temporal frequencies within natural stimuli and so equates their effective visibility.
Collapse
Affiliation(s)
- Peter J Bex
- Division of Visual Rehabilitation Research, The Institute of Ophthalmology, London, UK.
| | | | | |
Collapse
|
75
|
Floreano D, Suzuki M, Mattiussi D. Active vision and receptive field development in evolutionary robots. EVOLUTIONARY COMPUTATION 2005; 13:527-44. [PMID: 16297282 DOI: 10.1162/106365605774666912] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/05/2023]
Abstract
In this paper, we describe the artificial evolution of adaptive neural controllers for an outdoor mobile robot equipped with a mobile camera. The robot can dynamically select the gazing direction by moving the body and/or the camera. The neural control system, which maps visual information to motor commands, is evolved online by means of a genetic algorithm, but the synaptic connections (receptive fields) from visual photoreceptors to internal neurons can also be modified by Hebbian plasticity while the robot moves in the environment. We show that robots evolved in physics-based simulations with Hebbian visual plasticity display more robust adaptive behavior when transferred to real outdoor environments as compared to robots evolved without visual plasticity. We also show that the formation of visual receptive fields is significantly and consistently affected by active vision as compared to the formation of receptive fields with grid sample images in the environment of the robot. Finally, we show that the interplay between active vision and receptive field formation amounts to the selection and exploitation of a small and constant subset of visual features available to the robot.
Collapse
Affiliation(s)
- Dario Floreano
- Laboratory of Intelligent Systems, Swiss Federal Institute of Technology Lausanne (EPFL), CH-1015 Lausanne, Switzerland.
| | | | | |
Collapse
|
76
|
Salazar RF, König P, Kayser C. Directed interactions between visual areas and their role in processing image structure and expectancy. Eur J Neurosci 2004; 20:1391-401. [PMID: 15341611 DOI: 10.1111/j.1460-9568.2004.03579.x] [Citation(s) in RCA: 19] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
Abstract
During sensory processing, cortical areas continuously exchange information in different directions along the hierarchy. The functional role of such interactions, however, has been the subject of various proposals. Here, we investigate the role of bottom-up and top-down interactions in processing stimulus structure and their relation to expected events. Applying multivariate autoregressive methods to local field potentials recorded in alert cats, we quantify directed interactions between primary (A17/18) and higher (A21) visual areas. A trial-by-trial analysis yields the following findings. To assess the role of interareal interactions in processing stimulus structure, we recorded in naïve animals during stimulation with natural movies and pink noise stimuli. The overall interactions decrease compared with baseline for both stimuli. To investigate whether forthcoming events modulate interactions, we recorded in trained animals viewing two stimuli, one of which had been associated with a reward. Several results support such modulations. First, the interactions increase compared with baseline and this increase is not observed in a context where food was not delivered. Second, these stimuli have a differential effect on top-down and bottom-up components. This difference is emphasized during the stimulus presentation and is maximal shortly before the possible reward. Furthermore, a spectral decomposition of the interactions shows that this asymmetry is most dominant in the gamma frequency range. Concluding, these results support the notion that interareal interactions are more related to an expectancy state rather than to processing of stimulus structure.
Collapse
Affiliation(s)
- Rodrigo F Salazar
- Institute of Neuroinformatics, University Zürich, Winterthurerstrasse 190, 8057, Switzerland.
| | | | | |
Collapse
|