1
|
Laeng B, Nabil S, Kitaoka A. Tunnel motion: Pupil dilations to optic flow within illusory dark holes. Perception 2024; 53:730-745. [PMID: 39196577 PMCID: PMC11451077 DOI: 10.1177/03010066241270493] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2024] [Accepted: 06/28/2024] [Indexed: 08/29/2024]
Abstract
We showed to the same observers both dynamic and static 2D patterns that can both evoke distinctive perceptions of motion or optic flow, as if moving in a tunnel or into a dark hole. At all times pupil diameters were monitored with an infrared eye tracker. We found a converging set of results indicating stronger pupil dilations to expansive growth of shapes or optic flows evoking a forward motion into a dark tunnel. Multiple regression analyses showed that the pupil responses to the illusory expanding black holes of static patterns were predicted by the individuals' pupil response to optic flows showing spiraling motion or "free fall" into a black hole. Also, individuals' pupil responses to spiraling motion into dark tunnels predicted the individuals' sense of illusory expansion with the static, illusory expanding, dark holes. This correspondence across individuals between their pupil responses to both dynamic and static, illusory expanding, holes suggests that these percepts reflect a common perceptual mechanism, deriving motion from 2D scenes, and that the observers' pupil adjustments reflect the direction and strength of motion they perceive and the expected outcome of an increase in darkness.
Collapse
Affiliation(s)
- Bruno Laeng
- University of Oslo, Norway; RITMO Centre for Interdisciplinary Studies in Rhythm, Time and Motion, University of Oslo, Norway
| | - Shoaib Nabil
- University of Oslo, Norway; University of Sussex, UK
| | | |
Collapse
|
2
|
Durand JB, Marchand S, Nasres I, Laeng B, De Castro V. Illusory light drives pupil responses in primates. J Vis 2024; 24:14. [PMID: 39046721 PMCID: PMC11271809 DOI: 10.1167/jov.24.7.14] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2024] [Accepted: 06/07/2024] [Indexed: 07/25/2024] Open
Abstract
In humans, the eye pupils respond to both physical light sensed by the retina and mental representations of light produced by the brain. Notably, our pupils constrict when a visual stimulus is illusorily perceived brighter, even if retinal illumination is constant. However, it remains unclear whether such perceptual penetrability of pupil responses is an epiphenomenon unique to humans or whether it represents an adaptive mechanism shared with other animals to anticipate variations in retinal illumination between successive eye fixations. To address this issue, we measured the pupil responses of both humans and macaque monkeys exposed to three chromatic versions (cyan, magenta, and yellow) of the Asahi brightness illusion. We found that the stimuli illusorily perceived brighter or darker trigger differential pupil responses that are very similar in macaques and human participants. Additionally, we show that this phenomenon exhibits an analogous cyan bias in both primate species. Beyond evincing the macaque monkey as a relevant model to study the perceptual penetrability of pupil responses, our results suggest that this phenomenon is tuned to ecological conditions because the exposure to a "bright cyan-bluish sky" may be associated with increased risks of dazzle and retinal damages.
Collapse
Affiliation(s)
- Jean-Baptiste Durand
- Université de Toulouse, Centre de Recherche Cerveau et Cognition, Toulouse, France
- Centre National de la Recherche Scientifique, Toulouse, France
| | - Sarah Marchand
- Université de Toulouse, Centre de Recherche Cerveau et Cognition, Toulouse, France
- Centre National de la Recherche Scientifique, Toulouse, France
| | - Ilyas Nasres
- Université de Toulouse, Centre de Recherche Cerveau et Cognition, Toulouse, France
- Centre National de la Recherche Scientifique, Toulouse, France
| | - Bruno Laeng
- Department of Psychology, University of Oslo, Oslo, Norway
- RITMO Centre for Interdisciplinary Studies in Rhythm, Time and Motion, University of Oslo, Oslo, Norway
| | - Vanessa De Castro
- Université de Toulouse, Centre de Recherche Cerveau et Cognition, Toulouse, France
- Centre National de la Recherche Scientifique, Toulouse, France
| |
Collapse
|
3
|
Zhao Z, Wang Y, Zou Q, Xu T, Tao F, Zhang J, Wang X, Shi CJR, Luo J, Xie Y. The spike gating flow: A hierarchical structure-based spiking neural network for online gesture recognition. Front Neurosci 2022; 16:923587. [PMID: 36408382 PMCID: PMC9667043 DOI: 10.3389/fnins.2022.923587] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2022] [Accepted: 10/03/2022] [Indexed: 01/25/2023] Open
Abstract
Action recognition is an exciting research avenue for artificial intelligence since it may be a game changer in emerging industrial fields such as robotic visions and automobiles. However, current deep learning (DL) faces major challenges for such applications because of the huge computational cost and inefficient learning. Hence, we developed a novel brain-inspired spiking neural network (SNN) based system titled spiking gating flow (SGF) for online action learning. The developed system consists of multiple SGF units which are assembled in a hierarchical manner. A single SGF unit contains three layers: a feature extraction layer, an event-driven layer, and a histogram-based training layer. To demonstrate the capability of the developed system, we employed a standard dynamic vision sensor (DVS) gesture classification as a benchmark. The results indicated that we can achieve 87.5% of accuracy which is comparable with DL, but at a smaller training/inference data number ratio of 1.5:1. Only a single training epoch is required during the learning process. Meanwhile, to the best of our knowledge, this is the highest accuracy among the non-backpropagation based SNNs. Finally, we conclude the few-shot learning (FSL) paradigm of the developed network: 1) a hierarchical structure-based network design involves prior human knowledge; 2) SNNs for content-based global dynamic feature detection.
Collapse
Affiliation(s)
- Zihao Zhao
- School of Microelectronics, Fudan University, Shanghai, China
- Alibaba DAMO Academy, Shanghai, China
| | - Yanhong Wang
- School of Microelectronics, Fudan University, Shanghai, China
- Alibaba DAMO Academy, Shanghai, China
| | - Qiaosha Zou
- School of Microelectronics, Fudan University, Shanghai, China
| | - Tie Xu
- Alibaba Group, Hangzhou, China
| | | | | | - Xiaoan Wang
- BrainUp Research Laboratory, Shanghai, China
| | - C.-J. Richard Shi
- Department of Electrical and Computer Engineering, University of Washington, Seattle, WA, United States
| | - Junwen Luo
- Alibaba DAMO Academy, Shanghai, China
- BrainUp Research Laboratory, Shanghai, China
| | - Yuan Xie
- Alibaba DAMO Academy, Shanghai, China
| |
Collapse
|
4
|
Gomez-Villa A, Martín A, Vazquez-Corral J, Bertalmío M, Malo J. On the synthesis of visual illusions using deep generative models. J Vis 2022; 22:2. [PMID: 35833884 PMCID: PMC9290318 DOI: 10.1167/jov.22.8.2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Visual illusions expand our understanding of the visual system by imposing constraints in the models in two different ways: i) visual illusions for humans should induce equivalent illusions in the model, and ii) illusions synthesized from the model should be compelling for human viewers too. These constraints are alternative strategies to find good vision models. Following the first research strategy, recent studies have shown that artificial neural network architectures also have human-like illusory percepts when stimulated with classical hand-crafted stimuli designed to fool humans. In this work we focus on the second (less explored) strategy: we propose a framework to synthesize new visual illusions using the optimization abilities of current automatic differentiation techniques. The proposed framework can be used with classical vision models as well as with more recent artificial neural network architectures. This framework, validated by psychophysical experiments, can be used to study the difference between a vision model and the actual human perception and to optimize the vision model to decrease this difference.
Collapse
Affiliation(s)
- Alex Gomez-Villa
- Computer Vision Center, Universitat Autónoma de Barcelona, Barcelona, Spain.,
| | - Adrián Martín
- Department of Information and Communications Technologies, Universitat Pompeu Fabra, Barcelona, Spain.,
| | - Javier Vazquez-Corral
- Computer Science Department, Universitat Autónoma de Barcelona and Computer Vision Center, Barcelona, Spain.,
| | | | - Jesús Malo
- Image Processing Lab, Faculty of Physics, Universitat de Valéncia, Spain.,
| |
Collapse
|
5
|
Tang S, Wang K, Ogrey S, Villazon J, Khan S, Paul A, Ardolino N, Kubendran R, Cauwenberghs G. Unity Human Eye Model for Gaze Tracking with a Query-Driven Dynamic Vision Sensor. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2022; 2022:2194-2198. [PMID: 36085625 DOI: 10.1109/embc48229.2022.9871193] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Objective measurement of gaze pattern and eye movement during untethered activity has important applications for neuroscience research and neurological disease detection. Current commercial eye-tracking tools rely on desk-top devices with infrared emitters and conventional frame-based cameras. Although wearable options do exist, the large power-consumption from their conventional cameras limit true long-term mobile usage. The query-driven Dynamic Vision Sensor (qDVS) is a neuromorphic camera which dramatically reduces power consumption by outputting only intensity-change threshold events, as opposed to full frames of intensity data. However, such hardware has not yet been implemented for on-body eye-tracking, but the feasibility can be demonstrated using a mathematical simulator to evaluate the eye-tracking ca-pabilities of the qDVS under controlled conditions. Specifically, a framework utilizing a realistic human eye model in the 3D graphics engine, Unity, is presented to enable the controlled and direct comparison of image-based gaze tracking methods. Eye-tracking based on qDVS frames was compared against two different conventional frame eye-tracking methods - the traditional ellipse pupil-fitting algorithm and a deep learning neural network inference model. Gaze accuracy from qDVS frames achieved an average of 93.2% for movement along the primary horizontal axis (pitch angle) and 93.1 % for movement along the primary vertical axis (yaw angle) under 4 different illumination conditions, demonstrating the feasibility for using qDVS hardware cameras for such applications. The quantitative framework for the direct comparison of eye tracking algorithms presented here is made open-source and can be extended to include other eye parameters, such as pupil dilation, reflection, motion artifact, and more.
Collapse
|
6
|
Laeng B, Nabil S, Kitaoka A. The Eye Pupil Adjusts to Illusorily Expanding Holes. Front Hum Neurosci 2022; 16:877249. [PMID: 35706480 PMCID: PMC9190027 DOI: 10.3389/fnhum.2022.877249] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2022] [Accepted: 04/29/2022] [Indexed: 11/24/2022] Open
Abstract
Some static patterns evoke the perception of an illusory expanding central region or “hole.” We asked observers to rate the magnitudes of illusory motion or expansion of black holes, and these predicted the degree of dilation of the pupil, measured with an eye tracker. In contrast, when the “holes” were colored (including white), i.e., emitted light, these patterns constricted the pupils, but the subjective expansions were also weaker compared with the black holes. The change rates of pupil diameters were significantly related to the illusory motion phenomenology only with the black holes. These findings can be accounted for within a perceiving-the-present account of visual illusions, where both the illusory motion and the pupillary adjustments represent compensatory mechanisms to the perception of the next moment, based on shared experiences with the ecological regularities of light.
Collapse
Affiliation(s)
- Bruno Laeng
- Department of Psychology, University of Oslo, Oslo, Norway
- *Correspondence: Bruno Laeng,
| | - Shoaib Nabil
- Department of Psychology, University of Oslo, Oslo, Norway
| | | |
Collapse
|
7
|
|
8
|
Treccani C. The brain, the artificial neural network and the snake: why we see what we see. AI & SOCIETY 2021. [DOI: 10.1007/s00146-020-01065-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
9
|
Sedda G, Ostry DJ, Sanguineti V, Sabatini SP. Self-operated stimuli improve subsequent visual motion integration. J Vis 2021; 21:13. [PMID: 34529006 PMCID: PMC8447044 DOI: 10.1167/jov.21.10.13] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Evidences of perceptual changes that accompany motor activity have been limited primarily to audition and somatosensation. Here we asked whether motor learning results in changes to visual motion perception. We designed a reaching task in which participants were trained to make movements along several directions, while the visual feedback was provided by an intrinsically ambiguous moving stimulus directly tied to hand motion. We find that training improves coherent motion perception and that changes in movement are correlated with perceptual changes. No perceptual changes are observed in passive training even when observers were provided with an explicit strategy to facilitate single motion perception. A Bayesian model suggests that movement training promotes the fine-tuning of the internal representation of stimulus geometry. These results emphasize the role of sensorimotor interaction in determining the persistent properties in space and time that define a percept.
Collapse
Affiliation(s)
- Giulia Sedda
- Department of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genoa, Italy.,
| | - David J Ostry
- Department of Psychology, McGill University, Montreal, Canada.,Haskins Laboratories, New Haven, CT, USA.,
| | - Vittorio Sanguineti
- Department of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genoa, Italy.,
| | - Silvio P Sabatini
- Department of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genoa, Italy.,
| |
Collapse
|
10
|
Global and local interference effects in ensemble encoding are best explained by interactions between summary representations of the mean and the range. Atten Percept Psychophys 2021; 83:1106-1128. [PMID: 33506350 PMCID: PMC8049940 DOI: 10.3758/s13414-020-02224-7] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/30/2020] [Indexed: 11/16/2022]
Abstract
Through ensemble encoding, the visual system compresses redundant statistical properties from multiple items into a single summary metric (e.g., average size). Numerous studies have shown that global summary information is extracted quickly, does not require access to single-item representations, and often interferes with reports of single items from the set. Yet a thorough understanding of ensemble processing would benefit from a more extensive investigation at the local level. Thus, the purpose of this study was to provide a more critical inspection of global-local processing in ensemble perception. Taking inspiration from Navon (Cognitive Psychology, 9(3), 353-383, 1977), we employed a novel paradigm that independently manipulates the degree of interference at the global (mean) or local (single item) level of the ensemble. Initial results were consistent with reciprocal interference between global and local ensemble processing. However, further testing revealed that local interference effects were better explained by interference from another summary statistic, the range of the set. Furthermore, participants were unable to disambiguate single items from the ensemble display from other items that were within the ensemble range but, critically, were not actually present in the ensemble. Thus, it appears that local item values are likely inferred based on their relationship to higher-order summary statistics such as the range and the mean. These results conflict with claims that local information is captured alongside global information in summary representations. In such studies, successful identification of set members was not compared with misidentification of items within the range, but which were nevertheless not presented within the set.
Collapse
|
11
|
Abstract
Arguably the most foundational principle in perception research is that our experience of the world goes beyond the retinal image; we perceive the distal environment itself, not the proximal stimulation it causes. Shape may be the paradigm case of such "unconscious inference": When a coin is rotated in depth, we infer the circular object it truly is, discarding the perspectival ellipse projected on our eyes. But is this really the fate of such perspectival shapes? Or does a tilted coin retain an elliptical appearance even when we know it's circular? This question has generated heated debate from Locke and Hume to the present; but whereas extant arguments rely primarily on introspection, this problem is also open to empirical test. If tilted coins bear a representational similarity to elliptical objects, then a circular coin should, when rotated, impair search for a distal ellipse. Here, nine experiments demonstrate that this is so, suggesting that perspectival shapes persist in the mind far longer than traditionally assumed. Subjects saw search arrays of three-dimensional "coins," and simply had to locate a distally elliptical coin. Surprisingly, rotated circular coins slowed search for elliptical targets, even when subjects clearly knew the rotated coins were circular. This pattern arose with static and dynamic cues, couldn't be explained by strategic responding or unfamiliarity, generalized across shape classes, and occurred even with sustained viewing. Finally, these effects extended beyond artificial displays to real-world objects viewed in naturalistic, full-cue conditions. We conclude that objects have a remarkably persistent dual character: their objective shape "out there," and their perspectival shape "from here."
Collapse
|
12
|
Akbarinia A, Gil-Rodríguez R. Deciphering image contrast in object classification deep networks. Vision Res 2020; 173:61-76. [PMID: 32480109 DOI: 10.1016/j.visres.2020.04.015] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2019] [Revised: 04/24/2020] [Accepted: 04/28/2020] [Indexed: 11/16/2022]
Abstract
The ultimate goal of neuroscience is to explain how complex behaviour arises from neuronal activity. A comparable level of complexity also emerges in deep neural networks (DNNs) while exhibiting human-level performance in demanding visual tasks. Unlike in biological systems, all parameters and operations of DNNs are accessible. Therefore, in theory, it should be possible to decipher the exact mechanisms learnt by these artificial networks. Here, we investigate the concept of contrast invariance within the framework of DNNs. We start by discussing how a network can achieve robustness to changes in local and global image contrast. We used a technique from neuroscience-"kernel lesion"-to measure the degree of performance degradation when individual kernels are eliminated from a network. We further compared contrast normalisation, a mechanism used in biological systems, to the strategies that DNNs learn to cope with changes of contrast. The results of our analysis suggest that (i) contrast is a low-level feature for these networks, and it is encoded in the shallow layers; (ii) a handful of kernels appear to have a greater impact on this feature, and their removal causes a substantially larger accuracy loss for low-contrast images; (iii) edges are a distinct visual feature within the internal representation of object classification DNNs.
Collapse
Affiliation(s)
- Arash Akbarinia
- Department of General Psychology, Justus-Liebig University, D-35394 Giessen, Germany.
| | - Raquel Gil-Rodríguez
- Department of General Psychology, Justus-Liebig University, D-35394 Giessen, Germany
| |
Collapse
|
13
|
Ng CJ, Purves D. An Alternative Theory of Binocularity. Front Comput Neurosci 2019; 13:71. [PMID: 31649521 PMCID: PMC6794442 DOI: 10.3389/fncom.2019.00071] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2019] [Accepted: 09/20/2019] [Indexed: 11/13/2022] Open
Abstract
The fact that seeing with two eyes is universal among vertebrates raises a problem that has long challenged vision scientists: how do animals with overlapping visual fields combine non-identical right and left eye images to achieve fusion and the perception of depth that follows? Most theories address this problem in terms of matching corresponding images on the right and left retinas. Here we suggest an alternative theory of binocular vision based on anatomical correspondence that circumvents the correspondence problem and provides a rationale for ocular dominance.
Collapse
Affiliation(s)
- Cherlyn J. Ng
- Neuroscience and Behavioral Disorders Program, Duke-NUS Graduate Medical School Singapore, Singapore, Singapore
- Flaum Eye Institute, University of Rochester, Rochester, NY, United States
- Center for Visual Science, University of Rochester, Rochester, NY, United States
| | - Dale Purves
- Neuroscience and Behavioral Disorders Program, Duke-NUS Graduate Medical School Singapore, Singapore, Singapore
- Duke Institute for Brain Sciences, Duke University, Durham, NC, United States
| |
Collapse
|
14
|
Colorful glares: Effects of colors on brightness illusions measured with pupillometry. Acta Psychol (Amst) 2019; 198:102882. [PMID: 31288107 DOI: 10.1016/j.actpsy.2019.102882] [Citation(s) in RCA: 20] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2019] [Revised: 06/21/2019] [Accepted: 07/01/2019] [Indexed: 11/21/2022] Open
Abstract
We hypothesized that pupil constrictions to the glare illusion, where converging luminance gradients subjectively enhance the perception of brightness, would be stronger for 'blue' than for other colors. Such an expectation was based on reflections about the ecology of vision, where the experience of dazzling light is common when one happens to look directly at sunlight through some occluders. Thus, we hypothesized that pupil constrictions to 'blue' reflect an ecologically-based expectation of the visual system from the experience of sky's light and color, which also leads to interpret the blue gradients of illusory glare to act as effective cues to impending probable intense light. We therefore manipulated the gradients color of glare illusions and measured changes in subjective brightness of identical shape stimuli. We confirmed that the blue resulted in what was subjectively evaluated as the brightest condition, despite all colored stimuli were equiluminant. This enhanced brightness effect was observed both in a psychophysical adjustment task and in changes in pupil size, where the maximum pupil constriction peak was observed with the 'blue' converging gradients over and above to the pupil response to blue in other conditions (i.e., diverging gradients and homogeneous patches). Moreover, glare-related pupil constrictions for each participant were correlated to each individual's subjective brightness adjustments. Homogenous blue hues also constricted the pupil more than other hues, which represents a pupillometric analog of the Helmholtz-Kohlrausch effect on brightness perception. Together, these findings show that pupillometry constitutes an easy tool to assess individual differences in color brightness perception.
Collapse
|
15
|
Sama MA, Nestor A, Cant JS. Independence of viewpoint and identity in face ensemble processing. J Vis 2019; 19:2. [DOI: 10.1167/19.5.2] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Affiliation(s)
- Marco A. Sama
- Department of Psychology, University of Toronto Scarborough, Toronto, Canada
| | - Adrian Nestor
- Department of Psychology, University of Toronto Scarborough, Toronto, Canada
| | - Jonathan S. Cant
- Department of Psychology, University of Toronto Scarborough, Toronto, Canada
| |
Collapse
|
16
|
De Freitas J, Alvarez GA. Your visual system provides all the information you need to make moral judgments about generic visual events. Cognition 2018; 178:133-146. [DOI: 10.1016/j.cognition.2018.05.017] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2017] [Revised: 05/21/2018] [Accepted: 05/22/2018] [Indexed: 10/14/2022]
|
17
|
Abstract
Seeing-perception and vision-is implicitly the fundamental building block of the literature on rationality and cognition. Herbert Simon and Daniel Kahneman's arguments against the omniscience of economic agents-and the concept of bounded rationality-depend critically on a particular view of the nature of perception and vision. We propose that this framework of rationality merely replaces economic omniscience with perceptual omniscience. We show how the cognitive and social sciences feature a pervasive but problematic meta-assumption that is characterized by an "all-seeing eye." We raise concerns about this assumption and discuss different ways in which the all-seeing eye manifests itself in existing research on (bounded) rationality. We first consider the centrality of vision and perception in Simon's pioneering work. We then point to Kahneman's work-particularly his article "Maps of Bounded Rationality"-to illustrate the pervasiveness of an all-seeing view of perception, as manifested in the extensive use of visual examples and illusions. Similar assumptions about perception can be found across a large literature in the cognitive sciences. The central problem is the present emphasis on inverse optics-the objective nature of objects and environments, e.g., size, contrast, and color. This framework ignores the nature of the organism and perceiver. We argue instead that reality is constructed and expressed, and we discuss the species-specificity of perception, as well as perception as a user interface. We draw on vision science as well as the arts to develop an alternative understanding of rationality in the cognitive and social sciences. We conclude with a discussion of the implications of our arguments for the rationality and decision-making literature in cognitive psychology and behavioral economics, along with suggesting some ways forward.
Collapse
|
18
|
Zavagno D, Tommasi L, Laeng B. The Eye Pupil's Response to Static and Dynamic Illusions of Luminosity and Darkness. Iperception 2017; 8:2041669517717754. [PMID: 28835810 PMCID: PMC5555513 DOI: 10.1177/2041669517717754] [Citation(s) in RCA: 27] [Impact Index Per Article: 3.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
Pupil diameters were recorded with an eye-tracker while participants observed cruciform patterns of gray-scale gradients that evoked illusions of enhanced brightness (glare) or of enhanced darkness. The illusions were either presented as static images or as dynamic animations which initially appeared as a pattern of filled squares that—in a few seconds—gradually changed into gradients until the patterns were identical to the static ones. Gradients could either converge toward the center, resulting in a central region of enhanced, illusory, brightness or darkness, or oriented toward each side of the screen, resulting in the perception of a peripheral ring of illusory brightness or darkness. It was found that pupil responses to these illusions matched both the direction and intensity of perceived changes in light: Glare stimuli resulted in pupil constrictions, and darkness stimuli evoked dilations of the pupils. A second experiment found that gradients of brightness were most effective in constricting the pupils than isoluminant step-luminance, local, variations in luminance. This set of findings suggest that the eye strategically adjusts to reflect in a predictive manner, given that these brightness illusions only suggest a change in luminance when none has occurred, the content within brightness maps of the visual scene.
Collapse
Affiliation(s)
- Daniele Zavagno
- Department of Psychology, University of Milano-Bicocca, Italy
| | - Luca Tommasi
- Department of Psychological Science, Health and Territory, University of Chieti-Pescara, Italy
| | - Bruno Laeng
- Department of Psychology, University of Oslo, Norway
| |
Collapse
|
19
|
Abstract
The nature of perception has fascinated philosophers for centuries, and has more recently been the focus of research in psychology and neuroscience. Many psychiatric disorders are characterised by perceptual abnormalities, ranging from sensory distortions to illusions and hallucinations. The distinction between normal and abnormal perception is, however, hard to articulate. In this article we argue that the distinction between normal perception and abnormal perception is best seen as a quantitative one, resting on the degree to which the observer's prior expectations influence perceptual inference. We illustrate this point with an example taken from researchers at Google working on computer vision.
Collapse
Affiliation(s)
| | - Joseph M Nour
- Oxford University Clinical Academic Graduate School (OUCAGS), Oxford
| |
Collapse
|
20
|
Zhou L, Ooi TL, He ZJ. Intrinsic spatial knowledge about terrestrial ecology favors the tall for judging distance. SCIENCE ADVANCES 2016; 2:e1501070. [PMID: 27602402 PMCID: PMC5007070 DOI: 10.1126/sciadv.1501070] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/10/2015] [Accepted: 08/02/2016] [Indexed: 06/06/2023]
Abstract
Our sense of vision reliably directs and guides our everyday actions, such as reaching and walking. This ability is especially fascinating because the optical images of natural scenes that project into our eyes are insufficient to adequately form a perceptual space. It has been proposed that the brain makes up for this inadequacy by using its intrinsic spatial knowledge. However, it is unclear what constitutes intrinsic spatial knowledge and how it is acquired. We investigated this question and showed evidence of an ecological basis, which uses the statistical spatial relationship between the observer and the terrestrial environment, namely, the ground surface. We found that in dark and reduced-cue environments where intrinsic knowledge has a greater contribution, perceived target location is more accurate when referenced to the ground than to the ceiling. Furthermore, taller observers more accurately localized the target. Superior performance was also observed in the full-cue environment, even when we compensated for the observers' heights by having the taller observer sit on a chair and the shorter observers stand on a box. Although fascinating, this finding dovetails with the prediction of the ecological hypothesis for intrinsic spatial knowledge. It suggests that an individual's accumulated lifetime experiences of being tall and his or her constant interactions with ground-based objects not only determine intrinsic spatial knowledge but also endow him or her with an advantage in spatial ability in the intermediate distance range.
Collapse
Affiliation(s)
- Liu Zhou
- Key Laboratory of Brain Functional Genomics (Ministry of Education and Science and Technology Commission of Shanghai Municipality), Institute of Cognitive Neurosciences, School of Psychology and Cognitive Science, East China Normal University, Shanghai 200062, China
| | - Teng Leng Ooi
- College of Optometry, Ohio State University, Columbus, OH 43210, USA
| | - Zijiang J. He
- Key Laboratory of Brain Functional Genomics (Ministry of Education and Science and Technology Commission of Shanghai Municipality), Institute of Cognitive Neurosciences, School of Psychology and Cognitive Science, East China Normal University, Shanghai 200062, China
- Department of Psychological and Brain Sciences, University of Louisville, Louisville, KY 40292, USA
| |
Collapse
|
21
|
Emberson LL, Rubinstein DY. Statistical learning is constrained to less abstract patterns in complex sensory input (but not the least). Cognition 2016; 153:63-78. [PMID: 27139779 PMCID: PMC4905776 DOI: 10.1016/j.cognition.2016.04.010] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2014] [Revised: 04/12/2016] [Accepted: 04/15/2016] [Indexed: 11/06/2022]
Abstract
The influence of statistical information on behavior (either through learning or adaptation) is quickly becoming foundational to many domains of cognitive psychology and cognitive neuroscience, from language comprehension to visual development. We investigate a central problem impacting these diverse fields: when encountering input with rich statistical information, are there any constraints on learning? This paper examines learning outcomes when adult learners are given statistical information across multiple levels of abstraction simultaneously: from abstract, semantic categories of everyday objects to individual viewpoints on these objects. After revealing statistical learning of abstract, semantic categories with scrambled individual exemplars (Exp. 1), participants viewed pictures where the categories as well as the individual objects predicted picture order (e.g., bird1—dog1, bird2—dog2). Our findings suggest that participants preferentially encode the relationships between the individual objects, even in the presence of statistical regularities linking semantic categories (Exps. 2 and 3). In a final experiment we investigate whether learners are biased towards learning object-level regularities or simply construct the most detailed model given the data (and therefore best able to predict the specifics of the upcoming stimulus) by investigating whether participants preferentially learn from the statistical regularities linking individual snapshots of objects or the relationship between the objects themselves (e.g., bird_picture1— dog_picture1, bird_picture2—dog_picture2). We find that participants fail to learn the relationships between individual snapshots, suggesting a bias towards object-level statistical regularities as opposed to merely constructing the most complete model of the input. This work moves beyond the previous existence proofs that statistical learning is possible at both very high and very low levels of abstraction (categories vs. individual objects) and suggests that, at least with the current categories and type of learner, there are biases to pick up on statistical regularities between individual objects even when robust statistical information is present at other levels of abstraction. These findings speak directly to emerging theories about how systems supporting statistical learning and prediction operate in our structure-rich environments. Moreover, the theoretical implications of the current work across multiple domains of study is already clear: statistical learning cannot be assumed to be unconstrained even if statistical learning has previously been established at a given level of abstraction when that information is presented in isolation.
Collapse
Affiliation(s)
- Lauren L Emberson
- Brain and Cognitive Sciences Department, University of Rochester, USA; Psychology Department, Princeton University, USA.
| | - Dani Y Rubinstein
- Psychology Department, Cornell University, USA; Department of Neuroscience, Brown University, USA; Section on Integrative Neuroimaging, Clinical and Translational Neuroscience Branch, National Institute of Mental Health, NIH, Bethesda, MD, USA
| |
Collapse
|
22
|
Lacquaniti F, Bosco G, Gravano S, Indovina I, La Scaleia B, Maffei V, Zago M. Gravity in the Brain as a Reference for Space and Time Perception. Multisens Res 2016; 28:397-426. [PMID: 26595949 DOI: 10.1163/22134808-00002471] [Citation(s) in RCA: 38] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
Abstract
Moving and interacting with the environment require a reference for orientation and a scale for calibration in space and time. There is a wide variety of environmental clues and calibrated frames at different locales, but the reference of gravity is ubiquitous on Earth. The pull of gravity on static objects provides a plummet which, together with the horizontal plane, defines a three-dimensional Cartesian frame for visual images. On the other hand, the gravitational acceleration of falling objects can provide a time-stamp on events, because the motion duration of an object accelerated by gravity over a given path is fixed. Indeed, since ancient times, man has been using plumb bobs for spatial surveying, and water clocks or pendulum clocks for time keeping. Here we review behavioral evidence in favor of the hypothesis that the brain is endowed with mechanisms that exploit the presence of gravity to estimate the spatial orientation and the passage of time. Several visual and non-visual (vestibular, haptic, visceral) cues are merged to estimate the orientation of the visual vertical. However, the relative weight of each cue is not fixed, but depends on the specific task. Next, we show that an internal model of the effects of gravity is combined with multisensory signals to time the interception of falling objects, to time the passage through spatial landmarks during virtual navigation, to assess the duration of a gravitational motion, and to judge the naturalness of periodic motion under gravity.
Collapse
|
23
|
Purves D, Morgenstern Y, Wojtach WT. Perception and Reality: Why a Wholly Empirical Paradigm is Needed to Understand Vision. Front Syst Neurosci 2015; 9:156. [PMID: 26635546 PMCID: PMC4649043 DOI: 10.3389/fnsys.2015.00156] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2015] [Accepted: 10/29/2015] [Indexed: 11/13/2022] Open
Abstract
A central puzzle in vision science is how perceptions that are routinely at odds with physical measurements of real world properties can arise from neural responses that nonetheless lead to effective behaviors. Here we argue that the solution depends on: (1) rejecting the assumption that the goal of vision is to recover, however imperfectly, properties of the world; and (2) replacing it with a paradigm in which perceptions reflect biological utility based on past experience rather than objective features of the environment. Present evidence is consistent with the conclusion that conceiving vision in wholly empirical terms provides a plausible way to understand what we see and why.
Collapse
Affiliation(s)
- Dale Purves
- Duke Institute for Brain Sciences, Duke UniversityDurham, NC, USA
| | | | - William T. Wojtach
- Duke Institute for Brain Sciences, Duke UniversityDurham, NC, USA
- Duke-NUS Graduate Medical SchoolSingapore, Singapore
| |
Collapse
|
24
|
|
25
|
Purves D, Morgenstern Y, Wojtach WT. Will understanding vision require a wholly empirical paradigm? Front Psychol 2015; 6:1072. [PMID: 26283998 PMCID: PMC4519674 DOI: 10.3389/fpsyg.2015.01072] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/04/2015] [Accepted: 07/13/2015] [Indexed: 11/13/2022] Open
Abstract
Based on electrophysiological and anatomical studies, a prevalent conception is that the visual system recovers features of the world from retinal images to generate perceptions and guide behavior. This paradigm, however, is unable to explain why visual perceptions differ from physical measurements, or how behavior could routinely succeed on this basis. An alternative is that vision does not recover features of the world, but assigns perceptual qualities empirically by associating frequently occurring stimulus patterns with useful responses on the basis of survival and reproductive success. The purpose of the present article is to briefly describe this strategy of vision and the evidence for it.
Collapse
Affiliation(s)
- Dale Purves
- Duke Institute for Brain Sciences, Duke University, Durham, NC, USA
- Neuroscience and Behavioral Disorders Program, Duke-NUS Graduate Medical School Singapore, Singapore, Singapore
- Department of Neurobiology, Duke University, Durham, NC, USA
| | - Yaniv Morgenstern
- Neuroscience and Behavioral Disorders Program, Duke-NUS Graduate Medical School Singapore, Singapore, Singapore
| | - William T. Wojtach
- Duke Institute for Brain Sciences, Duke University, Durham, NC, USA
- Neuroscience and Behavioral Disorders Program, Duke-NUS Graduate Medical School Singapore, Singapore, Singapore
| |
Collapse
|
26
|
Prediction in speech and language processing. Cortex 2015; 68:1-7. [PMID: 26048658 DOI: 10.1016/j.cortex.2015.05.001] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2015] [Revised: 05/03/2015] [Accepted: 05/03/2015] [Indexed: 11/20/2022]
|
27
|
Kubilius J, Wagemans J, Op de Beeck HP. A conceptual framework of computations in mid-level vision. Front Comput Neurosci 2014; 8:158. [PMID: 25566044 PMCID: PMC4264474 DOI: 10.3389/fncom.2014.00158] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2014] [Accepted: 11/17/2014] [Indexed: 11/13/2022] Open
Abstract
If a picture is worth a thousand words, as an English idiom goes, what should those words-or, rather, descriptors-capture? What format of image representation would be sufficiently rich if we were to reconstruct the essence of images from their descriptors? In this paper, we set out to develop a conceptual framework that would be: (i) biologically plausible in order to provide a better mechanistic understanding of our visual system; (ii) sufficiently robust to apply in practice on realistic images; and (iii) able to tap into underlying structure of our visual world. We bring forward three key ideas. First, we argue that surface-based representations are constructed based on feature inference from the input in the intermediate processing layers of the visual system. Such representations are computed in a largely pre-semantic (prior to categorization) and pre-attentive manner using multiple cues (orientation, color, polarity, variation in orientation, and so on), and explicitly retain configural relations between features. The constructed surfaces may be partially overlapping to compensate for occlusions and are ordered in depth (figure-ground organization). Second, we propose that such intermediate representations could be formed by a hierarchical computation of similarity between features in local image patches and pooling of highly-similar units, and reestimated via recurrent loops according to the task demands. Finally, we suggest to use datasets composed of realistically rendered artificial objects and surfaces in order to better understand a model's behavior and its limitations.
Collapse
Affiliation(s)
- Jonas Kubilius
- Laboratory of Biological Psychology, Faculty of Psychology and Educational Sciences, KU LeuvenLeuven, Belgium
- Laboratory of Experimental Psychology, Faculty of Psychology and Educational Sciences, KU LeuvenLeuven, Belgium
| | - Johan Wagemans
- Laboratory of Experimental Psychology, Faculty of Psychology and Educational Sciences, KU LeuvenLeuven, Belgium
| | - Hans P. Op de Beeck
- Laboratory of Biological Psychology, Faculty of Psychology and Educational Sciences, KU LeuvenLeuven, Belgium
| |
Collapse
|
28
|
Morgenstern Y, Rukmini DV, Monson BB, Purves D. Properties of artificial neurons that report lightness based on accumulated experience with luminance. Front Comput Neurosci 2014; 8:134. [PMID: 25404912 PMCID: PMC4217489 DOI: 10.3389/fncom.2014.00134] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2013] [Accepted: 10/01/2014] [Indexed: 11/13/2022] Open
Abstract
The responses of visual neurons in experimental animals have been extensively characterized. To ask whether these responses are consistent with a wholly empirical concept of visual perception, we optimized simple neural networks that responded according to the cumulative frequency of occurrence of local luminance patterns in retinal images. Based on this estimation of accumulated experience, the neuron responses showed classical center-surround receptive fields, luminance gain control and contrast gain control, the key properties of early level visual neurons determined in animal experiments. These results imply that a major purpose of pre-cortical neuronal circuitry is to contend with the inherently uncertain significance of luminance values in natural stimuli.
Collapse
Affiliation(s)
- Yaniv Morgenstern
- Neuroscience and Behavioral Disorders Program, Duke-NUS Graduate Medical School Singapore, Singapore
| | - Dhara V Rukmini
- Neuroscience and Behavioral Disorders Program, Duke-NUS Graduate Medical School Singapore, Singapore
| | - Brian B Monson
- Neuroscience and Behavioral Disorders Program, Duke-NUS Graduate Medical School Singapore, Singapore
| | - Dale Purves
- Neuroscience and Behavioral Disorders Program, Duke-NUS Graduate Medical School Singapore, Singapore ; Department of Neurobiology, Duke University Medical Center Durham, NC, USA ; Duke Institute for Brain Sciences, Duke University Durham, NC, USA
| |
Collapse
|
29
|
Morgenstern Y, Rostami M, Purves D. Properties of artificial networks evolved to contend with natural spectra. Proc Natl Acad Sci U S A 2014; 111 Suppl 3:10868-72. [PMID: 25024184 PMCID: PMC4113924 DOI: 10.1073/pnas.1402669111] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Understanding why spectra that are physically the same appear different in different contexts (color contrast), whereas spectra that are physically different appear similar (color constancy) presents a major challenge in vision research. Here, we show that the responses of biologically inspired neural networks evolved on the basis of accumulated experience with spectral stimuli automatically generate contrast and constancy. The results imply that these phenomena are signatures of a strategy that biological vision uses to circumvent the inverse optics problem as it pertains to light spectra, and that double-opponent neurons in early-level vision evolve to serve this purpose. This strategy provides a way of understanding the peculiar relationship between the objective world and subjective color experience, as well as rationalizing the relevant visual circuitry without invoking feature detection or image representation.
Collapse
Affiliation(s)
- Yaniv Morgenstern
- Neuroscience and Behavioral Disorders Program, Duke-National University of Singapore Graduate Medical School, Singapore 169857; and
| | - Mohammad Rostami
- Neuroscience and Behavioral Disorders Program, Duke-National University of Singapore Graduate Medical School, Singapore 169857; and
| | - Dale Purves
- Neuroscience and Behavioral Disorders Program, Duke-National University of Singapore Graduate Medical School, Singapore 169857; andDuke Institute for Brain Sciences, Duke University, Durham, NC 27708
| |
Collapse
|