1
|
Carlini A, Bordeau C, Ambard M. Auditory localization: a comprehensive practical review. Front Psychol 2024; 15:1408073. [PMID: 39049946 PMCID: PMC11267622 DOI: 10.3389/fpsyg.2024.1408073] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2024] [Accepted: 06/17/2024] [Indexed: 07/27/2024] Open
Abstract
Auditory localization is a fundamental ability that allows to perceive the spatial location of a sound source in the environment. The present work aims to provide a comprehensive overview of the mechanisms and acoustic cues used by the human perceptual system to achieve such accurate auditory localization. Acoustic cues are derived from the physical properties of sound waves, and many factors allow and influence auditory localization abilities. This review presents the monaural and binaural perceptual mechanisms involved in auditory localization in the three dimensions. Besides the main mechanisms of Interaural Time Difference, Interaural Level Difference and Head Related Transfer Function, secondary important elements such as reverberation and motion, are also analyzed. For each mechanism, the perceptual limits of localization abilities are presented. A section is specifically devoted to reference systems in space, and to the pointing methods used in experimental research. Finally, some cases of misperception and auditory illusion are described. More than a simple description of the perceptual mechanisms underlying localization, this paper is intended to provide also practical information available for experiments and work in the auditory field.
Collapse
|
2
|
Stodt B, Neudek D, Getzmann S, Wascher E, Martin R. Comparing auditory distance perception in real and virtual environments and the role of the loudness cue: A study based on event-related potentials. Hear Res 2024; 444:108968. [PMID: 38350176 DOI: 10.1016/j.heares.2024.108968] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/30/2023] [Revised: 01/12/2024] [Accepted: 02/02/2024] [Indexed: 02/15/2024]
Abstract
The perception of the distance to a sound source is relevant in many everyday situations, not only in real spaces, but also in virtual reality (VR) environments. Where real rooms often reach their limits, VR offers far-reaching possibilities to simulate a wide range of acoustic scenarios. However, in virtual room acoustics a plausible reproduction of distance-related cues can be challenging. In the present study, we compared the detection of changes of the distance to a sound source and its neurocognitive correlates in a real and a virtual reverberant environment, using an active auditory oddball paradigm and EEG measures. The main goal was to test whether the experiments in the virtual and real environments produced equivalent behavioral and EEG results. Three loudspeakers were placed at ego-centric distances of 2 m (near), 4 m (center), and 8 m (far) in front of the participants (N = 20), each 66 cm below their ear level. Sequences of 500 ms noise stimuli were presented either from the center position (standards, 80 % of trials) or from the near or far position (targets, 10 % each). The participants had to indicate a target position via a joystick response ("near" or "far"). Sounds were emitted either by real loudspeakers in the real environment or rendered and played back for the corresponding positions via headphones in the virtual environment. In addition, within both environments, loudness of the auditory stimuli was either unaltered (natural loudness) or the loudness cue was manipulated, so that all three loudspeakers were perceived equally loud at the listener's position (matched loudness). The EEG analysis focused on the mismatch negativity (MMN), P3a, and P3b as correlates of deviance detection, attentional orientation, and context-updating/stimulus evaluation, respectively. Overall, behavioral data showed that detection of the target positions was reduced within the virtual environment, and especially when loudness was matched. Except for slight latency shifts in the virtual environment, EEG analysis indicated comparable patterns within both environments and independent of loudness settings. Thus, while the neurocognitive processing of changes in distance appears to be similar in virtual and real spaces, a proper representation of loudness appears to be crucial to achieve a good task performance in virtual acoustic environments.
Collapse
Affiliation(s)
- Benjamin Stodt
- Leibniz Research Centre for Working Environment and Human Factors at the TU Dortmund (IfADo), Ardeystraße 67, Dortmund 44139, Germany.
| | - Daniel Neudek
- Institute of Communication Acoustics, Ruhr-Universität Bochum, Universitätsstraße 150, Bochum 44780, Germany
| | - Stephan Getzmann
- Leibniz Research Centre for Working Environment and Human Factors at the TU Dortmund (IfADo), Ardeystraße 67, Dortmund 44139, Germany
| | - Edmund Wascher
- Leibniz Research Centre for Working Environment and Human Factors at the TU Dortmund (IfADo), Ardeystraße 67, Dortmund 44139, Germany
| | - Rainer Martin
- Institute of Communication Acoustics, Ruhr-Universität Bochum, Universitätsstraße 150, Bochum 44780, Germany
| |
Collapse
|
3
|
López-Mochales S, Aparicio-Terrés R, Díaz-Andreu M, Escera C. Acoustic perception and emotion evocation by rock art soundscapes of Altai (Russia). Front Psychol 2023; 14:1188567. [PMID: 37794915 PMCID: PMC10546042 DOI: 10.3389/fpsyg.2023.1188567] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2023] [Accepted: 09/04/2023] [Indexed: 10/06/2023] Open
Abstract
The major goal of psychoarchaeoacoustics is to understand the psychology behind motivations and emotions of past communities when selecting certain acoustic environments to set activities involving the production of paintings and carvings. Within this framework, the present study seeks to explore whether a group of archaeological rock art sites in Altai (Siberia, Russia) are distinguished by particular acoustic imprints that elicit distinct reactions on listeners, in perceptual and emotional terms. Sixty participants were presented with a series of natural sounds convolved with six impulse responses from Altai, three of them recorded in locations in front of rock art panels and three of them in front of similar locations but without any trace of rock art. Participants were interrogated about their subjective perception of the sounds presented, using 10 psychoacoustic and emotional scales. The mixed ANOVA analyses carried out revealed that feelings of "presence," "closeness," and "tension" evoked by all sounds were significantly influenced by the location. These effects were attributed to the differences in reverberation between the locations with and without rock art. Despite results are not consistent across all the studied rock art sites, and acknowledging the presence of several limitations, this study highlights the significance of its methodology. It stresses the crucial aspect of incorporating the limitations encountered in shaping future research endeavors.
Collapse
Affiliation(s)
- Samantha López-Mochales
- Brainlab – Cognitive Neuroscience Research Group, Department of Clinical Psychology and Psychobiology, Universitat de Barcelona (UB), Barcelona, Spain
- Institut de Neurociències, Universitat de Barcelona (UB), Barcelona, Spain
| | - Raquel Aparicio-Terrés
- Brainlab – Cognitive Neuroscience Research Group, Department of Clinical Psychology and Psychobiology, Universitat de Barcelona (UB), Barcelona, Spain
- Institut de Neurociències, Universitat de Barcelona (UB), Barcelona, Spain
| | - Margarita Díaz-Andreu
- Departament d’Història i Arqueologia, Universitat de Barcelona (UB), Barcelona, Spain
- Institució Catalana de Recerca i Estudis Avançats (ICREA), Barcelona, Spain
- Institut d’Arqueologia de la Universitat de Barcelona (IUAB), Universitat de Barcelona (UB), Barcelona, Spain
| | - Carles Escera
- Brainlab – Cognitive Neuroscience Research Group, Department of Clinical Psychology and Psychobiology, Universitat de Barcelona (UB), Barcelona, Spain
- Institut de Neurociències, Universitat de Barcelona (UB), Barcelona, Spain
- Institució Catalana de Recerca i Estudis Avançats (ICREA), Barcelona, Spain
- Institut de Recerca Sant Joan de Déu (IRSJD), Esplugues de Llobregat, Spain
| |
Collapse
|
4
|
Higgins NC, Pupo DA, Ozmeral EJ, Eddins DA. Head movement and its relation to hearing. Front Psychol 2023; 14:1183303. [PMID: 37448716 PMCID: PMC10338176 DOI: 10.3389/fpsyg.2023.1183303] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2023] [Accepted: 06/07/2023] [Indexed: 07/15/2023] Open
Abstract
Head position at any point in time plays a fundamental role in shaping the auditory information that reaches a listener, information that continuously changes as the head moves and reorients to different listening situations. The connection between hearing science and the kinesthetics of head movement has gained interest due to technological advances that have increased the feasibility of providing behavioral and biological feedback to assistive listening devices that can interpret movement patterns that reflect listening intent. Increasing evidence also shows that the negative impact of hearing deficits on mobility, gait, and balance may be mitigated by prosthetic hearing device intervention. Better understanding of the relationships between head movement, full body kinetics, and hearing health, should lead to improved signal processing strategies across a range of assistive and augmented hearing devices. The purpose of this review is to introduce the wider hearing community to the kinesiology of head movement and to place it in the context of hearing and communication with the goal of expanding the field of ecologically-specific listener behavior.
Collapse
Affiliation(s)
- Nathan C. Higgins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
| | - Daniel A. Pupo
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
- School of Aging Studies, University of South Florida, Tampa, FL, United States
| | - Erol J. Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
| | - David A. Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
| |
Collapse
|
5
|
Lohse M, Zimmer-Harwood P, Dahmen JC, King AJ. Integration of somatosensory and motor-related information in the auditory system. Front Neurosci 2022; 16:1010211. [PMID: 36330342 PMCID: PMC9622781 DOI: 10.3389/fnins.2022.1010211] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2022] [Accepted: 09/28/2022] [Indexed: 11/30/2022] Open
Abstract
An ability to integrate information provided by different sensory modalities is a fundamental feature of neurons in many brain areas. Because visual and auditory inputs often originate from the same external object, which may be located some distance away from the observer, the synthesis of these cues can improve localization accuracy and speed up behavioral responses. By contrast, multisensory interactions occurring close to the body typically involve a combination of tactile stimuli with other sensory modalities. Moreover, most activities involving active touch generate sound, indicating that stimuli in these modalities are frequently experienced together. In this review, we examine the basis for determining sound-source distance and the contribution of auditory inputs to the neural encoding of space around the body. We then consider the perceptual consequences of combining auditory and tactile inputs in humans and discuss recent evidence from animal studies demonstrating how cortical and subcortical areas work together to mediate communication between these senses. This research has shown that somatosensory inputs interface with and modulate sound processing at multiple levels of the auditory pathway, from the cochlear nucleus in the brainstem to the cortex. Circuits involving inputs from the primary somatosensory cortex to the auditory midbrain have been identified that mediate suppressive effects of whisker stimulation on auditory thalamocortical processing, providing a possible basis for prioritizing the processing of tactile cues from nearby objects. Close links also exist between audition and movement, and auditory responses are typically suppressed by locomotion and other actions. These movement-related signals are thought to cancel out self-generated sounds, but they may also affect auditory responses via the associated somatosensory stimulation or as a result of changes in brain state. Together, these studies highlight the importance of considering both multisensory context and movement-related activity in order to understand how the auditory cortex operates during natural behaviors, paving the way for future work to investigate auditory-somatosensory interactions in more ecological situations.
Collapse
|
6
|
Agrawal T, Schachner A. Hearing water temperature: characterizing the development of nuanced perception of sound sources. Dev Sci 2022; 26:e13321. [PMID: 36068928 DOI: 10.1111/desc.13321] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2022] [Revised: 08/18/2022] [Accepted: 08/22/2022] [Indexed: 11/30/2022]
Abstract
Without conscious thought, listeners link events in the world to sounds they hear. We study one surprising example: Adults can judge the temperature of water simply from hearing it being poured. We test development of the ability to hear water temperature, with the goal of informing developmental theories regarding the origins and cognitive bases of nuanced sound source judgments. We first confirmed that adults accurately distinguished the sounds of hot and cold water (pre-registered Exps. 1, 2; total N = 384), even though many were unaware or uncertain of this ability. By contrast, children showed protracted development of this skill over the course of middle childhood (Exps. 2, 3; total N = 178). In spite of accurately identifying other sounds and hot/cold images, older children (7-11 years) but not younger children (3-6 years) reliably distinguished the sounds of hot and cold water. Accuracy increased with age; 11 year old's performance was similar to adults'. Adults also showed individual differences in accuracy that were predicted by their amount of prior relevant experience (Exp. 1). Experience may similarly play a role in children's performance; differences in auditory sensitivity and multimodal integration may also contribute to young children's failures. The ability to hear water temperature develops slowly over childhood, such that nuanced auditory information that is easily and quickly accessible to adults is not available to guide young children's behavior. Adults can make nuanced judgments from sound, including accurately judging the temperature of water from the sound of it being poured. Children showed protracted development of this skill over the course of middle childhood, such that 7-11 year-olds reliably succeeded while 3-6 year-olds performed at chance. Developmental changes may be due to experience (adults with greater relevant experience showed higher accuracy) and development of multimodal integration and auditory sensitivity. Young children may not detect subtle auditory information that adults easily perceive. This article is protected by copyright. All rights reserved.
Collapse
Affiliation(s)
| | - Adena Schachner
- Department of Psychology, University of California, San Diego, USA
| |
Collapse
|
7
|
An Algorithm for Generating Virtual Sources in Dynamic Virtual Auditory Display Based on Tensor Decomposition of Head-Related Impulse Responses. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12157715] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Dynamic virtual auditory displays (VADs) are increasingly used for generating various auditory objects and scenes in virtual and augmented reality. Dynamic VADs are required to generate virtual sources in various directions and distances by using HRTF- or HRIR-based binaural synthesis. In the present work, an algorithm for improving the efficiency and performance of binaural synthesis in dynamic VAD is proposed. Based on tensor decomposition, a full set of near-field HRIRs is decomposed as a combination of distance-, direction-, and time-related modes. Then, binaural synthesis in VAD can be implemented by a common set of time mode-related convolvers or filters associated with direction- and distance-related weights. Dynamic binaural signals are created by updating the weights rather than updating the HRIR-based convolvers, which enables the independent control of virtual source distance and direction and avoids the audible artifact caused by updating the HRIR-based convolvers. An example of implementation indicates that a set of eight common convolvers or filters for each ear is enough to synthesize the binaural signals with sufficient accuracy. The computational efficiency of simultaneously generating multiple virtual sources is improved when the number of virtual sources is larger than eight. A virtual-source localization experiment validates the algorithm.
Collapse
|
8
|
Hu X, Song A, Wei Z, Zeng H. StereoPilot: A Wearable Target Location System for Blind and Visually Impaired Using Spatial Audio Rendering. IEEE Trans Neural Syst Rehabil Eng 2022; 30:1621-1630. [PMID: 35696467 DOI: 10.1109/tnsre.2022.3182661] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Vision loss severely impacts object recognition and spatial cognition for limited vision individuals. It is a challenge to compensate for this using other sensory modalities, such as touch or hearing. This paper introduces StereoPilot, a wearable target location system to facilitate the spatial cognition of BVI. Through wearing a head-mounted RGB-D camera, the 3D spatial information of the environment is measured and processed into navigation cues. Leveraging spatial audio rendering (SAR) technology, it allows the navigation cues to be transmitted in a type of 3D sound from which the sound orientation can be distinguished by the sound localization instincts in humans. Three haptic and auditory display strategies were compared with SAR through experiments with three BVI and four sighted subjects. Compared with mainstream speech instructional feedback, the experimental results of the Fitts' law test showed that SAR increases the information transfer rate (ITR) by a factor of three for spatial navigation, while the positioning error is reduced by 40%. Furthermore, SAR has a lower learning effect than other sonification approaches such as vOICe. In desktop manipulation experiments, StereoPilot was able to obtain precise localization of desktop objects while reducing the completion time of target grasping tasks in half as compared to the voice instruction method. In summary, StereoPilot provides an innovative wearable target location solution that swiftly and intuitively transmits environmental information to BVI individuals in the real world.
Collapse
|
9
|
Chua SFA, Liu Y, Harris JM, Otto TU. No selective integration required: A race model explains responses to audiovisual motion-in-depth. Cognition 2022; 227:105204. [PMID: 35753178 DOI: 10.1016/j.cognition.2022.105204] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2021] [Revised: 06/02/2022] [Accepted: 06/08/2022] [Indexed: 11/03/2022]
Abstract
Looming motion is an ecologically salient signal that often signifies danger. In both audition and vision, humans show behavioral biases in response to perceiving looming motion, which is suggested to indicate an adaptation for survival. However, it is an open question whether such biases occur also in the combined processing of multisensory signals. Towards this aim, Cappe, Thut, Romei, and Murraya (2009) found that responses to audiovisual signals were faster for congruent looming motion compared to receding motion or incongruent combinations. They considered this as evidence for selective integration of multisensory looming signals. To test this proposal, here, we successfully replicate the behavioral results by Cappe et al. (2009). We then show that the redundant signals effect (RSE - a speedup of multisensory compared to unisensory responses) is not distinct for congruent looming motion. Instead, as predicted by a simple probability summation rule, the RSE is primarily modulated by the looming bias in audition, which suggests that multisensory processing inherits a unisensory effect. Finally, we compare a large set of so-called race models that implement probability summation, but that allow for interference between auditory and visual processing. The best-fitting model, selected by the Akaike Information Criterion (AIC), virtually perfectly explained the RSE across conditions with interference parameters that were either constant or varied only with auditory motion. In the absence of effects jointly caused by auditory and visual motion, we conclude that selective integration is not required to explain the behavioral benefits that occur with audiovisual looming motion.
Collapse
Affiliation(s)
- S F Andrew Chua
- School of Psychology & Neuroscience, University of St Andrews, St Mary's Quad, South Street, St Andrews KY16 9JP, United Kingdom.
| | - Yue Liu
- School of Psychology & Neuroscience, University of St Andrews, St Mary's Quad, South Street, St Andrews KY16 9JP, United Kingdom
| | - Julie M Harris
- School of Psychology & Neuroscience, University of St Andrews, St Mary's Quad, South Street, St Andrews KY16 9JP, United Kingdom
| | - Thomas U Otto
- School of Psychology & Neuroscience, University of St Andrews, St Mary's Quad, South Street, St Andrews KY16 9JP, United Kingdom.
| |
Collapse
|
10
|
Russell MK. Age and Auditory Spatial Perception in Humans: Review of Behavioral Findings and Suggestions for Future Research. Front Psychol 2022; 13:831670. [PMID: 35250777 PMCID: PMC8888835 DOI: 10.3389/fpsyg.2022.831670] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2021] [Accepted: 01/24/2022] [Indexed: 11/13/2022] Open
Abstract
It has been well documented, and fairly well known, that concomitant with an increase in chronological age is a corresponding increase in sensory impairment. As most people realize, our hearing suffers as we get older; hence, the increased need for hearing aids. The first portion of the present paper is how the change in age apparently affects auditory judgments of sound source position. A summary of the literature evaluating the changes in the perception of sound source location and the perception of sound source motion as a function of chronological age is presented. The review is limited to empirical studies with behavioral findings involving humans. It is the view of the author that we have an immensely limited understanding of how chronological age affects perception of space when based on sound. In the latter part of the paper, discussion is given to how auditory spatial perception is traditionally conducted in the laboratory. Theoretically, beneficial reasons exist for conducting research in the manner it has been. Nonetheless, from an ecological perspective, the vast majority of previous research can be considered unnatural and greatly lacking in ecological validity. Suggestions for an alternative and more ecologically valid approach to the investigation of auditory spatial perception are proposed. It is believed an ecological approach to auditory spatial perception will enhance our understanding of the extent to which individuals perceive sound source location and how those perceptual judgments change with an increase in chronological age.
Collapse
|
11
|
Auditory distance perception in front and rear space. Hear Res 2022; 417:108468. [DOI: 10.1016/j.heares.2022.108468] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/15/2021] [Revised: 01/22/2022] [Accepted: 02/12/2022] [Indexed: 11/21/2022]
|
12
|
Wen W, Okon Y, Yamashita A, Asama H. The over-estimation of distance for self-voice versus other-voice. Sci Rep 2022; 12:420. [PMID: 35013503 PMCID: PMC8748720 DOI: 10.1038/s41598-021-04437-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2021] [Accepted: 12/22/2021] [Indexed: 11/08/2022] Open
Abstract
Self-related stimuli are important cues for people to recognize themselves in the external world and hold a special status in our perceptual system. Self-voice plays an important role in daily social communication and is also a frequent input for self-identification. Although many studies have been conducted on the acoustic features of self-voice, no research has ever examined the spatial aspect, although the spatial perception of voice is important for humans. This study proposes a novel perspective for studying self-voice. We investigated people's distance perception of their own voice when the voice was heard from an external position. Participants heard their own voice from one of four speakers located either 90 or 180 cm from their sitting position, either immediately after uttering a short vowel (i.e., active session) or hearing the replay of their own pronunciation (i.e., replay session). They were then asked to indicate which speaker they heard the voice from. Their voices were either pitch-shifted by ± 4 semitones (i.e., other-voice condition) or unaltered (i.e., self-voice condition). The results of spatial judgment showed that self-voice from the closer speakers was misattributed to that from the speakers further away at a significantly higher proportion than other-voice. This phenomenon was also observed when the participants remained silent and heard prerecorded voices. Additional structural equation modeling using participants' schizotypal scores showed that the effect of self-voice on distance perception was significantly associated with the score of delusional thoughts (Peters Delusion Inventory) and distorted body image (Perceptual Aberration Scale) in the active speaking session but not in the replay session. The findings of this study provide important insights for understanding how people process self-related stimuli when there is a small distortion and how this may be linked to the risk of psychosis.
Collapse
Affiliation(s)
- Wen Wen
- Research Into Artifacts, Center for Engineering, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo, 113-8656, Japan.
- Department of Precision Engineering, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo, 113-8656, Japan.
| | - Yuta Okon
- Department of Precision Engineering, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo, 113-8656, Japan
| | - Atsushi Yamashita
- Department of Precision Engineering, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo, 113-8656, Japan
| | - Hajime Asama
- Research Into Artifacts, Center for Engineering, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo, 113-8656, Japan
- Department of Precision Engineering, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo, 113-8656, Japan
| |
Collapse
|
13
|
Kolarik AJ, Moore BCJ, Cirstea S, Aggius-Vella E, Gori M, Campus C, Pardhan S. Factors Affecting Auditory Estimates of Virtual Room Size: Effects of Stimulus, Level, and Reverberation. Perception 2021; 50:646-663. [PMID: 34053354 DOI: 10.1177/03010066211020598] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
Abstract
When vision is unavailable, auditory level and reverberation cues provide important spatial information regarding the environment, such as the size of a room. We investigated how room-size estimates were affected by stimulus type, level, and reverberation. In Experiment 1, 15 blindfolded participants estimated room size after performing a distance bisection task in virtual rooms that were either anechoic (with level cues only) or reverberant (with level and reverberation cues) with a relatively short reverberation time of T60 = 400 milliseconds. Speech, noise, or clicks were presented at distances between 1.9 and 7.1 m. The reverberant room was judged to be significantly larger than the anechoic room (p < .05) for all stimuli. In Experiment 2, only the reverberant room was used and the overall level of all sounds was equalized, so only reverberation cues were available. Ten blindfolded participants took part. Room-size estimates were significantly larger for speech than for clicks or noise. The results show that when level and reverberation cues are present, reverberation increases judged room size. Even relatively weak reverberation cues provide room-size information, which could potentially be used by blind or visually impaired individuals encountering novel rooms.
Collapse
Affiliation(s)
- Andrew J Kolarik
- Anglia Ruskin University, Cambridge, UK.,Anglia Ruskin University, Cambridge, UK
| | - Brian C J Moore
- Anglia Ruskin University, Cambridge, UK; University of Cambridge, Cambridge, UK.,Anglia Ruskin University, Cambridge, UK
| | - Silvia Cirstea
- Anglia Ruskin University, Cambridge, UK.,Anglia Ruskin University, Cambridge, UK
| | - Elena Aggius-Vella
- Fondazione Istituto Italiano di Tecnologia, Genoa, Italy; Institute for Mind, Brain and Technology, Herzeliya, Israel.,Anglia Ruskin University, Cambridge, UK
| | | | - Claudio Campus
- Fondazione Istituto Italiano di Tecnologia, Genoa, Italy.,Anglia Ruskin University, Cambridge, UK
| | | |
Collapse
|
14
|
Causal inference in environmental sound recognition. Cognition 2021; 214:104627. [PMID: 34044231 DOI: 10.1016/j.cognition.2021.104627] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/13/2020] [Revised: 01/28/2021] [Accepted: 02/05/2021] [Indexed: 11/23/2022]
Abstract
Sound is caused by physical events in the world. Do humans infer these causes when recognizing sound sources? We tested whether the recognition of common environmental sounds depends on the inference of a basic physical variable - the source intensity (i.e., the power that produces a sound). A source's intensity can be inferred from the intensity it produces at the ear and its distance, which is normally conveyed by reverberation. Listeners could thus use intensity at the ear and reverberation to constrain recognition by inferring the underlying source intensity. Alternatively, listeners might separate these acoustic cues from their representation of a sound's identity in the interest of invariant recognition. We compared these two hypotheses by measuring recognition accuracy for sounds with typically low or high source intensity (e.g., pepper grinders vs. trucks) that were presented across a range of intensities at the ear or with reverberation cues to distance. The recognition of low-intensity sources (e.g., pepper grinders) was impaired by high presentation intensities or reverberation that conveyed distance, either of which imply high source intensity. Neither effect occurred for high-intensity sources. The results suggest that listeners implicitly use the intensity at the ear along with distance cues to infer a source's power and constrain its identity. The recognition of real-world sounds thus appears to depend upon the inference of their physical generative parameters, even generative parameters whose cues might otherwise be separated from the representation of a sound's identity.
Collapse
|
15
|
Cuevas-Rodriguez M, Gonzalez-Toledo D, Reyes-Lecuona A, Picinali L. Impact of non-individualised head related transfer functions on speech-in-noise performances within a synthesised virtual environment. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:2573. [PMID: 33940900 DOI: 10.1121/10.0004220] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/28/2020] [Accepted: 03/19/2021] [Indexed: 06/12/2023]
Abstract
When performing binaural spatialisation, it is widely accepted that the choice of the head related transfer functions (HRTFs), and in particular the use of individually measured ones, can have an impact on localisation accuracy, externalization, and overall realism. Yet the impact of HRTF choices on speech-in-noise performances in cocktail party-like scenarios has not been investigated in depth. This paper introduces a study where 22 participants were presented with a frontal speech target and two lateral maskers, spatialised using a set of non-individual HRTFs. Speech reception threshold (SRT) was measured for each HRTF. Furthermore, using the SRT predicted by an existing speech perception model, the measured values were compensated in the attempt to remove overall HRTF-specific benefits. Results show significant overall differences among the SRTs measured using different HRTFs, consistently with the results predicted by the model. Individual differences between participants related to their SRT performances using different HRTFs could also be found, but their significance was reduced after the compensation. The implications of these findings are relevant to several research areas related to spatial hearing and speech perception, suggesting that when testing speech-in-noise performances within binaurally rendered virtual environments, the choice of the HRTF for each individual should be carefully considered.
Collapse
Affiliation(s)
- Maria Cuevas-Rodriguez
- Departamento de Tecnología Electrónica, Universidad de Málaga, ETSI Telecomunicación, 29010 Málaga, Spain
| | - Daniel Gonzalez-Toledo
- Departamento de Tecnología Electrónica, Universidad de Málaga, ETSI Telecomunicación, 29010 Málaga, Spain
| | - Arcadio Reyes-Lecuona
- Departamento de Tecnología Electrónica, Universidad de Málaga, ETSI Telecomunicación, 29010 Málaga, Spain
| | - Lorenzo Picinali
- Dyson School of Design Engineering, Imperial College London, London SW7 2DB, United Kingdom
| |
Collapse
|
16
|
Preserving Human Perspectives in Cultural Heritage Acoustics: Distance Cues and Proxemics in Aural Heritage Fieldwork. ACOUSTICS 2021. [DOI: 10.3390/acoustics3010012] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
We examine the praxis implications of our working definition of aural heritage: spatial acoustics as physically experienced by humans in cultural contexts; aligned with the aims of anthropological archaeology (the study of human life from materials). Here we report on human-centered acoustical data collection strategies from our project “Digital Preservation and Access to Aural Heritage via a Scalable, Extensible Method,” supported by the National Endowment for the Humanities (NEH) in the USA. The documentation and accurate translation of human sensory perspectives is fundamental to the ecological validity of cultural heritage fieldwork and the preservation of heritage acoustics. Auditory distance cues, which enable and constrain sonic communication, relate to proxemics, contextualized understandings of distance relationships that are fundamental to human social interactions. We propose that source–receiver locations in aural heritage measurements should be selected to represent a comprehensive range of proxemics according to site-contextualized spatial-use scenarios, and we identify and compare acoustical metrics for auditory distance cues from acoustical fieldwork we conducted using this strategy in three contrasting case-study heritage sites. This conceptual shift from architectural acoustical sampling to aural heritage sampling prioritizes culturally and physically plausible human auditory/sound-sensing perspectives and relates them to spatial proxemics as scaled architecturally.
Collapse
|
17
|
Ptito M, Bleau M, Djerourou I, Paré S, Schneider FC, Chebat DR. Brain-Machine Interfaces to Assist the Blind. Front Hum Neurosci 2021; 15:638887. [PMID: 33633557 PMCID: PMC7901898 DOI: 10.3389/fnhum.2021.638887] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2020] [Accepted: 01/19/2021] [Indexed: 12/31/2022] Open
Abstract
The loss or absence of vision is probably one of the most incapacitating events that can befall a human being. The importance of vision for humans is also reflected in brain anatomy as approximately one third of the human brain is devoted to vision. It is therefore unsurprising that throughout history many attempts have been undertaken to develop devices aiming at substituting for a missing visual capacity. In this review, we present two concepts that have been prevalent over the last two decades. The first concept is sensory substitution, which refers to the use of another sensory modality to perform a task that is normally primarily sub-served by the lost sense. The second concept is cross-modal plasticity, which occurs when loss of input in one sensory modality leads to reorganization in brain representation of other sensory modalities. Both phenomena are training-dependent. We also briefly describe the history of blindness from ancient times to modernity, and then proceed to address the means that have been used to help blind individuals, with an emphasis on modern technologies, invasive (various type of surgical implants) and non-invasive devices. With the advent of brain imaging, it has become possible to peer into the neural substrates of sensory substitution and highlight the magnitude of the plastic processes that lead to a rewired brain. Finally, we will address the important question of the value and practicality of the available technologies and future directions.
Collapse
Affiliation(s)
- Maurice Ptito
- École d’Optométrie, Université de Montréal, Montréal, QC, Canada
- Department of Nuclear Medicine, University of Southern Denmark, Odense, Denmark
- Department of Neuroscience, University of Copenhagen, Copenhagen, Denmark
| | - Maxime Bleau
- École d’Optométrie, Université de Montréal, Montréal, QC, Canada
| | - Ismaël Djerourou
- École d’Optométrie, Université de Montréal, Montréal, QC, Canada
| | - Samuel Paré
- École d’Optométrie, Université de Montréal, Montréal, QC, Canada
| | - Fabien C. Schneider
- TAPE EA7423 University of Lyon-Saint Etienne, Saint Etienne, France
- Neuroradiology Unit, University Hospital of Saint-Etienne, Saint-Etienne, France
| | - Daniel-Robert Chebat
- Visual and Cognitive Neuroscience Laboratory (VCN Lab), Department of Psychology, Faculty of Social Sciences and Humanities, Ariel University, Ariel, Israël
- Navigation and Accessibility Research Center of Ariel University (NARCA), Ariel, Israël
| |
Collapse
|
18
|
Rajguru C, Obrist M, Memoli G. Spatial Soundscapes and Virtual Worlds: Challenges and Opportunities. Front Psychol 2020; 11:569056. [PMID: 33262723 PMCID: PMC7685989 DOI: 10.3389/fpsyg.2020.569056] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/02/2020] [Accepted: 04/16/2020] [Indexed: 12/03/2022] Open
Abstract
There is increasing effort to characterize the soundscapes around us so that we can design more compelling and immersive experiences. This review paper focuses on the challenges and opportunities around sound perception, with a particular focus on spatial sound perception in a virtual reality (VR) cityscape. We review how research on temporal aspects has recently been extended to evaluating spatial factors when designing soundscapes. In particular, we discuss key findings on the human capability of localizing and distinguishing spatial sound cues for different technical setups. We highlight studies carried out in both real-world and virtual reality settings to evaluate spatial sound perception. We conclude this review by highlighting the opportunities offered by VR technology and the remaining open questions for virtual soundscape designers, especially with the advances in spatial sound stimulation.
Collapse
Affiliation(s)
| | | | - Gianluca Memoli
- School of Engineering and Informatics, University of Sussex, Falmer, United Kingdom
| |
Collapse
|
19
|
Courtois G, Grimaldi V, Lissek H, Estoppey P, Georganti E. Perception of Auditory Distance in Normal-Hearing and Moderate-to-Profound Hearing-Impaired Listeners. Trends Hear 2020; 23:2331216519887615. [PMID: 31774032 PMCID: PMC6887817 DOI: 10.1177/2331216519887615] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
The auditory system allows the estimation of the distance to sound-emitting objects using multiple spatial cues. In virtual acoustics over headphones, a prerequisite to render auditory distance impression is sound externalization, which denotes the perception of synthesized stimuli outside of the head. Prior studies have found that listeners with mild-to-moderate hearing loss are able to perceive auditory distance and are sensitive to externalization. However, this ability may be degraded by certain factors, such as non-linear amplification in hearing aids or the use of a remote wireless microphone. In this study, 10 normal-hearing and 20 moderate-to-profound hearing-impaired listeners were instructed to estimate the distance of stimuli processed with different methods yielding various perceived auditory distances in the vicinity of the listeners. Two different configurations of non-linear amplification were implemented, and a novel feature aiming to restore a sense of distance in wireless microphone systems was tested. The results showed that the hearing-impaired listeners, even those with a profound hearing loss, were able to discriminate nearby and far sounds that were equalized in level. Their perception of auditory distance was however more contracted than in normal-hearing listeners. Non-linear amplification was found to distort the original spatial cues, but no adverse effect on the ratings of auditory distance was evident. Finally, it was shown that the novel feature was successful in allowing the hearing-impaired participants to perceive externalized sounds with wireless microphone systems.
Collapse
Affiliation(s)
- Gilles Courtois
- Swiss Federal Institute of Technology (EPFL), Signal Processing Laboratory (LTS2), Lausanne, Switzerland.,Sonova AG, Stäfa, Switzerland
| | - Vincent Grimaldi
- Swiss Federal Institute of Technology (EPFL), Signal Processing Laboratory (LTS2), Lausanne, Switzerland
| | - Hervé Lissek
- Swiss Federal Institute of Technology (EPFL), Signal Processing Laboratory (LTS2), Lausanne, Switzerland
| | | | | |
Collapse
|
20
|
Wagenhäuser PA, Wiegrebe L, Baier AL. Biosonar spatial resolution along the distance axis: revisiting the clutter interference zone. J Exp Biol 2020; 223:jeb224311. [PMID: 32843361 DOI: 10.1242/jeb.224311] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2020] [Accepted: 08/18/2020] [Indexed: 11/20/2022]
Abstract
Many echolocating bats forage close to vegetation - a chaotic arrangement of prey and foliage where multiple targets are positioned behind one another. Bats excel at determining distance: they measure the delay between the outgoing call and the returning echo. In their auditory cortex, delay-sensitive neurons form a topographic map, suggesting that bats can resolve echoes of multiple targets along the distance axis - a skill crucial for the forage-amongst-foliage scenario. We tested this hypothesis combining an auditory virtual reality with formal psychophysics: we simulated a prey item embedded in two foliage elements, one in front of and one behind the prey. The simulated spacing between 'prey' (target) and 'foliage' (maskers) was defined by the inter-masker delay (IMD). We trained Phyllostomus discolor bats to detect the target in the presence of the maskers, systematically varying both loudness and spacing of the maskers. We show that target detection is impaired when maskers are closely spaced (IMD<1 ms), but remarkably improves when the spacing is increased: the release from masking is approximately 5 dB for intermediate IMDs (1-3 ms) and increases to over 15 dB for large IMDs (≥9 ms). These results are comparable to those from earlier work on the clutter interference zone of bats (Simmons et al., 1988). They suggest that prey would enjoy considerable acoustic protection from closely spaced foliage, but also that the range resolution of bats would let them 'peek into gaps'. Our study puts target ranging into a meaningful context and highlights the limitations of computational topographic maps.
Collapse
Affiliation(s)
- Peter A Wagenhäuser
- Department of Biology II, Ludwig Maximilians University Munich, Großhaderner Str. 2, 82152 Planegg-Martinsried, Germany
| | - Lutz Wiegrebe
- Department of Biology II, Ludwig Maximilians University Munich, Großhaderner Str. 2, 82152 Planegg-Martinsried, Germany
| | - A Leonie Baier
- Department of Biology II, Ludwig Maximilians University Munich, Großhaderner Str. 2, 82152 Planegg-Martinsried, Germany
- Chair of Zoology, Technical University Munich, Liesel-Beckmann-Str. 4, 85354 Freising-Weihenstephan, Germany
| |
Collapse
|
21
|
Abstract
Being able to pick out particular sounds, such as speech, against a background of other sounds represents one of the key tasks performed by the auditory system. Understanding how this happens is important because speech recognition in noise is particularly challenging for older listeners and for people with hearing impairments. Central to this ability is the capacity of neurons to adapt to the statistics of sounds reaching the ears, which helps to generate noise-tolerant representations of sounds in the brain. In more complex auditory scenes, such as a cocktail party — where the background noise comprises other voices, sound features associated with each source have to be grouped together and segregated from those belonging to other sources. This depends on precise temporal coding and modulation of cortical response properties when attending to a particular speaker in a multi-talker environment. Furthermore, the neural processing underlying auditory scene analysis is shaped by experience over multiple timescales.
Collapse
|
22
|
Prud'homme L, Lavandier M. Do we need two ears to perceive the distance of a virtual frontal sound source? THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 148:1614. [PMID: 33003836 DOI: 10.1121/10.0001954] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/27/2020] [Accepted: 08/26/2020] [Indexed: 06/11/2023]
Abstract
The present study investigated whether the perception of virtual auditory distance is binaural, monaural, or both. Listeners evaluated the distance of a frontal source of pink noise simulated in a room via headphones. Experiment 1 was performed with eyes closed in a soundproof booth. Experiment 2 was performed with eyes open in the room used to create the stimuli. Individualized and non-individualized stimuli were compared. Different conditions for controlling sound level were tested. The amount of binaural information in the stimuli was varied by mixing the left and right ear signals in different proportions. Results showed that the use of non-individualized stimuli did not impair distance perception. Binaural information was not used by naive listeners to evaluate distance, both with and without visual information available. However, for some listeners, a complete absence of binaural information could disrupt distance evaluation with headphones. Sound level was a dominant cue used by listeners to judge for distance, and some listeners could also reliably use reverberation-related changes in spectral content. In the absence of specific training, artificial manipulation of sound level greatly altered distance judgments.
Collapse
Affiliation(s)
- Luna Prud'homme
- Univ. Lyon, ENTPE, Laboratoire Génie Civil et Bâtiment, Rue M. Audin, Vaulx-en-Velin Cedex, 69518, France
| | - Mathieu Lavandier
- Univ. Lyon, ENTPE, Laboratoire Génie Civil et Bâtiment, Rue M. Audin, Vaulx-en-Velin Cedex, 69518, France
| |
Collapse
|
23
|
Yamasaki D, Ashida H. Size-Distance Scaling With Absolute and Relative Auditory Distance Information. Multisens Res 2020; 33:109-126. [PMID: 31648194 DOI: 10.1163/22134808-20191467] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2019] [Accepted: 08/10/2019] [Indexed: 11/19/2022]
Abstract
In the dynamic 3D space, it is critical for survival to perceive size of an object and rescale it with distance from an observer. Humans can perceive distance via not only vision but also audition, which plays an important role in the localization of objects, especially in visually ambiguous environments. However, whether and how auditory distance information contributes to visual size perception is not well understood. To address this issue, we investigated the efficiency of size-distance scaling by using auditory distance information that was conveyed by binaurally recorded auditory stimuli. We examined the effects of absolute distance information of a single sound sequence (Experiment 1) and relative distance information between two sound sequences (Experiment 2) on visual size estimation performances in darkened and well-lit environments. We demonstrated that humans could perform size-distance disambiguation by using auditory distance information even in darkness. Curiously, relative distance information was more efficient in size-distance scaling than absolute distance information, suggesting a high reliance on relative auditory distance information in our visual spatial experiences. The results highlight a benefit of audiovisual interaction for size-distance processing and calibration of external events under visually degraded situations.
Collapse
Affiliation(s)
- Daiki Yamasaki
- 1Graduate School of Letters, Kyoto University, Japan.,2Japan Society for the Promotion of Science, Japan
| | | |
Collapse
|
24
|
Badajoz-Davila J, Buchholz JM, Van-Hoesel R. Effect of noise and reverberation on speech intelligibility for cochlear implant recipients in realistic sound environments. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:3538. [PMID: 32486825 DOI: 10.1121/10.0001259] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/09/2019] [Accepted: 04/28/2020] [Indexed: 06/11/2023]
Abstract
Previous studies have suggested a strong effect of reverberation on speech intelligibility (SI) in cochlear implant (CI) recipients. In many of them, different reverberation conditions were obtained by altering the acoustic absorption of a single room, thereby omitting the effect of the room volume. In addition, studies that have investigated the combined effects of reverberation and noise on SI have overlooked the effect of reverberation on the modulation of the noise. In the present study, SI was measured unilaterally in 12 CI recipients in quiet and in noise using a three-dimensional loudspeaker array. Target speech was convolved with room impulse responses (RIRs) recorded at three talker-to-listener distances in five physical rooms with distinct reverberation times. Noise consisted of four two-talker dialogues convolved with RIRs measured at fixed positions around the listener. Results in quiet suggest that a significant drop in SI occurs mainly at long talker-to-listener distances, and small reverberant rooms affect SI the most. In noise, the most detrimental type of noise is anechoic as it is the most modulated. Overall, the results suggest that at fixed signal-to-noise ratios the effects of noise and reverberation are smallest at short distances in large rooms or in small rooms with some reverberation.
Collapse
Affiliation(s)
- Javier Badajoz-Davila
- Department of Linguistics, Audiology Section, Macquarie University, Australian Hearing Hub, 16 University Avenue, New South Wales 2109, Australia
| | - Jörg M Buchholz
- Department of Linguistics, Audiology Section, Macquarie University, Australian Hearing Hub, 16 University Avenue, New South Wales 2109, Australia
| | - Richard Van-Hoesel
- The Hearing Cooperative Research Centre (CRC), 550 Swanston Street, Carlton, Victoria 3053, Australia
| |
Collapse
|
25
|
Kolarik AJ, Raman R, Moore BCJ, Cirstea S, Gopalakrishnan S, Pardhan S. The accuracy of auditory spatial judgments in the visually impaired is dependent on sound source distance. Sci Rep 2020; 10:7169. [PMID: 32346036 PMCID: PMC7189236 DOI: 10.1038/s41598-020-64306-8] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2019] [Accepted: 04/13/2020] [Indexed: 11/09/2022] Open
Abstract
Blindness leads to substantial enhancements in many auditory abilities, and deficits in others. It is unknown how severe visual losses need to be before changes in auditory abilities occur, or whether the relationship between severity of visual loss and changes in auditory abilities is proportional and systematic. Here we show that greater severity of visual loss is associated with increased auditory judgments of distance and room size. On average participants with severe visual losses perceived sounds to be twice as far away, and rooms to be three times larger, than sighted controls. Distance estimates for sighted controls were most accurate for closer sounds and least accurate for farther sounds. As the severity of visual impairment increased, accuracy decreased for closer sounds and increased for farther sounds. However, it is for closer sounds that accurate judgments are needed to guide rapid motor responses to auditory events, e.g. planning a safe path through a busy street to avoid collisions with other people, and falls. Interestingly, greater visual impairment severity was associated with more accurate room size estimates. The results support a new hypothesis that crossmodal calibration of audition by vision depends on the severity of visual loss.
Collapse
Affiliation(s)
- Andrew J Kolarik
- Vision and Eye Research Institute, School of Medicine, Anglia Ruskin University, Cambridge, United Kingdom. .,Department of Psychology, University of Cambridge, Cambridge, United Kingdom.
| | - Rajiv Raman
- Vision and Eye Research Institute, School of Medicine, Anglia Ruskin University, Cambridge, United Kingdom.,Shri Bhagwan Mahavir Vitreoretinal Services, Sankara Nethralaya Eye Hospital, Chennai, India
| | - Brian C J Moore
- Vision and Eye Research Institute, School of Medicine, Anglia Ruskin University, Cambridge, United Kingdom.,Department of Psychology, University of Cambridge, Cambridge, United Kingdom
| | - Silvia Cirstea
- Vision and Eye Research Institute, School of Medicine, Anglia Ruskin University, Cambridge, United Kingdom.,School of Computing and Information Science, Anglia Ruskin University, Cambridge, United Kingdom
| | - Sarika Gopalakrishnan
- Faculty of Low Vision Care, Elite School of Optometry, Chennai, India.,Low Vision Care Department, Sankara Nethralaya Eye Hospital, Chennai, India
| | - Shahina Pardhan
- Vision and Eye Research Institute, School of Medicine, Anglia Ruskin University, Cambridge, United Kingdom
| |
Collapse
|
26
|
Kopco N, Doreswamy KK, Huang S, Rossi S, Ahveninen J. Cortical auditory distance representation based on direct-to-reverberant energy ratio. Neuroimage 2020; 208:116436. [PMID: 31809885 PMCID: PMC6997045 DOI: 10.1016/j.neuroimage.2019.116436] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/03/2019] [Revised: 11/30/2019] [Accepted: 12/02/2019] [Indexed: 11/26/2022] Open
Abstract
Auditory distance perception and its neuronal mechanisms are poorly understood, mainly because 1) it is difficult to separate distance processing from intensity processing, 2) multiple intensity-independent distance cues are often available, and 3) the cues are combined in a context-dependent way. A recent fMRI study identified human auditory cortical area representing intensity-independent distance for sources presented along the interaural axis (Kopco et al. PNAS, 109, 11019-11024). For these sources, two intensity-independent cues are available, interaural level difference (ILD) and direct-to-reverberant energy ratio (DRR). Thus, the observed activations may have been contributed by not only distance-related, but also direction-encoding neuron populations sensitive to ILD. Here, the paradigm from the previous study was used to examine DRR-based distance representation for sounds originating in front of the listener, where ILD is not available. In a virtual environment, we performed behavioral and fMRI experiments, combined with computational analyses to identify the neural representation of distance based on DRR. The stimuli varied in distance (15-100 cm) while their received intensity was varied randomly and independently of distance. Behavioral performance showed that intensity-independent distance discrimination is accurate for frontal stimuli, even though it is worse than for lateral stimuli. fMRI activations for sounds varying in frontal distance, as compared to varying only in intensity, increased bilaterally in the posterior banks of Heschl's gyri, the planum temporale, and posterior superior temporal gyrus regions. Taken together, these results suggest that posterior human auditory cortex areas contain neuron populations that are sensitive to distance independent of intensity and of binaural cues relevant for directional hearing.
Collapse
Affiliation(s)
- Norbert Kopco
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Harvard Medical School/Massachusetts General Hospital, Charlestown, MA, 02129, USA; Institute of Computer Science, P. J. Šafárik University, Košice, 04001, Slovakia; Hearing Research Center, Boston University, Boston, MA, 02215, USA.
| | - Keerthi Kumar Doreswamy
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Harvard Medical School/Massachusetts General Hospital, Charlestown, MA, 02129, USA; Institute of Computer Science, P. J. Šafárik University, Košice, 04001, Slovakia
| | - Samantha Huang
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Harvard Medical School/Massachusetts General Hospital, Charlestown, MA, 02129, USA
| | - Stephanie Rossi
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Harvard Medical School/Massachusetts General Hospital, Charlestown, MA, 02129, USA
| | - Jyrki Ahveninen
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Harvard Medical School/Massachusetts General Hospital, Charlestown, MA, 02129, USA
| |
Collapse
|
27
|
Devos P, Aletta F, Thomas P, Petrovic M, Vander Mynsbrugge T, Van de Velde D, De Vriendt P, Botteldooren D. Designing Supportive Soundscapes for Nursing Home Residents with Dementia. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2019; 16:ijerph16244904. [PMID: 31817300 PMCID: PMC6950055 DOI: 10.3390/ijerph16244904] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/31/2019] [Revised: 11/22/2019] [Accepted: 11/28/2019] [Indexed: 12/12/2022]
Abstract
Sound and its resulting soundscape is a major appraisal component of the living environment. Where environmental sounds (e.g., outdoor traffic sounds) are often perceived as negative, a soundscape (e.g., containing natural sounds) can also have a positive effect on health and well-being. This supportive effect of a soundscape is getting increasing attention for use in practice. This paper addresses the design of a supportive sonic environment for persons with dementia in nursing homes. Starting from a review of key mechanisms related to sonic perception, cognitive deficits and related behavior, a framework is derived for the composition of a sonic environment for persons with dementia. The proposed framework is centered around using acoustic stimuli for influencing mood, stimulating the feeling of safety and triggering a response in a person. These stimuli are intended to be deployed as added sounds in a nursing home to improve the well-being and behavior of the residents.
Collapse
Affiliation(s)
- Paul Devos
- Department of Information Technology, Ghent University, 9052 Ghent, Belgium; (F.A.); (P.T.); (D.B.)
- Correspondence:
| | - Francesco Aletta
- Department of Information Technology, Ghent University, 9052 Ghent, Belgium; (F.A.); (P.T.); (D.B.)
- Institute for Environmental Design and Engineering, University College London, London WC1H0NN, UK
| | - Pieter Thomas
- Department of Information Technology, Ghent University, 9052 Ghent, Belgium; (F.A.); (P.T.); (D.B.)
| | - Mirko Petrovic
- Department of Internal Medicine and Paediatrics, Ghent University, 9000 Ghent, Belgium;
| | - Tara Vander Mynsbrugge
- Department of Occupational Therapy, Artevelde University College, 9000 Ghent, Belgium; (T.V.M.); (D.V.d.V.); (P.D.V.)
| | - Dominique Van de Velde
- Department of Occupational Therapy, Artevelde University College, 9000 Ghent, Belgium; (T.V.M.); (D.V.d.V.); (P.D.V.)
- Department of Occupational Therapy, Ghent University, 9000 Ghent, Belgium
| | - Patricia De Vriendt
- Department of Occupational Therapy, Artevelde University College, 9000 Ghent, Belgium; (T.V.M.); (D.V.d.V.); (P.D.V.)
- Department of Occupational Therapy, Ghent University, 9000 Ghent, Belgium
| | - Dick Botteldooren
- Department of Information Technology, Ghent University, 9052 Ghent, Belgium; (F.A.); (P.T.); (D.B.)
| |
Collapse
|
28
|
Zhang YD, Liu W. A study of auditory localization mechanism based on thought experiments. Phys Life Rev 2019; 31:206-213. [DOI: 10.1016/j.plrev.2019.01.005] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2018] [Revised: 06/09/2018] [Accepted: 01/15/2019] [Indexed: 10/27/2022]
|
29
|
Kaplanis N, Bech S, Lokki T, van Waterschoot T, Holdt Jensen S. Perception and preference of reverberation in small listening rooms for multi-loudspeaker reproduction. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:3562. [PMID: 31795724 DOI: 10.1121/1.5135582] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/03/2019] [Accepted: 11/06/2019] [Indexed: 06/10/2023]
Abstract
An experiment was conducted to identify the perceptual effects of acoustical properties of domestic listening environments, in a stereophonic reproduction scenario. Nine sound fields, originating from four rooms, were captured and spatially reproduced over a three-dimensional loudspeaker array. A panel of ten expert assessors identified and quantified the perceived differences of those sound fields using their own perceptual attributes. A multivariate analysis revealed two principal dimensions that could summarize the sound fields of this investigation. Four perceptual constructs seem to characterize the sensory properties of these dimensions, relating to Reverberance, Width & Envelopment, Proximity, and Bass. Overall, the results signify the importance of reverberation in residential listening environments on the perceived sensory experience, and as a consequence, the assessors' preferences towards certain decay times.
Collapse
Affiliation(s)
| | - Søren Bech
- Bang & Olufsen a/s, Bang & Olufsen Allé 1, Struer, DK-7600, Denmark
| | - Tapio Lokki
- Department of Computer Science, Aalto University, P.O. Box 13000, FI-00076 Aalto, Finland
| | - Toon van Waterschoot
- Department of Electrical Engineering (ESAT-STADIUS/ETC), KU Leuven, Kasteelpark Arenberg 10, 3001 Leuven, Belgium
| | - Søren Holdt Jensen
- Department of Electronic Systems, Aalborg University, 9220 Aalborg, Denmark
| |
Collapse
|
30
|
Yin TC, Smith PH, Joris PX. Neural Mechanisms of Binaural Processing in the Auditory Brainstem. Compr Physiol 2019; 9:1503-1575. [DOI: 10.1002/cphy.c180036] [Citation(s) in RCA: 26] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
|
31
|
3D Tune-In Toolkit: An open-source library for real-time binaural spatialisation. PLoS One 2019; 14:e0211899. [PMID: 30856198 PMCID: PMC6411112 DOI: 10.1371/journal.pone.0211899] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2018] [Accepted: 01/22/2019] [Indexed: 12/05/2022] Open
Abstract
The 3D Tune-In Toolkit (3DTI Toolkit) is an open-source standard C++ library which includes a binaural spatialiser. This paper presents the technical details of this renderer, outlining its architecture and describing the processes implemented in each of its components. In order to put this description into context, the basic concepts behind binaural spatialisation are reviewed through a chronology of research milestones in the field in the last 40 years. The 3DTI Toolkit renders the anechoic signal path by convolving sound sources with Head Related Impulse Responses (HRIRs), obtained by interpolating those extracted from a set that can be loaded from any file in a standard audio format. Interaural time differences are managed separately, in order to be able to customise the rendering according the head size of the listener, and to reduce comb-filtering when interpolating between different HRIRs. In addition, geometrical and frequency-dependent corrections for simulating near-field sources are included. Reverberation is computed separately using a virtual loudspeakers Ambisonic approach and convolution with Binaural Room Impulse Responses (BRIRs). In all these processes, special care has been put in avoiding audible artefacts produced by changes in gains and audio filters due to the movements of sources and of the listener. The 3DTI Toolkit performance, as well as some other relevant metrics such as non-linear distortion, are assessed and presented, followed by a comparison between the features offered by the 3DTI Toolkit and those found in other currently available open- and closed-source binaural renderers.
Collapse
|
32
|
Personalized HRTF Modeling Based on Deep Neural Network Using Anthropometric Measurements and Images of the Ear. APPLIED SCIENCES-BASEL 2018. [DOI: 10.3390/app8112180] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
This paper proposes a personalized head-related transfer function (HRTF) estimation method based on deep neural networks by using anthropometric measurements and ear images. The proposed method consists of three sub-networks for representing personalized features and estimating the HRTF. As input features for neural networks, the anthropometric measurements regarding the head and torso are used for a feedforward deep neural network (DNN), and the ear images are used for a convolutional neural network (CNN). After that, the outputs of these two sub-networks are merged into another DNN for estimation of the personalized HRTF. To evaluate the performance of the proposed method, objective and subjective evaluations are conducted. For the objective evaluation, the root mean square error (RMSE) and the log spectral distance (LSD) between the reference HRTF and the estimated one are measured. Consequently, the proposed method provides the RMSE of −18.40 dB and LSD of 4.47 dB, which are lower by 0.02 dB and higher by 0.85 dB than the DNN-based method using anthropometric data without pinna measurements, respectively. Next, a sound localization test is performed for the subjective evaluation. As a result, it is shown that the proposed method can localize sound sources with higher accuracy of around 11% and 6% than the average HRTF method and DNN-based method, respectively. In addition, the reductions of the front/back confusion rate by 12.5% and 2.5% are achieved by the proposed method, compared to the average HRTF method and DNN-based method, respectively.
Collapse
|
33
|
Lundbeck M, Hartog L, Grimm G, Hohmann V, Bramsløw L, Neher T. Influence of Multi-microphone Signal Enhancement Algorithms on the Acoustics and Detectability of Angular and Radial Source Movements. Trends Hear 2018; 22:2331216518779719. [PMID: 29900799 PMCID: PMC6024528 DOI: 10.1177/2331216518779719] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022] Open
Abstract
Hearing-impaired listeners are known to have difficulties not only with understanding speech in noise but also with judging source distance and movement, and these deficits are related to perceived handicap. It is possible that the perception of spatially dynamic sounds can be improved with hearing aids (HAs), but so far this has not been investigated. In a previous study, older hearing-impaired listeners showed poorer detectability for virtual left-right (angular) and near-far (radial) source movements due to lateral interfering sounds and reverberation, respectively. In the current study, potential ways of improving these deficits with HAs were explored. Using stimuli very similar to before, detailed acoustic analyses were carried out to examine the influence of different HA algorithms for suppressing noise and reverberation on the acoustic cues previously shown to be associated with source movement detectability. For an algorithm that combined unilateral directional microphones with binaural coherence-based noise reduction and for a bilateral beamformer with binaural cue preservation, movement-induced changes in spectral coloration, signal-to-noise ratio, and direct-to-reverberant energy ratio were greater compared with no HA processing. To evaluate these two algorithms perceptually, aided measurements of angular and radial source movement detectability were performed with 20 older hearing-impaired listeners. The analyses showed that, in the presence of concurrent interfering sounds and reverberation, the bilateral beamformer could restore source movement detectability in both spatial dimensions, whereas the other algorithm only improved detectability in the near-far dimension. Together, these results provide a basis for improving the detectability of spatially dynamic sounds with HAs.
Collapse
Affiliation(s)
- Micha Lundbeck
- 1 Medizinische Physik and Cluster of Excellence "Hearing4all", Oldenburg University, Germany.,2 HörTech gGmbH, Oldenburg, Germany
| | - Laura Hartog
- 1 Medizinische Physik and Cluster of Excellence "Hearing4all", Oldenburg University, Germany.,2 HörTech gGmbH, Oldenburg, Germany
| | - Giso Grimm
- 1 Medizinische Physik and Cluster of Excellence "Hearing4all", Oldenburg University, Germany.,2 HörTech gGmbH, Oldenburg, Germany
| | - Volker Hohmann
- 1 Medizinische Physik and Cluster of Excellence "Hearing4all", Oldenburg University, Germany.,2 HörTech gGmbH, Oldenburg, Germany
| | - Lars Bramsløw
- 3 Eriksholm Research Centre, Oticon A/S, Snekkersten, Denmark
| | - Tobias Neher
- 1 Medizinische Physik and Cluster of Excellence "Hearing4all", Oldenburg University, Germany.,4 Institute of Clinical Research, University of Southern Denmark, Odense, Denmark
| |
Collapse
|
34
|
Kates JM, Arehart KH, Muralimanohar RK, Sommerfeldt K. Externalization of remote microphone signals using a structural binaural model of the head and pinna. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 143:2666. [PMID: 29857749 DOI: 10.1121/1.5032326] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
In a remote microphone (RM) system, a talker speaks into a microphone and the signal is transmitted to the hearing aids worn by the hearing-impaired listener. A difficulty with remote microphones, however, is that the signal received at the hearing aid bypasses the head and pinna, so the acoustic cues needed to externalize the sound source are missing. The objective of this paper is to process the RM signal to improve externalization when listening through earphones. The processing is based on a structural binaural model, which uses a cascade of processing modules to simulate the interaural level difference, interaural time difference, pinna reflections, ear-canal resonance, and early room reflections. The externalization results for the structural binaural model are compared to a left-right signal blend, the listener's own anechoic head-related impulse response (HRIR), and the listener's own HRIR with room reverberation. The azimuth is varied from straight ahead to 90° to one side. The results show that the structural binaural model is as effective as the listener's own HRIR plus reverberation in producing an externalized acoustic image, and that there is no significant difference in externalization between hearing-impaired and normal-hearing listeners.
Collapse
Affiliation(s)
- James M Kates
- Department of Speech, Language, and Hearing Sciences, University of Colorado, Boulder, Colorado 80309, USA
| | - Kathryn H Arehart
- Department of Speech, Language, and Hearing Sciences, University of Colorado, Boulder, Colorado 80309, USA
| | - Ramesh Kumar Muralimanohar
- Department of Speech, Language, and Hearing Sciences, University of Colorado, Boulder, Colorado 80309, USA
| | - Kristin Sommerfeldt
- Department of Speech, Language, and Hearing Sciences, University of Colorado, Boulder, Colorado 80309, USA
| |
Collapse
|
35
|
Abstract
Distance is important: From an ecological perspective, knowledge about the distance to either prey or predator is vital. However, the distance of an unknown sound source is particularly difficult to assess, especially in anechoic environments. In vision, changes in perspective resulting from observer motion produce a reliable, consistent, and unambiguous impression of depth known as motion parallax. Here we demonstrate with formal psychophysics that humans can exploit auditory motion parallax, i.e., the change in the dynamic binaural cues elicited by self-motion, to assess the relative depths of two sound sources. Our data show that sensitivity to relative depth is best when subjects move actively; performance deteriorates when subjects are moved by a motion platform or when the sound sources themselves move. This is true even though the dynamic binaural cues elicited by these three types of motion are identical. Our data demonstrate a perceptual strategy to segregate intermittent sound sources in depth and highlight the tight interaction between self-motion and binaural processing that allows assessment of the spatial layout of complex acoustic scenes.
Collapse
|
36
|
Shestopalova LB, Petropavlovskaia EA, Semenova VV, Nikitin NI. Mismatch negativity and psychophysical detection of rising and falling intensity sounds. Biol Psychol 2018; 133:99-111. [PMID: 29421188 DOI: 10.1016/j.biopsycho.2018.01.018] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2017] [Revised: 01/26/2018] [Accepted: 01/31/2018] [Indexed: 10/18/2022]
Abstract
Human subjects demonstrate a perceptual priority for rising level sounds compared with falling level sounds. The aim of the present study was to investigate whether or not the perceptual preference for rising intensity can be found in the preattentive processing indexed by mismatch negativity (MMN). Reversed oddball stimulation was used to produce MMNs and to test the behavioral discrimination of rising, falling and constant level sounds. Three types of stimuli served as standards or deviants in different blocks: constant level sounds and two kinds of rising/falling sounds with gradual or stepwise change of intensity. The MMN amplitudes were calculated by subtracting ERPs to identical stimuli presented as standard in one block and deviant in another block. Both rising and falling level deviants elicited MMNs which peaked after 250 ms and did not overlap with N1 waves. MMN was elicited by level changes even when the deviants were not discriminated behaviorally. Most importantly, we found dissociation between earlier and later stages of auditory processing: the MMN responses to the level changes were mostly affected by the direction of deviance (increment or decrement) in the sequence, whereas behavioral performance depended on the direction of the level change within the stimuli (rising or falling).
Collapse
Affiliation(s)
- Lidia B Shestopalova
- I.P. Pavlov Institute of Physiology, Russian Academy of Sciences, Saint-Petersburg, Russia.
| | | | - Varvara V Semenova
- I.P. Pavlov Institute of Physiology, Russian Academy of Sciences, Saint-Petersburg, Russia
| | - Nikolai I Nikitin
- I.P. Pavlov Institute of Physiology, Russian Academy of Sciences, Saint-Petersburg, Russia
| |
Collapse
|
37
|
Spence C, Lee J, Van der Stoep N. Responding to sounds from unseen locations: crossmodal attentional orienting in response to sounds presented from the rear. Eur J Neurosci 2017; 51:1137-1150. [PMID: 28973789 DOI: 10.1111/ejn.13733] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2017] [Revised: 09/27/2017] [Accepted: 09/27/2017] [Indexed: 11/28/2022]
Abstract
To date, most of the research on spatial attention has focused on probing people's responses to stimuli presented in frontal space. That is, few researchers have attempted to assess what happens in the space that is currently unseen (essentially rear space). In a sense, then, 'out of sight' is, very much, 'out of mind'. In this review, we highlight what is presently known about the perception and processing of sensory stimuli (focusing on sounds) whose source is not currently visible. We briefly summarize known differences in the localizability of sounds presented from different locations in 3D space, and discuss the consequences for the crossmodal attentional and multisensory perceptual interactions taking place in various regions of space. The latest research now clearly shows that the kinds of crossmodal interactions that take place in rear space are very often different in kind from those that have been documented in frontal space. Developing a better understanding of how people respond to unseen sound sources in naturalistic environments by integrating findings emerging from multiple fields of research will likely lead to the design of better warning signals in the future. This review highlights the need for neuroscientists interested in spatial attention to spend more time researching what happens (in terms of the covert and overt crossmodal orienting of attention) in rear space.
Collapse
Affiliation(s)
- Charles Spence
- Crossmodal Research Laboratory, Department of Experimental Psychology, Oxford University, Oxford, OX1 3UD, UK
| | - Jae Lee
- Crossmodal Research Laboratory, Department of Experimental Psychology, Oxford University, Oxford, OX1 3UD, UK
| | - Nathan Van der Stoep
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands
| |
Collapse
|
38
|
Direct-location versus verbal report methods for measuring auditory distance perception in the far field. Behav Res Methods 2017; 50:1234-1247. [PMID: 28786043 DOI: 10.3758/s13428-017-0939-x] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
In this study we evaluated whether a method of direct location is an appropriate response method for measuring auditory distance perception of far-field sound sources. We designed an experimental set-up that allows participants to indicate the distance at which they perceive the sound source by moving a visual marker. We termed this method Cross-Modal Direct Location (CMDL) since the response procedure involves the visual modality while the stimulus is presented through the auditory modality. Three experiments were conducted with sound sources located from 1 to 6 m. The first one compared the perceived distances obtained using either the CMDL device or verbal report (VR), which is the response method more frequently used for reporting auditory distance in the far field, and found differences on response compression and bias. In Experiment 2, participants reported visual distance estimates to the visual marker that were found highly accurate. Then, we asked the same group of participants to report VR estimates of auditory distance and found that the spatial visual information, obtained from the previous task, did not influence their reports. Finally, Experiment 3 compared the same responses that Experiment 1 but interleaving the methods, showing a weak, but complex, mutual influence. However, the estimates obtained with each method remained statistically different. Our results show that the auditory distance psychophysical functions obtained with the CMDL method are less susceptible to previously reported underestimation for distances over 2 m.
Collapse
|
39
|
Spiousas I, Etchemendy PE, Eguia MC, Calcagno ER, Abregú E, Vergara RO. Sound Spectrum Influences Auditory Distance Perception of Sound Sources Located in a Room Environment. Front Psychol 2017; 8:969. [PMID: 28690556 PMCID: PMC5479918 DOI: 10.3389/fpsyg.2017.00969] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2016] [Accepted: 05/26/2017] [Indexed: 12/03/2022] Open
Abstract
Previous studies on the effect of spectral content on auditory distance perception (ADP) focused on the physically measurable cues occurring either in the near field (low-pass filtering due to head diffraction) or when the sound travels distances >15 m (high-frequency energy losses due to air absorption). Here, we study how the spectrum of a sound arriving from a source located in a reverberant room at intermediate distances (1–6 m) influences the perception of the distance to the source. First, we conducted an ADP experiment using pure tones (the simplest possible spectrum) of frequencies 0.5, 1, 2, and 4 kHz. Then, we performed a second ADP experiment with stimuli consisting of continuous broadband and bandpass-filtered (with center frequencies of 0.5, 1.5, and 4 kHz and bandwidths of 1/12, 1/3, and 1.5 octave) pink-noise clips. Our results showed an effect of the stimulus frequency on the perceived distance both for pure tones and filtered noise bands: ADP was less accurate for stimuli containing energy only in the low-frequency range. Analysis of the frequency response of the room showed that the low accuracy observed for low-frequency stimuli can be explained by the presence of sparse modal resonances in the low-frequency region of the spectrum, which induced a non-monotonic relationship between binaural intensity and source distance. The results obtained in the second experiment suggest that ADP can also be affected by stimulus bandwidth but in a less straightforward way (i.e., depending on the center frequency, increasing stimulus bandwidth could have different effects). Finally, the analysis of the acoustical cues suggests that listeners judged source distance using mainly changes in the overall intensity of the auditory stimulus with distance rather than the direct-to-reverberant energy ratio, even for low-frequency noise bands (which typically induce high amount of reverberation). The results obtained in this study show that, depending on the spectrum of the auditory stimulus, reverberation can degrade ADP rather than improve it.
Collapse
Affiliation(s)
- Ignacio Spiousas
- Laboratorio de Dinámica Sensomotora, Departamento de Ciencia y Tecnología, CONICET, Universidad Nacional de QuilmesBernal, Argentina
| | - Pablo E Etchemendy
- Laboratorio de Acústica y Percepción Sonora, Escuela Universitaria de Artes, CONICET, Universidad Nacional de QuilmesBernal, Argentina
| | - Manuel C Eguia
- Laboratorio de Acústica y Percepción Sonora, Escuela Universitaria de Artes, CONICET, Universidad Nacional de QuilmesBernal, Argentina
| | - Esteban R Calcagno
- Laboratorio de Acústica y Percepción Sonora, Escuela Universitaria de Artes, CONICET, Universidad Nacional de QuilmesBernal, Argentina
| | - Ezequiel Abregú
- Laboratorio de Acústica y Percepción Sonora, Escuela Universitaria de Artes, CONICET, Universidad Nacional de QuilmesBernal, Argentina
| | - Ramiro O Vergara
- Laboratorio de Acústica y Percepción Sonora, Escuela Universitaria de Artes, CONICET, Universidad Nacional de QuilmesBernal, Argentina
| |
Collapse
|
40
|
Hearing Scenes: A Neuromagnetic Signature of Auditory Source and Reverberant Space Separation. eNeuro 2017; 4:eN-NWR-0007-17. [PMID: 28451630 PMCID: PMC5394928 DOI: 10.1523/eneuro.0007-17.2017] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2016] [Revised: 02/03/2017] [Accepted: 02/06/2017] [Indexed: 11/21/2022] Open
Abstract
Perceiving the geometry of surrounding space is a multisensory process, crucial to contextualizing object perception and guiding navigation behavior. Humans can make judgments about surrounding spaces from reverberation cues, caused by sounds reflecting off multiple interior surfaces. However, it remains unclear how the brain represents reverberant spaces separately from sound sources. Here, we report separable neural signatures of auditory space and source perception during magnetoencephalography (MEG) recording as subjects listened to brief sounds convolved with monaural room impulse responses (RIRs). The decoding signature of sound sources began at 57 ms after stimulus onset and peaked at 130 ms, while space decoding started at 138 ms and peaked at 386 ms. Importantly, these neuromagnetic responses were readily dissociable in form and time: while sound source decoding exhibited an early and transient response, the neural signature of space was sustained and independent of the original source that produced it. The reverberant space response was robust to variations in sound source, and vice versa, indicating a generalized response not tied to specific source-space combinations. These results provide the first neuromagnetic evidence for robust, dissociable auditory source and reverberant space representations in the human brain and reveal the temporal dynamics of how auditory scene analysis extracts percepts from complex naturalistic auditory signals.
Collapse
|
41
|
Human Exploration of Enclosed Spaces through Echolocation. J Neurosci 2017; 37:1614-1627. [PMID: 28073936 DOI: 10.1523/jneurosci.1566-12.2016] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2012] [Revised: 11/23/2016] [Accepted: 12/01/2016] [Indexed: 11/21/2022] Open
Abstract
Some blind humans have developed echolocation, as a method of navigation in space. Echolocation is a truly active sense because subjects analyze echoes of dedicated, self-generated sounds to assess space around them. Using a special virtual space technique, we assess how humans perceive enclosed spaces through echolocation, thereby revealing the interplay between sensory and vocal-motor neural activity while humans perform this task. Sighted subjects were trained to detect small changes in virtual-room size analyzing real-time generated echoes of their vocalizations. Individual differences in performance were related to the type and number of vocalizations produced. We then asked subjects to estimate virtual-room size with either active or passive sounds while measuring their brain activity with fMRI. Subjects were better at estimating room size when actively vocalizing. This was reflected in the hemodynamic activity of vocal-motor cortices, even after individual motor and sensory components were removed. Activity in these areas also varied with perceived room size, although the vocal-motor output was unchanged. In addition, thalamic and auditory-midbrain activity was correlated with perceived room size; a likely result of top-down auditory pathways for human echolocation, comparable with those described in echolocating bats. Our data provide evidence that human echolocation is supported by active sensing, both behaviorally and in terms of brain activity. The neural sensory-motor coupling complements the fundamental acoustic motor-sensory coupling via the environment in echolocation.SIGNIFICANCE STATEMENT Passive listening is the predominant method for examining brain activity during echolocation, the auditory analysis of self-generated sounds. We show that sighted humans perform better when they actively vocalize than during passive listening. Correspondingly, vocal motor and cerebellar activity is greater during active echolocation than vocalization alone. Motor and subcortical auditory brain activity covaries with the auditory percept, although motor output is unchanged. Our results reveal behaviorally relevant neural sensory-motor coupling during echolocation.
Collapse
|
42
|
Lundbeck M, Grimm G, Hohmann V, Laugesen S, Neher T. Sensitivity to Angular and Radial Source Movements as a Function of Acoustic Complexity in Normal and Impaired Hearing. Trends Hear 2017; 21:2331216517717152. [PMID: 28675088 PMCID: PMC5548306 DOI: 10.1177/2331216517717152] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2016] [Revised: 05/16/2017] [Accepted: 05/23/2017] [Indexed: 11/15/2022] Open
Abstract
In contrast to static sounds, spatially dynamic sounds have received little attention in psychoacoustic research so far. This holds true especially for acoustically complex (reverberant, multisource) conditions and impaired hearing. The current study therefore investigated the influence of reverberation and the number of concurrent sound sources on source movement detection in young normal-hearing (YNH) and elderly hearing-impaired (EHI) listeners. A listening environment based on natural environmental sounds was simulated using virtual acoustics and rendered over headphones. Both near-far ('radial') and left-right ('angular') movements of a frontal target source were considered. The acoustic complexity was varied by adding static lateral distractor sound sources as well as reverberation. Acoustic analyses confirmed the expected changes in stimulus features that are thought to underlie radial and angular source movements under anechoic conditions and suggested a special role of monaural spectral changes under reverberant conditions. Analyses of the detection thresholds showed that, with the exception of the single-source scenarios, the EHI group was less sensitive to source movements than the YNH group, despite adequate stimulus audibility. Adding static sound sources clearly impaired the detectability of angular source movements for the EHI (but not the YNH) group. Reverberation, on the other hand, clearly impaired radial source movement detection for the EHI (but not the YNH) listeners. These results illustrate the feasibility of studying factors related to auditory movement perception with the help of the developed test setup.
Collapse
Affiliation(s)
- Micha Lundbeck
- Medizinische Physik and Cluster of Excellence ‘Hearing4all,’ Department of Medical Physics and Acoustics, Oldenburg University, Germany
- HörTech gGmbH, Oldenburg, Germany
| | - Giso Grimm
- Medizinische Physik and Cluster of Excellence ‘Hearing4all,’ Department of Medical Physics and Acoustics, Oldenburg University, Germany
- HörTech gGmbH, Oldenburg, Germany
| | - Volker Hohmann
- Medizinische Physik and Cluster of Excellence ‘Hearing4all,’ Department of Medical Physics and Acoustics, Oldenburg University, Germany
- HörTech gGmbH, Oldenburg, Germany
| | | | - Tobias Neher
- Medizinische Physik and Cluster of Excellence ‘Hearing4all,’ Department of Medical Physics and Acoustics, Oldenburg University, Germany
| |
Collapse
|
43
|
Mendonça C, Mandelli P, Pulkki V. Modeling the Perception of Audiovisual Distance: Bayesian Causal Inference and Other Models. PLoS One 2016; 11:e0165391. [PMID: 27959919 PMCID: PMC5154506 DOI: 10.1371/journal.pone.0165391] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2016] [Accepted: 10/11/2016] [Indexed: 11/23/2022] Open
Abstract
Studies of audiovisual perception of distance are rare. Here, visual and auditory cue interactions in distance are tested against several multisensory models, including a modified causal inference model. In this causal inference model predictions of estimate distributions are included. In our study, the audiovisual perception of distance was overall better explained by Bayesian causal inference than by other traditional models, such as sensory dominance and mandatory integration, and no interaction. Causal inference resolved with probability matching yielded the best fit to the data. Finally, we propose that sensory weights can also be estimated from causal inference. The analysis of the sensory weights allows us to obtain windows within which there is an interaction between the audiovisual stimuli. We find that the visual stimulus always contributes by more than 80% to the perception of visual distance. The visual stimulus also contributes by more than 50% to the perception of auditory distance, but only within a mobile window of interaction, which ranges from 1 to 4 m.
Collapse
Affiliation(s)
- Catarina Mendonça
- Department of Signal Processing and Acoustics, Aalto University, Espoo, Finland
- * E-mail:
| | - Pietro Mandelli
- School of Industrial and Information Engineering, Polytechnic University of Milan, Milan, Italy
| | - Ville Pulkki
- Department of Signal Processing and Acoustics, Aalto University, Espoo, Finland
| |
Collapse
|
44
|
Gil-Carvajal JC, Cubick J, Santurette S, Dau T. Spatial Hearing with Incongruent Visual or Auditory Room Cues. Sci Rep 2016; 6:37342. [PMID: 27853290 PMCID: PMC5112595 DOI: 10.1038/srep37342] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2016] [Accepted: 10/26/2016] [Indexed: 11/09/2022] Open
Abstract
In day-to-day life, humans usually perceive the location of sound sources as outside their heads. This externalized auditory spatial perception can be reproduced through headphones by recreating the sound pressure generated by the source at the listener’s eardrums. This requires the acoustical features of the recording environment and listener’s anatomy to be recorded at the listener’s ear canals. Although the resulting auditory images can be indistinguishable from real-world sources, their externalization may be less robust when the playback and recording environments differ. Here we tested whether a mismatch between playback and recording room reduces perceived distance, azimuthal direction, and compactness of the auditory image, and whether this is mostly due to incongruent auditory cues or to expectations generated from the visual impression of the room. Perceived distance ratings decreased significantly when collected in a more reverberant environment than the recording room, whereas azimuthal direction and compactness remained room independent. Moreover, modifying visual room-related cues had no effect on these three attributes, while incongruent auditory room-related cues between the recording and playback room did affect distance perception. Consequently, the external perception of virtual sounds depends on the degree of congruency between the acoustical features of the environment and the stimuli.
Collapse
Affiliation(s)
- Juan C Gil-Carvajal
- Hearing Systems, Department of Electrical Engineering, Technical University of Denmark, Ørsteds Plads 352, 2800, Kgs. Lyngby, Denmark
| | - Jens Cubick
- Hearing Systems, Department of Electrical Engineering, Technical University of Denmark, Ørsteds Plads 352, 2800, Kgs. Lyngby, Denmark
| | - Sébastien Santurette
- Hearing Systems, Department of Electrical Engineering, Technical University of Denmark, Ørsteds Plads 352, 2800, Kgs. Lyngby, Denmark
| | - Torsten Dau
- Hearing Systems, Department of Electrical Engineering, Technical University of Denmark, Ørsteds Plads 352, 2800, Kgs. Lyngby, Denmark
| |
Collapse
|
45
|
Statistics of natural reverberation enable perceptual separation of sound and space. Proc Natl Acad Sci U S A 2016; 113:E7856-E7865. [PMID: 27834730 DOI: 10.1073/pnas.1612524113] [Citation(s) in RCA: 33] [Impact Index Per Article: 4.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
In everyday listening, sound reaches our ears directly from a source as well as indirectly via reflections known as reverberation. Reverberation profoundly distorts the sound from a source, yet humans can both identify sound sources and distinguish environments from the resulting sound, via mechanisms that remain unclear. The core computational challenge is that the acoustic signatures of the source and environment are combined in a single signal received by the ear. Here we ask whether our recognition of sound sources and spaces reflects an ability to separate their effects and whether any such separation is enabled by statistical regularities of real-world reverberation. To first determine whether such statistical regularities exist, we measured impulse responses (IRs) of 271 spaces sampled from the distribution encountered by humans during daily life. The sampled spaces were diverse, but their IRs were tightly constrained, exhibiting exponential decay at frequency-dependent rates: Mid frequencies reverberated longest whereas higher and lower frequencies decayed more rapidly, presumably due to absorptive properties of materials and air. To test whether humans leverage these regularities, we manipulated IR decay characteristics in simulated reverberant audio. Listeners could discriminate sound sources and environments from these signals, but their abilities degraded when reverberation characteristics deviated from those of real-world environments. Subjectively, atypical IRs were mistaken for sound sources. The results suggest the brain separates sound into contributions from the source and the environment, constrained by a prior on natural reverberation. This separation process may contribute to robust recognition while providing information about spaces around us.
Collapse
|
46
|
Grimm G, Luberadzka J, Hohmann V. Virtual acoustic environments for comprehensive evaluation of model-based hearing devices<sup/>. Int J Audiol 2016; 57:S112-S117. [PMID: 27813439 DOI: 10.1080/14992027.2016.1247501] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
Abstract
OBJECTIVE Create virtual acoustic environments (VAEs) with interactive dynamic rendering for applications in audiology. DESIGN A toolbox for creation and rendering of dynamic virtual acoustic environments (TASCAR) that allows direct user interaction was developed for application in hearing aid research and audiology. The software architecture and the simulation methods used to produce VAEs are outlined. Example environments are described and analysed. CONCLUSION With the proposed software, a tool for simulation of VAEs is available. A set of VAEs rendered with the proposed software was described.
Collapse
Affiliation(s)
- Giso Grimm
- a Medizinische Physik and Cluster of Excellence Hearing4all , Universität Oldenburg , Oldenburg , Germany and.,b HörTech gGmbH , Oldenburg , Germany
| | - Joanna Luberadzka
- a Medizinische Physik and Cluster of Excellence Hearing4all , Universität Oldenburg , Oldenburg , Germany and
| | - Volker Hohmann
- a Medizinische Physik and Cluster of Excellence Hearing4all , Universität Oldenburg , Oldenburg , Germany and.,b HörTech gGmbH , Oldenburg , Germany
| |
Collapse
|
47
|
Auditory distance perception in humans: a review of cues, development, neuronal bases, and effects of sensory loss. Atten Percept Psychophys 2016; 78:373-95. [PMID: 26590050 PMCID: PMC4744263 DOI: 10.3758/s13414-015-1015-1] [Citation(s) in RCA: 95] [Impact Index Per Article: 11.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/05/2022]
Abstract
Auditory distance perception plays a major role in spatial awareness, enabling location of objects and avoidance of obstacles in the environment. However, it remains under-researched relative to studies of the directional aspect of sound localization. This review focuses on the following four aspects of auditory distance perception: cue processing, development, consequences of visual and auditory loss, and neurological bases. The several auditory distance cues vary in their effective ranges in peripersonal and extrapersonal space. The primary cues are sound level, reverberation, and frequency. Nonperceptual factors, including the importance of the auditory event to the listener, also can affect perceived distance. Basic internal representations of auditory distance emerge at approximately 6 months of age in humans. Although visual information plays an important role in calibrating auditory space, sensorimotor contingencies can be used for calibration when vision is unavailable. Blind individuals often manifest supranormal abilities to judge relative distance but show a deficit in absolute distance judgments. Following hearing loss, the use of auditory level as a distance cue remains robust, while the reverberation cue becomes less effective. Previous studies have not found evidence that hearing-aid processing affects perceived auditory distance. Studies investigating the brain areas involved in processing different acoustic distance cues are described. Finally, suggestions are given for further research on auditory distance perception, including broader investigation of how background noise and multiple sound sources affect perceived auditory distance for those with sensory loss.
Collapse
|
48
|
Room Volume Estimation Based on Ambiguity of Short-Term Interaural Phase Differences Using Humanoid Robot Head. ROBOTICS 2016. [DOI: 10.3390/robotics5030016] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
|
49
|
Van der Stoep N, Van der Stigchel S, Nijboer TCW, Van der Smagt MJ. Audiovisual integration in near and far space: effects of changes in distance and stimulus effectiveness. Exp Brain Res 2016; 234:1175-88. [PMID: 25788009 PMCID: PMC4828496 DOI: 10.1007/s00221-015-4248-2] [Citation(s) in RCA: 18] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2014] [Accepted: 03/03/2015] [Indexed: 12/19/2022]
Abstract
A factor that is often not considered in multisensory research is the distance from which information is presented. Interestingly, various studies have shown that the distance at which information is presented can modulate the strength of multisensory interactions. In addition, our everyday multisensory experience in near and far space is rather asymmetrical in terms of retinal image size and stimulus intensity. This asymmetry is the result of the relation between the stimulus-observer distance and its retinal image size and intensity: an object that is further away is generally smaller on the retina as compared to the same object when it is presented nearer. Similarly, auditory intensity decreases as the distance from the observer increases. We investigated how each of these factors alone, and their combination, affected audiovisual integration. Unimodal and bimodal stimuli were presented in near and far space, with and without controlling for distance-dependent changes in retinal image size and intensity. Audiovisual integration was enhanced for stimuli that were presented in far space as compared to near space, but only when the stimuli were not corrected for visual angle and intensity. The same decrease in intensity and retinal size in near space did not enhance audiovisual integration, indicating that these results cannot be explained by changes in stimulus efficacy or an increase in distance alone, but rather by an interaction between these factors. The results are discussed in the context of multisensory experience and spatial uncertainty, and underline the importance of studying multisensory integration in the depth space.
Collapse
Affiliation(s)
- N Van der Stoep
- Department of Experimental Psychology, Helmholtz Institute, Utrecht University, Heidelberglaan 1, 3584 CS, Utrecht, The Netherlands.
| | - S Van der Stigchel
- Department of Experimental Psychology, Helmholtz Institute, Utrecht University, Heidelberglaan 1, 3584 CS, Utrecht, The Netherlands
| | - T C W Nijboer
- Department of Experimental Psychology, Helmholtz Institute, Utrecht University, Heidelberglaan 1, 3584 CS, Utrecht, The Netherlands
- Brain Center Rudolf Magnus, and Center of Excellence for Rehabilitation Medicine, De Hoogstraat Rehabilitation, University Medical Center Utrecht and De Hoogstraat Rehabilitation Center, Utrecht, The Netherlands
| | - M J Van der Smagt
- Department of Experimental Psychology, Helmholtz Institute, Utrecht University, Heidelberglaan 1, 3584 CS, Utrecht, The Netherlands
| |
Collapse
|
50
|
Kuusinen A, Lokki T. Investigation of auditory distance perception and preferences in concert halls by using virtual acoustics. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 138:3148-3159. [PMID: 26627788 DOI: 10.1121/1.4935388] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Virtual acoustics with multichannel sound reproduction was used to study auditory distance perception in four concert halls with multiple sound sources on stage. Eight subjects reported apparent auditory distances in five seating positions from 10 to 26 m to the middle of the sources on stage. The distance estimates were collected by absolute distance estimation procedure as well as a free modulus estimation procedure including both within and between halls evaluations. In addition, pairwise preferences were collected for two positions within each hall and for one position between halls. Results reveal that the perception of distance is dependent on the hall acoustics and show how the strength factor G and direct-to-reverberant energy ratio covary in relation to perceptual distances in these halls. The results also indicate that in such large spaces the overestimation of short distances may continue up to and further than 10 m from the sound sources. Preference results show that closer seats were liked more than further ones and that the strength of this preference is associated with the difference in perceptual distances.
Collapse
Affiliation(s)
- Antti Kuusinen
- Department of Computer Science, Aalto University School of Science, P.O. Box 15500, Otaniementie 17, FI-00076 Aalto, Finland
| | - Tapio Lokki
- Department of Computer Science, Aalto University School of Science, P.O. Box 15500, Otaniementie 17, FI-00076 Aalto, Finland
| |
Collapse
|