1
|
Carlini A, Bordeau C, Ambard M. Auditory localization: a comprehensive practical review. Front Psychol 2024; 15:1408073. [PMID: 39049946 PMCID: PMC11267622 DOI: 10.3389/fpsyg.2024.1408073] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2024] [Accepted: 06/17/2024] [Indexed: 07/27/2024] Open
Abstract
Auditory localization is a fundamental ability that allows to perceive the spatial location of a sound source in the environment. The present work aims to provide a comprehensive overview of the mechanisms and acoustic cues used by the human perceptual system to achieve such accurate auditory localization. Acoustic cues are derived from the physical properties of sound waves, and many factors allow and influence auditory localization abilities. This review presents the monaural and binaural perceptual mechanisms involved in auditory localization in the three dimensions. Besides the main mechanisms of Interaural Time Difference, Interaural Level Difference and Head Related Transfer Function, secondary important elements such as reverberation and motion, are also analyzed. For each mechanism, the perceptual limits of localization abilities are presented. A section is specifically devoted to reference systems in space, and to the pointing methods used in experimental research. Finally, some cases of misperception and auditory illusion are described. More than a simple description of the perceptual mechanisms underlying localization, this paper is intended to provide also practical information available for experiments and work in the auditory field.
Collapse
|
2
|
Abstract
Humans and other animals use spatial hearing to rapidly localize events in the environment. However, neural encoding of sound location is a complex process involving the computation and integration of multiple spatial cues that are not represented directly in the sensory organ (the cochlea). Our understanding of these mechanisms has increased enormously in the past few years. Current research is focused on the contribution of animal models for understanding human spatial audition, the effects of behavioural demands on neural sound location encoding, the emergence of a cue-independent location representation in the auditory cortex, and the relationship between single-source and concurrent location encoding in complex auditory scenes. Furthermore, computational modelling seeks to unravel how neural representations of sound source locations are derived from the complex binaural waveforms of real-life sounds. In this article, we review and integrate the latest insights from neurophysiological, neuroimaging and computational modelling studies of mammalian spatial hearing. We propose that the cortical representation of sound location emerges from recurrent processing taking place in a dynamic, adaptive network of early (primary) and higher-order (posterior-dorsal and dorsolateral prefrontal) auditory regions. This cortical network accommodates changing behavioural requirements and is especially relevant for processing the location of real-life, complex sounds and complex auditory scenes.
Collapse
|
3
|
Auditory Localization in Low-Bitrate Compressed Ambisonic Scenes. APPLIED SCIENCES-BASEL 2019. [DOI: 10.3390/app9132618] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
The increasing popularity of Ambisonics as a spatial audio format for streaming services poses new challenges to existing audio coding techniques. Immersive audio delivered to mobile devices requires an efficient bitrate compression that does not affect the spatial quality of the content. Good localizability of virtual sound sources is one of the key elements that must be preserved. This study was conducted to investigate the localization precision of virtual sound source presentations within Ambisonic scenes encoded with Opus low-bitrate compression at different bitrates and Ambisonic orders (1st, 3rd, and 5th). The test stimuli were reproduced over a 50-channel spherical loudspeaker configuration and binaurally using individually measured and generic Head-Related Transfer Functions (HRTFs). Participants were asked to adjust the position of a virtual acoustic pointer to match the position of virtual sound source within the bitrate-compressed Ambisonic scene. Results show that auditory localization in low-bitrate compressed Ambisonic scenes is not significantly affected by codec parameters. The key factors influencing localization are the rendering method and Ambisonic order truncation. This suggests that efficient perceptual coding might be successfully used for mobile spatial audio delivery.
Collapse
|
4
|
Lundbeck M, Grimm G, Hohmann V, Bramsløw L, Neher T. Effects of directional hearing aid settings on different laboratory measures of spatial awareness perception. Audiol Res 2018; 8:215. [PMID: 30581544 PMCID: PMC6275462 DOI: 10.4081/audiores.2018.215] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2018] [Accepted: 09/25/2018] [Indexed: 11/22/2022] Open
Abstract
Hearing loss can negatively influence the spatial hearing abilities of hearing-impaired listeners, not only in static but also in dynamic auditory environments. Therefore, ways of addressing these deficits with advanced hearing aid algorithms need to be investigated. In a previous study based on virtual acoustics and a computer simulation of different bilateral hearing aid fittings, we investigated auditory source movement detectability in older hearing- impaired (OHI) listeners. We found that two directional processing algorithms could substantially improve the detectability of left-right and near-far source movements in the presence of reverberation and multiple interfering sounds. In the current study, we carried out similar measurements with a loudspeaker-based setup and wearable hearing aids. We fitted a group of 15 OHI listeners with bilateral behind-the-ear devices that were programmed to have three different directional processing settings. Apart from source movement detectability, we assessed two other aspects of spatial awareness perception. Using a street scene with up to five environmental sound sources, the participants had to count the number of presented sources or to indicate the movement direction of a single target signal. The data analyses showed a clear influence of the number of concurrent sound sources and the starting position of the moving target signal on the participants' performance, but no influence of the different hearing aid settings. Complementary artificial head recordings showed that the acoustic differences between the three hearing aid settings were rather small. Another explanation for the lack of effects of the tested hearing aid settings could be that the simulated street scenario was not sufficiently sensitive. Possible ways of improving the sensitivity of the laboratory measures while maintaining high ecological validity and complexity are discussed.
Collapse
Affiliation(s)
- Micha Lundbeck
- Medizinische Physik and Cluster of Excellence “Hearing4all”, Oldenburg University, Oldenburg, Germany
- HörTech gGmbH, Oldenburg, Germany
| | - Giso Grimm
- Medizinische Physik and Cluster of Excellence “Hearing4all”, Oldenburg University, Oldenburg, Germany
- HörTech gGmbH, Oldenburg, Germany
| | - Volker Hohmann
- Medizinische Physik and Cluster of Excellence “Hearing4all”, Oldenburg University, Oldenburg, Germany
- HörTech gGmbH, Oldenburg, Germany
| | | | - Tobias Neher
- Institute of Clinical Research, University of Southern Denmark, Odense, Denmark
| |
Collapse
|
5
|
May KR, Walker BN. The effects of distractor sounds presented through bone conduction headphones on the localization of critical environmental sounds. APPLIED ERGONOMICS 2017; 61:144-158. [PMID: 28237013 DOI: 10.1016/j.apergo.2017.01.009] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/20/2016] [Revised: 01/04/2017] [Accepted: 01/15/2017] [Indexed: 06/06/2023]
Abstract
Bone conduction headphones are devices that transmit sound through the bones of a listener's head rather than through the air in their outer ear. They have been marketed as a safer way to enjoy audio content while walking, jogging, or cycling. However, listening to distracting sounds over bone conduction may still disrupt a listener's awareness of their auditory environment. The present study investigated the nature of this interference with the faculty of sound source localization-a key prerequisite for generating situation awareness through audio. Participants sat in the middle of a circle of loudspeakers and listened for target sounds played from different directions. Each time they heard a sound, they responded by indicating what direction they judged the sound to have come from. Meanwhile, participants listened to distractor sounds played through bone conduction headphones. Participants heard (1) no distractor sounds, (2) a spoken story that they were instructed to ignore, and (3) the same spoken story that they were instructed to attend to. For conditions (2) and (3), some participants heard a version of the story with background music, while others heard the spoken story without the music. Participants had greater localization error in the distractor-present conditions. Additionally, participants who heard the spoken story with music exhibited greater localization error. However, there was no effect of whether participants ignored or attended to distractors. This pattern was attributed to masking effects, and was more pronounced for narrow-band targets compared to broadband targets. Post-hoc analyses found evidence of a 'pulling' effect, in which localization judgments were systematically biased toward the apparent direction of the bone conducted distractors. These results indicate that using bone conduction headphones can be expected to cause a decline in a person's awareness of their environment, in a subtle way that a jogger or cyclist might not be actively aware of, even if their attention is directed to the environment and environmental sounds are readily detectible.
Collapse
Affiliation(s)
- Keenan R May
- Georgia Institute of Technology, 654 Cherry Street, Atlanta GA 30332, United States.
| | - Bruce N Walker
- Georgia Institute of Technology, 654 Cherry Street, Atlanta GA 30332, United States.
| |
Collapse
|
6
|
Zhong X, Yost WA. How many images are in an auditory scene? THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 141:2882. [PMID: 28464690 PMCID: PMC6909977 DOI: 10.1121/1.4981118] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
If an auditory scene consists of many spatially separated sound sources, how many sound sources can be processed by the auditory system? Experiment I determined how many speech sources could be localized simultaneously on the azimuth plane. Different words were played from multiple loudspeakers, and listeners reported the total number of sound sources and their individual locations. In experiment II the accuracy of localizing one speech source in a mixture of multiple speech sources was determined. An extra sound source was added to an existing set of sound sources, and the task was to localize that extra source. In experiment III the setup and task were the same as in experiment I, except that the sounds were tones. The results showed that the maximum number of sound sources that listeners could perceive was limited to approximately four spatially separated speech signals and three for tonal signals. The localization errors increased along with the increase of total number of sound sources. When four or more speech sources already existed, the accuracy in localizing an additional source was near chance.
Collapse
Affiliation(s)
- Xuan Zhong
- Department of Speech and Hearing Science, Arizona State University, Tempe, Arizona 85287, USA
| | - William A Yost
- Department of Speech and Hearing Science, Arizona State University, Tempe, Arizona 85287, USA
| |
Collapse
|
7
|
Mao B, Aytekin M, Wilkinson GS, Moss CF. Big brown bats (Eptesicus fuscus) reveal diverse strategies for sonar target tracking in clutter. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 140:1839. [PMID: 27914429 PMCID: PMC6909987 DOI: 10.1121/1.4962496] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/21/2016] [Revised: 08/18/2016] [Accepted: 08/20/2016] [Indexed: 05/15/2023]
Abstract
Bats actively adjust the acoustic features of their sonar calls to control echo information specific to a given task and environment. A previous study investigated how bats adapted their echolocation behavior when tracking a moving target in the presence of a stationary distracter at different distances and angular offsets. The use of only one distracter, however, left open the possibility that a bat could reduce the interference of the distracter by turning its head. Here, bats tracked a moving target in the presence of one or two symmetrically placed distracters to investigate adaptive echolocation behavior in a situation where vocalizing off-axis would result in increased interference from distracter echoes. Both bats reduced bandwidth and duration but increased sweep rate in more challenging distracter conditions, and surprisingly, made more head turns in the two-distracter condition compared to one, but only when distracters were placed at large angular offsets. However, for most variables examined, subjects showed distinct strategies to reduce clutter interference, either by (1) changing spectral or temporal features of their calls, or (2) producing large numbers of sonar sound groups and consistent head-turning behavior. The results suggest that individual bats can use different strategies for target tracking in cluttered environments.
Collapse
Affiliation(s)
- Beatrice Mao
- Department of Biology, University of Maryland, College Park, Maryland 20742, USA
| | - Murat Aytekin
- Department of Psychology, Institute for Systems Research, University of Maryland, College Park, Maryland 20742, USA
| | - Gerald S Wilkinson
- Department of Biology, University of Maryland, College Park, Maryland 20742, USA
| | - Cynthia F Moss
- Department of Psychology, Institute for Systems Research, University of Maryland, College Park, Maryland 20742, USA
| |
Collapse
|
8
|
Josupeit A, Kopčo N, Hohmann V. Modeling of speech localization in a multi-talker mixture using periodicity and energy-based auditory features. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 139:2911. [PMID: 27250183 DOI: 10.1121/1.4950699] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
A recent study showed that human listeners are able to localize a short speech target simultaneously masked by four speech tokens in reverberation [Kopčo, Best, and Carlile (2010). J. Acoust. Soc. Am. 127, 1450-1457]. Here, an auditory model for solving this task is introduced. The model has three processing stages: (1) extraction of the instantaneous interaural time difference (ITD) information, (2) selection of target-related ITD information ("glimpses") using a template-matching procedure based on periodicity, spectral energy, or both, and (3) target location estimation. The model performance was compared to the human data, and to the performance of a modified model using an ideal binary mask (IBM) at stage (2). The IBM-based model performed similarly to the subjects, indicating that the binaural model is able to accurately estimate source locations. Template matching using spectral energy and using a combination of spectral energy and periodicity achieved good results, while using periodicity alone led to poor results. Particularly, the glimpses extracted from the initial portion of the signal were critical for good performance. Simulation data show that the auditory features investigated here are sufficient to explain human performance in this challenging listening condition and thus may be used in models of auditory scene analysis.
Collapse
Affiliation(s)
- Angela Josupeit
- Medizinische Physik, Cluster of Excellence Hearing4all, Universität Oldenburg, 26111 Oldenburg, Germany
| | - Norbert Kopčo
- Institute of Computer Science, Faculty of Science, P. J. Šafárik University, Jesenná 5, 04001 Košice, Slovakia
| | - Volker Hohmann
- Medizinische Physik, Cluster of Excellence Hearing4all, Universität Oldenburg, 26111 Oldenburg, Germany
| |
Collapse
|
9
|
Kayser H, Hohmann V, Ewert SD, Kollmeier B, Anemüller J. Robust auditory localization using probabilistic inference and coherence-based weighting of interaural cues. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 138:2635-2648. [PMID: 26627742 DOI: 10.1121/1.4932588] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Robust sound source localization is performed by the human auditory system even in challenging acoustic conditions and in previously unencountered, complex scenarios. Here a computational binaural localization model is proposed that possesses mechanisms for handling of corrupted or unreliable localization cues and generalization across different acoustic situations. Central to the model is the use of interaural coherence, measured as interaural vector strength (IVS), to dynamically weight the importance of observed interaural phase (IPD) and level (ILD) differences in frequency bands up to 1.4 kHz. This is accomplished through formulation of a probabilistic model in which the ILD and IPD distributions pertaining to a specific source location are dependent on observed interaural coherence. Bayesian computation of the direction-of-arrival probability map naturally leads to coherence-weighted integration of location cues across frequency and time. Results confirm the model's validity through statistical analyses of interaural parameter values. Simulated localization experiments show that even data points with low reliability (i.e., low IVS) can be exploited to enhance localization performance. A temporal integration length of at least 200 ms is required to gain a benefit; this is in accordance with previous psychoacoustic findings on temporal integration of spatial cues in the human auditory system.
Collapse
Affiliation(s)
- Hendrik Kayser
- Medizinische Physik and Cluster of Excellence Hearing4all, Universität Oldenburg, 26111 Oldenburg, Germany
| | - Volker Hohmann
- Medizinische Physik and Cluster of Excellence Hearing4all, Universität Oldenburg, 26111 Oldenburg, Germany
| | - Stephan D Ewert
- Medizinische Physik and Cluster of Excellence Hearing4all, Universität Oldenburg, 26111 Oldenburg, Germany
| | - Birger Kollmeier
- Medizinische Physik and Cluster of Excellence Hearing4all, Universität Oldenburg, 26111 Oldenburg, Germany
| | - Jörn Anemüller
- Medizinische Physik and Cluster of Excellence Hearing4all, Universität Oldenburg, 26111 Oldenburg, Germany
| |
Collapse
|
10
|
Parrington L, MacMahon C, Ball K. How Task Complexity and Stimulus Modality Affect Motor Execution: Target Accuracy, Response Timing and Hesitations. J Mot Behav 2015; 47:343-51. [DOI: 10.1080/00222895.2014.984649] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
|
11
|
Durin V, Carlile S, Guillon P, Best V, Kalluri S. Acoustic analysis of the directional information captured by five different hearing aid styles. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2014; 136:818-828. [PMID: 25096115 DOI: 10.1121/1.4883372] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
This study compared the head-related transfer functions (HRTFs) recorded from the bare ear of a mannequin for 393 spatial locations and for five different hearing aid styles: Invisible-in-the-canal (IIC), completely-in-the-canal (CIC), in-the-canal (ITC), in-the-ear (ITE), and behind-the-ear (BTE). The spectral distortions of each style compared to the bare ear were described qualitatively in terms of the gain and frequency characteristics of the prominent spectral notch and two peaks in the HRTFs. Two quantitative measures of the differences between the HRTF sets and a measure of the dissimilarity of the HRTFs within each set were also computed. In general, the IIC style was most similar and the BTE most dissimilar to the bare ear recordings. The relative similarities among the CIC, ITC, and ITE styles depended on the metric employed. The within-style spectral dissimilarities were comparable for the bare ear, IIC, CIC, and ITC with increasing ambiguity for the ITE and BTE styles. When the analysis bandwidth was limited to 8 kHz, the HRTFs within each set became much more similar.
Collapse
Affiliation(s)
- Virginie Durin
- VAST Audio Pty Ltd., 4 Cornwallis Street, Eveleigh, New South Wales 2015, Australia
| | - Simon Carlile
- Bosh Institute and School of Medical Sciences, Anderson Stuart Building (F13), University of Sydney, New South Wales 2006, Australia
| | - Pierre Guillon
- Computing and Audio Research Laboratory, School of Electrical and Information Engineering, University of Sydney, New South Wales 2006, Australia
| | - Virginia Best
- Bosh Institute and School of Medical Sciences, Anderson Stuart Building (F13), University of Sydney, New South Wales 2006, Australia
| | - Sridhar Kalluri
- Starkey Hearing Research Center, 2150 Shattuck Avenue, Suite 408, Berkeley, California 94704-1345
| |
Collapse
|
12
|
Robinson PW, Pätynen J, Lokki T, Jang HS, Jeon JY, Xiang N. The role of diffusive architectural surfaces on auditory spatial discrimination in performance venues. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 133:3940-50. [PMID: 23742348 DOI: 10.1121/1.4803846] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/22/2023]
Abstract
In musical or theatrical performance, some venues allow listeners to individually localize and segregate individual performers, while others produce a well blended ensemble sound. The room acoustic conditions that make this possible, and the psycho-acoustic effects at work are not fully understood. This research utilizes auralizations from measured and simulated performance venues to investigate spatial discrimination of multiple acoustic sources in rooms. Signals were generated from measurements taken in a small theater, and listeners in the audience area were asked to distinguish pairs of speech sources on stage with various spatial separations. This experiment was repeated with the proscenium splay walls treated to be flat, diffusive, or absorptive. Similar experiments were conducted in a simulated hall, utilizing 11 early reflections with various characteristics, and measured late reverberation. The experiments reveal that discriminating the lateral arrangement of two sources is possible at narrower separation angles when reflections come from flat or absorptive rather than diffusive surfaces.
Collapse
Affiliation(s)
- Philip W Robinson
- Department of Media Technology, Aalto University School of Science, P.O. Box 15500, FI-00076 Aalto, Finland.
| | | | | | | | | | | |
Collapse
|
13
|
Yost WA, Brown CA. Localizing the sources of two independent noises: role of time varying amplitude differences. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 133:2301-13. [PMID: 23556597 PMCID: PMC3631260 DOI: 10.1121/1.4792155] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/25/2012] [Revised: 01/07/2013] [Accepted: 01/24/2013] [Indexed: 05/25/2023]
Abstract
Listeners localized the free-field sources of either one or two simultaneous and independently generated noise bursts. Listeners' localization performance was better when localizing one rather than two sound sources. With two sound sources, localization performance was better when the listener was provided prior information about the location of one of them. Listeners also localized two simultaneous noise bursts that had sinusoidal amplitude modulation (AM) applied, in which the modulation envelope was in-phase across the two source locations or was 180° out-of-phase. The AM was employed to investigate a hypothesis as to what process listeners might use to localize multiple sound sources. The results supported the hypothesis that localization of two sound sources might be based on temporal-spectral regions of the combined waveform in which the sound from one source was more intense than that from the other source. The interaural information extracted from such temporal-spectral regions might provide reliable estimates of the sound source location that produced the more intense sound in that temporal-spectral region.
Collapse
Affiliation(s)
- William A Yost
- Spatial Hearing Laboratory, Department of Speech and Hearing Science, Arizona State University, P.O. Box 870102, Tempe, Arizona 85287-0102, USA.
| | | |
Collapse
|
14
|
Mueller MF, Kegel A, Schimmel SM, Dillier N, Hofbauer M. Localization of virtual sound sources with bilateral hearing aids in realistic acoustical scenes. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2012; 131:4732-4742. [PMID: 22712946 DOI: 10.1121/1.4705292] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
Sound localization with hearing aids has traditionally been investigated in artificial laboratory settings. These settings are not representative of environments in which hearing aids are used. With individual Head-Related Transfer Functions (HRTFs) and room simulations, realistic environments can be reproduced and the performance of hearing aid algorithms can be evaluated. In this study, four different environments with background noise have been implemented in which listeners had to localize different sound sources. The HRTFs were measured inside the ear canals of the test subjects and by the microphones of Behind-The-Ear (BTEs) hearing aids. In the first experiment the system for virtual acoustics was evaluated by comparing perceptual sound localization results for the four scenes in a real room with a simulated one. In the second experiment, sound localization with three BTE algorithms, an omnidirectional microphone, a monaural cardioid-shaped beamformer and a monaural noise canceler, was examined. The results showed that the system for generating virtual environments is a reliable tool to evaluate sound localization with hearing aids. With BTE hearing aids localization performance decreased and the number of front-back confusions was at chance level. The beamformer, due to its directivity characteristics, allowed the listener to resolve the front-back ambiguity.
Collapse
Affiliation(s)
- Martin F Mueller
- Laboratory for Experimental Audiology, University Hospital Zurich, Zurich 8091, Switzerland.
| | | | | | | | | |
Collapse
|
15
|
The auditory dorsal pathway: Orienting vision. Neurosci Biobehav Rev 2011; 35:2162-73. [PMID: 21530585 DOI: 10.1016/j.neubiorev.2011.04.005] [Citation(s) in RCA: 60] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2010] [Revised: 03/16/2011] [Accepted: 04/10/2011] [Indexed: 11/24/2022]
|
16
|
Moss CF, Chiu C, Surlykke A. Adaptive vocal behavior drives perception by echolocation in bats. Curr Opin Neurobiol 2011; 21:645-52. [PMID: 21705213 DOI: 10.1016/j.conb.2011.05.028] [Citation(s) in RCA: 47] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2011] [Revised: 05/19/2011] [Accepted: 05/27/2011] [Indexed: 11/29/2022]
Abstract
Echolocation operates through adaptive sensorimotor systems that collectively enable the bat to localize and track sonar objects as it flies. The features of sonar signals used by a bat to probe its surroundings determine the information available to its acoustic imaging system. In turn, the bat's perception of a complex scene guides its active adjustments in the features of subsequent sonar vocalizations. Here, we propose that the bat's active vocal-motor behaviors play directly into its representation of a dynamic auditory scene.
Collapse
Affiliation(s)
- Cynthia F Moss
- University of Maryland, Department of Psychology, Biology-Psychology Building, College Park, MD 20742, United States.
| | | | | |
Collapse
|
17
|
Martin RL, Flanagan P, McAnally KI, Eberle G. Memory for the locations of environmental sounds. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2011; 129:3873-3883. [PMID: 21682410 DOI: 10.1121/1.3589254] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
The accuracy with which a single source of sound can be localized has been examined in many studies, but very few studies have examined the ability of participants to determine the absolute locations of multiple sources of sound. The current study assessed participants' abilities to determine and remember the locations of up to six sources of environmental sound that were positioned at a range of azimuths and elevations in virtual auditory space. In experiment 1, a sequence of one to six sounds was presented one, three, or five times in each trial and the target sound was nominated following presentation of the last sequence. In experiment 2, memory load was held constant by nominating the target sound prior to a single sequence presentation. Localization accuracy was observed to decrease as the number of sounds was increased to three or more under the conditions of experiment 1, but not those of experiment 2. In experiment 1, localization was more accurate when sequences were presented more than once. Pronounced primacy and recency effects were observed for the six sound conditions in experiment 1. An analysis of errors for those conditions indicated that immediate temporal errors, but not immediate spatial errors, were over-represented.
Collapse
Affiliation(s)
- Russell L Martin
- Air Operations Division, Defence Science and Technology Organisation, Fishermans Bend 3207, Australia.
| | | | | | | |
Collapse
|
18
|
Brungart DS, Simpson BD. Effects of bandwidth on auditory localization with a noise masker. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2009; 126:3199-3208. [PMID: 20000933 DOI: 10.1121/1.3243309] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/28/2023]
Abstract
Although high-frequency content is known to be critically important for the accurate location of isolated sounds, relatively little is known about the importance of high-frequency spectral content for the localization of sounds in the presence of a masker. In this experiment, listeners were asked to identify the location of a pulsed-noise target in the presence of a randomly located continuous noise masker. Both the target and masker were low-pass filtered at one of eight cutoff frequencies ranging from 1 to 16 kHz, and the signal-to-noise ratio was varied from -12 to +12 dB. The results confirm the importance of high frequencies for the localization of isolated sounds, and show that high-frequency content remains critical in cases where the target sound is masked by a spatially separated masker. In fact, when two sources of the same level are randomly located in space, these results show that a decrease in stimulus bandwidth from 16 to 12 kHz might result in a 30% increase in overall localization error.
Collapse
Affiliation(s)
- Douglas S Brungart
- Air Force Research Laboratory, 2610 Seventh Street, WPAFB, Ohio 45433, USA.
| | | |
Collapse
|
19
|
Populin LC. Human sound localization: measurements in untrained, head-unrestrained subjects using gaze as a pointer. Exp Brain Res 2008; 190:11-30. [PMID: 18575853 DOI: 10.1007/s00221-008-1445-2] [Citation(s) in RCA: 25] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2007] [Accepted: 05/25/2008] [Indexed: 11/30/2022]
Abstract
Studies of sound localization in humans have used various behavioral measures to quantify the observers' perceptions; a non-comprehensive list includes verbal reports, head pointing, gun pointing, stylus pointing, and laser aiming. Comparison of localization performance reveals that in humans, just as in animals, different results are obtained with different experimental tasks. Accordingly, to circumvent problems associated with task selection and training, this study used gaze, an ethologically valid behavior for spatial pointing in species with a specialized area of the fovea, to measure sound localization perception of human subjects. Orienting using gaze as a pointer does not require training, preserves the natural link between perception and action, and allows for direct behavioral comparisons across species. The results revealed, unexpectedly, a large degree of variability across subjects in both accuracy and precision. The magnitude of the average angular localization errors for the most eccentric horizontal targets, however, were very similar to those documented in studies that used head pointing, whereas the magnitude of the localization errors for the frontal targets were considerably larger. In addition, an overall improvement in sound localization in the context of the memory-saccade task, as well as a lack of effect of initial eye and head position on perceived sound location were documented.
Collapse
Affiliation(s)
- Luis C Populin
- Department of Anatomy, Neuroscience Training Program, and UW Eye Research Institute, University of Wisconsin-Madison, 1300 University Ave, Madison, WI 53706, USA.
| |
Collapse
|
20
|
Steven Colburn H, Shinn-Cunningham B, Kidd G, Durlach N. The perceptual consequences of binaural hearing. Int J Audiol 2007; 45 Suppl 1:S34-44. [PMID: 16938773 DOI: 10.1080/14992020600782642] [Citation(s) in RCA: 36] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
Abstract
Binaural processing in normal hearing activities is based on the ability of listeners to use the information provided by the differences between the signals at the two ears. The most prominent differences are the interaural time difference and the interaural level difference, both of which depend on frequency. This paper describes the stages by which these differences are estimated by the physiological structures of the auditory system, summarizes the sensitivity of the human listener to these differences, and reviews the nature of the interaural differences in realistic environments.
Collapse
Affiliation(s)
- H Steven Colburn
- Hearing Research Center, Boston University, Boston, Massachusetts, USA.
| | | | | | | |
Collapse
|
21
|
Branstetter BK, Mevissen SJ, Pack AA, Herman LM, Roberts SR, Carsrud LK. Dolphin (Tursiops truncatus) echoic angular discrimination: effects of object separation and complexity. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 121:626-35. [PMID: 17297816 DOI: 10.1121/1.2400664] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
A bottlenose dolphin was tested on its ability to echoically discriminate horizontal angular differences between arrays of vertically oriented air-filled PVC rods. The blindfolded dolphin was required to station in a submerged hoop 2 radial m from the stimuli and indicate if an array with two rods (S+) was to the right or the left of a single rod (S-). The angular separation between the two rods (thetaw) was held constant within each experiment while the angle between the S+ and the S-stimuli (thetab) varied to produce angular differences (deltatheta= thetab-thetaw) ranging from 0.25 to 4 degrees. In experiment I, thetaw was maintained at 2 degrees and in experiment II, thetaw was maintained at 4 degrees. Resulting 75% correct thresholds (method of constant stimuli) were 1.5 and 0.7 degrees, respectively. The two main findings of this study are: (1) decreasing the number of targets does not aid in localization, and (2) increasing the space between the rods enhances localization. Taken as a whole, the experiments suggest dolphins have a well-developed ability to resolve spatial information through sonar.
Collapse
Affiliation(s)
- Brian K Branstetter
- Psychology Department, University of Hawaii, Manoa, Honolulu, Hawaii 96822-2294, USA.
| | | | | | | | | | | |
Collapse
|
22
|
Carlile S, Martin R, McAnally K. Spectral Information in Sound Localization. INTERNATIONAL REVIEW OF NEUROBIOLOGY 2005; 70:399-434. [PMID: 16472641 DOI: 10.1016/s0074-7742(05)70012-x] [Citation(s) in RCA: 43] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
Affiliation(s)
- Simon Carlile
- Auditory Neuroscience Laboratory, School of Medical Sciences University of Sydney, Sydney NSW 2006, Australia
| | | | | |
Collapse
|
23
|
Faller C, Merimaa J. Source localization in complex listening situations: selection of binaural cues based on interaural coherence. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2004; 116:3075-3089. [PMID: 15603153 DOI: 10.1121/1.1791872] [Citation(s) in RCA: 48] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/24/2023]
Abstract
In everyday complex listening situations, sound emanating from several different sources arrives at the ears of a listener both directly from the sources and as reflections from arbitrary directions. For localization of the active sources, the auditory system needs to determine the direction of each source, while ignoring the reflections and superposition effects of concurrently arriving sound. A modeling mechanism with these desired properties is proposed. Interaural time difference (ITD) and interaural level difference (ILD) cues are only considered at time instants when only the direct sound of a single source has non-negligible energy in the critical band and, thus, when the evoked ITD and ILD represent the direction of that source. It is shown how to identify such time instants as a function of the interaural coherence (IC). The source directions suggested by the selected ITD and ILD cues are shown to imply the results of a number of published psychophysical studies related to source localization in the presence of distracters, as well as in precedence effect conditions.
Collapse
Affiliation(s)
- Christof Faller
- Mobile Terminals Division, Agere Systems, Allentown, Pennsylvania, USA.
| | | |
Collapse
|
24
|
Macpherson EA, Middlebrooks JC. Vertical-plane sound localization probed with ripple-spectrum noise. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2003; 114:430-445. [PMID: 12880054 DOI: 10.1121/1.1582174] [Citation(s) in RCA: 18] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/24/2023]
Abstract
Ripple-spectrum stimuli were used to investigate the scale of spectral detail used by listeners in interpreting spectral cues for vertical-plane localization. In three experiments, free-field localization judgments were obtained for 250-ms, 0.6-16-kHz noise bursts with log-ripple spectra that varied in ripple density, peak-to-trough depth, and phase. When ripple density was varied and depth was held constant at 40 dB, listeners' localization error rates increased most (relative to rates for flat-spectrum targets) for densities of 0.5-2 ripples/oct. When depth was varied and density was held constant at 1 ripple/oct, localization accuracy was degraded only for ripple depths > or = 20 dB. When phase was varied and density was held constant at 1 ripple/oct and depth at 40 dB, three of five listeners made errors at consistent locations unrelated to the ripple phase, whereas two listeners made errors at locations systematically modulated by ripple phase. Although the reported upper limit for ripple discrimination is 10 ripples/oct [Supin et al., J. Acoust. Soc. Am. 106, 2800-2804 (1999)], present results indicate that details finer than 2 ripples/oct or coarser than 0.5 ripples/oct do not strongly influence processing of spectral cues for sound localization. The low spectral-frequency limit suggests that broad-scale spectral variation is discounted, even though components at this scale are among those contributing the most to the shapes of directional transfer functions.
Collapse
Affiliation(s)
- Ewan A Macpherson
- Kresge Hearing Research Institute, University of Michigan, 1301 East Ann Street, Ann Arbor, Michigan 48109-0506, USA.
| | | |
Collapse
|
25
|
Polyakov A, Pratt H. Electrophysiological correlates of azimuth and elevation cues for sound localization in human middle latency auditory evoked potentials. Ear Hear 2003; 24:143-55. [PMID: 12677111 DOI: 10.1097/01.aud.0000058109.44006.47] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVE To study, in humans, the effects of sound source azimuth and elevation on primary auditory cortex binaural activity associated with sound localization. DESIGN Middle Latency Auditory Evoked Potentials (MLAEPs) were recorded from three channels, in response to alternating polarity clicks, presented at a rate of 5/sec, at nine virtual spatial locations with different azimuths and elevations. Equivalent dipoles of Binaural Interaction Components (BICs) of MLAEPs were derived from 15 normally and symmetrically hearing adults by subtracting the response to binaural clicks at each spatial location from the algebraic sum of responses to stimulation of each ear alone. The amplified potentials were averaged over 4000 repetitions using a dwell time of 78 micro sec/address/channel. Variations in magnitudes, latencies and orientations of the dipole equivalents of cortical activity were noted in response to the nine spatial locations. RESULTS Middle-latency BICs included six major components corresponding in latency to the vertex-neck recorded components of MLAEP. A significant decrease of equivalent dipole magnitude was observed for two of the components: Pa2 in response to clicks in the backward positions (medium and no elevation); and Nb in response to clicks in the back and front positions (medium and no elevation) in the midsagittal plane. In the coronal plane, Pa2 equivalent dipole magnitude significantly decreased in response to right-horizontal (no elevation) clicks. Significant effects on equivalent dipole latencies of Pa2 were found for backward positions (no elevation) in the midsagittal plane. No significant effects on Pa2 and Nb equivalent dipole orientations were found across stimulus conditions. CONCLUSIONS The changes in equivalent dipole magnitudes and latencies of MLAEP BICs across stimulus conditions may reflect spectral tuning in binaural primary auditory cortex neurons processing the frequency cues for sound localization.
Collapse
Affiliation(s)
- Andrey Polyakov
- Evoked Potentials Laboratory, Technion--Israel Institute of Technology, Haifa
| | | |
Collapse
|
26
|
Polyakov A, Pratt H. Electrophysiologic correlates of direction and elevation cues for sound localization in the human brainstem. Int J Audiol 2003; 42:140-51. [PMID: 12705779 DOI: 10.3109/14992020309090423] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
Abstract
Our objective was to study the effects of sound source direction and elevation on human brainstem electrical activity associated with sound localization. The subjects comprised 15 normal-hearing and symmetrically hearing adults Auditory brainstem evoked potentials (ABEPs) were recorded from three channels, in response to alternating-polarity clicks, presented at a rate of 21.1/s, at nine virtual spatial locations with different direction and elevation attributes Equivalent dipoles of the binaural interaction components (BICs) of ABEPs were derived by subtracting the response to binaural clicks at each spatial location from the algebraic sum of monaural responses to stimulation of each ear in turn. The BICs included two major components corresponding in latency to the vertex-neck-recorded components V and VI of ABEP. A significant decrease of the first BIC's equivalent dipole magnitude was observed for clicks in the horizontal-frontal position (no elevation) in the midsagittal plane, and for clicks in the left-horizontal (no elevation) and right diagonally above the head (medium elevation) positions in the coronal plane, compared to clicks positioned directly above the head. Significant effects on equivalent dipole latencies of this component were found for front-back positions in the midsagittal plane and left-right positions in the coronal plane, compared to clicks positioned directly above the head. The most remarkable finding was a significant change in equivalent dipole orientations across stimulus conditions. We conclude that the changes in BIC equivalent dipole latency, amplitude and orientation across stimulus conditions reflect differences in the distribution of binaural pontine activity evoked by sounds in different spatial locations.
Collapse
Affiliation(s)
- Andrey Polyakov
- Evoked Potentials Laboratory, Technion-Israel Institute of Technology, Haifa, Israel
| | | |
Collapse
|
27
|
Langendijk EHA, Bronkhorst AW. Contribution of spectral cues to human sound localization. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2002; 112:1583-1596. [PMID: 12398464 DOI: 10.1121/1.1501901] [Citation(s) in RCA: 69] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/24/2023]
Abstract
The contribution of spectral cues to human sound localization was investigated by removing cues in 1/2-, 1- or 2-octave bands in the frequency range above 4 kHz. Localization responses were given by placing an acoustic pointer at the same apparent position as a virtual target. The pointer was generated by filtering a 100-ms harmonic complex with equalized head-related transfer functions (HRTFs). Listeners controlled the pointer via a hand-held stick that rotated about a fixed point. In the baseline condition, the target, a 200-ms noise burst, was filtered with the same HRTFs as the pointer. In other conditions, the spectral information within a certain frequency band was removed by replacing the directional transfer function within this band with the average transfer of this band. Analysis of the data showed that removing cues in 1/2-octave bands did not affect localization, whereas for the 2-octave band correct localization was virtually impossible. The results obtained for the 1-octave bands indicate that up-down cues are located mainly in the 6-12-kHz band, and front-back cues in the 8-16-kHz band. The interindividual spread in response patterns suggests that different listeners use different localization cues. The response patterns in the median plane can be predicted using a model based on spectral comparison of directional transfer functions for target and response directions.
Collapse
|
28
|
Litovsky RY, Delgutte B. Neural correlates of the precedence effect in the inferior colliculus: effect of localization cues. J Neurophysiol 2002; 87:976-94. [PMID: 11826062 DOI: 10.1152/jn.00568.2001] [Citation(s) in RCA: 33] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
The precedence effect (PE) is an auditory phenomenon involved in suppressing the perception of echoes in reverberant environments, and is thought to facilitate accurate localization of sound sources. We investigated physiological correlates of the PE in the inferior colliculus (IC) of anesthetized cats, with a focus on directional mechanisms for this phenomenon. We used a virtual space (VS) technique, where two clicks (a "lead" and a "lag") separated by a brief time delay were each filtered through head-related transfer functions (HRTFs). For nearly all neurons, the response to the lag was suppressed for short delays and recovered at long delays. In general, both the time course and the directional patterns of suppression resembled those reported in free-field studies in many respects, suggesting that our VS simulation contained the essential cues for studying PE phenomena. The relationship between the directionality of the response to the lead and that of its suppressive effect on the lag varied a great deal among IC neurons. For a majority of units, both excitation produced by the lead and suppression of the lag response were highly directional, and the two were similar to one another. For these neurons, the long-lasting inhibitory inputs thought to be responsible for suppression seem to have similar spatial tuning as the inputs that determine the excitatory response to the lead. Further, the behavior of these neurons is consistent with psychophysical observations that the PE is strongest when the lead and the lag originate from neighboring spatial locations. For other neurons, either there was no obvious relationship between the directionality of the excitatory lead response and the directionality of suppression, or the suppression was highly directional whereas the excitation was not, or vice versa. For these neurons, the excitation and the suppression produced by the lead seem to depend on different mechanisms. Manipulation of the directional cues (such as interaural time and level differences) contained in the lead revealed further dissociations between excitation and suppression. Specifically, for about one-third of the neurons, suppression depended on different directional cues than did the response to the lead, even though the directionality of suppression was similar to that of the lead response when all cues were present. This finding suggests that the inhibitory inputs causing suppression may originate in part from subcollicular auditory nuclei processing different directional cues than the inputs that determine the excitatory response to the lead. Neurons showing such dissociations may play an important role in the PE when the lead and the lag originate from very different directions.
Collapse
Affiliation(s)
- R Y Litovsky
- Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA 02143, USA.
| | | |
Collapse
|