1
|
Fontana F, Järveläinen H, Papetti S, De Pra Y. Acoustic cues of keyboard mechanics enable auditory localization of upright piano tones. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 156:164-175. [PMID: 38958583 DOI: 10.1121/10.0026484] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/28/2023] [Accepted: 06/13/2024] [Indexed: 07/04/2024]
Abstract
Piano tone localization at the performer's listening point is a multisensory process involving audition, vision, and upper limb proprioception. The consequent representation of the auditory scene, especially in experienced pianists, is likely also influenced by their memory about the instrument keyboard. Disambiguating such components is not obvious, and first requires an analysis of the acoustic tone localization process to assess the role of auditory feedback in forming this scene. This analysis is complicated by the acoustic behavior of the piano, which does not guarantee the activation of the auditory precedence effect during a tone attack, nor can it provide robust interaural differences during the subsequent free evolution of the sound. In a tone localization task using a Disklavier upright piano (which can be operated remotely and configured to have its hammers hit a damper instead of producing a tone), twenty-three expert musicians, including pianists, successfully recognized the angular position of seven evenly distributed notes across the keyboard. The experiment involved listening to either full piano tones or just the key mechanical noise, with no additional feedback from other senses. This result suggests that the key mechanical noise alone activated the localization process without support from vision and/or limb proprioception. Since the same noise is present in the onset of the full tones, the key mechanics of our piano created a touch precursor in such tones that may be responsible of their correct angular localization by means of the auditory precedence effect. However, the significance of pitch cues arriving at a listener after the touch precursor was not measured when full tones were presented. As these cues characterize a note and, hence, the corresponding key position comprehensively, an open question remains regarding the contribution of pianists' spatial memory of the instrument keyboard to tone localization.
Collapse
Affiliation(s)
- Federico Fontana
- Department of Mathematics, Computer Science and Physics, Università di Udine, 206 via delle Scienze, Udine 33100, Italy
| | - Hanna Järveläinen
- Institute for Computer Music and Sound Technology, Zürcher Hochschule der Künste, 96 Pfingstweidstrasse, Zurich 8048, Switzerland
| | - Stefano Papetti
- Institute for Computer Music and Sound Technology, Zürcher Hochschule der Künste, 96 Pfingstweidstrasse, Zurich 8048, Switzerland
| | - Yuri De Pra
- Department of Energy Technologies and Renewable Sources, Agenzia nazionale per le nuove tecnologie, l'energia e lo sviluppo economico sostenibile, 301 via Anguillarese, Rome 00123, Italy
| |
Collapse
|
2
|
Zurek PM, Freyman RL, Najem F. Investigating the Utility of a Compact Loudspeaker Array for Audiometric Testing. Am J Audiol 2024; 33:476-491. [PMID: 38668699 DOI: 10.1044/2024_aja-23-00199] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/05/2024] Open
Abstract
PURPOSE This project addressed the uses of a loudspeaker array for audiometric measurements. It sought to evaluate a prototype compact array in terms of the reliability of test results across sound booths. METHOD A seven-loudspeaker array was developed to deliver sounds from -60° to +60° on an arc with a radius of 0.5 m. The system was equipped with a head position sensing system to maintain the listener's head near the optimal test position. Three array systems were distributed to each of the two test sites for within-subject assessments of booth equivalence on tests of sound localization, speech reception in noise, and threshold detection. A total of 36 subjects participated, 18 at each test site. RESULTS Results showed excellent interbooth consistency on tests of sound localization using speech and noise signals, including conditions in which one or both ears were covered with a muff. Booth consistency was also excellent on sound field threshold measurements for detecting quasi-diffuse noise bands. Nonequivalence was observed in some cases of speech-in-noise tests, particularly with a small one-person booth. Acoustic analyses of in situ loudspeaker responses indicated that some of the nonequivalent comparisons on speech-in-noise tests could be traced to the effects of reflections. CONCLUSIONS Overall, the results demonstrate the utility and reliability of a compact array for the assessment of localization ability, speech reception in noise, and sound field thresholds. However, the results indicate that researchers and clinicians should be aware of the reflection effects that can influence the results of sound field tests in which signal and noise levels from separate loudspeakers are critical.
Collapse
Affiliation(s)
| | - Richard L Freyman
- Department of Speech, Language, and Hearing Sciences, University of Massachusetts Amherst
| | - Fadi Najem
- Hearing and Balance Center, University of the Pacific, San Francisco, CA
| |
Collapse
|
3
|
Fumero MJ, Marrufo-Pérez MI, Eustaquio-Martín A, Lopez-Poveda EA. Factors that can affect divided speech intelligibility. Hear Res 2024; 441:108917. [PMID: 38061268 DOI: 10.1016/j.heares.2023.108917] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/26/2023] [Revised: 11/10/2023] [Accepted: 11/16/2023] [Indexed: 12/27/2023]
Abstract
Previous studies have shown that in challenging listening situations, people find it hard to equally divide their attention between two simultaneous talkers and tend to favor one talker over the other. The aim here was to investigate whether talker onset/offset, sex and location determine the favored talker. Fifteen people with normal hearing were asked to recognize as many words as possible from two sentences uttered by two talkers located at 45° and +45° azimuth, respectively. The sentences were from the same corpus, were time-centered and had equal sound level. In Conditions 1 and 2, the talkers had different sexes (male at +45°), sentence duration was not controlled for, and sentences were presented at 65 and 35 dB SPL, respectively. Listeners favored the male over the female talker, even more so at 35 dB SPL (62 % vs 43 % word recognition, respectively) than at 65 dB SPL (74 % vs 64 %, respectively). The greater asymmetry in intelligibility at the lower level supports that divided listening is harder and more 'asymmetric' in challenging acoustic scenarios. Listeners continued to favor the male talker when the experiment was repeated with sentences of equal average duration for the two talkers (Condition 3). This suggests that the earlier onset or later offset of male sentences (52 ms on average) was not the reason for the asymmetric intelligibility in Conditions 1 or 2. When the location of the talkers was switched (Condition 4) or the two talkers were the same woman (Condition 5), listeners continued to favor the talker to their right albeit non-significantly. Altogether, results confirm that in hard divided listening situations, listeners tend to favor the talker to their right. This preference is not affected by talker onset/offset delays less than 52 ms on average. Instead, the preference seems to be modulated by the voice characteristics of the talkers.
Collapse
Affiliation(s)
- Milagros J Fumero
- Instituto de Neurociencias de Castilla y León, Universidad de Salamanca, Salamanca 37007, Spain; Instituto de Investigación Biomédica de Salamanca, Universidad de Salamanca, Salamanca 37007, Spain
| | - Miriam I Marrufo-Pérez
- Instituto de Neurociencias de Castilla y León, Universidad de Salamanca, Salamanca 37007, Spain; Instituto de Investigación Biomédica de Salamanca, Universidad de Salamanca, Salamanca 37007, Spain
| | - Almudena Eustaquio-Martín
- Instituto de Neurociencias de Castilla y León, Universidad de Salamanca, Salamanca 37007, Spain; Instituto de Investigación Biomédica de Salamanca, Universidad de Salamanca, Salamanca 37007, Spain
| | - Enrique A Lopez-Poveda
- Instituto de Neurociencias de Castilla y León, Universidad de Salamanca, Salamanca 37007, Spain; Instituto de Investigación Biomédica de Salamanca, Universidad de Salamanca, Salamanca 37007, Spain; Departamento de Cirugía, Facultad de Medicina, Universidad de Salamanca, Salamanca 37007, Spain.
| |
Collapse
|
4
|
Luo TZ, Kim TD, Gupta D, Bondy AG, Kopec CD, Elliot VA, DePasquale B, Brody CD. Transitions in dynamical regime and neural mode underlie perceptual decision-making. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.10.15.562427. [PMID: 37904994 PMCID: PMC10614809 DOI: 10.1101/2023.10.15.562427] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/02/2023]
Abstract
Perceptual decision-making is the process by which an animal uses sensory stimuli to choose an action or mental proposition. This process is thought to be mediated by neurons organized as attractor networks 1,2 . However, whether attractor dynamics underlie decision behavior and the complex neuronal responses remains unclear. Here we use an unsupervised, deep learning-based method to discover decision-related dynamics from the simultaneous activity of neurons in frontal cortex and striatum of rats while they accumulate pulsatile auditory evidence. We show that contrary to prevailing hypotheses, attractors play a role only after a transition from a regime in the dynamics that is strongly driven by inputs to one dominated by the intrinsic dynamics. The initial regime mediates evidence accumulation, and the subsequent intrinsic-dominant regime subserves decision commitment. This regime transition is coupled to a rapid reorganization in the representation of the decision process in the neural population (a change in the "neural mode" along which the process develops). A simplified model approximating the coupled transition in the dynamics and neural mode allows inferring, from each trial's neural activity, the internal decision commitment time in that trial, and captures diverse and complex single-neuron temporal profiles, such as ramping and stepping 3-5 . It also captures trial-averaged curved trajectories 6-8 , and reveals distinctions between brain regions. Our results show that the formation of a perceptual choice involves a rapid, coordinated transition in both the dynamical regime and the neural mode of the decision process, and suggest pairing deep learning and parsimonious models as a promising approach for understanding complex data.
Collapse
|
5
|
Fantini D, Presti G, Geronazzo M, Bona R, Privitera AG, Avanzini F. Co-Immersion in Audio Augmented Virtuality: The Case Study of a Static and Approximated Late Reverberation Algorithm. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2023; 29:4472-4482. [PMID: 37782609 DOI: 10.1109/tvcg.2023.3320213] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/04/2023]
Abstract
In immersive Audio Augmented Reality, a virtual sound source should be indistinguishable from the existing real ones. This property can be evaluated with the co-immersion criterion, which encompasses scenes constituted by arbitrary configurations of real and virtual objects. Thus, we introduce the term Audio Augmented Virtuality (AAV) to describe a fully virtual environment consisting of auditory content captured from the real world, augmented by synthetic sound generation. We propose an experimental design in AAV investigating how simplified late reverberation (LR) affects the co-immersion of a sound source. Participants listened to simultaneous virtual speakers dynamically rendered through spatial Room Impulse Responses, and were asked to detect the presence of an impostor, i.e., a speaker rendered with one of two simplified LR conditions. Detection rates were found to be close to chance level, especially for one condition, suggesting a limited influence on co-immersion of the simplified LR in the evaluated AAV scenes. This methodology can be straightforwardly extended and applied to different acoustics scenes, complexities, i.e., the number of simultaneous speakers, and rendering parameters in order to further investigate the requirements for immersive audio technologies in AAR and AAV applications.
Collapse
|
6
|
Andrejková G, Best V, Kopčo N. Time scales of adaptation to context in horizontal sound localization. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 154:2191-2202. [PMID: 37815410 PMCID: PMC10567122 DOI: 10.1121/10.0021304] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/24/2022] [Revised: 08/01/2023] [Accepted: 09/19/2023] [Indexed: 10/11/2023]
Abstract
Psychophysical experiments explored how the repeated presentation of a context, consisting of an adaptor and a target, induces plasticity in the localization of an identical target presented alone on interleaved trials. The plasticity, and its time course, was examined both in a classroom and in an anechoic chamber. Adaptors and targets were 2 ms noise clicks and listeners were tasked with localizing the targets while ignoring the adaptors (when present). The context was either simple, consisting of a single-click adaptor and a target, or complex, containing either a single-click or an eight-click adaptor that varied from trial to trial. The adaptor was presented either from a frontal or a lateral location, fixed within a run. The presence of context caused responses to the isolated targets to be displaced up to 14° away from the adaptor location. This effect was stronger and slower if the context was complex, growing over the 5 min duration of the runs. Additionally, the simple context buildup had a slower onset in the classroom. Overall, the results illustrate that sound localization is subject to slow adaptive processes that depend on the spatial and temporal structure of the context and on the level of reverberation in the environment.
Collapse
Affiliation(s)
- Gabriela Andrejková
- Institute of Computer Science, Faculty of Science, P. J. Šafárik University, Košice, 04001, Slovakia
| | - Virginia Best
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - Norbert Kopčo
- Institute of Computer Science, Faculty of Science, P. J. Šafárik University, Košice, 04001, Slovakia
| |
Collapse
|
7
|
Popov VV, Nechaev DI, Supin AY, Sysueva EV. Monaural and dichotic forward masking in the dolphin's auditory system. J Comp Physiol A Neuroethol Sens Neural Behav Physiol 2023:10.1007/s00359-023-01675-1. [PMID: 37770756 DOI: 10.1007/s00359-023-01675-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/23/2023] [Revised: 09/08/2023] [Accepted: 09/09/2023] [Indexed: 09/30/2023]
Abstract
Short-latency auditory-evoked potentials (AEPs) were recorded non-invasively in the bottlenose dolphin Tursiops truncatus. The stimuli were two sound clicks that were played either monaurally (both clicks to one and the same acoustic window) or dichotically (the leading stimulus (masker) to one acoustic window and the delayed stimulus (test) to the other window). The ratio of the levels of the two stimuli was 0, 10, or 20 dB (at 10 and 20 dB, the leading stimulus was of a higher level). The inter-stimulus intervals (ISIs) varied from 0.15 to 10 ms. The test response magnitude was assessed by correlation analysis as a percentage of the control (non-masked) response. At monaural stimulation, the test response was of a constant magnitude (5-6% of the control) at ISIs of 0.15-0.3 ms and recovered at longer ISIs. At dichotic stimulation, the deepest suppression of the test response occurred at ISIs of 0.5-0.7 ms. The response was slightly suppressed at short ISIs (0.15-0.3 ms) and recovered at ISIs longer than 0.5-0.7 ms. The relation of parameters of the forward masking to echolocation in dolphins is discussed.
Collapse
Affiliation(s)
- Vladimir V Popov
- A.N. Severtsov Institute of Ecology and Evolution, Russian Academy of Sciences, Leninsky Prospekt, 33, 119071, Moscow, Russia
| | - Dmitry I Nechaev
- A.N. Severtsov Institute of Ecology and Evolution, Russian Academy of Sciences, Leninsky Prospekt, 33, 119071, Moscow, Russia
| | - Alexander Ya Supin
- A.N. Severtsov Institute of Ecology and Evolution, Russian Academy of Sciences, Leninsky Prospekt, 33, 119071, Moscow, Russia
| | - Evgeniya V Sysueva
- A.N. Severtsov Institute of Ecology and Evolution, Russian Academy of Sciences, Leninsky Prospekt, 33, 119071, Moscow, Russia.
| |
Collapse
|
8
|
Higgins NC, Pupo DA, Ozmeral EJ, Eddins DA. Head movement and its relation to hearing. Front Psychol 2023; 14:1183303. [PMID: 37448716 PMCID: PMC10338176 DOI: 10.3389/fpsyg.2023.1183303] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2023] [Accepted: 06/07/2023] [Indexed: 07/15/2023] Open
Abstract
Head position at any point in time plays a fundamental role in shaping the auditory information that reaches a listener, information that continuously changes as the head moves and reorients to different listening situations. The connection between hearing science and the kinesthetics of head movement has gained interest due to technological advances that have increased the feasibility of providing behavioral and biological feedback to assistive listening devices that can interpret movement patterns that reflect listening intent. Increasing evidence also shows that the negative impact of hearing deficits on mobility, gait, and balance may be mitigated by prosthetic hearing device intervention. Better understanding of the relationships between head movement, full body kinetics, and hearing health, should lead to improved signal processing strategies across a range of assistive and augmented hearing devices. The purpose of this review is to introduce the wider hearing community to the kinesiology of head movement and to place it in the context of hearing and communication with the goal of expanding the field of ecologically-specific listener behavior.
Collapse
Affiliation(s)
- Nathan C. Higgins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
| | - Daniel A. Pupo
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
- School of Aging Studies, University of South Florida, Tampa, FL, United States
| | - Erol J. Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
| | - David A. Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
| |
Collapse
|
9
|
Brown AD, Hayward T, Portfors CV, Coffin AB. On the value of diverse organisms in auditory research: From fish to flies to humans. Hear Res 2023; 432:108754. [PMID: 37054531 PMCID: PMC10424633 DOI: 10.1016/j.heares.2023.108754] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/16/2022] [Revised: 02/28/2023] [Accepted: 03/27/2023] [Indexed: 03/31/2023]
Abstract
Historically, diverse organisms have contributed to our understanding of auditory function. In recent years, the laboratory mouse has become the prevailing non-human model in auditory research, particularly for biomedical studies. There are many questions in auditory research for which the mouse is the most appropriate (or the only) model system available. But mice cannot provide answers for all auditory problems of basic and applied importance, nor can any single model system provide a synthetic understanding of the diverse solutions that have evolved to facilitate effective detection and use of acoustic information. In this review, spurred by trends in funding and publishing and inspired by parallel observations in other domains of neuroscience, we highlight a few examples of the profound impact and lasting benefits of comparative and basic organismal research in the auditory system. We begin with the serendipitous discovery of hair cell regeneration in non-mammalian vertebrates, a finding that has fueled an ongoing search for pathways to hearing restoration in humans. We then turn to the problem of sound source localization - a fundamental task that most auditory systems have been compelled to solve despite large variation in the magnitudes and kinds of spatial acoustic cues available, begetting varied direction-detecting mechanisms. Finally, we consider the power of work in highly specialized organisms to reveal exceptional solutions to sensory problems - and the diverse returns of deep neuroethological inquiry - via the example of echolocating bats. Throughout, we consider how discoveries made possible by comparative and curiosity-driven organismal research have driven fundamental scientific, biomedical, and technological advances in the auditory field.
Collapse
Affiliation(s)
- Andrew D Brown
- Department of Speech and Hearing Sciences, University of Washington, 1417 NE 42nd St, Seattle, WA, 98105 USA; Virginia-Merrill Bloedel Hearing Research Center, University of Washington, 1701 NE Columbia Rd, Seattle, WA, 98195 USA.
| | - Tamasen Hayward
- College of Arts and Sciences, Washington State University, 14204 NE Salmon Creek Ave, Vancouver, WA 98686 USA
| | - Christine V Portfors
- School of Biological Sciences, Washington State University, 14204 NE Salmon Creek Ave, Vancouver, WA 98686 USA
| | - Allison B Coffin
- College of Arts and Sciences, Washington State University, 14204 NE Salmon Creek Ave, Vancouver, WA 98686 USA; School of Biological Sciences, Washington State University, 14204 NE Salmon Creek Ave, Vancouver, WA 98686 USA; Department of Integrative Physiology and Neuroscience, Washington State University, 14204 NE Salmon Creek Ave, Vancouver, WA 98686 USA.
| |
Collapse
|
10
|
Willmore BDB, King AJ. Adaptation in auditory processing. Physiol Rev 2023; 103:1025-1058. [PMID: 36049112 PMCID: PMC9829473 DOI: 10.1152/physrev.00011.2022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/21/2023] Open
Abstract
Adaptation is an essential feature of auditory neurons, which reduces their responses to unchanging and recurring sounds and allows their response properties to be matched to the constantly changing statistics of sounds that reach the ears. As a consequence, processing in the auditory system highlights novel or unpredictable sounds and produces an efficient representation of the vast range of sounds that animals can perceive by continually adjusting the sensitivity and, to a lesser extent, the tuning properties of neurons to the most commonly encountered stimulus values. Together with attentional modulation, adaptation to sound statistics also helps to generate neural representations of sound that are tolerant to background noise and therefore plays a vital role in auditory scene analysis. In this review, we consider the diverse forms of adaptation that are found in the auditory system in terms of the processing levels at which they arise, the underlying neural mechanisms, and their impact on neural coding and perception. We also ask what the dynamics of adaptation, which can occur over multiple timescales, reveal about the statistical properties of the environment. Finally, we examine how adaptation to sound statistics is influenced by learning and experience and changes as a result of aging and hearing loss.
Collapse
Affiliation(s)
- Ben D. B. Willmore
- Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, United Kingdom
| | - Andrew J. King
- Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
11
|
Dirks CE, Nelson PB, Oxenham AJ. No Benefit of Deriving Cochlear-Implant Maps From Binaural Temporal-Envelope Sensitivity for Speech Perception or Spatial Hearing Under Single-Sided Deafness. Ear Hear 2022; 43:310-322. [PMID: 34291758 PMCID: PMC8770730 DOI: 10.1097/aud.0000000000001094] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES This study tested whether speech perception and spatial acuity improved in people with single-sided deafness and a cochlear implant (SSD+CI) when the frequency allocation table (FAT) of the CI was adjusted to optimize frequency-dependent sensitivity to binaural disparities. DESIGN Nine SSD+CI listeners with at least 6 months of CI listening experience participated. Individual experimental FATs were created to best match the frequency-to-place mapping across ears using either sensitivity to binaural temporal-envelope disparities or estimated insertion depth. Spatial localization ability was measured, along with speech perception in spatially collocated or separated noise, first with the clinical FATs and then with the experimental FATs acutely and at 2-month intervals for 6 months. Listeners then returned to the clinical FATs and were retested acutely and after 1 month to control for long-term learning effects. RESULTS The experimental FAT varied between listeners, differing by an average of 0.15 octaves from the clinical FAT. No significant differences in performance were observed in any of the measures between the experimental FAT after 6 months and the clinical FAT one month later, and no clear relationship was found between the size of the frequency-allocation shift and perceptual changes. CONCLUSION Adjusting the FAT to optimize sensitivity to interaural temporal-envelope disparities did not improve localization or speech perception. The clinical frequency-to-place alignment may already be sufficient, given the inherently poor spectral resolution of CIs. Alternatively, other factors, such as temporal misalignment between the two ears, may need to be addressed before any benefits of spectral alignment can be observed.
Collapse
Affiliation(s)
- Coral E Dirks
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, Minnesota, USA
- Center for Applied and Translational Sensory Sciences, Department of Psychology, University of Minnesota, Minneapolis, Minnesota, USA
| | - Peggy B Nelson
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, Minnesota, USA
- Center for Applied and Translational Sensory Sciences, Department of Psychology, University of Minnesota, Minneapolis, Minnesota, USA
| | - Andrew J Oxenham
- Center for Applied and Translational Sensory Sciences, Department of Psychology, University of Minnesota, Minneapolis, Minnesota, USA
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota, USA
| |
Collapse
|
12
|
Deep neural network models of sound localization reveal how perception is adapted to real-world environments. Nat Hum Behav 2022; 6:111-133. [PMID: 35087192 PMCID: PMC8830739 DOI: 10.1038/s41562-021-01244-z] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2020] [Accepted: 10/29/2021] [Indexed: 11/15/2022]
Abstract
Mammals localize sounds using information from their two ears.
Localization in real-world conditions is challenging, as echoes provide
erroneous information, and noises mask parts of target sounds. To better
understand real-world localization we equipped a deep neural network with human
ears and trained it to localize sounds in a virtual environment. The resulting
model localized accurately in realistic conditions with noise and reverberation.
In simulated experiments, the model exhibited many features of human spatial
hearing: sensitivity to monaural spectral cues and interaural time and level
differences, integration across frequency, biases for sound onsets, and limits
on localization of concurrent sources. But when trained in unnatural
environments without either reverberation, noise, or natural sounds, these
performance characteristics deviated from those of humans. The results show how
biological hearing is adapted to the challenges of real-world environments and
illustrate how artificial neural networks can reveal the real-world constraints
that shape perception.
Collapse
|
13
|
Neidhardt A, Schneiderwind C, Klein F. Perceptual Matching of Room Acoustics for Auditory Augmented Reality in Small Rooms - Literature Review and Theoretical Framework. Trends Hear 2022; 26:23312165221092919. [PMID: 35505625 PMCID: PMC9073123 DOI: 10.1177/23312165221092919] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/04/2022] Open
Abstract
For the realization of auditory augmented reality (AAR), it is important that the room acoustical properties of the virtual elements are perceived in agreement with the acoustics of the actual environment. This perceptual matching of room acoustics is the subject reviewed in this paper. Realizations of AAR that fulfill the listeners’ expectations were achieved based on pre-characterization of the room acoustics, for example, by measuring acoustic impulse responses or creating detailed room models for acoustic simulations. For future applications, the goal is to realize an online adaptation in (close to) real-time. Perfect physical matching is hard to achieve with these practical constraints. For this reason, an understanding of the essential psychoacoustic cues is of interest and will help to explore options for simplifications. This paper reviews a broad selection of previous studies and derives a theoretical framework to examine possibilities for psychoacoustical optimization of room acoustical matching.
Collapse
|
14
|
Peng ZE, Pausch F, Fels J. Spatial release from masking in reverberation for school-age children. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 150:3263. [PMID: 34852617 PMCID: PMC8730369 DOI: 10.1121/10.0006752] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/15/2021] [Revised: 09/29/2021] [Accepted: 09/29/2021] [Indexed: 05/06/2023]
Abstract
Understanding speech in noisy environments, such as classrooms, is a challenge for children. When a spatial separation is introduced between the target and masker, as compared to when both are co-located, children demonstrate intelligibility improvement of the target speech. Such intelligibility improvement is known as spatial release from masking (SRM). In most reverberant environments, binaural cues associated with the spatial separation are distorted; the extent to which such distortion will affect children's SRM is unknown. Two virtual acoustic environments with reverberation times between 0.4 s and 1.1 s were compared. SRM was measured using a spatial separation with symmetrically displaced maskers to maximize access to binaural cues. The role of informational masking in modulating SRM was investigated through voice similarity between the target and masker. Results showed that, contradictory to previous developmental findings on free-field SRM, children's SRM in reverberation has not yet reached maturity in the 7-12 years age range. When reducing reverberation, an SRM improvement was seen in adults but not in children. Our findings suggest that, even though school-age children have access to binaural cues that are distorted in reverberation, they demonstrate immature use of such cues for speech-in-noise perception, even in mild reverberation.
Collapse
Affiliation(s)
- Z Ellen Peng
- Institute for Hearing Technology and Acoustics, RWTH Aachen University, Kopernikusstrasse 5, 52074 Aachen, Germany
| | - Florian Pausch
- Institute for Hearing Technology and Acoustics, RWTH Aachen University, Kopernikusstrasse 5, 52074 Aachen, Germany
| | - Janina Fels
- Institute for Hearing Technology and Acoustics, RWTH Aachen University, Kopernikusstrasse 5, 52074 Aachen, Germany
| |
Collapse
|
15
|
Legett HD, Aihara I, Bernal XE. The dual benefits of synchronized mating signals in a Japanese treefrog: attracting mates and manipulating predators. Philos Trans R Soc Lond B Biol Sci 2021; 376:20200340. [PMID: 34420389 PMCID: PMC8380967 DOI: 10.1098/rstb.2020.0340] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 03/08/2021] [Indexed: 12/31/2022] Open
Abstract
In dense mating aggregations, such as leks and choruses, acoustic signals produced by competing male conspecifics often overlap in time. When signals overlap at a fine temporal scale the ability of females to discriminate between individual signals is reduced. Yet, despite this cost, males of some species deliberately overlap their signals with those of conspecifics, synchronizing signal production in the chorus. Here, we investigate two hypotheses of synchronized mating signals in a Japanese treefrog (Buergeria japonica): (1) increased female attraction to the chorus (the beacon effect hypothesis) and (2) reduced attraction of eavesdropping predators (the eavesdropper avoidance hypothesis). Our results from playback experiments on female frogs and eavesdropping micropredators (midges and mosquitoes) support both hypotheses. Signal transmission and female phonotaxis experiments suggest that away from the chorus, synchronized calls are more attractive to females than unsynchronized calls. At the chorus, however, eavesdroppers are less attracted to calls that closely follow an initial call, while female attraction to individual signals is not affected. Therefore, synchronized signalling likely benefits male B. japonica by both increasing attraction of females to the chorus and reducing eavesdropper attacks. These findings highlight how multiple selective pressures likely promoted the evolution and maintenance of this behaviour. This article is part of the theme issue 'Synchrony and rhythm interaction: from the brain to behavioural ecology'.
Collapse
Affiliation(s)
- Henry D. Legett
- Department of Biological Sciences, Purdue University, West Lafayette, IN 47907, USA
| | - Ikkyu Aihara
- Graduate School of Systems and Information Engineering, University of Tsukuba, Ibaraki 305-8573, Japan
| | - X. E. Bernal
- Department of Biological Sciences, Purdue University, West Lafayette, IN 47907, USA
- Smithsonian Tropical Research Institute, Apartado Postal 0843-03092, Balboa, Ancón, Panama
| |
Collapse
|
16
|
Baltzell LS, Best V. High-resolution temporal weighting of interaural time differences in speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 150:1311. [PMID: 34470281 PMCID: PMC8561715 DOI: 10.1121/10.0005934] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/02/2021] [Revised: 07/29/2021] [Accepted: 07/30/2021] [Indexed: 06/13/2023]
Abstract
Previous studies have shown that for high-rate click trains and low-frequency pure tones, interaural time differences (ITDs) at the onset of stimulus contribute most strongly to the overall lateralization percept (receive the largest perceptual weight). Previous studies have also shown that when these stimuli are modulated, ITDs during the rising portion of the modulation cycle receive increased perceptual weight. Baltzell, Cho, Swaminathan, and Best [(2020). J. Acoust. Soc. Am. 147, 3883-3894] measured perceptual weights for a pair of spoken words ("two" and "eight"), and found that word-initial phonemes receive larger weight than word-final phonemes, suggesting a "word-onset dominance" for speech. Generalizability of this conclusion was limited by a coarse temporal resolution and limited stimulus set. In the present study, temporal weighting functions (TWFs) were measured for four spoken words ("two," "eight," "six," and "nine"). Stimuli were partitioned into 30-ms bins, ITDs were applied independently to each bin, and lateralization judgements were obtained. TWFs were derived using a hierarchical regression model. Results suggest that "word-initial" onset dominance does not generalize across words and that TWFs depend in part on acoustic changes throughout the stimulus. Two model-based predictions were generated to account for observed TWFs, but neither could fully account for the perceptual data.
Collapse
Affiliation(s)
- Lucas S Baltzell
- Department of Speech, Language, and Hearing Sciences, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| | - Virginia Best
- Department of Speech, Language, and Hearing Sciences, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| |
Collapse
|
17
|
Sensitivity to interaural time differences in the inferior colliculus of cochlear implanted rats with or without hearing experience. Hear Res 2021; 408:108305. [PMID: 34315027 DOI: 10.1016/j.heares.2021.108305] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/13/2021] [Revised: 06/24/2021] [Accepted: 06/29/2021] [Indexed: 01/11/2023]
Abstract
For deaf patients cochlear implants (CIs) can restore substantial amounts of functional hearing. However, binaural hearing, and in particular, the perception of interaural time differences (ITDs) with current CIs has been found to be notoriously poor, especially in the event of early hearing loss. One popular hypothesis for these deficits posits that a lack of early binaural experience may be a principal cause of poor ITD perception in pre-lingually deaf CI patients. This is supported by previous electrophysiological studies done in neonatally deafened, bilateral CI-stimulated animals showing reduced ITD sensitivity. However, we have recently demonstrated that neonatally deafened CI rats can quickly learn to discriminate microsecond ITDs under optimized stimulation conditions which suggests that the inability of human CI users to make use of ITDs is not due to lack of binaural hearing experience during development. In the study presented here, we characterized ITD sensitivity and tuning of inferior colliculus neurons under bilateral CI stimulation of neonatally deafened and hearing experienced rats. The hearing experienced rats were not deafened prior to implantation. Both cohorts were implanted bilaterally between postnatal days 64-77 and recorded immediately following surgery. Both groups showed comparably large proportions of ITD sensitive multi-units in the inferior colliculus (Deaf: 84.8%, Hearing: 82.5%), and the strength of ITD tuning, quantified as mutual information between response and stimulus ITD, was independent of hearing experience. However, the shapes of tuning curves differed substantially between both groups. We observed four main clusters of tuning curves - trough, contralateral, central, and ipsilateral tuning. Interestingly, over 90% of multi-units for hearing experienced rats showed predominantly contralateral tuning, whereas as many as 50% of multi-units in neonatally deafened rats were centrally tuned. However, when we computed neural d' scores to predict likely limits on performance in sound lateralization tasks, we did not find that these differences in tuning shapes predicted worse psychoacoustic performance for the neonatally deafened animals. We conclude that, at least in rats, substantial amounts of highly precise, "innate" ITD sensitivity can be found even after profound hearing loss throughout infancy. However, ITD tuning curve shapes appear to be strongly influenced by auditory experience although substantial lateralization encoding is present even in its absence.
Collapse
|
18
|
Brown AD, Tollin DJ. Effects of interaural decoherence on sensitivity to interaural level differences across frequency. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:4630. [PMID: 34241434 PMCID: PMC8249038 DOI: 10.1121/10.0005123] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/22/2020] [Revised: 04/27/2021] [Accepted: 05/10/2021] [Indexed: 05/19/2023]
Abstract
The interaural level difference (ILD) is a robust indicator of sound source azimuth, and human ILD sensitivity persists under conditions that degrade normally-dominant interaural time difference (ITD) cues. Nonetheless, ILD sensitivity varies somewhat with both stimulus frequency and interaural correlation (coherence). To further investigate the combined binaural perceptual influence of these variables, the present study assessed ILD sensitivity at frequencies 250-4000 Hz using stimuli of varied interaural correlation. In the first of two experiments, ILD discrimination thresholds were modestly elevated, and subjective lateralization slightly reduced, for both half-correlated and uncorrelated narrowband noise tokens relative to correlated tokens. Different from thresholds in the correlated condition, which were worst at 1000 Hz [Grantham, D.W. (1984). J. Acoust. Soc. Am. 75, 1191-1194], thresholds in the decorrelated conditions were independent of frequency. However, intrinsic envelope fluctuations in narrowband stimuli caused moment-to-moment variation of the nominal ILD, complicating interpretation of measured thresholds. Thus, a second experiment employed low-fluctuation noise tokens, revealing a clear effect of interaural decoherence per se that was strongly frequency-dependent, decreasing in magnitude from low to high frequencies. Measurements are consistent with known integration times in ILD-sensitive neurons and also suggest persistent influences of covert ITD cues in putative "ILD" tasks.
Collapse
Affiliation(s)
- Andrew D Brown
- Department of Speech and Hearing Sciences, University of Washington, 1417 Northeast 42nd Street, Seattle, Washington 98105, USA
| | - Daniel J Tollin
- Department of Physiology and Biophysics, University of Colorado School of Medicine, Aurora, Colorado 80045, USA
| |
Collapse
|
19
|
Does Double Biofeedback Affect Functional Hemispheric Asymmetry and Activity? A Pilot Study. Symmetry (Basel) 2021. [DOI: 10.3390/sym13060937] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
In the current pilot study, we attempt to find out how double neurofeedback influences functional hemispheric asymmetry and activity. We examined 30 healthy participants (8 males; 22 females, mean age = 29; SD = 8). To measure functional hemispheric asymmetry and activity, we used computer laterometry in the ‘two-source’ lead-lag dichotic paradigm. Double biofeedback included 8 min of EEG oscillation recording with five minutes of basic mode. During the basic mode, the current amplitude of the EEG oscillator gets transformed into feedback sounds while the current amplitude of alpha EEG oscillator is used to modulate the intensity of light signals. Double neurofeedback did not directly influence the asymmetry itself but accelerated individual sound perception characteristics during dichotic listening in the preceding effect paradigm. Further research is needed to investigate the effect of double neurofeedback training on functional brain activity and asymmetry, taking into account participants’ age, gender, and motivation.
Collapse
|
20
|
Tirado C, Gerdfeldter B, Nilsson ME. Individual differences in the ability to access spatial information in lag-clicks. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:2963. [PMID: 34241133 DOI: 10.1121/10.0004821] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/07/2020] [Accepted: 04/11/2021] [Indexed: 06/13/2023]
Abstract
It may be difficult to determine whether a dichotic lag-click points to the left or right when preceded by a diotic lead-click. Previous research suggests that this loss of spatial information is most prominent at inter-click intervals (ICIs) <10 ms. However, a study by Nilsson, Tirado, and Szychowska [(2019). J. Acoust. Soc. Am. 145, 512-524] found support for loss of spatial information in lag-clicks at much longer ICIs using a stimulus setup differing from those in previous research. The present study used a setup similar to that of the Nilsson, Tirado, and Szychowska study [(2019). J. Acoust. Soc. Am. 145, 512-524] to measure 13 listeners' ability to lateralize (left versus right) and detect (present versus absent) the lag-click in lead-lag click pairs with ICIs of 6-48 ms. The main finding was distinct individual differences in performance. Some listeners could lateralize lag-clicks all the way down to their detection threshold, whereas others had lateralization thresholds substantially higher than their detection thresholds, i.e., they could not lateralize lag-clicks that they could easily detect. Two such listeners trained for 30 days and managed to improve their lateralization thresholds to reach their detection thresholds, but only at longer ICIs (>20 ms), suggesting different mechanisms underlying lag-click lateralization at short versus long ICIs.
Collapse
Affiliation(s)
- Carlos Tirado
- Gösta Ekman Laboratory, Department of Psychology, Stockholm University, Stockholm, Sweden
| | - Billy Gerdfeldter
- Gösta Ekman Laboratory, Department of Psychology, Stockholm University, Stockholm, Sweden
| | - Mats E Nilsson
- Gösta Ekman Laboratory, Department of Psychology, Stockholm University, Stockholm, Sweden
| |
Collapse
|
21
|
Haywood NR, Undurraga JA, McAlpine D. The influence of envelope shape on the lateralization of amplitude-modulated, low-frequency sound. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:3133. [PMID: 34241105 DOI: 10.1121/10.0004788] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/18/2020] [Accepted: 04/06/2021] [Indexed: 06/13/2023]
Abstract
For abruptly gated sound, interaural time difference (ITD) cues at onset carry greater perceptual weight than those following. This research explored how envelope shape influences such carrier ITD weighting. Experiment 1 assessed the perceived lateralization of a tonal binaural beat that transitioned through ITD (diotic envelope, mean carrier frequency of 500 Hz). Listeners' left/right lateralization judgments were compared to those for static-ITD tones. For an 8 Hz sinusoidally amplitude-modulated envelope, ITD cues 24 ms after onset well-predicted reported sidedness. For an equivalent-duration "abrupt" envelope, which was unmodulated besides 20-ms onset/offset ramps, reported sidedness corresponded to ITDs near onset (e.g., 6 ms). However, unlike for sinusoidal amplitude modulation, ITDs toward offset seemingly also influenced perceived sidedness. Experiment 2 adjusted the duration of the offset ramp (25-75 ms) and found evidence for such offset weighting only for the most abrupt ramp tested. In experiment 3, an ITD was imposed on a brief segment of otherwise diotic filtered noise. Listeners discriminated right- from left-leading ITDs. In sinusoidal amplitude modulation, thresholds were lowest when the ITD segment occurred during rising amplitude. For the abrupt envelope, the lowest thresholds were observed when the segment occurred at either onset or offset. These experiments demonstrate the influence of envelope profile on carrier ITD sensitivity.
Collapse
Affiliation(s)
- Nicholas R Haywood
- Department of Linguistics, Faculty of Medicine, Health and Human Sciences, Macquarie Hearing, Macquarie University, Sydney, New South Wales 2109, Australia
| | - Jaime A Undurraga
- Department of Linguistics, Faculty of Medicine, Health and Human Sciences, Macquarie Hearing, Macquarie University, Sydney, New South Wales 2109, Australia
| | - David McAlpine
- Department of Linguistics, Faculty of Medicine, Health and Human Sciences, Macquarie Hearing, Macquarie University, Sydney, New South Wales 2109, Australia
| |
Collapse
|
22
|
Auditory Brainstem Models: Adapting Cochlear Nuclei Improve Spatial Encoding by the Medial Superior Olive in Reverberation. J Assoc Res Otolaryngol 2021; 22:289-318. [PMID: 33861395 DOI: 10.1007/s10162-021-00797-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2020] [Accepted: 03/22/2021] [Indexed: 10/21/2022] Open
Abstract
Listeners typically perceive a sound as originating from the direction of its source, even as direct sound is followed milliseconds later by reflected sound from multiple different directions. Early-arriving sound is emphasised in the ascending auditory pathway, including the medial superior olive (MSO) where binaural neurons encode the interaural-time-difference (ITD) cue for spatial location. Perceptually, weighting of ITD conveyed during rising sound energy is stronger at 600 Hz than at 200 Hz, consistent with the minimum stimulus rate for binaural adaptation, and with the longer reverberation times at 600 Hz, compared with 200 Hz, in many natural outdoor environments. Here, we computationally explore the combined efficacy of adaptation prior to the binaural encoding of ITD cues, and excitatory binaural coincidence detection within MSO neurons, in emphasising ITDs conveyed in early-arriving sound. With excitatory inputs from adapting, nonlinear model spherical bushy cells (SBCs) of the bilateral cochlear nuclei, a nonlinear model MSO neuron with low-threshold potassium channels reproduces the rate-dependent emphasis of rising vs. peak sound energy in ITD encoding; adaptation is equally effective in the model MSO. Maintaining adaptation in model SBCs, and adjusting membrane speed in model MSO neurons, 'left' and 'right' populations of computationally efficient, linear model SBCs and MSO neurons reproduce this stronger weighting of ITD conveyed during rising sound energy at 600 Hz compared to 200 Hz. This hemispheric population model demonstrates a link between strong weighting of spatial information during rising sound energy, and correct unambiguous lateralisation of a speech source in reverberation.
Collapse
|
23
|
Engel I, Henry C, Amengual Garí SV, Robinson PW, Picinali L. Perceptual implications of different Ambisonics-based methods for binaural reverberation. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:895. [PMID: 33639797 DOI: 10.1121/10.0003437] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/16/2020] [Accepted: 01/11/2021] [Indexed: 06/12/2023]
Abstract
Reverberation is essential for the realistic auralisation of enclosed spaces. However, it can be computationally expensive to render with high fidelity and, in practice, simplified models are typically used to lower costs while preserving perceived quality. Ambisonics-based methods may be employed to this purpose as they allow us to render a reverberant sound field more efficiently by limiting its spatial resolution. The present study explores the perceptual impact of two simplifications of Ambisonics-based binaural reverberation that aim to improve efficiency. First, a "hybrid Ambisonics" approach is proposed in which the direct sound path is generated by convolution with a spatially dense head related impulse response set, separately from reverberation. Second, the reverberant virtual loudspeaker method (RVL) is presented as a computationally efficient approach to dynamically render binaural reverberation for multiple sources with the potential limitation of inaccurately simulating listener's head rotations. Numerical and perceptual evaluations suggest that the perceived quality of hybrid Ambisonics auralisations of two measured rooms ceased to improve beyond the third order, which is a lower threshold than what was found by previous studies in which the direct sound path was not processed separately. Additionally, RVL is shown to produce auralisations with comparable perceived quality to Ambisonics renderings.
Collapse
Affiliation(s)
- Isaac Engel
- Dyson School of Design Engineering, Imperial College London, London SW7 2DB, United Kingdom
| | - Craig Henry
- Dyson School of Design Engineering, Imperial College London, London SW7 2DB, United Kingdom
| | | | | | - Lorenzo Picinali
- Dyson School of Design Engineering, Imperial College London, London SW7 2DB, United Kingdom
| |
Collapse
|
24
|
van Bentum GC, van Wanrooij MM, van Opstal AJ. Spatiotemporal factors influence sound-source segregation in localization behavior. J Neurophysiol 2020; 125:556-567. [PMID: 33378250 DOI: 10.1152/jn.00184.2020] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
To program a goal-directed response in the presence of acoustic reflections, the audio-motor system should suppress the detection of time-delayed sources. We examined the effects of spatial separation and interstimulus delay on the ability of human listeners to localize a pair of broadband sounds in the horizontal plane. Participants indicated how many sounds were heard and where these were perceived by making one or two head-orienting localization responses. Results suggest that perceptual fusion of the two sounds depends on delay and spatial separation. Leading and lagging stimuli in close spatial proximity required longer stimulus delays to be perceptually separated than those further apart. Whenever participants heard one sound, their localization responses for synchronous sounds were oriented to a weighted average of both source locations. For short delays, responses were directed toward the leading stimulus location. Increasing spatial separation enhanced this effect. For longer delays, responses were again directed toward a weighted average. When participants perceived two sounds, the first and the second response were directed to either of the leading and lagging source locations. Perceived locations were interchanged often in their temporal order (in ∼40% of trials). We show that the percept of two sounds occurring requires sufficient spatiotemporal separation, after which localization can be performed with high accuracy. We propose that the percept of temporal order of two concurrent sounds results from a different process than localization and discuss how dynamic lateral excitatory-inhibitory interactions within a spatial sensorimotor map could explain the findings.NEW & NOTEWORTHY Sound localization requires spectral and temporal processing of implicit acoustic cues, and is seriously challenged when multiple sources coincide closely in space and time. We systematically varied spatial-temporal disparities for two sounds and instructed listeners to generate goal-directed head movements. We found that even when the auditory system has accurate representations of both sources, it still has trouble to decide whether the scene contained one or two sounds, and in which order they appeared.
Collapse
Affiliation(s)
- Guus C van Bentum
- Department of Biophysics, Donders Center for Neuroscience, Donders Institute for Brain, Cognition, and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands
| | - Marc M van Wanrooij
- Department of Biophysics, Donders Center for Neuroscience, Donders Institute for Brain, Cognition, and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands
| | - A John van Opstal
- Department of Biophysics, Donders Center for Neuroscience, Donders Institute for Brain, Cognition, and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands
| |
Collapse
|
25
|
Haqqee Z, Valdizón-Rodríguez R, Faure PA. High frequency sensitivity to interaural onset time differences in the bat inferior colliculus. Hear Res 2020; 400:108133. [PMID: 33340969 DOI: 10.1016/j.heares.2020.108133] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/01/2020] [Revised: 11/20/2020] [Accepted: 11/25/2020] [Indexed: 02/02/2023]
Abstract
Many neurons in the auditory midbrain are tuned to binaural cues. Two prominent binaural cues are the interaural level difference (ILD) and the interaural time difference (ITD). The ITD cue can further be subdivided into the ongoing envelope ITD cues and transient onset ITD cues. More is known about the sensitivity of single neurons to ongoing envelope ITDs compared to transient onset ITDs in the mammalian auditory system, particularly in bats. The current study examines the response properties of single neurons in the inferior colliculus (IC) of the big brown bat, Eptesicus fuscus, to onset ITDs in response to high frequency pure tones. Measures of neurons' dynamic ITD response revealed an average change of 36% of its maximum response within the behaviorally relevant range of ITDs (±50 µs). Across all IC neurons, we measured an average time-intensity trading ratio of 30 µs/dB in the sensitivity of the ITD response function to changing ILDs. Minimum and maximum ITD responses were clustered within a narrow range of ITDs. The average peak in the ITD response function was at 268 µs, a finding that is consistent with other non-echolocating mammals. Some ITD-sensitive neurons also showed weak facilitation of maximum response during binaural stimulation, compared to monaural stimulation. These results suggest that echolocating bats possess the potential to use onset ITD cues to assist in the azimuthal sound localization of ultrasonic frequencies.
Collapse
Affiliation(s)
- Zeeshan Haqqee
- Department of Psychology, Neuroscience & Behaviour, McMaster University, Hamilton, ON, L8S 4K1, Canada
| | | | - Paul A Faure
- Department of Psychology, Neuroscience & Behaviour, McMaster University, Hamilton, ON, L8S 4K1, Canada.
| |
Collapse
|
26
|
Attribute capture underlying the precedence effect in rats. Hear Res 2020; 400:108096. [PMID: 33212323 DOI: 10.1016/j.heares.2020.108096] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/17/2020] [Revised: 09/19/2020] [Accepted: 10/13/2020] [Indexed: 11/22/2022]
Abstract
In a reverberant environment, humans with normal hearing can perceptually fuse the soundwave from a source with its reflections off nearby surfaces into a single auditory image, whose location appears to be around the source. This phenomenon is called the precedence effect, which is based on the perceptual capture of the reflected (lagging) sounds' attributes by the direct wave from the source. Using the paradigm of attentional modulation of the prepulse inhibition (PPI) of the startle reflex, with both the prepulse-feature specificity and the perceived-prepulse-location specificity, this study was to examine whether the perceptual attribute capture underlying the precedence effect occurs in rats. One broadband continuous noise was delivered by each of two spatially separated left and right loudspeakers with a 1-ms inter-loudspeaker delay. A silent gap was embedded in one of the two noises as the prepulse stimulus. The results showed that regardless of whether the gap was physically in the leading or lagging noise when the leading noise was either the left or right one, fear conditioning the gap enhanced PPI only when the leading noise was delivered from the loudspeaker that was the leading but not the lagging loudspeaker during the conditioning, indicating that due to the spatial specificity (either left or right) in the attentional enhancement of PPI, the perceived location of the conditioned gap was always on the leading side even though the gap was physically on the lagging side. Thus, rats have the same perceptual ability of attribute capture, thereby experiencing the auditory precedence effect as humans.
Collapse
|
27
|
Pavão R, Sussman ES, Fischer BJ, Peña JL. Natural ITD statistics predict human auditory spatial perception. eLife 2020; 9:e51927. [PMID: 33043884 PMCID: PMC7661036 DOI: 10.7554/elife.51927] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2019] [Accepted: 10/09/2020] [Indexed: 11/28/2022] Open
Abstract
A neural code adapted to the statistical structure of sensory cues may optimize perception. We investigated whether interaural time difference (ITD) statistics inherent in natural acoustic scenes are parameters determining spatial discriminability. The natural ITD rate of change across azimuth (ITDrc) and ITD variability over time (ITDv) were combined in a Fisher information statistic to assess the amount of azimuthal information conveyed by this sensory cue. We hypothesized that natural ITD statistics underlie the neural code for ITD and thus influence spatial perception. To test this hypothesis, sounds with invariant statistics were presented to measure human spatial discriminability and spatial novelty detection. Human auditory spatial perception showed correlation with natural ITD statistics, supporting our hypothesis. Further analysis showed that these results are consistent with classic models of ITD coding and can explain the ITD tuning distribution observed in the mammalian brainstem.
Collapse
Affiliation(s)
- Rodrigo Pavão
- Dominick P. Purpura Department of Neuroscience - Albert Einstein College of MedicineNew YorkUnited States
- Centro de Matemática, Computação e Cognição - Universidade Federal do ABCSanto AndréBrazil
| | - Elyse S Sussman
- Dominick P. Purpura Department of Neuroscience - Albert Einstein College of MedicineNew YorkUnited States
| | - Brian J Fischer
- Department of Mathematics - Seattle UniversitySeattleUnited States
| | - José L Peña
- Dominick P. Purpura Department of Neuroscience - Albert Einstein College of MedicineNew YorkUnited States
| |
Collapse
|
28
|
Dean K, Grose JH. The Binaural Interaction Component of the Auditory Brainstem Response Under Precedence Effect Conditions. Trends Hear 2020; 24:2331216520946133. [PMID: 32808860 PMCID: PMC7436800 DOI: 10.1177/2331216520946133] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022] Open
Abstract
The purpose of this study was to measure the binaural interaction component (BIC) derived from click-evoked auditory brainstem responses (ABRs) using stimuli configured to elicit the Precedence Effect. The hypothesis was that the contribution of binaural processing to echo suppression can be evidenced by a diminished or absent BIC associated with the echo. Ten normal-hearing young adults provided ABRs generated by sequences of click pairs. Results showed that BICs elicited by diotic clicks in isolation were obliterated when those diotic clicks were preceded by a click pair having an interaural time difference of 400 µs and where the interclick interval was 8.4 ms. The presence of the leading click pair increased the latency of the ABR generated by the lagging diotic click pair but did not decrease its amplitude. The results were interpreted as indicating a contribution of binaural processing at the level of the brainstem to echo suppression, at least for the conditions tested here.
Collapse
Affiliation(s)
- Kelly Dean
- Department of Otolaryngology-Head and Neck Surgery, University of North Carolina at Chapel Hill
| | - John H Grose
- Department of Otolaryngology-Head and Neck Surgery, University of North Carolina at Chapel Hill
| |
Collapse
|
29
|
Visentin C, Pellegatti M, Prodi N. Effect of a single lateral diffuse reflection on spatial percepts and speech intelligibility. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 148:122. [PMID: 32752761 DOI: 10.1121/10.0001527] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/01/2020] [Accepted: 06/20/2020] [Indexed: 06/11/2023]
Abstract
This study examines the influence of an early lateral reflection on spatial perceptual attributes and speech reception. To this aim, a diffuse reflection is compared with a specular one. Although diffusive surfaces have widespread applications in room acoustics design, the knowledge of the perceptual and behavioral outcomes of these surfaces has yet to be fully developed. Two experiments were conducted to investigate how the reflection type, its temporal delay, and its azimuth affect spatial percepts (source distance, width, and focus) and speech intelligibility (SI) in diffuse stationary noise. The experimental setup included ecological elements: field measurements, a speaker-like source directivity, and real flat and diffusive surfaces. The results indicate that the presence of a single diffuse reflection reduces the perceived distance of a frontal speech source and makes it clearer. SI is higher with a diffuse reflection than with a specular one. Perceptual and behavioral outcomes both depend on the angle of reflection given the frequency- and angular-dependent properties of the diffusing surface and the directivity of the speech source. The results are interpreted with reference to loudness and binaural cues and to the precedence effect. Implications of the findings for acoustic design are also discussed.
Collapse
Affiliation(s)
- Chiara Visentin
- Department of Engineering, University of Ferrara, Ferrara, Italy
| | | | - Nicola Prodi
- Department of Engineering, University of Ferrara, Ferrara, Italy
| |
Collapse
|
30
|
Baltzell LS, Cho AY, Swaminathan J, Best V. Spectro-temporal weighting of interaural time differences in speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:3883. [PMID: 32611137 PMCID: PMC7297545 DOI: 10.1121/10.0001418] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/23/2019] [Revised: 05/06/2020] [Accepted: 05/18/2020] [Indexed: 05/19/2023]
Abstract
Numerous studies have demonstrated that the perceptual weighting of interaural time differences (ITDs) is non-uniform in time and frequency, leading to reports of spectral and temporal "dominance" regions. It is unclear however, how these dominance regions apply to spectro-temporally complex stimuli such as speech. The authors report spectro-temporal weighting functions for ITDs in a pair of naturally spoken speech tokens ("two" and "eight"). Each speech token was composed of two phonemes, and was partitioned into eight frequency regions over two time bins (one time bin for each phoneme). To derive lateralization weights, ITDs for each time-frequency bin were drawn independently from a normal distribution with a mean of 0 and a standard deviation of 200 μs, and listeners were asked to indicate whether the speech token was presented from the left or right. ITD thresholds were also obtained for each of the 16 time-frequency bins in isolation. The results suggest that spectral dominance regions apply to speech, and that ITDs carried by phonemes in the first position of the syllable contribute more strongly to lateralization judgments than ITDs carried by phonemes in the second position. The results also show that lateralization judgments are partially accounted for by ITD sensitivity across time-frequency bins.
Collapse
Affiliation(s)
- Lucas S Baltzell
- Department of Speech, Language, and Hearing Sciences, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| | - Adrian Y Cho
- Department of Speech, Language, and Hearing Sciences, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| | - Jayaganesh Swaminathan
- Department of Speech, Language, and Hearing Sciences, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| | - Virginia Best
- Department of Speech, Language, and Hearing Sciences, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| |
Collapse
|
31
|
Charaziak KK, Dong W, Altoè A, Shera CA. Asymmetry and Microstructure of Temporal-Suppression Patterns in Basilar-Membrane Responses to Clicks: Relation to Tonal Suppression and Traveling-Wave Dispersion. J Assoc Res Otolaryngol 2020; 21:151-170. [PMID: 32166602 DOI: 10.1007/s10162-020-00747-2] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2019] [Accepted: 02/13/2020] [Indexed: 10/24/2022] Open
Abstract
The cochlea's wave-based signal processing allows it to efficiently decompose a complex acoustic waveform into frequency components. Because cochlear responses are nonlinear, the waves arising from one frequency component of a complex sound can be altered by the presence of others that overlap with it in time and space (e.g., two-tone suppression). Here, we investigate the suppression of basilar-membrane (BM) velocity responses to a transient signal (a test click) by another click or tone. We show that the BM response to the click can be reduced when the stimulus is shortly preceded or followed by another (suppressor) click. More surprisingly, the data reveal two curious dependencies on the interclick interval, Δt. First, the temporal suppression curve (amount of suppression vs. Δt) manifests a pronounced and nearly periodic microstructure. Second, temporal suppression is generally strongest not when the two clicks are presented simultaneously (Δt = 0), but when the suppressor click precedes the test click by a time interval corresponding to one to two periods of the best frequency (BF) at the measurement location. By systematically varying the phase of the suppressor click, we demonstrate that the suppression microstructure arises from alternating constructive and destructive interference between the BM responses to the two clicks. And by comparing temporal and tonal suppression in the same animals, we test the hypothesis that the asymmetry of the temporal-suppression curve around Δt = 0 stems from cochlear dispersion and the well-known asymmetry of tonal suppression around the BF. Just as for two-tone suppression, BM responses to clicks are most suppressed by tones at frequencies just above the BF of the measurement location. On average, the frequency place of maximal suppressibility of the click response predicted from temporal-suppression data agrees with the frequency at which tonal suppression peaks, consistent with our hypothesis.
Collapse
Affiliation(s)
- Karolina K Charaziak
- Caruso Department of Otolaryngology, University of Southern California, Los Angeles, CA, USA.
| | - Wei Dong
- Research Service, VA Loma Linda Healthcare System, Loma Linda, CA, USA.,Department of Otolaryngology-Head & Neck Surgery, Loma Linda University Health, Loma Linda, USA
| | - Alessandro Altoè
- Caruso Department of Otolaryngology, University of Southern California, Los Angeles, CA, USA
| | - Christopher A Shera
- Caruso Department of Otolaryngology, University of Southern California, Los Angeles, CA, USA.,Department of Physics and Astronomy, University of Southern California, Los Angeles, CA, USA
| |
Collapse
|
32
|
Ausili SA, Backus B, Agterberg MJH, van Opstal AJ, van Wanrooij MM. Sound Localization in Real-Time Vocoded Cochlear-Implant Simulations With Normal-Hearing Listeners. Trends Hear 2019; 23:2331216519847332. [PMID: 31088265 PMCID: PMC6535744 DOI: 10.1177/2331216519847332] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/26/2022] Open
Abstract
Bilateral cochlear-implant (CI) users and single-sided deaf listeners with a CI are less effective at localizing sounds than normal-hearing (NH) listeners. This performance gap is due to the degradation of binaural and monaural sound localization cues, caused by a combination of device-related and patient-related issues. In this study, we targeted the device-related issues by measuring sound localization performance of 11 NH listeners, listening to free-field stimuli processed by a real-time CI vocoder. The use of a real-time vocoder is a new approach, which enables testing in a free-field environment. For the NH listening condition, all listeners accurately and precisely localized sounds according to a linear stimulus–response relationship with an optimal gain and a minimal bias both in the azimuth and in the elevation directions. In contrast, when listening with bilateral real-time vocoders, listeners tended to orient either to the left or to the right in azimuth and were unable to determine sound source elevation. When listening with an NH ear and a unilateral vocoder, localization was impoverished on the vocoder side but improved toward the NH side. Localization performance was also reflected by systematic variations in reaction times across listening conditions. We conclude that perturbation of interaural temporal cues, reduction of interaural level cues, and removal of spectral pinna cues by the vocoder impairs sound localization. Listeners seem to ignore cues that were made unreliable by the vocoder, leading to acute reweighting of available localization cues. We discuss how current CI processors prevent CI users from localizing sounds in everyday environments.
Collapse
Affiliation(s)
- Sebastian A Ausili
- 1 Department of Biophysics, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, the Netherlands
| | | | - Martijn J H Agterberg
- 1 Department of Biophysics, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, the Netherlands.,3 Department of Otorhinolaryngology, Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen Medical Center, the Netherlands
| | - A John van Opstal
- 1 Department of Biophysics, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, the Netherlands
| | - Marc M van Wanrooij
- 1 Department of Biophysics, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, the Netherlands
| |
Collapse
|
33
|
Deshpande N, Braasch J. Detection of early reflections from a binaural activity map using neural networks. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:2529. [PMID: 31671956 DOI: 10.1121/1.5129129] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/07/2019] [Accepted: 09/24/2019] [Indexed: 06/10/2023]
Abstract
Human listeners localize sounds to their sources despite competing directional cues from early room reflections. Binaural activity maps computed from a running signal can provide useful information about the presence of room reflections, but must be inspected visually to estimate auditory cues. A model was constructed using machine learning to validate the presence of and perform the extraction of these cues. The model uses the running signal output of a binaurally integrated cross-correlation/autocorrelation mechanism (BICAM) to analyze a lead/lag stimulus and generate a binaural activity map. System reflections are visually presented on the binaural display as correlation peaks with increased amplitude. Three independent neural networks estimate the location of the direct sound, the time delay of the reflection, and the location of the reflection from binaural activity maps displayed by BICAM. Depending on the task, neural network accuracies on test data sets vary from 84.1% to 98.5%.
Collapse
Affiliation(s)
- Nikhil Deshpande
- School of Architecture, Rensselaer Polytechnic Institute, Troy, New York 12180, USA
| | - Jonas Braasch
- School of Architecture, Rensselaer Polytechnic Institute, Troy, New York 12180, USA
| |
Collapse
|
34
|
Denk F, Ewert SD, Kollmeier B. On the limitations of sound localization with hearing devices. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:1732. [PMID: 31590539 DOI: 10.1121/1.5126521] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/02/2019] [Accepted: 08/28/2019] [Indexed: 06/10/2023]
Abstract
Limited abilities to localize sound sources and other reduced spatial hearing capabilities remain a largely unsolved issue in hearing devices like hearing aids or hear-through headphones. Hence, the impact of the microphone location, signal bandwidth, different equalization approaches, as well as processing delays in superposition with direct sound leaking through a vent was addressed in this study. A localization experiment was performed with normal-hearing subjects using individual binaural synthesis to separately assess the above-mentioned potential limiting issues for localization in the horizontal and vertical plane with linear hearing devices. To this end, listening through hearing devices was simulated utilizing transfer functions for six different microphone locations, measured both individually and on a dummy head. Results show that the microphone location is the governing factor for localization abilities with linear hearing devices, and non-optimal microphone locations have a disruptive influence on localization in the vertical domain, and an effect on lateral sound localization. Processing delays cause additional detrimental effects for lateral sound localization; and diffuse-field equalization to the open-ear response leads to better localization performance than free-field equalization. Stimuli derived from dummy head measurements are unsuited for evaluating individual localization abilities with a hearing device.
Collapse
Affiliation(s)
- Florian Denk
- Medizinische Physik and Cluster of Excellence "Hearing4all," Universität Oldenburg, Küpkersweg 74, 26129 Oldenburg, Germany
| | - Stephan D Ewert
- Medizinische Physik and Cluster of Excellence "Hearing4all," Universität Oldenburg, Küpkersweg 74, 26129 Oldenburg, Germany
| | - Birger Kollmeier
- Medizinische Physik and Cluster of Excellence "Hearing4all," Universität Oldenburg, Küpkersweg 74, 26129 Oldenburg, Germany
| |
Collapse
|
35
|
Pastore MT, Braasch J. The impact of peripheral mechanisms on the precedence effect. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:425. [PMID: 31370612 PMCID: PMC6658214 DOI: 10.1121/1.5116680] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/29/2018] [Revised: 06/20/2019] [Accepted: 06/25/2019] [Indexed: 06/10/2023]
Abstract
When two similar sounds are presented from different locations, with one (the lead) preceding the other (the lag) by a small delay, listeners typically report hearing one sound near the location of the lead sound source-this is called the precedence effect (PE). Several questions about the underlying mechanisms that produce the PE are asked. (1) How might listeners' relative weighting of cues at onset versus ongoing stimulus portions affect perceived lateral position of long-duration lead/lag noise stimuli? (2) What are the factors that influence this weighting? (3) Are the mechanisms invoked to explain the PE for transient stimuli applicable to long-duration stimuli? To answer these questions, lead/lag noise stimuli are presented with a range of durations, onset slopes, and lag-to-lead level ratios over headphones. Monaural, peripheral mechanisms, and binaural cue extraction are modeled to estimate the cues available for determination of perceived laterality. Results showed that all three stimulus manipulations affect the relative weighting of onset and ongoing cues and that mechanisms invoked to explain the PE for transient stimuli are also applicable to the PE, in terms of both onset and ongoing segments of long-duration, lead/lag stimuli.
Collapse
Affiliation(s)
- M Torben Pastore
- Spatial Hearing Laboratory, College of Health Solutions, Arizona State University, Tempe, Arizona 85287, USA
| | - Jonas Braasch
- School of Architecture & Cognitive and Immersive Systems Laboratory (CISL), Rensselaer Polytechnic Institute, Troy, New York 12180, USA
| |
Collapse
|
36
|
Bernstein JGW, Stakhovskaya OA, Schuchman GI, Jensen KK, Goupell MJ. Interaural Time-Difference Discrimination as a Measure of Place of Stimulation for Cochlear-Implant Users With Single-Sided Deafness. Trends Hear 2019; 22:2331216518765514. [PMID: 29623771 PMCID: PMC5894906 DOI: 10.1177/2331216518765514] [Citation(s) in RCA: 16] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/30/2022] Open
Abstract
Current clinical practice in programming a cochlear implant (CI) for individuals with single-sided deafness (SSD) is to maximize the transmission of speech information via the implant, with the implicit assumption that this will also result in improved spatial-hearing abilities. However, binaural sensitivity is reduced by interaural place-of-stimulation mismatch, a likely occurrence with a standard CI frequency-to-electrode allocation table (FAT). As a step toward reducing interaural mismatch, this study investigated whether a test of interaural-time-difference (ITD) discrimination could be used to estimate the acoustic frequency yielding the best place match for a given CI electrode. ITD-discrimination performance was measured by presenting 300-ms bursts of 100-pulses-per-second electrical pulse trains to a single CI electrode and band-limited pulse trains with variable carrier frequencies to the acoustic ear. Listeners discriminated between two reference intervals (four bursts each with constant ITD) and a moving target interval (four bursts with variable ITD). For 17 out of the 26 electrodes tested across eight listeners, the function describing the relationship between ITD-discrimination performance and carrier frequency had a discernable peak where listeners achieved 70% to 100% performance. On average, this peak occurred 1.15 octaves above the CI manufacturer’s default FAT. ITD discrimination shows promise as a method of estimating the cochlear place of stimulation for a given electrode, thereby providing information to optimize the FAT for SSD-CI listeners.
Collapse
Affiliation(s)
- Joshua G W Bernstein
- 1 National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, Bethesda, MD, USA
| | - Olga A Stakhovskaya
- 1 National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, Bethesda, MD, USA.,2 Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| | - Gerald I Schuchman
- 1 National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, Bethesda, MD, USA
| | - Kenneth K Jensen
- 1 National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, Bethesda, MD, USA
| | - Matthew J Goupell
- 2 Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| |
Collapse
|
37
|
Schutte M, Ewert SD, Wiegrebe L. The percept of reverberation is not affected by visual room impression in virtual environments. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 145:EL229. [PMID: 31067971 DOI: 10.1121/1.5093642] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/03/2018] [Accepted: 02/21/2019] [Indexed: 06/09/2023]
Abstract
Humans possess mechanisms to suppress distracting early sound reflections, summarized as the precedence effect. Recent work shows that precedence is affected by visual stimulation. This paper investigates possible effects of visual stimulation on the perception of later reflections, i.e., reverberation. In a highly immersive audio-visual virtual reality environment, subjects were asked to quantify reverberation in conditions where simultaneously presented auditory and visual stimuli either match in room identity, sound source azimuth, and sound source distance, or diverge in one of these aspects. While subjects reliably judged reverberation across acoustic environments, the visual room impression did not affect reverberation estimates.
Collapse
Affiliation(s)
- Michael Schutte
- Division of Neurobiology, Department Biology II and Graduate School of Systemic Neurosciences, Ludwig-Maximilians-Universität München, Germany
| | - Stephan D Ewert
- Medical Physics and Cluster of Excellence Hearing4all, University of Oldenburg, , ,
| | - Lutz Wiegrebe
- Division of Neurobiology, Department Biology II and Graduate School of Systemic Neurosciences, Ludwig-Maximilians-Universität München, Germany
| |
Collapse
|
38
|
Reichert MS, Ronacher B. Temporal integration of conflicting directional cues in sound localization. J Exp Biol 2019; 222:jeb.208751. [DOI: 10.1242/jeb.208751] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/10/2019] [Accepted: 11/12/2019] [Indexed: 01/13/2023]
Abstract
Sound localization is fundamental to hearing. In nature, sound degradation and noise erode directional cues and can generate conflicting directional perceptions across different subcomponents of sounds. Little is known about how sound localization is achieved in the face of conflicting directional cues in non-human animals, although this is relevant for many species in which sound localization in noisy conditions mediates mate finding or predator avoidance. We studied the effects of conflicting directional cues in male grasshoppers, Chorthippus biguttulus, which orient towards signaling females. We presented playbacks varying in the number and temporal position of song syllables providing directional cues in the form of either time or amplitude differences between two speakers. Males oriented towards the speaker broadcasting a greater number of leading or louder syllables. For a given number of syllables providing directional information, syllables with timing differences at the song's beginning were weighted most heavily, while syllables with intensity differences were weighted most heavily when they were in the middle of the song. When timing and intensity cues conflicted, the magnitude and temporal position of each cue determined their relative influence on lateralization, and males sometimes quickly corrected their directional responses. We discuss our findings with respect to similar results from humans.
Collapse
Affiliation(s)
- Michael S. Reichert
- Department of Integrative Biology, Oklahoma State University, 501 Life Sciences West, Stillwater, OK, 74078 USA
- Institut für Biologie, Abteilung Verhaltensphysiologie, Humboldt-Universität zu Berlin, Philippstraße 13, Haus 18, 10099 Berlin, Germany
| | - Bernhard Ronacher
- Institut für Biologie, Abteilung Verhaltensphysiologie, Humboldt-Universität zu Berlin, Philippstraße 13, Haus 18, 10099 Berlin, Germany
| |
Collapse
|
39
|
Nilsson ME, Tirado C, Szychowska M. Psychoacoustic evidence for stronger discrimination suppression of spatial information conveyed by lag-click interaural time than interaural level differences. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 145:512. [PMID: 30710980 DOI: 10.1121/1.5087707] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/08/2018] [Accepted: 12/29/2018] [Indexed: 06/09/2023]
Abstract
Listeners have limited access to spatial information in lagging sound, a phenomenon known as discrimination suppression. It is unclear whether discrimination suppression works differently for interaural time differences (ITDs) and interaural level differences (ILDs). To explore this, three listeners assessed the lateralization (left or right) and detection (present or not) of lag clicks with a large fixed ITD (350 μs) or ILD (10 dB) following a diotic lead click, with inter-click intervals (ICIs) of 0.125-256 ms. Performance was measured on a common scale for both cues: the lag-lead amplitude ratio [dB] at 75% correct answers. The main finding was that the lateralization thresholds, but not detection thresholds, were more strongly elevated for ITD-only than ILD-only clicks at intermediate ICIs (1-8 ms) in which previous research has found the strongest discrimination suppression effects. Altogether, these findings suggest that discrimination suppression involves mechanisms that make spatial information conveyed by lag-click ITDs less accessible to listeners than spatial information conveyed by lag-click ILDs.
Collapse
Affiliation(s)
- Mats E Nilsson
- Gösta Ekman Laboratory, Department of Psychology, Stockholm University, Stockholm, Sweden
| | - Carlos Tirado
- Gösta Ekman Laboratory, Department of Psychology, Stockholm University, Stockholm, Sweden
| | - Malina Szychowska
- Gösta Ekman Laboratory, Department of Psychology, Stockholm University, Stockholm, Sweden
| |
Collapse
|
40
|
Jones PR. The development of perceptual averaging: Efficiency metrics in children and adults using a multiple-observation sound-localization task. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 144:228. [PMID: 30075655 DOI: 10.1121/1.5043394] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
This study examined the ability of older children to integrate spatial information across sequential observations of bandpass noise. In experiment I, twelve adults and twelve 8-14 yr olds localized 1-5 sounds, all presented at the same location along a 34° speaker array. Rate of gain in response precision (as a function of N observations) was used to measure integration efficiency. Children were no worse at localizing a single sound than adults, and-unexpectedly-were no less efficient at integrating information across observations. Experiment II repeated the task using a Reverse Correlation paradigm. The number of observations was fixed (N = 5), and the location of each sound was independently randomly jittered. Relative weights were computed for each observation interval. Distance from the ideal weight-vector was used to index integration efficiency. The data showed that children were significantly less efficient integrators than adults: only reaching adult-like performance by around 11 yrs. The developmental effect was small, however, relative to the amount of individual variability, with some younger children exhibiting greater efficiency than some adults. This work indicates that sensory integration continues to mature into late childhood, but that this development is relatively gradual.
Collapse
Affiliation(s)
- Pete R Jones
- Institute of Ophthalmology, University College London (UCL), 11-43 Bath Street, London EC1V 9EL, United Kingdom
| |
Collapse
|
41
|
Ege R, van Opstal AJ, Bremen P, van Wanrooij MM. Testing the Precedence Effect in the Median Plane Reveals Backward Spatial Masking of Sound. Sci Rep 2018; 8:8670. [PMID: 29875363 PMCID: PMC5989261 DOI: 10.1038/s41598-018-26834-2] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2018] [Accepted: 05/17/2018] [Indexed: 11/17/2022] Open
Abstract
Two synchronous sounds at different locations in the midsagittal plane induce a fused percept at a weighted-average position, with weights depending on relative sound intensities. In the horizontal plane, sound fusion (stereophony) disappears with a small onset asynchrony of 1-4 ms. The leading sound then fully determines the spatial percept (the precedence effect). Given that accurate localisation in the median plane requires an analysis of pinna-related spectral-shape cues, which takes ~25-30 ms of sound input to complete, we wondered at what time scale a precedence effect for elevation would manifest. Listeners localised the first of two sounds, with spatial disparities between 10-80 deg, and inter-stimulus delays between 0-320 ms. We demonstrate full fusion (averaging), and largest response variability, for onset asynchronies up to at least 40 ms for all spatial disparities. Weighted averaging persisted, and gradually decayed, for delays >160 ms, suggesting considerable backward masking. Moreover, response variability decreased with increasing delays. These results demonstrate that localisation undergoes substantial spatial blurring in the median plane by lagging sounds. Thus, the human auditory system, despite its high temporal resolution, is unable to spatially dissociate sounds in the midsagittal plane that co-occur within a time window of at least 160 ms.
Collapse
Affiliation(s)
- Rachel Ege
- Biophysics Department, Donders Institute for Brain, Cognition, and Behaviour, Radboud University, 6525 AJ, Nijmegen, The Netherlands
| | - A John van Opstal
- Biophysics Department, Donders Institute for Brain, Cognition, and Behaviour, Radboud University, 6525 AJ, Nijmegen, The Netherlands.
| | - Peter Bremen
- Biophysics Department, Donders Institute for Brain, Cognition, and Behaviour, Radboud University, 6525 AJ, Nijmegen, The Netherlands
- Department of Neuroscience, Erasmus Medical Center, P.O. Box 2040, 3000 CA, Rotterdam, The Netherlands
| | - Marc M van Wanrooij
- Biophysics Department, Donders Institute for Brain, Cognition, and Behaviour, Radboud University, 6525 AJ, Nijmegen, The Netherlands.
| |
Collapse
|
42
|
Nilsson ME. Learning to extract a large inter-aural level difference in lag clicks. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 143:EL456. [PMID: 29960434 DOI: 10.1121/1.5041467] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Many blind people learn to use sound reflections to localize objects. However, precedence-effect research has reported evidence both for and against the possibility to improve lateralization of lag clicks preceded by lead clicks. This training study used stimuli more relevant to human echolocation than did previous training studies. One participant, the author, practiced lateralizing a lag-click inter-aural level difference (ILD) of 10 dB for 60 days, with performance measured in the lag-lead peak amplitude ratio at threshold. Clear improvements were observed at interclick intervals of 2-18 ms, suggesting that extracting a large lag-click ILD may improve with practice.
Collapse
Affiliation(s)
- Mats E Nilsson
- Department of Psychology, Gösta Ekman Laboratory, Stockholm University, SE-106 91, Stockholm, Sweden
| |
Collapse
|
43
|
Schepman A, Rodway P, Cornmell L, Smith B, de Sa SL, Borwick C, Belfon-Thompson E. Right-ear precedence and vocal emotion contagion: The role of the left hemisphere. Laterality 2018; 23:290-317. [DOI: 10.1080/1357650x.2017.1360902] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Affiliation(s)
- Astrid Schepman
- Department of Psychology, University of Chester, Chester, UK
| | - Paul Rodway
- Department of Psychology, University of Chester, Chester, UK
| | - Louise Cornmell
- Department of Psychology, University of Chester, Chester, UK
| | - Bethany Smith
- Department of Psychology, University of Chester, Chester, UK
| | | | - Ciara Borwick
- Department of Psychology, University of Chester, Chester, UK
| | | |
Collapse
|
44
|
Stecker GC, Moore TM. Reverberation enhances onset dominance in sound localization. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 143:786. [PMID: 29495688 PMCID: PMC5805551 DOI: 10.1121/1.5023221] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/04/2017] [Revised: 01/10/2018] [Accepted: 01/17/2018] [Indexed: 05/19/2023]
Abstract
Temporal variation in sensitivity to sound-localization cues was measured in anechoic conditions and in simulated reverberation using the temporal weighting function (TWF) paradigm [Stecker and Hafter (2002). J. Acoust. Soc. Am. 112, 1046-1057]. Listeners judged the locations of Gabor click trains (4 kHz center frequency, 5-ms interclick interval) presented from an array of loudspeakers spanning 360° azimuth. Targets ranged ±56.25° across trials. Individual clicks within each train varied by an additional ±11.25° to allow TWF calculation by multiple regression. In separate conditions, sounds were presented directly or in the presence of simulated reverberation: 13 orders of lateral reflection were computed for a 10 m × 10 m room ( RT60≊300 ms) and mapped to the appropriate locations in the loudspeaker array. Results reveal a marked increase in perceptual weight applied to the initial click in reverberation, along with a reduction in the impact of late-arriving sound. In a second experiment, target stimuli were preceded by trains of "conditioner" sounds with or without reverberation. Effects were modest and limited to the first few clicks in a train, suggesting that impacts of reverberant pre-exposure on localization may be limited to the processing of information from early reflections.
Collapse
Affiliation(s)
- G Christopher Stecker
- Department of Hearing and Speech Sciences, Vanderbilt University School of Medicine, 1215 21st Avenue South, Nashville, Tennessee 37232, USA
| | - Travis M Moore
- Department of Hearing and Speech Sciences, Vanderbilt University School of Medicine, 1215 21st Avenue South, Nashville, Tennessee 37232, USA
| |
Collapse
|
45
|
Stecker GC. Temporal weighting functions for interaural time and level differences. V. Modulated noise carriers. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 143:686. [PMID: 29495689 PMCID: PMC5800884 DOI: 10.1121/1.5022785] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/09/2017] [Revised: 01/09/2018] [Accepted: 01/11/2018] [Indexed: 05/29/2023]
Abstract
Sound onsets dominate spatial judgments of many types of periodic sound. Conversely, ongoing cues often dominate in spatial judgments of aperiodic noise. This study quantified onset dominance as a function of both the bandwidth and the temporal regularity of stimuli by measuring temporal weighting functions (TWF) from Stecker, Ostreicher, and Brown [(2013) J. Acoust. Soc. Am. 134, 1242-1252] for lateralization of periodic and aperiodic noise-burst trains. Stimuli consisted of 16 noise bursts (1 ms each) repeating at an interval of 2 or 5 ms. TWFs were calculated by multiple regression of lateralization judgments onto interaural time and level differences, which varied independently ( ±100 μs, ±2 dB) across bursts. Noise tokens were either refreshed on each burst (aperiodic) or repeated across sets of 2, 4, 8, or 16 bursts. TWFs revealed strong onset dominance for periodic noise-burst trains (16 repeats per token), which was markedly reduced in aperiodic trains. A second experiment measured TWFs for periodic but sinusoidally amplitude-modulated noise burst trains, revealing greater weight on the earliest and least intense bursts of the rising envelope slope. The results support the view that envelope fluctuations drive access to binaural information in both periodic and aperiodic sounds.
Collapse
Affiliation(s)
- G Christopher Stecker
- Department of Hearing and Speech Sciences, Vanderbilt University School of Medicine, 1215 21st Avenue South, Nashville, Tennessee 37232, USA
| |
Collapse
|
46
|
Slow Temporal Integration Enables Robust Neural Coding and Perception of a Cue to Sound Source Location. J Neurosci 2017; 36:9908-21. [PMID: 27656028 DOI: 10.1523/jneurosci.1421-16.2016] [Citation(s) in RCA: 34] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2016] [Accepted: 08/07/2016] [Indexed: 11/21/2022] Open
Abstract
UNLABELLED In mammals, localization of sound sources in azimuth depends on sensitivity to interaural differences in sound timing (ITD) and level (ILD). Paradoxically, while typical ILD-sensitive neurons of the auditory brainstem require millisecond synchrony of excitatory and inhibitory inputs for the encoding of ILDs, human and animal behavioral ILD sensitivity is robust to temporal stimulus degradations (e.g., interaural decorrelation due to reverberation), or, in humans, bilateral clinical device processing. Here we demonstrate that behavioral ILD sensitivity is only modestly degraded with even complete decorrelation of left- and right-ear signals, suggesting the existence of a highly integrative ILD-coding mechanism. Correspondingly, we find that a majority of auditory midbrain neurons in the central nucleus of the inferior colliculus (of chinchilla) effectively encode ILDs despite complete decorrelation of left- and right-ear signals. We show that such responses can be accounted for by relatively long windows of bilateral excitatory-inhibitory interaction, which we explicitly measure using trains of narrowband clicks. Neural and behavioral data are compared with the outputs of a simple model of ILD processing with a single free parameter, the duration of excitatory-inhibitory interaction. Behavioral, neural, and modeling data collectively suggest that ILD sensitivity depends on binaural integration of excitation and inhibition within a ≳3 ms temporal window, significantly longer than observed in lower brainstem neurons. This relatively slow integration potentiates a unique role for the ILD system in spatial hearing that may be of particular importance when informative ITD cues are unavailable. SIGNIFICANCE STATEMENT In mammalian hearing, interaural differences in the timing (ITD) and level (ILD) of impinging sounds carry critical information about source location. However, natural sounds are often decorrelated between the ears by reverberation and background noise, degrading the fidelity of both ITD and ILD cues. Here we demonstrate that behavioral ILD sensitivity (in humans) and neural ILD sensitivity (in single neurons of the chinchilla auditory midbrain) remain robust under stimulus conditions that render ITD cues undetectable. This result can be explained by "slow" temporal integration arising from several-millisecond-long windows of excitatory-inhibitory interaction evident in midbrain, but not brainstem, neurons. Such integrative coding can account for the preservation of ILD sensitivity despite even extreme temporal degradations in ecological acoustic stimuli.
Collapse
|
47
|
Yost WA. Sound source localization identification accuracy: Envelope dependencies. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:173. [PMID: 28764438 PMCID: PMC5509470 DOI: 10.1121/1.4990656] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/13/2016] [Revised: 06/12/2017] [Accepted: 06/15/2017] [Indexed: 06/07/2023]
Abstract
Sound source localization accuracy as measured in an identification procedure in a front azimuth sound field was studied for click trains, modulated noises, and a modulated tonal carrier. Sound source localization accuracy was determined as a function of the number of clicks in a 64 Hz click train and click rate for a 500 ms duration click train. The clicks were either broadband or high-pass filtered. Sound source localization accuracy was also measured for a single broadband filtered click and compared to a similar broadband filtered, short-duration noise. Sound source localization accuracy was determined as a function of sinusoidal amplitude modulation and the "transposed" process of modulation of filtered noises and a 4 kHz tone. Different rates (16 to 512 Hz) of modulation (including unmodulated conditions) were used. Providing modulation for filtered click stimuli, filtered noises, and the 4 kHz tone had, at most, a very small effect on sound source localization accuracy. These data suggest that amplitude modulation, while providing information about interaural time differences in headphone studies, does not have much influence on sound source localization accuracy in a sound field.
Collapse
Affiliation(s)
- William A Yost
- Speech and Hearing Science, Arizona State University, P.O. Box 870102, Tempe, Arizona 85287, USA
| |
Collapse
|
48
|
Hassager HG, Wiinberg A, Dau T. Effects of hearing-aid dynamic range compression on spatial perception in a reverberant environment. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 141:2556. [PMID: 28464692 DOI: 10.1121/1.4979783] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
This study investigated the effects of fast-acting hearing-aid compression on normal-hearing and hearing-impaired listeners' spatial perception in a reverberant environment. Three compression schemes-independent compression at each ear, linked compression between the two ears, and "spatially ideal" compression operating solely on the dry source signal-were considered using virtualized speech and noise bursts. Listeners indicated the location and extent of their perceived sound images on the horizontal plane. Linear processing was considered as the reference condition. The results showed that both independent and linked compression resulted in more diffuse and broader sound images as well as internalization and image splits, whereby more image splits were reported for the noise bursts than for speech. Only the spatially ideal compression provided the listeners with a spatial percept similar to that obtained with linear processing. The same general pattern was observed for both listener groups. An analysis of the interaural coherence and direct-to-reverberant ratio suggested that the spatial distortions associated with independent and linked compression resulted from enhanced reverberant energy. Thus, modifications of the relation between the direct and the reverberant sound should be avoided in amplification strategies that attempt to preserve the natural sound scene while restoring loudness cues.
Collapse
Affiliation(s)
- Henrik Gert Hassager
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, DK-2800 Kongens Lyngby, Denmark
| | - Alan Wiinberg
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, DK-2800 Kongens Lyngby, Denmark
| | - Torsten Dau
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, DK-2800 Kongens Lyngby, Denmark
| |
Collapse
|
49
|
Grosse J, van de Par S, Trahiotis C. Stimulus coherence influences sound-field localization and fusion/segregation of leading and lagging sounds. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 141:2673. [PMID: 28464616 DOI: 10.1121/1.4980137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
The ability to localize sound sources in reverberant environments is dependent upon first-arriving information, an outcome commonly termed "the precedence effect." For example, in laboratory experiments, the combination of a leading (direct) sound followed by a lagging (reflected) sound is localized in the direction of the leading sound. This study was designed to measure the degree to which stimulus compactness/diffuseness (i.e., coherence as represented by interaural cross correlation) of leading and lagging sounds influences performance. The compactness/diffuseness of leading or lagging sounds was varied by either presenting a sound from a single loudspeaker or by presenting mutually uncorrelated versions of similar sounds from nine adjacent loudspeakers. In separate experiments, the listener's task was to point to the perceived location of leading and lagging 10-ms long low-pass filtered white noises or 2-s long tokens of speech. The leading and lagging stimuli were presented either from speakers located directly in front of the listeners or from speakers located ±45° to the right or left. The results indicate that leading compact (coherent) sounds influence perceived location more so than do leading diffuse (incoherent) sounds. This was true independent of whether the sounds were Gaussian noises or tokens of speech.
Collapse
Affiliation(s)
- Julian Grosse
- Acoustics Group, Cluster of Excellence "Hearing4all," Carl von Ossietzky University, 26111 Oldenburg, Germany
| | - Steven van de Par
- Acoustics Group, Cluster of Excellence "Hearing4all," Carl von Ossietzky University, 26111 Oldenburg, Germany
| | - Constantine Trahiotis
- Department of Neuroscience, University of Connecticut Health Center, 263 Farmington Avenue, Farmington, Connecticut 06030, USA
| |
Collapse
|
50
|
Hu H, Ewert SD, McAlpine D, Dietz M. Differences in the temporal course of interaural time difference sensitivity between acoustic and electric hearing in amplitude modulated stimuli. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 141:1862. [PMID: 28372072 DOI: 10.1121/1.4977014] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Previous studies have shown that normal-hearing (NH) listeners' spatial perception of non-stationary interaural time differences (ITDs) is dominated by the carrier ITD during rising amplitude segments. Here, ITD sensitivity throughout the amplitude-modulation cycle in NH listeners and bilateral cochlear implant (CI) subjects is compared, the latter by means of direct stimulation of a single electrode pair. The data indicate that, while NH listeners are most sensitive to ITDs applied toward the beginning of a modulation cycle at 600 Hz, NH listeners at 200 Hz and especially bilateral CI subjects at 200 pulses per second (pps) are more sensitive to ITDs applied to the modulation maximum. This has implications for spatial-hearing in complex environments: NH listeners' dominant 600-Hz ITD information from the rising amplitude segments comprises direct sound information. The 200-pps low rate required to get ITD sensitivity in CI users results in a higher weight of pulses later in the modulation cycle where the source ITDs are more likely corrupted by reflections. This indirectly indicates that even if future binaural CI processors are able to provide perceptually exploitable ITD information, CI users will likely not get the full benefit from such pulse-based ITD cues in reverberant and other complex environments.
Collapse
Affiliation(s)
- Hongmei Hu
- Medizinische Physik and Cluster of Excellence "Hearing4all," Universität Oldenburg, D-26111 Oldenburg, Germany
| | - Stephan D Ewert
- Medizinische Physik and Cluster of Excellence "Hearing4all," Universität Oldenburg, D-26111 Oldenburg, Germany
| | - David McAlpine
- Department of Linguistics, Australian Hearing Hub, Macquarie University, New South Wales 2109, Australia
| | - Mathias Dietz
- Medizinische Physik and Cluster of Excellence "Hearing4all," Universität Oldenburg, D-26111 Oldenburg, Germany
| |
Collapse
|