1
|
Low-dimensional interference of mid-level sound statistics predicts human speech recognition in natural environmental noise. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.02.13.579526. [PMID: 38405870 PMCID: PMC10888804 DOI: 10.1101/2024.02.13.579526] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/27/2024]
Abstract
Recognizing speech in noise, such as in a busy street or restaurant, is an essential listening task where the task difficulty varies across acoustic environments and noise levels. Yet, current cognitive models are unable to account for changing real-world hearing sensitivity. Here, using natural and perturbed background sounds we demonstrate that spectrum and modulations statistics of environmental backgrounds drastically impact human word recognition accuracy and they do so independently of the noise level. These sound statistics can facilitate or hinder recognition - at the same noise level accuracy can range from 0% to 100%, depending on the background. To explain this perceptual variability, we optimized a biologically grounded hierarchical model, consisting of frequency-tuned cochlear filters and subsequent mid-level modulation-tuned filters that account for central auditory tuning. Low-dimensional summary statistics from the mid-level model accurately predict single trial perceptual judgments, accounting for more than 90% of the perceptual variance across backgrounds and noise levels, and substantially outperforming a cochlear model. Furthermore, perceptual transfer functions in the mid-level auditory space identify multi-dimensional natural sound features that impact recognition. Thus speech recognition in natural backgrounds involves interference of multiple summary statistics that are well described by an interpretable, low-dimensional auditory model. Since this framework relates salient natural sound cues to single trial perceptual judgements, it may improve outcomes for auditory prosthetics and clinical measurements of real-world hearing sensitivity.
Collapse
|
2
|
Looking for the Edge of the World: How 3D Immersive Audio Produces a Shift from an Internalised Inner Voice to Unsymbolised Affect-Driven Ways of Thinking and Heightened Sensory Awareness. Behav Sci (Basel) 2023; 13:858. [PMID: 37887508 PMCID: PMC10604218 DOI: 10.3390/bs13100858] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2023] [Revised: 09/30/2023] [Accepted: 10/12/2023] [Indexed: 10/28/2023] Open
Abstract
In this practice-based case study, we investigate the subjective aesthetic and affective responses to a shift from 2D stereo-based modelling to 3D object-based Dolby Atmos in an audio installation artwork. Dolby Atmos is an infinite object-based audio format released in 2012 but only recently incorporated into more public-facing formats. Our analysis focuses on the artist Sadia Sadia's 30-channel audio installation 'Notes to an Unknown Lover', based on her book of free verse poetry of the same title, which was rebuilt and reformatted in a Dolby Atmos specified studio. We examine what effect altered spatiality with an infinite number of 'placements' has on the psychoacoustic and neuroaesthetic response to the text. The effectiveness of three-dimensional (3D) object-based audio is interrogated against more traditional stereo and two-dimensional (2D) formats regarding the expression and communication of emotion and what effect altered spatiality with an infinite number of placements has on the psychoacoustic and neuroaesthetic response to the text. We provide a unique examination of the consequences of a shift from 2D to wholly encompassing object-based audio in a text-based artist's audio installation work. These findings may also have promising applications for health and well-being issues.
Collapse
|
3
|
Subjectively preferred octave size is resolved at the late stages of cerebral auditory processing. Eur J Neurosci 2023; 58:3686-3704. [PMID: 37752605 DOI: 10.1111/ejn.16150] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2023] [Revised: 09/06/2023] [Accepted: 09/11/2023] [Indexed: 09/28/2023]
Abstract
Human listeners prefer octave intervals slightly above the exact 2:1 frequency ratio. To study the neural underpinnings of this subjective preference, called the octave enlargement phenomenon, we compared neural responses between exact, slightly enlarged, oversized, and compressed octaves (or their multiples). The first experiment (n = 20) focused on the N1 and P2 event-related potentials (ERPs) elicited in EEG 50-250 ms after the second tone onset during passive listening of one-octave intervals. In the second experiment (n = 20) applying four-octave intervals, musician participants actively rated the different octave types as 'low', 'good' and 'high'. The preferred slightly enlarged octave was individually determined prior to the second experiment. In both experiments, N1-P2 peak-to-peak amplitudes attenuated for the exact and slightly enlarged octave intervals compared with compressed and oversized intervals, suggesting overlapping neural representations of tones an octave (or its multiples) apart. While there were no differences between the N1-P2 amplitudes to the exact and preferred enlarged octaves, ERP amplitudes differed after 500 ms from onset of the second tone of the pair. In the multivariate pattern analysis (MVPA) of the second experiment, the different octave types were distinguishable (spatial classification across electroencephalography [EEG] channels) 200 ms after second tone onset. Temporal classification within channels suggested two separate discrimination processes peaking around 300 and 700 ms. These findings appear to be related to active listening, as no multivariate results were found in the first, passive listening experiment. The present results suggest that the subjectively preferred octave size is resolved at the late stages of auditory processing.
Collapse
|
4
|
Acoustic perception and emotion evocation by rock art soundscapes of Altai (Russia). Front Psychol 2023; 14:1188567. [PMID: 37794915 PMCID: PMC10546042 DOI: 10.3389/fpsyg.2023.1188567] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2023] [Accepted: 09/04/2023] [Indexed: 10/06/2023] Open
Abstract
The major goal of psychoarchaeoacoustics is to understand the psychology behind motivations and emotions of past communities when selecting certain acoustic environments to set activities involving the production of paintings and carvings. Within this framework, the present study seeks to explore whether a group of archaeological rock art sites in Altai (Siberia, Russia) are distinguished by particular acoustic imprints that elicit distinct reactions on listeners, in perceptual and emotional terms. Sixty participants were presented with a series of natural sounds convolved with six impulse responses from Altai, three of them recorded in locations in front of rock art panels and three of them in front of similar locations but without any trace of rock art. Participants were interrogated about their subjective perception of the sounds presented, using 10 psychoacoustic and emotional scales. The mixed ANOVA analyses carried out revealed that feelings of "presence," "closeness," and "tension" evoked by all sounds were significantly influenced by the location. These effects were attributed to the differences in reverberation between the locations with and without rock art. Despite results are not consistent across all the studied rock art sites, and acknowledging the presence of several limitations, this study highlights the significance of its methodology. It stresses the crucial aspect of incorporating the limitations encountered in shaping future research endeavors.
Collapse
|
5
|
Development of an Arabic "Command in Noise" Hearing Test to Assess Fitness for Duty. Noise Health 2023; 25:104-112. [PMID: 37203127 DOI: 10.4103/nah.nah_69_22] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/20/2023] Open
Abstract
Objective The goal is to implement the developed speech material in a hearing test to assess auditory fitness for duty (AFFD), specifically in areas where the intelligibility of spoken commands is essential. Design In study 1, a speech corpus with equal intelligibility was constructed using constant stimuli to test each target word's psychometric functions. Study 2 used an adaptive interleaving procedure to maximize equalized terms. Study 3 used Monte Carlo simulations to determine speech test accuracy. Study sample Study 1 (n = 24) and study 2 (n = 20) were completed by civilians with normal hearing. Study 3 ran 10,000 simulations per condition across various conditions varying in slopes and speech recognition thresholds (SRTs). Results Studies 1 and 2 produced three 8-word wordlists. The mean, standard deviation in dB SNR is -13.1 1.2 for wordlist 1, -13.7 1.6 for wordlist 2, and -13.7 1.3 for wordlist 3, with word SRTs within 3.4 dB SNR. Study 3 revealed that a 6 dB SNR range is appropriate for equally understandable speech using a closed-set adaptive technique. Conclusion The developed speech corpus may be used in an AFFD measure. Concerning the homogeneity of the speech in noise test material, care should be taken when generalizing and using ranges and standard deviations from multiple tests.
Collapse
|
6
|
Corrigendum: Effects of acute ischemic stroke on binaural perception. Front Neurosci 2023; 17:1143063. [PMID: 36816110 PMCID: PMC9930201 DOI: 10.3389/fnins.2023.1143063] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2023] [Accepted: 01/17/2023] [Indexed: 02/04/2023] Open
Abstract
[This corrects the article DOI: 10.3389/fnins.2022.1022354.].
Collapse
|
7
|
Effects of acute ischemic stroke on binaural perception. Front Neurosci 2022; 16:1022354. [PMID: 36620448 PMCID: PMC9817147 DOI: 10.3389/fnins.2022.1022354] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2022] [Accepted: 11/28/2022] [Indexed: 12/24/2022] Open
Abstract
Stroke-induced lesions at different locations in the brain can affect various aspects of binaural hearing, including spatial perception. Previous studies found impairments in binaural hearing, especially in patients with temporal lobe tumors or lesions, but also resulting from lesions all along the auditory pathway from brainstem nuclei up to the auditory cortex. Currently, structural magnetic resonance imaging (MRI) is used in the clinical treatment routine of stroke patients. In combination with structural imaging, an analysis of binaural hearing enables a better understanding of hearing-related signaling pathways and of clinical disorders of binaural processing after a stroke. However, little data are currently available on binaural hearing in stroke patients, particularly for the acute phase of stroke. Here, we sought to address this gap in an exploratory study of patients in the acute phase of ischemic stroke. We conducted psychoacoustic measurements using two tasks of binaural hearing: binaural tone-in-noise detection, and lateralization of stimuli with interaural time- or level differences. The location of the stroke lesion was established by previously acquired MRI data. An additional general assessment included three-frequency audiometry, cognitive assessments, and depression screening. Fifty-five patients participated in the experiments, on average 5 days after their stroke onset. Patients whose lesions were in different locations were tested, including lesions in brainstem areas, basal ganglia, thalamus, temporal lobe, and other cortical and subcortical areas. Lateralization impairments were found in most patients with lesions within the auditory pathway. Lesioned areas at brainstem levels led to distortions of lateralization in both hemifields, thalamus lesions were correlated with a shift of the whole auditory space, whereas some cortical lesions predominantly affected the lateralization of stimuli contralateral to the lesion and resulted in more variable responses. Lateralization performance was also found to be affected by lesions of the right, but not the left, basal ganglia, as well as by lesions in non-auditory cortical areas. In general, altered lateralization was common in the stroke group. In contrast, deficits in tone-in-noise detection were relatively scarce in our sample of lesion patients, although a significant number of patients with multiple lesion sites were not able to complete the task.
Collapse
|
8
|
Left Frontal White Matter Links to Rhythm Processing Relevant to Speech Production in Apraxia of Speech. NEUROBIOLOGY OF LANGUAGE (CAMBRIDGE, MASS.) 2022; 3:515-537. [PMID: 37215340 PMCID: PMC10158569 DOI: 10.1162/nol_a_00075] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/12/2021] [Accepted: 06/03/2022] [Indexed: 05/24/2023]
Abstract
Recent mechanistic models argue for a key role of rhythm processing in both speech production and speech perception. Patients with the non-fluent variant (NFV) of primary progressive aphasia (PPA) with apraxia of speech (AOS) represent a specific study population in which this link can be examined. Previously, we observed impaired rhythm processing in NFV with AOS. We hypothesized that a shared neurocomputational mechanism structures auditory input (sound and speech) and output (speech production) in time, a "temporal scaffolding" mechanism. Since considerable white matter damage is observed in NFV, we test here whether white matter changes are related to impaired rhythm processing. Forty-seven participants performed a psychoacoustic test battery: 12 patients with NFV and AOS, 11 patients with the semantic variant of PPA, and 24 cognitively intact age- and education-matched controls. Deformation-based morphometry was used to test whether white matter volume correlated to rhythmic abilities. In 34 participants, we also obtained tract-based metrics of the left Aslant tract, which is typically damaged in patients with NFV. Nine out of 12 patients with NFV displayed impaired rhythmic processing. Left frontal white matter atrophy adjacent to the supplementary motor area (SMA) correlated with poorer rhythmic abilities. The structural integrity of the left Aslant tract also correlated with rhythmic abilities. A colocalized and perhaps shared white matter substrate adjacent to the SMA is associated with impaired rhythmic processing and motor speech impairment. Our results support the existence of a temporal scaffolding mechanism structuring perceptual input and speech output.
Collapse
|
9
|
Effects of Abacus Training on Auditory Spatial Maturation in Children with Normal Hearing. Int Arch Otorhinolaryngol 2022; 27:e56-e66. [PMID: 36714899 PMCID: PMC9879648 DOI: 10.1055/s-0041-1741434] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2021] [Accepted: 09/11/2021] [Indexed: 02/01/2023] Open
Abstract
Introduction The spatial auditory system, though developed at birth, attains functional maturity in the late childhood (12 years). Spatial changes during childhood affect navigation in the environment and source segregation. Accommodation of a new skill through learning, especially during childhood, can expedite this process. Objective To explore the auditory spatial benefits of abacus training on psychoacoustic metrics in children. The study also aimed to identify the most sensitive metric to abacus training related changes in spatial processing, and utilize this metric for a detailed spatial error profiling. Methods A standard group comparison analysis with 90 participants divided into three groups: I: children with abacus training (C-AT); II: children with no training (C-UT); III: adults with no training (A-UT). The groups underwent a series of psychoacoustic tests, such as interaural time difference (ITD), interaural level difference (ILD), and virtual auditory space identification (VASI), as well as perceptual tests such as the Kannada version of the speech, spatial, and quality questionnaire (K-SSQ). Results Significant group differences were observed in the multivariate analysis of variance (MANOVA) and post-hoc tests, with the C-AT group showing significantly lower ILD scores ( p = 0.01) and significantly higher VASI scores ( p <0.001) compared to the C-UT group, which is indicative of better spatial processing abilities in the former group. The discriminant function (DF) analyses showed that the VASI was the most sensitive metric for training-related changes, based on which elaborate error analyses were performed. Conclusions Despite the physiological limits of the immature neural framework, the performance of the C-AT group was equivalent to that of untrained adults on psychoacoustic tests, which is reflective of the positive role of abacus training in expediting auditory spatial maturation.
Collapse
|
10
|
Experimental Enhancement of Feelings of Transcendence, Tenderness, and Expressiveness by Music in Christian Liturgical Spaces. Front Psychol 2022; 13:844029. [PMID: 35360627 PMCID: PMC8960987 DOI: 10.3389/fpsyg.2022.844029] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2021] [Accepted: 02/16/2022] [Indexed: 11/20/2022] Open
Abstract
In western cultures, when it comes to places of worship and liturgies, music, acoustics and architecture go hand in hand. In the present study, we aimed to investigate whether the emotions evoked by music are enhanced by the acoustics of the space where the music was composed to be played on. We explored whether the emotional responses of western naïve listeners to two vocal pieces from the Renaissance, one liturgical and one secular, convolved with the impulse responses of four Christian temples from the United Kingdom, were modulated by the appropriate piece/space matching. In an alternative forced choice task where participants had to indicate their preference for the original recording of the piece (not convolved with any temple-like acoustics) vs. the convolved one, no significant differences were found. However, in the tasks where participants rated their emotional in response to each piece and acoustic condition, the factorial ANCOVA analyses performed on the results revealed significant effects. We observed that, across pieces and spaces, participants found the temple-like acoustics as more transcendent, compared to the acoustics of the original version of the pieces. In addition, they rated the secular piece as more tender and the liturgical piece as more expressive in its original versions, compared to the convolved ones. We conclude that the acoustic signature of the four Christian temples causes an exaltation of certain emotions on listeners, although this effect is not associated to one or another musical piece.
Collapse
|
11
|
Interval and Ratio Scaling of Spectral Audio Descriptors. Front Psychol 2022; 13:835401. [PMID: 35432077 PMCID: PMC9007158 DOI: 10.3389/fpsyg.2022.835401] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2021] [Accepted: 03/07/2022] [Indexed: 11/13/2022] Open
Abstract
Two experiments were conducted for the derivation of psychophysical scales of the following audio descriptors: spectral centroid, spectral spread, spectral skewness, odd-to-even harmonic ratio, spectral deviation, and spectral slope. The stimulus sets of each audio descriptor were synthesized and (wherever possible) independently controlled through appropriate synthesis techniques. Partition scaling methods were used in both experiments, and the scales were constructed by fitting well-behaving functions to the listeners' ratings. In the first experiment, the listeners' task was the estimation of the relative differences between successive levels of a particular audio descriptor. The median values of listeners' ratings increased with increasing feature values, which confirmed listeners' abilities to estimate intervals. However, there was a large variability in the reliability of the derived interval scales depending on the stimulus spacing in each trial. In the second experiment, listeners had control over the stimulus values and were asked to divide the presented range of values into perceptually equal intervals, which provides a ratio scale. For every descriptor, the reliability of the derived ratio scales was excellent. The unit of a particular ratio scale was assigned empirically so as to facilitate qualitative comparisons between the scales of all audio descriptors. The construction of psychophysical scales based on univariate stimuli allowed for the establishment of cause-and-effect relations between audio descriptors and perceptual dimensions, contrary to past research that has relied on multivariate stimuli and has only examined the correlations between the two. Most importantly, this study provides an understanding of the ways in which the sensation magnitudes of several audio descriptors are apprehended.
Collapse
|
12
|
Purkyně's Opistophone: the hearing 'Deaf', auditory attention and organic subjectivity in Prague psychophysical experiments, ca 1850s. ANNALS OF SCIENCE 2022; 79:60-80. [PMID: 34962446 DOI: 10.1080/00033790.2021.2010806] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/21/2021] [Accepted: 11/22/2021] [Indexed: 06/14/2023]
Abstract
ABSTRACTThe paper examines the little-known experiments in audition performed by the prominent experimental physiologist Jan Purkyně in Prague in the 1850s. Purkyně's original research on spatial hearing and auditory attention is studied against the backdrop of the nineteenth century research on binaural audition and the nascent field of psychophysics. The article revolves around an acoustic research instrument of Purkyně's own making, the opistophone, in which hearing became both an object of investigation and an instrument of scientific inquiry. It argues that Purkyně's understanding of auditory attention, which combined acoustic stimulation, physiological conditions, and sensory training, preceded a similar approach to hearing in psychophysical debates in the second half of the nineteenth century. Purkyně was the first scholar to experimentally investigate intracranial sounds, which he studied in his experiments with the inmates of the Prague Institute of Deaf-Mutes. This research on intracranial hearing was part of Purkyně's study of so-called organic subjectivity, in which subjective hearing experience was interpreted as the result of the interaction between individual perception and objective acoustic phenomena.
Collapse
|
13
|
Abstract
We tested whether sensitivity to acoustic spectrotemporal modulations can be observed from reaction times for normal-hearing and impaired-hearing conditions. In a manual reaction-time task, normal-hearing listeners had to detect the onset of a ripple (with density between 0–8 cycles/octave and a fixed modulation depth of 50%), that moved up or down the log-frequency axis at constant velocity (between 0–64 Hz), in an otherwise-unmodulated broadband white-noise. Spectral and temporal modulations elicited band-pass filtered sensitivity characteristics, with fastest detection rates around 1 cycle/oct and 32 Hz for normal-hearing conditions. These results closely resemble data from other studies that typically used the modulation-depth threshold as a sensitivity criterion. To simulate hearing-impairment, stimuli were processed with a 6-channel cochlear-implant vocoder, and a hearing-aid simulation that introduced separate spectral smearing and low-pass filtering. Reaction times were always much slower compared to normal hearing, especially for the highest spectral densities. Binaural performance was predicted well by the benchmark race model of binaural independence, which models statistical facilitation of independent monaural channels. For the impaired-hearing simulations this implied a “best-of-both-worlds” principle in which the listeners relied on the hearing-aid ear to detect spectral modulations, and on the cochlear-implant ear for temporal-modulation detection. Although singular-value decomposition indicated that the joint spectrotemporal sensitivity matrix could be largely reconstructed from independent temporal and spectral sensitivity functions, in line with time-spectrum separability, a substantial inseparable spectral-temporal interaction was present in all hearing conditions. These results suggest that the reaction-time task yields a valid and effective objective measure of acoustic spectrotemporal-modulation sensitivity.
Collapse
|
14
|
Differences in Auditory Perception Between Young and Older Adults When Controlling for Differences in Hearing Loss and Cognition. Trends Hear 2022; 26:23312165211066180. [PMID: 34989641 PMCID: PMC8753078 DOI: 10.1177/23312165211066180] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
This study was designed to examine age effects on various auditory perceptual skills using a large group of listeners (155 adults, 121 aged 60-88 years and 34 aged 18-30 years), while controlling for the factors of hearing loss and working memory (WM). All subjects completed 3 measures of WM, 7 psychoacoustic tasks (24 conditions) and a hearing assessment. Psychophysical measures were selected to tap phenomena thought to be mediated by higher-level auditory function and included modulation detection, modulation detection interference, informational masking (IM), masking level difference (MLD), anisochrony detection, harmonic mistuning, and stream segregation. Principal-components analysis (PCA) was applied to each psychoacoustic test. For 6 of the 7 tasks, a single component represented performance across the multiple stimulus conditions well, whereas the modulation-detection interference (MDI) task required two components to do so. The effect of age was analyzed using a general linear model applied to each psychoacoustic component. Once hearing loss and WM were accounted for as covariates in the analyses, estimated marginal mean thresholds were lower for older adults on tasks based on temporal processing. When evaluated separately, hearing loss led to poorer performance on roughly 1/2 the tasks and declines in WM accounted for poorer performance on 6 of the 8 psychoacoustic components. These results make clear the need to interpret age-group differences in performance on psychoacoustic tasks in light of cognitive declines commonly associated with aging, and point to hearing loss and cognitive declines as negatively influencing auditory perceptual skills.
Collapse
|
15
|
Abstract
Late reverberation involves the superposition of many sound reflections, approaching the properties of a diffuse sound field. Since the spatially resolved perception of individual late reflections is impossible, simplifications can potentially be made for modelling late reverberation in room acoustics simulations with reduced spatial resolution. Such simplifications are desired for interactive, real-time virtual acoustic environments with applications in hearing research and for the evaluation of hearing supportive devices. In this context, the number and spatial arrangement of loudspeakers used for playback additionally affect spatial resolution. The current study assessed the minimum number of spatially evenly distributed virtual late reverberation sources required to perceptually approximate spatially highly resolved isotropic and anisotropic late reverberation and to technically approximate a spherically isotropic sound field. The spatial resolution of the rendering was systematically reduced by using subsets of the loudspeakers of an 86-channel spherical loudspeaker array in an anechoic chamber, onto which virtual reverberation sources were mapped using vector base amplitude panning. It was tested whether listeners can distinguish lower spatial resolutions of reproduction of late reverberation from the highest achievable spatial resolution in different simulated rooms. The rendering of early reflections remained unchanged. The coherence of the sound field across a pair of microphones at ear and behind-the-ear hearing device distance was assessed to separate the effects of number of virtual sources and loudspeaker array geometry. Results show that between 12 and 24 reverberation sources are required for the rendering of late reverberation in virtual acoustic environments.
Collapse
|
16
|
Verification of a Mobile Psychoacoustic Test System. Audiol Res 2021; 11:673-690. [PMID: 34940019 PMCID: PMC8698855 DOI: 10.3390/audiolres11040061] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2021] [Revised: 11/25/2021] [Accepted: 12/08/2021] [Indexed: 11/20/2022] Open
Abstract
Many hearing difficulties can be explained as a loss of audibility, a problem easily detected and treated using standard audiological procedures. Yet, hearing can be much poorer (or more impaired) than audibility predicts because of deficits in the suprathreshold mechanisms that encode the rapidly changing, spectral, temporal, and binaural aspects of the sound. The ability to evaluate these mechanisms requires well-defined stimuli and strict adherence to rigorous psychometric principles. This project reports on the comparison between a laboratory-based and a mobile system's results for psychoacoustic assessment in adult listeners with normal hearing. A description of both systems employed is provided. Psychoacoustic tests include frequency discrimination, amplitude modulation detection, binaural encoding, and temporal gap detection. Results reported by the mobile system were not significantly different from those collected with the laboratory-based system for most of the tests and were consistent with those reported in the literature. The mobile system has the potential to be a feasible option for the assessment of suprathreshold auditory encoding abilities.
Collapse
|
17
|
Speech Perception in Noise Predicts Oral Narrative Comprehension in Children With Developmental Language Disorder. Front Psychol 2021; 12:735026. [PMID: 34744907 PMCID: PMC8566731 DOI: 10.3389/fpsyg.2021.735026] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/01/2021] [Accepted: 09/17/2021] [Indexed: 11/13/2022] Open
Abstract
We examined the relative contribution of auditory processing abilities (tone perception and speech perception in noise) after controlling for short-term memory capacity and vocabulary, to narrative language comprehension in children with developmental language disorder. Two hundred and sixteen children with developmental language disorder, ages 6 to 9 years (Mean = 7; 6), were administered multiple measures. The dependent variable was children's score on the narrative comprehension scale of the Test of Narrative Language. Predictors were auditory processing abilities, phonological short-term memory capacity, and language (vocabulary) factors, with age, speech perception in quiet, and non-verbal IQ as covariates. Results showed that narrative comprehension was positively correlated with the majority of the predictors. Regression analysis suggested that speech perception in noise contributed uniquely to narrative comprehension in children with developmental language disorder, over and above all other predictors; however, tone perception tasks failed to explain unique variance. The relative importance of speech perception in noise over tone-perception measures for language comprehension reinforces the need for the assessment and management of listening in noise deficits and makes a compelling case for the functional implications of complex listening situations for children with developmental language disorder.
Collapse
|
18
|
Behavioral Measures of Cochlear Gain Reduction Depend on Precursor Frequency, Bandwidth, and Level. Front Neurosci 2021; 15:716689. [PMID: 34671236 PMCID: PMC8520990 DOI: 10.3389/fnins.2021.716689] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2021] [Accepted: 09/06/2021] [Indexed: 11/25/2022] Open
Abstract
Sensory systems adjust to the environment to maintain sensitivity to change. In the auditory system, the medial olivocochlear reflex (MOCR) is a known physiological mechanism capable of such adjustment. The MOCR provides efferent feedback between the brainstem and cochlea, reducing cochlear gain in response to sound. The perceptual effects of the MOCR are not well understood, such as how gain reduction depends on elicitor characteristics in human listeners. Physiological and behavioral data suggest that ipsilateral MOCR tuning is only slightly broader than it is for afferent fibers, and that the fibers feed back to the frequency region of the cochlea that stimulated them. However, some otoacoustic emission (OAE) data suggest that noise is a more effective elicitor than would be consistent with sharp tuning, and that a broad region of the cochlea may be involved in elicitation. If the elicitor is processed in a cochlear channel centered at the signal frequency, the growth of gain reduction with elicitor level would be expected to depend on the frequency content of the elicitor. In the current study, the effects of the frequency content and level of a preceding sound (called a precursor) on signal threshold was examined. The results show that signal threshold increased with increasing precursor level at a shallower slope for a tonal precursor at the signal frequency than for a tonal precursor nearly an octave below the signal frequency. A broadband noise was only slightly more effective than a tone at the signal frequency, with a relatively shallow slope similar to that of the tonal precursor at the signal frequency. Overall, these results suggest that the excitation at the signal cochlear place, regardless of elicitor frequency, determines the magnitude of ipsilateral cochlear gain reduction, and that it increases with elicitor level.
Collapse
|
19
|
The Influence of the Psychophysical Assessment Paradigm on Pitch Discrimination for Adults (and a Pilot Sample of Children). Percept Mot Skills 2021; 128:2582-2604. [PMID: 34474624 DOI: 10.1177/00315125211044063] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Pitch discrimination ability has been of research interest due to its potential relationship to language and literacy. However, assessment protocols for pitch discrimination have varied widely. Prior studies with both children and adults have produced conflicting performance findings across different pitch discrimination research paradigms, though they have consistently shown that discrimination accuracy is based on the psychophysical assessment method applied. In the present study, we examined pitch discrimination performance among convenience samples of 19 adult women and ten female children across six different adaptive psychophysical measurement conditions. We found pitch discrimination performance in both groups to be impacted by the measurement paradigm such that, while adults exhibited significantly better discrimination thresholds than did children, the pattern of performance across the six conditions was similar for both the adults and the children.
Collapse
|
20
|
Abstract
OBJECTIVES When one ear of an individual can hear significantly better than the other ear, evaluating the worse ear with loud probe tones may require delivering masking noise to the better ear to prevent the probe tones from inadvertently being heard by the better ear. Current masking protocols are confusing, laborious, and time consuming. Adding a standardized masking protocol to an active machine learning audiogram procedure could potentially alleviate all of these drawbacks by dynamically adapting the masking as needed for each individual. The goal of this study is to determine the accuracy and efficiency of automated machine learning masking for obtaining true hearing thresholds. DESIGN Dynamically masked automated audiograms were collected for 29 participants between the ages of 21 and 83 (mean 43, SD 20) with a wide range of hearing abilities. Normal-hearing listeners were given unmasked and masked machine learning audiogram tests. Listeners with hearing loss were given a standard audiogram test by an audiologist, with masking stimuli added as clinically determined, followed by a masked machine learning audiogram test. The hearing thresholds estimated for each pair of techniques were compared at standard audiogram frequencies (i.e., 0.25, 0.5, 1, 2, 4, 8 kHz). RESULTS Masked and unmasked machine learning audiogram threshold estimates matched each other well in normal-hearing listeners, with a mean absolute difference between threshold estimates of 3.4 dB. Masked machine learning audiogram thresholds also matched well the thresholds determined by a conventional masking procedure, with a mean absolute difference between threshold estimates for listeners with low asymmetry and high asymmetry between the ears, respectively, of 4.9 and 2.6 dB. Notably, out of 6200 masked machine learning audiogram tone deliveries for this study, no instances of tones detected by the nontest ear were documented. The machine learning methods were also generally faster than the manual methods, and for some listeners, substantially so. CONCLUSIONS Dynamically masked audiograms achieve accurate true threshold estimates and reduce test time compared with current clinical masking procedures. Dynamic masking is a compelling alternative to the methods currently used to evaluate individuals with highly asymmetric hearing, yet can also be used effectively and efficiently for anyone.
Collapse
|
21
|
The "Missing 6 dB" Revisited: Influence of Room Acoustics and Binaural Parameters on the Loudness Mismatch Between Headphones and Loudspeakers. Front Psychol 2021; 12:623670. [PMID: 33841255 PMCID: PMC8032957 DOI: 10.3389/fpsyg.2021.623670] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2020] [Accepted: 03/05/2021] [Indexed: 11/13/2022] Open
Abstract
Generations of researchers observed a mismatch between headphone and loudspeaker presentation: the sound pressure level at the eardrum generated by a headphone has to be about 6 dB higher compared to the level created by a loudspeaker that elicits the same loudness. While it has been shown that this effect vanishes if the same waveforms are generated at the eardrum in a blind comparison, the origin of the mismatch is still unclear. We present new data on the issue that systematically characterize this mismatch under variation of the stimulus frequency, presentation room, and binaural parameters of the headphone presentation. Subjects adjusted the playback level of a headphone presentation to equal loudness as loudspeaker presentation, and the levels at the eardrum were determined through appropriate transfer function measurements. Identical experiments were conducted at Oldenburg and Aachen with 40 normal-hearing subjects including 14 that passed through both sites. Our data verify a mismatch between loudspeaker and binaural headphone presentation, especially at low frequencies. This mismatch depends on the room acoustics, and on the interaural coherence in both presentation modes. It vanishes for high frequencies and broadband signals if individual differences in the sound transfer to the eardrums are accounted for. Moreover, small acoustic and non-acoustic differences in an anechoic reference environment (Oldenburg vs. Aachen) exert a large effect on the recorded loudness mismatch, whereas not such a large effect of the respective room is observed across moderately reverberant rooms at both sites. Hence, the non-conclusive findings from the literature appear to be related to the experienced disparity between headphone and loudspeaker presentation, where even small differences in (anechoic) room acoustics significantly change the response behavior of the subjects. Moreover, individual factors like loudness summation appear to be only loosely connected to the observed mismatch, i.e., no direct prediction is possible from individual binaural loudness summation to the observed mismatch. These findings – even though not completely explainable by the yet limited amount of parameter variations performed in this study – have consequences for the comparability of experiments using loudspeakers with conditions employing headphones or other ear-level hearing devices.
Collapse
|
22
|
Word and Mystery: The Acoustics of Cultural Transmission During the Protestant Reformation. Front Psychol 2021; 12:564542. [PMID: 33737893 PMCID: PMC7960663 DOI: 10.3389/fpsyg.2021.564542] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2020] [Accepted: 02/02/2021] [Indexed: 11/28/2022] Open
Abstract
To a first-order approximation we can place most worship services on a continuum between clarity and mystery, depending on the setting and content of the service. This liturgical space can be thought of as a combination of the physical acoustics of the worship space and the qualities of the sound created during the worship service. A very clear acoustic channel emphasizes semantic content, especially speech intelligibility. An immersive, reverberant acoustic emphasizes mystery and music. One of the chief challenges in acoustical design is the fact that both clarity and immersion are subjectively preferred by audiences, yet these two goals are almost mutually exclusive of one another. The movement along this continuum in liturgical space can also be seen in the religious contexts for many of the worship spaces constructed in the West in the last two millennia. In the case of religious ceremony, a free field acoustic environment provides more clarity and precision in the spoken word received from God and given to the congregation. Yet a diffuse field environment provides an embodied, otherworldly sense of the supernatural: the mystery of the faith received which cannot merely be put into words. This tension is perceptible in many of the religious controversies in the West during this time period. This article examines the history of the spaces used by early Western Catholic Christians as well as those of the traditions—Lutheran and Calvinist—that left the Catholic faith during the 16th century Reformation. By considering the stated goals of these traditions alongside the architectural and liturgical innovations they created, it can be seen that emergent liturgical spaces mirror the assumptions of their respective traditions regarding the proper balance between semantic and aesthetic communication during the worship service. The Reformed faiths' emphasis on the power of the Word is reflected in the liturgical space of their services, while the Catholic faith gave greater priority to the role of Mystery, in their liturgical space as well as their explicit theology. Once constructed, these spaces also aid the cultural transmission of the sung or spoken liturgy of each tradition to future generations.
Collapse
|
23
|
Microsecond interaural time difference discrimination restored by cochlear implants after neonatal deafness. eLife 2021; 10:59300. [PMID: 33427644 PMCID: PMC7815311 DOI: 10.7554/elife.59300] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2020] [Accepted: 01/07/2021] [Indexed: 01/03/2023] Open
Abstract
Spatial hearing in cochlear implant (CI) patients remains a major challenge, with many early deaf users reported to have no measurable sensitivity to interaural time differences (ITDs). Deprivation of binaural experience during an early critical period is often hypothesized to be the cause of this shortcoming. However, we show that neonatally deafened (ND) rats provided with precisely synchronized CI stimulation in adulthood can be trained to lateralize ITDs with essentially normal behavioral thresholds near 50 μs. Furthermore, comparable ND rats show high physiological sensitivity to ITDs immediately after binaural implantation in adulthood. Our result that ND-CI rats achieved very good behavioral ITD thresholds, while prelingually deaf human CI patients often fail to develop a useful sensitivity to ITD raises urgent questions concerning the possibility that shortcomings in technology or treatment, rather than missing input during early development, may be behind the usually poor binaural outcomes for current CI patients.
Collapse
|
24
|
Time-dependent discrimination advantages for harmonic sounds suggest efficient coding for memory. Proc Natl Acad Sci U S A 2020; 117:32169-32180. [PMID: 33262275 PMCID: PMC7749397 DOI: 10.1073/pnas.2008956117] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Perceptual systems have finite memory resources and must store incoming signals in compressed formats. To explore whether representations of a sound's pitch might derive from this need for compression, we compared discrimination of harmonic and inharmonic sounds across delays. In contrast to inharmonic spectra, harmonic spectra can be summarized, and thus compressed, using their fundamental frequency (f0). Participants heard two sounds and judged which was higher. Despite being comparable for sounds presented back-to-back, discrimination was better for harmonic than inharmonic stimuli when sounds were separated in time, implicating memory representations unique to harmonic sounds. Patterns of individual differences (correlations between thresholds in different conditions) indicated that listeners use different representations depending on the time delay between sounds, directly comparing the spectra of temporally adjacent sounds, but transitioning to comparing f0s across delays. The need to store sound in memory appears to determine reliance on f0-based pitch and may explain its importance in music, in which listeners must extract relationships between notes separated in time.
Collapse
|
25
|
Maximum Expected Information Approach for Improving Efficiency of Categorical Loudness Scaling. Front Psychol 2020; 11:578352. [PMID: 33281677 PMCID: PMC7705216 DOI: 10.3389/fpsyg.2020.578352] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2020] [Accepted: 10/29/2020] [Indexed: 11/13/2022] Open
Abstract
Categorical loudness scaling (CLS) measures provide useful information about an individual's loudness perception across the dynamic range of hearing. A probability model of CLS categories has previously been described as a multi-category psychometric function (MCPF). In the study, a representative "catalog" of potential listener MCPFs was used in conjunction with maximum-likelihood estimation to derive CLS functions for participants with normal hearing and with hearing loss. The approach of estimating MCPFs for each listener has the potential to improve the accuracy of the CLS measurements, particularly when a relatively low number of data points are available. The present study extends the MCPF approach by using Bayesian inference to select stimulus parameters that are predicted to yield maximum expected information (MEI) during data collection. The accuracy and reliability of the MCPF-MEI approach were compared to the standardized CLS measurement procedure (ISO 16832:2006, 2006). A non-adaptive, fixed-level, paradigm served as a "gold-standard" for this comparison. The test time required to obtain measurements in the standard procedure is a major barrier to its clinical uptake. Test time was reduced from approximately 15 min to approximately 3 min with the MEI-adaptive procedure. Results indicated that the test-retest reliability and accuracy of the MCPF-MEI adaptive procedures were similar to the standardized CLS procedure. Computer simulations suggest that the reliability and accuracy of the MEI procedure were limited by intrinsic uncertainty of the listeners represented in the MCPF catalog. In other words, the MCPF provided insufficient predictive power to significantly improve adaptive-tracking efficiency under practical conditions. Concurrent optimization of both the MCPF catalog and the MEI-adaptive procedure have the potential to produce better results. Regardless of the adaptive-tracking method used in the CLS procedure, the MCPF catalog remains clinically useful for enabling maximum-likelihood determination of loudness categories.
Collapse
|
26
|
Bioacoustics in cognitive research: Applications, considerations, and recommendations. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2020; 11:e1538. [PMID: 32548958 DOI: 10.1002/wcs.1538] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/08/2019] [Revised: 05/15/2020] [Accepted: 05/18/2020] [Indexed: 12/23/2022]
Abstract
The multifaceted ability to produce, transmit, receive, and respond to acoustic signals is widespread in animals and forms the basis of the interdisciplinary science of bioacoustics. Bioacoustics research methods, including sound recording and playback experiments, are applicable in cognitive research that centers around the processing of information from the acoustic environment. We provide an overview of bioacoustics techniques in the context of cognitive studies and make the case for the importance of bioacoustics in the study of cognition by outlining some of the major cognitive processes in which acoustic signals are involved. We also describe key considerations associated with the recording of sound and its use in cognitive applications. Based on these considerations, we provide a set of recommendations for best practices in the recording and use of acoustic signals in cognitive studies. Our aim is to demonstrate that acoustic recordings and stimuli are valuable tools for cognitive researchers when used appropriately. In doing so, we hope to stimulate opportunities for innovative cognitive research that incorporates robust recording protocols. This article is categorized under: Neuroscience > Cognition Psychology > Theory and Methods Neuroscience > Behavior Neuroscience > Cognition.
Collapse
|
27
|
Sound Localization in Real-Time Vocoded Cochlear-Implant Simulations With Normal-Hearing Listeners. Trends Hear 2019; 23:2331216519847332. [PMID: 31088265 PMCID: PMC6535744 DOI: 10.1177/2331216519847332] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/26/2022] Open
Abstract
Bilateral cochlear-implant (CI) users and single-sided deaf listeners with a CI are less effective at localizing sounds than normal-hearing (NH) listeners. This performance gap is due to the degradation of binaural and monaural sound localization cues, caused by a combination of device-related and patient-related issues. In this study, we targeted the device-related issues by measuring sound localization performance of 11 NH listeners, listening to free-field stimuli processed by a real-time CI vocoder. The use of a real-time vocoder is a new approach, which enables testing in a free-field environment. For the NH listening condition, all listeners accurately and precisely localized sounds according to a linear stimulus–response relationship with an optimal gain and a minimal bias both in the azimuth and in the elevation directions. In contrast, when listening with bilateral real-time vocoders, listeners tended to orient either to the left or to the right in azimuth and were unable to determine sound source elevation. When listening with an NH ear and a unilateral vocoder, localization was impoverished on the vocoder side but improved toward the NH side. Localization performance was also reflected by systematic variations in reaction times across listening conditions. We conclude that perturbation of interaural temporal cues, reduction of interaural level cues, and removal of spectral pinna cues by the vocoder impairs sound localization. Listeners seem to ignore cues that were made unreliable by the vocoder, leading to acute reweighting of available localization cues. We discuss how current CI processors prevent CI users from localizing sounds in everyday environments.
Collapse
|
28
|
Modeling active sensing reveals echo detection even in large groups of bats. Proc Natl Acad Sci U S A 2019; 116:26662-26668. [PMID: 31822613 DOI: 10.1073/pnas.1821722116] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Active sensing animals perceive their surroundings by emitting probes of energy and analyzing how the environment modulates these probes. However, the probes of conspecifics can jam active sensing, which should cause problems for groups of active sensing animals. This problem was termed the cocktail party nightmare for echolocating bats: as bats listen for the faint returning echoes of their loud calls, these echoes will be masked by the loud calls of other close-by bats. Despite this problem, many bats echolocate in groups and roost socially. Here, we present a biologically parametrized framework to quantify echo detection in groups. Incorporating properties of echolocation, psychoacoustics, acoustics, and group flight, we quantify how well bats flying in groups can detect each other despite jamming. A focal bat in the center of a group can detect neighbors in group sizes of up to 100 bats. With increasing group size, fewer and only the closest and frontal neighbors are detected. Neighbor detection is improved by longer call intervals, shorter call durations, denser groups, and more variable flight and sonar beam directions. Our results provide a quantification of the sensory input of echolocating bats in collective group flight, such as mating swarms or emergences. Our results further generate predictions on the sensory strategies bats may use to reduce jamming in the cocktail party nightmare. Lastly, we suggest that the spatially limited sensory field of echolocators leads to limited interactions within a group, so that collective behavior is achieved by following only nearest neighbors.
Collapse
|
29
|
Perception of Ultrasonic Vocalizations by Socially Housed and Isolated Mice. eNeuro 2019; 6:ENEURO.0049-19.2019. [PMID: 31570420 PMCID: PMC6794080 DOI: 10.1523/eneuro.0049-19.2019] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2019] [Revised: 09/06/2019] [Accepted: 09/10/2019] [Indexed: 11/22/2022] Open
Abstract
It is currently unclear whether mice use their ultrasonic vocalizations (USVs) for communication purposes. It is also unknown whether mice require previous experience with USVs to understand conspecifics. There is some evidence that experience changes the perception of juvenile USVs; however, it is unclear whether similar plasticity also occurs for adult USVs. To examine whether social exposure or deprivation throughout development leads to changes in USV perception, eleven female CBA/CaJ mice were trained to discriminate between 18 USVs of three different categories using operant conditioning procedures. Mice were group housed with four females or housed individually from weaning for the duration of the experiment. Socially housed and isolated mice differed in initial training times on pure tones, suggesting isolated mice had a more difficult time learning the task. Both groups completed USV discrimination conditions quicker at the end of the testing phases relative to the beginning. The overall discrimination of USVs did not differ between the two housing conditions, but a multidimensional scaling analysis revealed that socially experienced and isolated mice perceive some USVs differently, illustrated by differences in locations of USVs on the scaling maps from the two groups. Finally, a negative correlation was found between spectrotemporal similarity and percent discrimination, and analyses support the idea that mice may show categorical perception of at least two of the three USV categories. Thus, experience with USVs changes USV perception.
Collapse
|
30
|
Effects of Low Frequency Residual Hearing on Music Perception and Psychoacoustic Abilities in Pediatric Cochlear Implant Recipients. Front Neurosci 2019; 13:924. [PMID: 31551687 PMCID: PMC6733978 DOI: 10.3389/fnins.2019.00924] [Citation(s) in RCA: 20] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2019] [Accepted: 08/19/2019] [Indexed: 12/02/2022] Open
Abstract
Studies have demonstrated the benefits of low frequency residual hearing in music perception and for psychoacoustic abilities of adult cochlear implant (CI) users, but less is known about these effects in the pediatric group. Understanding the contribution of combined electric and acoustic stimulation in this group can help to gain a better perspective on decisions regarding bilateral implantation. We evaluated the performance of six unilaterally implanted children between 9 and 13 years of age with contralateral residual hearing using the Clinical Assessment of Music Perception (CAMP), spectral ripple discrimination (SRD), and temporal modulation transfer function (TMTF) tests and compared findings with previous research. Our study sample performed similarly to normal hearing subjects in pitch direction discrimination (0.81 semitones) and performed well above typical CI users in melody recognition (43.37%). The performance difference was less in timbre recognition (48.61%), SRD (1.47 ripple/octave), and TMTF for four modulation frequencies. These findings suggest that the combination of low frequency acoustic hearing with the broader frequency range of electric hearing can help to increase clinical CI benefit in pediatric users and decisions regarding second-side implantation should consider these factors.
Collapse
|
31
|
A Mathematical Approach to Correlating Objective Spectro-Temporal Features of Non-linguistic Sounds With Their Subjective Perceptions in Humans. Front Neurosci 2019; 13:794. [PMID: 31417350 PMCID: PMC6685481 DOI: 10.3389/fnins.2019.00794] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/11/2019] [Accepted: 07/16/2019] [Indexed: 11/13/2022] Open
Abstract
Non-linguistic sounds (NLSs) are a core feature of our everyday life and many evoke powerful cognitive and emotional outcomes. The subjective perception of NLSs by humans has occasionally been defined for single percepts, e.g., their pleasantness, whereas many NLSs evoke multiple perceptions. There has also been very limited attempt to determine if NLS perceptions are predicted from objective spectro-temporal features. We therefore examined three human perceptions well-established in previous NLS studies ("Complexity," "Pleasantness," and "Familiarity"), and the accuracy of identification, for a large NLS database and related these four measures to objective spectro-temporal NLS features, defined using rigorous mathematical descriptors including stimulus entropic and algorithmic complexity measures, peaks-related measures, fractal dimension estimates, and various spectral measures (mean spectral centroid, power in discrete frequency ranges, harmonicity, spectral flatness, and spectral structure). We mapped the perceptions to the spectro-temporal measures individually and in combinations, using complex multivariate analyses including principal component analyses and agglomerative hierarchical clustering.
Collapse
|
32
|
The effects of age and sex on the detection of pure tones by adult CBA/CaJ mice (Mus musculus). J Neurosci Res 2019; 98:1731-1744. [PMID: 31304616 DOI: 10.1002/jnr.24496] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2019] [Revised: 05/23/2019] [Accepted: 06/28/2019] [Indexed: 12/28/2022]
Abstract
Age-related hearing loss (ARHL) is a neurodegenerative disorder characterized by a gradual decrease in hearing sensitivity. Previous electrophysiological and behavioral studies have demonstrated that the CBA/CaJ mouse strain is an appropriate model for the late-onset hearing loss found in humans. However, few studies have characterized hearing in these mice behaviorally using longitudinal methodologies. The goal of this research was to utilize a longitudinal design and operant conditioning procedures with positive reinforcement to construct audiograms and temporal integration functions in aging CBA/CaJ mice. In the first experiment, thresholds were collected for 8, 16, 24, 42, and 64 kHz pure tones in 30 male and 35 female CBA/CaJ mice. Similar to humans, mice had higher thresholds for high frequency tones than for low frequency pure tones across the lifespan. Female mice had better hearing acuity than males after 645 days of age. In the second experiment, temporal integration functions were constructed for 18 male and 18 female mice for 16 and 64 kHz tones varying in duration. Mice showed an increase in thresholds for tones shorter than 200 ms, reaching peak performance at shorter durations than other rodent species. Overall, CBA/CaJ mice experience ARHL for pure tones of different frequencies and durations, making them a good model for studies on hearing loss. These findings highlight the importance of using a wide range of stimuli and a longitudinal design when comparing presbycusis across different species.
Collapse
|
33
|
Environmental Sounds Influence the Multisensory Perception of Chocolate Gelati. Foods 2019; 8:E124. [PMID: 30991748 PMCID: PMC6517920 DOI: 10.3390/foods8040124] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2019] [Revised: 04/08/2019] [Accepted: 04/10/2019] [Indexed: 11/20/2022] Open
Abstract
Recently, it has been shown that various auditory stimuli modulate flavour perception. The present study attempts to understand the effects of environmental sounds (park, food court, fast food restaurant, cafe, and bar sounds) on the perception of chocolate gelato (specifically, sweet, bitter, milky, creamy, cocoa, roasted, and vanilla notes) using the Temporal Check-All-That-Apply (TCATA) method. Additionally, affective ratings of the auditory stimuli were obtained using the Self-Assessment Manikin (SAM) in terms of their valence, arousal, and dominance. In total, 58 panellists rated the sounds and chocolate gelato in a sensory laboratory. The results revealed that bitterness, roasted, and cocoa notes were more evident when the bar, fast food, and food court sounds were played. Meanwhile, sweetness was cited more in the early mastication period when listening to park and café sounds. The park sound was significantly higher in valence, while the bar sound was significantly higher in arousal. Dominance was significantly higher for the fast food restaurant, food court, and bar sound conditions. Intriguingly, the valence evoked by the pleasant park sound was positively correlated with the sweetness of the gelato. Meanwhile, the arousal associated with bar sounds was positively correlated with bitterness, roasted, and cocoa attributes. Taken together, these results clearly demonstrate that people's perception of the flavour of gelato varied with the different real-world sounds used in this study.
Collapse
|
34
|
Abstract
Tinnitus masking and residual inhibition (RI) are two well-known psychoacoustic measures of tinnitus. While it has long been suggested that they may provide diagnostic and prognostic information, these measures are still rarely performed in clinics, as they are too time consuming. Given this issue, the main goal of the present study was to validate a new method for assessing these measures. An acoustic sequence made of pulsed stimuli, which included a fixed stimulus duration and interstimulus interval, was applied to 68 tinnitus patients at two testing sites. First, the minimum masking level (MML) was measured by raising the stimulus intensity until the tinnitus was unheard during the stimulus presentation. Second, the level of the stimulus was further increased until the tinnitus was suppressed during the silence interval between the acoustic pulses. This level was called the minimum residual inhibition level (MRIL). The sequential measurement of MML and MRIL from the same stimulus condition offers several advantages such as time efficiency and the ability to compare results between the MRIL and MML. Our study confirms that, from this new approach, MML and MRIL can be easily and quickly obtained from a wide variety of patients displaying either normal hearing or different hearing loss configurations. Indeed, MML was obtained in all patients except one (98.5%), and some level of MRIL was found on 59 patients (86.7%). More so, this approach allows the categorization of tinnitus patients into different subgroups based on the properties of their MRIL.
Collapse
|
35
|
Effects of Acquired Aphasia on the Recognition of Speech Under Energetic and Informational Masking Conditions. Trends Hear 2019; 23:2331216519884480. [PMID: 31694486 PMCID: PMC7000861 DOI: 10.1177/2331216519884480] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2019] [Revised: 09/24/2019] [Accepted: 10/01/2019] [Indexed: 11/16/2022] Open
Abstract
Persons with aphasia (PWA) often report difficulty understanding spoken language in noisy environments that require listeners to identify and selectively attend to target speech while ignoring competing background sounds or “maskers.” This study compared the performance of PWA and age-matched healthy controls (HC) on a masked speech identification task and examined the consequences of different types of masking on performance. Twelve PWA and 12 age-matched HC completed a speech identification task comprising three conditions designed to differentiate between the effects of energetic and informational masking on receptive speech processing. The target and masker speech materials were taken from a closed-set matrix-style corpus, and a forced-choice word identification task was used. Target and maskers were spatially separated from one another in order to simulate real-world listening environments and allow listeners to make use of binaural cues for source segregation. Individualized frequency-specific gain was applied to compensate for the effects of hearing loss. Although both groups showed similar susceptibility to the effects of energetic masking, PWA were more susceptible than age-matched HC to the effects of informational masking. Results indicate that this increased susceptibility cannot be attributed to age, hearing loss, or comprehension deficits and is therefore a consequence of acquired cognitive-linguistic impairments associated with aphasia. This finding suggests that aphasia may result in increased difficulty segregating target speech from masker speech, which in turn may have implications for the ability of PWA to comprehend target speech in multitalker environments, such as restaurants, family gatherings, and other everyday situations.
Collapse
|
36
|
Abstract
Frequency analysis of sound by the cochlea is the most fundamental property of the auditory system. Despite its importance, the resolution of this frequency analysis in humans remains controversial. The controversy persists because the methods used to estimate tuning in humans are indirect and have not all been independently validated in other species. Some data suggest that human cochlear tuning is considerably sharper than that of laboratory animals, while others suggest little or no difference between species. We show here in a single species (ferret) that behavioral estimates of tuning bandwidths obtained using perceptual masking methods, and objective estimates obtained using otoacoustic emissions, both also employed in humans, agree closely with direct physiological measurements from single auditory-nerve fibers. Combined with human behavioral data, this outcome indicates that the frequency analysis performed by the human cochlea is of significantly higher resolution than found in common laboratory animals. This finding raises important questions about the evolutionary origins of human cochlear tuning, its role in the emergence of speech communication, and the mechanisms underlying our ability to separate and process natural sounds in complex acoustic environments.
Collapse
|
37
|
On the Pitch Strength of Bandpass Noise in Normal-Hearing and Hearing-Impaired Listeners. Trends Hear 2018; 22:2331216518787067. [PMID: 30009682 PMCID: PMC6048668 DOI: 10.1177/2331216518787067] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
The psychoacoustic measure pitch strength describes the strength of the tonal sensation evoked by a sound on a scale from weak to strong. For normal-hearing listeners, it was shown in the literature that pitch strength of bandpass noise (relative to the pitch strength of a sinusoid at its center frequency) decreases with increasing bandwidth. This decrease also depends on the center frequency. These effects were often attributed to the frequency selectivity of the auditory system. The present study investigated the relative pitch strength of bandpass noise in hearing-impaired listeners and for comparison in a normal-hearing control group. For the normal-hearing listeners, pitch strength was measured at sound pressure levels of 30 and 70 dB SPL for bandwidths between 5 and 1620 Hz and center frequencies of 375, 750, and 1500 Hz. In addition, two ways of generating the stimuli (filtering in frequency or time domain) were used to compare the data with previous results. Apart from the known effect of center frequency on the change of relative pitch strength with increasing bandwidth, stimulus generation also had a significant influence on the results. Relative pitch strength of bandpass noise in hearing-impaired listeners was measured for bandwidths from 5 to 1620 Hz; the center frequency was 1500 Hz. Compared with the corresponding results of the normal hearing, relative pitch strength was altered in the hearing-impaired listeners. These alterations, however, could not be explained by altered spectral processing in the damaged cochlea alone.
Collapse
|
38
|
Olivocochlear Efferents in Animals and Humans: From Anatomy to Clinical Relevance. Front Neurol 2018; 9:197. [PMID: 29632514 PMCID: PMC5879449 DOI: 10.3389/fneur.2018.00197] [Citation(s) in RCA: 81] [Impact Index Per Article: 13.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2018] [Accepted: 03/13/2018] [Indexed: 11/13/2022] Open
Abstract
Olivocochlear efferents allow the central auditory system to adjust the functioning of the inner ear during active and passive listening. While many aspects of efferent anatomy, physiology and function are well established, others remain controversial. This article reviews the current knowledge on olivocochlear efferents, with emphasis on human medial efferents. The review covers (1) the anatomy and physiology of olivocochlear efferents in animals; (2) the methods used for investigating this auditory feedback system in humans, their limitations and best practices; (3) the characteristics of medial-olivocochlear efferents in humans, with a critical analysis of some discrepancies across human studies and between animal and human studies; (4) the possible roles of olivocochlear efferents in hearing, discussing the evidence in favor and against their role in facilitating the detection of signals in noise and in protecting the auditory system from excessive acoustic stimulation; and (5) the emerging association between abnormal olivocochlear efferent function and several health conditions. Finally, we summarize some open issues and introduce promising approaches for investigating the roles of efferents in human hearing using cochlear implants.
Collapse
|
39
|
Spatio-Temporal Analysis of Urban Acoustic Environments with Binaural Psycho-Acoustical Considerations for IoT-Based Applications. SENSORS 2018; 18:s18030690. [PMID: 29495407 PMCID: PMC5877108 DOI: 10.3390/s18030690] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/18/2018] [Revised: 02/14/2018] [Accepted: 02/21/2018] [Indexed: 12/03/2022]
Abstract
Sound pleasantness or annoyance perceived in urban soundscapes is a major concern in environmental acoustics. Binaural psychoacoustic parameters are helpful to describe generic acoustic environments, as it is stated within the ISO 12913 framework. In this paper, the application of a Wireless Acoustic Sensor Network (WASN) to evaluate the spatial distribution and the evolution of urban acoustic environments is described. Two experiments are presented using an indoor and an outdoor deployment of a WASN with several nodes using an Internet of Things (IoT) environment to collect audio data and calculate meaningful parameters such as the sound pressure level, binaural loudness and binaural sharpness. A chunk of audio is recorded in each node periodically with a microphone array and the binaural rendering is conducted by exploiting the estimated directional characteristics of the incoming sound by means of DOA estimation. Each node computes the parameters in a different location and sends the values to a cloud-based broker structure that allows spatial statistical analysis through Kriging techniques. A cross-validation analysis is also performed to confirm the usefulness of the proposed system.
Collapse
|
40
|
Four Distinctions for the Auditory "Wastebasket" of Timbre. Front Psychol 2017; 8:1747. [PMID: 29046659 PMCID: PMC5632649 DOI: 10.3389/fpsyg.2017.01747] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2017] [Accepted: 09/21/2017] [Indexed: 11/13/2022] Open
|
41
|
Spoken Word Recognition Enhancement Due to Preceding Synchronized Beats Compared to Unsynchronized or Unrhythmic Beats. Front Neurosci 2017; 11:415. [PMID: 28769752 PMCID: PMC5513984 DOI: 10.3389/fnins.2017.00415] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2017] [Accepted: 07/04/2017] [Indexed: 11/16/2022] Open
Abstract
The relation between rhythm and language has been investigated over the last decades, with evidence that these share overlapping perceptual mechanisms emerging from several different strands of research. The dynamic Attention Theory posits that neural entrainment to musical rhythm results in synchronized oscillations in attention, enhancing perception of other events occurring at the same rate. In this study, this prediction was tested in 10 year-old children by means of a psychoacoustic speech recognition in babble paradigm. It was hypothesized that rhythm effects evoked via a short isochronous sequence of beats would provide optimal word recognition in babble when beats and word are in sync. We compared speech recognition in babble performance in the presence of isochronous and in sync vs. non-isochronous or out of sync sequence of beats. Results showed that (a) word recognition was the best when rhythm and word were in sync, and (b) the effect was not uniform across syllables and gender of subjects. Our results suggest that pure tone beats affect speech recognition at early levels of sensory or phonemic processing.
Collapse
|
42
|
Sound Spectrum Influences Auditory Distance Perception of Sound Sources Located in a Room Environment. Front Psychol 2017; 8:969. [PMID: 28690556 PMCID: PMC5479918 DOI: 10.3389/fpsyg.2017.00969] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2016] [Accepted: 05/26/2017] [Indexed: 12/03/2022] Open
Abstract
Previous studies on the effect of spectral content on auditory distance perception (ADP) focused on the physically measurable cues occurring either in the near field (low-pass filtering due to head diffraction) or when the sound travels distances >15 m (high-frequency energy losses due to air absorption). Here, we study how the spectrum of a sound arriving from a source located in a reverberant room at intermediate distances (1–6 m) influences the perception of the distance to the source. First, we conducted an ADP experiment using pure tones (the simplest possible spectrum) of frequencies 0.5, 1, 2, and 4 kHz. Then, we performed a second ADP experiment with stimuli consisting of continuous broadband and bandpass-filtered (with center frequencies of 0.5, 1.5, and 4 kHz and bandwidths of 1/12, 1/3, and 1.5 octave) pink-noise clips. Our results showed an effect of the stimulus frequency on the perceived distance both for pure tones and filtered noise bands: ADP was less accurate for stimuli containing energy only in the low-frequency range. Analysis of the frequency response of the room showed that the low accuracy observed for low-frequency stimuli can be explained by the presence of sparse modal resonances in the low-frequency region of the spectrum, which induced a non-monotonic relationship between binaural intensity and source distance. The results obtained in the second experiment suggest that ADP can also be affected by stimulus bandwidth but in a less straightforward way (i.e., depending on the center frequency, increasing stimulus bandwidth could have different effects). Finally, the analysis of the acoustical cues suggests that listeners judged source distance using mainly changes in the overall intensity of the auditory stimulus with distance rather than the direct-to-reverberant energy ratio, even for low-frequency noise bands (which typically induce high amount of reverberation). The results obtained in this study show that, depending on the spectrum of the auditory stimulus, reverberation can degrade ADP rather than improve it.
Collapse
|
43
|
Corrigendum: A Mixed-Methods Trial of Broad Band Noise and Nature Sounds for Tinnitus Therapy: Group and Individual Responses Modeled under the Adaptation Level Theory of Tinnitus. Front Aging Neurosci 2017; 9:116. [PMID: 28491033 PMCID: PMC5422423 DOI: 10.3389/fnagi.2017.00116] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2017] [Accepted: 04/10/2017] [Indexed: 11/25/2022] Open
|
44
|
Stimulus Pauses and Perturbations Differentially Delay or Promote the Segregation of Auditory Objects: Psychoacoustics and Modeling. Front Neurosci 2017; 11:198. [PMID: 28473747 PMCID: PMC5397483 DOI: 10.3389/fnins.2017.00198] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2016] [Accepted: 03/23/2017] [Indexed: 11/21/2022] Open
Abstract
Segregating distinct sound sources is fundamental for auditory perception, as in the cocktail party problem. In a process called the build-up of stream segregation, distinct sound sources that are perceptually integrated initially can be segregated into separate streams after several seconds. Previous research concluded that abrupt changes in the incoming sounds during build-up—for example, a step change in location, loudness or timing—reset the percept to integrated. Following this reset, the multisecond build-up process begins again. Neurophysiological recordings in auditory cortex (A1) show fast (subsecond) adaptation, but unified mechanistic explanations for the bias toward integration, multisecond build-up and resets remain elusive. Combining psychoacoustics and modeling, we show that initial unadapted A1 responses bias integration, that the slowness of build-up arises naturally from competition downstream, and that recovery of adaptation can explain resets. An early bias toward integrated perceptual interpretations arising from primary cortical stages that encode low-level features and feed into competition downstream could also explain similar phenomena in vision. Further, we report a previously overlooked class of perturbations that promote segregation rather than integration. Our results challenge current understanding for perturbation effects on the emergence of sound source segregation, leading to a new hypothesis for differential processing downstream of A1. Transient perturbations can momentarily redirect A1 responses as input to downstream competition units that favor segregation.
Collapse
|
45
|
A Mixed-Methods Trial of Broad Band Noise and Nature Sounds for Tinnitus Therapy: Group and Individual Responses Modeled under the Adaptation Level Theory of Tinnitus. Front Aging Neurosci 2017; 9:44. [PMID: 28337139 PMCID: PMC5343046 DOI: 10.3389/fnagi.2017.00044] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2016] [Accepted: 02/20/2017] [Indexed: 12/12/2022] Open
Abstract
Objectives: A randomized cross-over trial in 18 participants tested the hypothesis that nature sounds, with unpredictable temporal characteristics and high valence would yield greater improvement in tinnitus than constant, emotionally neutral broadband noise. Study Design: The primary outcome measure was the Tinnitus Functional Index (TFI). Secondary measures were: loudness and annoyance ratings, loudness level matches, minimum masking levels, positive and negative emotionality, attention reaction and discrimination time, anxiety, depression and stress. Each sound was administered using MP3 players with earbuds for 8 continuous weeks, with a 3 week wash-out period before crossing over to the other treatment sound. Measurements were undertaken for each arm at sound fitting, 4 and 8 weeks after administration. Qualitative interviews were conducted at each of these appointments. Results: From a baseline TFI score of 41.3, sound therapy resulted in TFI scores at 8 weeks of 35.6; broadband noise resulted in significantly greater reduction (8.2 points) after 8 weeks of sound therapy use than nature sounds (3.2 points). The positive effect of sound on tinnitus was supported by secondary outcome measures of tinnitus, emotion, attention, and psychological state, but not interviews. Tinnitus loudness level match was higher for BBN at 8 weeks; while there was little change in loudness level matches for nature sounds. There was no change in minimum masking levels following sound therapy administration. Self-reported preference for one sound over another did not correlate with changes in tinnitus. Conclusions: Modeled under an adaptation level theory framework of tinnitus perception, the results indicate that the introduction of broadband noise shifts internal adaptation level weighting away from the tinnitus signal, reducing tinnitus magnitude. Nature sounds may modify the affective components of tinnitus via a secondary, residual pathway, but this appears to be less important for sound effectiveness. The different rates of adaptation to broadband noise and nature sound by the auditory system may explain the different tinnitus loudness level matches. In addition to group effects there also appears to be a great deal of individual variation. A sound therapy framework based on adaptation level theory is proposed that accounts for individual variation in preference and response to sound. Clinical Trial Registration: www.anzctr.org.au, identifier #12616000742471.
Collapse
|
46
|
Corrigendum: Auditory and cognitive factors underlying individual differences in aided speech-understanding among older adults. Front Syst Neurosci 2016; 10:91. [PMID: 27843432 PMCID: PMC5105598 DOI: 10.3389/fnsys.2016.00091] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2016] [Accepted: 10/28/2016] [Indexed: 11/25/2022] Open
|
47
|
Statistics of natural reverberation enable perceptual separation of sound and space. Proc Natl Acad Sci U S A 2016; 113:E7856-E7865. [PMID: 27834730 DOI: 10.1073/pnas.1612524113] [Citation(s) in RCA: 58] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
In everyday listening, sound reaches our ears directly from a source as well as indirectly via reflections known as reverberation. Reverberation profoundly distorts the sound from a source, yet humans can both identify sound sources and distinguish environments from the resulting sound, via mechanisms that remain unclear. The core computational challenge is that the acoustic signatures of the source and environment are combined in a single signal received by the ear. Here we ask whether our recognition of sound sources and spaces reflects an ability to separate their effects and whether any such separation is enabled by statistical regularities of real-world reverberation. To first determine whether such statistical regularities exist, we measured impulse responses (IRs) of 271 spaces sampled from the distribution encountered by humans during daily life. The sampled spaces were diverse, but their IRs were tightly constrained, exhibiting exponential decay at frequency-dependent rates: Mid frequencies reverberated longest whereas higher and lower frequencies decayed more rapidly, presumably due to absorptive properties of materials and air. To test whether humans leverage these regularities, we manipulated IR decay characteristics in simulated reverberant audio. Listeners could discriminate sound sources and environments from these signals, but their abilities degraded when reverberation characteristics deviated from those of real-world environments. Subjectively, atypical IRs were mistaken for sound sources. The results suggest the brain separates sound into contributions from the source and the environment, constrained by a prior on natural reverberation. This separation process may contribute to robust recognition while providing information about spaces around us.
Collapse
|
48
|
Relating approach-to-target and detection tasks in animal psychoacoustics. Behav Neurosci 2016; 130:393-405. [PMID: 27196623 PMCID: PMC4961071 DOI: 10.1037/bne0000143] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/26/2015] [Revised: 03/15/2016] [Accepted: 03/17/2016] [Indexed: 11/08/2022]
Abstract
Psychophysical experiments seek to measure the limits of perception. While straightforward in humans, in animals they are time consuming. Choosing an appropriate task and interpreting measurements can be challenging. We investigated the localization of high-frequency auditory signals in noise using an "approach-to-target" task in ferrets, how task performance should be interpreted in terms of perception, and how the measurements relate to other types of tasks. To establish their general ability to localize, animals were first trained to discriminate broadband noise from 12 locations. Subsequently we tested their ability to discriminate between band-limited targets at 2 or 3 more widely spaced locations, in a continuous background noise. The ability to discriminate between 3 possible locations (-90°, 0°, 90°) of a 10-kHz pure tone decreased gradually over a wide range (>30 dB) of signal-to-noise ratios (SNRs). Location discrimination ability was better for wide band noise targets (0.5 and 2 octave). These results were consistent with localization ability limiting performance for pure tones. Discrimination of pure tones at 2 locations (-90/left, 90/right) was robust at positive SNRs, yielding psychometric functions which fell steeply at negative SNRs. Thresholds for discrimination were similar to previous tone-in-noise thresholds measured in ferrets using a yes/no task. Thus, using an approach-to-target task, sound "localization" in noise can reflect detectability or the ability to localize, depending on the stimulus configuration. Signal-detection-theory-based models were able to account for the results when discriminating between pure tones from 2- and 3-source locations. (PsycINFO Database Record
Collapse
|
49
|
Selecting Appropriate Tests to Assess the Benefits of Bilateral Amplification With Hearing Aids. Trends Hear 2016; 20:20/0/2331216516658239. [PMID: 27460871 PMCID: PMC4964154 DOI: 10.1177/2331216516658239] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/02/2022] Open
Abstract
The aim of this study was to investigate the effect of bilateral hearing aids (HA) in subjects with mild and moderate-to-severe hearing loss. This study was designed as a within-subject feasibility study. Bilateral HA use was assessed using different laboratory tests on speech reception, listening effort, noise tolerance, and localization. All data were evaluated with bilateral and unilateral HA fittings. Forty experienced bilateral HA users were included with hearing impairment ranging from mild to moderate-to-severe. Subjects were stratified into two groups based on the degree of hearing loss. Speech reception in noise, listening effort, and localization tests showed a bilateral benefit for the moderate-to-severely hearing-impaired subjects. A bilateral benefit was also observed for listening effort in the mildly hearing-impaired group. The assessment of listening effort shows promise as a measure of bilateral HA benefit for mild hearing impairment. Localization and speech reception in noise tests provide additional value for larger losses. The next step is to compare experienced unilateral with bilateral HA users.
Collapse
|
50
|
Abstract
Previous research on harmony perception has mainly been concerned with horizontal aspects of harmony, turning less attention to how listeners perceive psychoacoustic qualities and emotions in single isolated chords. A recent study found mild dissonances to be more preferred than consonances in single chord perception, although the authors did not systematically vary register and consonance in their study; these omissions were explored here. An online empirical experiment was conducted where participants (N = 410) evaluated chords on the dimensions of Valence, Tension, Energy, Consonance, and Preference; 15 different chords were played with piano timbre across two octaves. The results suggest significant differences on all dimensions across chord types, and a strong correlation between perceived dissonance and tension. The register and inversions contributed to the evaluations significantly, nonmusicians distinguishing between triadic inversions similarly to musicians. The mildly dissonant minor ninth, major ninth, and minor seventh chords were rated highest for preference, regardless of musical sophistication. The role of theoretical explanations such as aggregate dyadic consonance, the inverted-U hypothesis, and psychoacoustic roughness, harmonicity, and sharpness will be discussed to account for the preference of mild dissonance over consonance in single chord perception.
Collapse
|