1
|
Suresh CH, Krishnan A. Frequency-Following Response to Steady-State Vowel in Quiet and Background Noise Among Marching Band Participants With Normal Hearing. Am J Audiol 2022; 31:719-736. [PMID: 35944059 DOI: 10.1044/2022_aja-21-00226] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022] Open
Abstract
OBJECTIVE Human studies enrolling individuals at high risk for cochlear synaptopathy (CS) have reported difficulties in speech perception in adverse listening conditions. The aim of this study is to determine if these individuals show a degradation in the neural encoding of speech in quiet and in the presence of background noise as reflected in neural phase-locking to both envelope periodicity and temporal fine structure (TFS). To our knowledge, there are no published reports that have specifically examined the neural encoding of both envelope periodicity and TFS of speech stimuli (in quiet and in adverse listening conditions) among a sample with loud-sound exposure history who are at risk for CS. METHOD Using scalp-recorded frequency-following response (FFR), the authors evaluated the neural encoding of envelope periodicity (FFRENV) and TFS (FFRTFS) for a steady-state vowel (English back vowel /u/) in quiet and in the presence of speech-shaped noise presented at +5- and 0 dB SNR. Participants were young individuals with normal hearing who participated in the marching band for at least 5 years (high-risk group) and non-marching band group with low-noise exposure history (low-risk group). RESULTS The results showed no group differences in the neural encoding of either the FFRENV or the first formant (F1) in the FFRTFS in quiet and in noise. Paradoxically, the high-risk group demonstrated enhanced representation of F2 harmonics across all stimulus conditions. CONCLUSIONS These results appear to be in line with a music experience-dependent enhancement of F2 harmonics. However, due to sound overexposure in the high-risk group, the role of homeostatic central compensation cannot be ruled out. A larger scale data set with different noise exposure background, longitudinal measurements with an array of behavioral and electrophysiological tests is needed to disentangle the nature of the complex interaction between the effects of central compensatory gain and experience-dependent enhancement.
Collapse
Affiliation(s)
- Chandan H Suresh
- Department of Communication Disorders, California State University, Los Angeles
| | | |
Collapse
|
2
|
Chauvette L, Fournier P, Sharp A. The frequency-following response to assess the neural representation of spectral speech cues in older adults. Hear Res 2022; 418:108486. [DOI: 10.1016/j.heares.2022.108486] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/01/2021] [Revised: 03/12/2022] [Accepted: 03/15/2022] [Indexed: 11/04/2022]
|
3
|
Shukla B, Bidelman GM. Enhanced brainstem phase-locking in low-level noise reveals stochastic resonance in the frequency-following response (FFR). Brain Res 2021; 1771:147643. [PMID: 34473999 PMCID: PMC8490316 DOI: 10.1016/j.brainres.2021.147643] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2021] [Revised: 08/23/2021] [Accepted: 08/28/2021] [Indexed: 11/29/2022]
Abstract
In nonlinear systems, the inclusion of low-level noise can paradoxically improve signal detection, a phenomenon known as stochastic resonance (SR). SR has been observed in human hearing whereby sensory thresholds (e.g., signal detection and discrimination) are enhanced in the presence of noise. Here, we asked whether subcortical auditory processing (neural phase locking) shows evidence of SR. We recorded brainstem frequency-following-responses (FFRs) in young, normal-hearing listeners to near-electrophysiological-threshold (40 dB SPL) complex tones composed of 10 iso-amplitude harmonics of 150 Hz fundamental frequency (F0) presented concurrent with low-level noise (+20 to -20 dB SNRs). Though variable and weak across ears, some listeners showed improvement in auditory detection thresholds with subthreshold noise confirming SR psychophysically. At the neural level, low-level FFRs were initially eradicated by noise (expected masking effect) but were surprisingly reinvigorated at select masker levels (local maximum near ∼ 35 dB SPL). These data suggest brainstem phase-locking to near threshold periodic stimuli is enhanced in optimal levels of noise, the hallmark of SR. Our findings provide novel evidence for stochastic resonance in the human auditory brainstem and suggest that under some circumstances, noise can actually benefit both the behavioral and neural encoding of complex sounds.
Collapse
Affiliation(s)
- Bhanu Shukla
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA
| | - Gavin M Bidelman
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA; University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA.
| |
Collapse
|
4
|
Mai G, Howell P. Causal Relationship between the Right Auditory Cortex and Speech-Evoked Envelope-Following Response: Evidence from Combined Transcranial Stimulation and Electroencephalography. Cereb Cortex 2021; 32:1437-1454. [PMID: 34424956 PMCID: PMC8971082 DOI: 10.1093/cercor/bhab298] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2021] [Revised: 07/26/2021] [Accepted: 07/27/2021] [Indexed: 11/27/2022] Open
Abstract
Speech-evoked envelope-following response (EFR) reflects brain encoding of speech periodicity that serves as a biomarker for pitch and speech perception and various auditory and language disorders. Although EFR is thought to originate from the subcortex, recent research illustrated a right-hemispheric cortical contribution to EFR. However, it is unclear whether this contribution is causal. This study aimed to establish this causality by combining transcranial direct current stimulation (tDCS) and measurement of EFR (pre- and post-tDCS) via scalp-recorded electroencephalography. We applied tDCS over the left and right auditory cortices in right-handed normal-hearing participants and examined whether altering cortical excitability via tDCS causes changes in EFR during monaural listening to speech syllables. We showed significant changes in EFR magnitude when tDCS was applied over the right auditory cortex compared with sham stimulation for the listening ear contralateral to the stimulation site. No such effect was found when tDCS was applied over the left auditory cortex. Crucially, we further observed a hemispheric laterality where aftereffect was significantly greater for tDCS applied over the right than the left auditory cortex in the contralateral ear condition. Our finding thus provides the first evidence that validates the causal relationship between the right auditory cortex and EFR.
Collapse
Affiliation(s)
- Guangting Mai
- Hearing Theme, National Institute for Health Research Nottingham Biomedical Research Centre, Nottingham NG1 5DU, UK.,Division of Clinical Neuroscience, School of Medicine, University of Nottingham, Nottingham NG7 2UH, UK.,Department of Experimental Psychology, University College London, London WC1H 0AP, UK
| | - Peter Howell
- Department of Experimental Psychology, University College London, London WC1H 0AP, UK
| |
Collapse
|
5
|
Subcortical rather than cortical sources of the frequency-following response (FFR) relate to speech-in-noise perception in normal-hearing listeners. Neurosci Lett 2021; 746:135664. [PMID: 33497718 DOI: 10.1016/j.neulet.2021.135664] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2020] [Revised: 12/22/2020] [Accepted: 01/13/2021] [Indexed: 12/27/2022]
Abstract
Scalp-recorded frequency-following responses (FFRs) reflect a mixture of phase-locked activity across the auditory pathway. FFRs have been widely used as a neural barometer of complex listening skills, especially speech-in noise (SIN) perception. Applying individually optimized source reconstruction to speech-FFRs recorded via EEG (FFREEG), we assessed the relative contributions of subcortical [auditory nerve (AN), brainstem/midbrain (BS)] and cortical [bilateral primary auditory cortex, PAC] source generators with the aim of identifying which source(s) drive the brain-behavior relation between FFRs and SIN listening skills. We found FFR strength declined precipitously from AN to PAC, consistent with diminishing phase-locking along the ascending auditory neuroaxis. FFRs to the speech fundamental (F0) were robust to noise across sources, but were largest in subcortical sources (BS > AN > PAC). PAC FFRs were only weakly observed above the noise floor and only at the low pitch of speech (F0≈100 Hz). Brain-behavior regressions revealed (i) AN and BS FFRs were sufficient to describe listeners' QuickSIN scores and (ii) contrary to neuromagnetic (MEG) FFRs, neither left nor right PAC FFREEG related to SIN performance. Our findings suggest subcortical sources not only dominate the electrical FFR but also the link between speech-FFRs and SIN processing in normal-hearing adults as observed in previous EEG studies.
Collapse
|
6
|
Kessler DM, Ananthakrishnan S, Smith SB, D'Onofrio K, Gifford RH. Frequency Following Response and Speech Recognition Benefit for Combining a Cochlear Implant and Contralateral Hearing Aid. Trends Hear 2020; 24:2331216520902001. [PMID: 32003296 PMCID: PMC7257083 DOI: 10.1177/2331216520902001] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/15/2022] Open
Abstract
Multiple studies have shown significant speech recognition benefit when acoustic hearing is combined with a cochlear implant (CI) for a bimodal hearing configuration. However, this benefit varies greatly between individuals. There are few clinical measures correlated with bimodal benefit and those correlations are driven by extreme values prohibiting data-driven, clinical counseling. This study evaluated the relationship between neural representation of fundamental frequency (F0) and temporal fine structure via the frequency following response (FFR) in the nonimplanted ear as well as spectral and temporal resolution of the nonimplanted ear and bimodal benefit for speech recognition in quiet and noise. Participants included 14 unilateral CI users who wore a hearing aid (HA) in the nonimplanted ear. Testing included speech recognition in quiet and in noise with the HA-alone, CI-alone, and in the bimodal condition (i.e., CI + HA), measures of spectral and temporal resolution in the nonimplanted ear, and FFR recording for a 170-ms/da/stimulus in the nonimplanted ear. Even after controlling for four-frequency pure-tone average, there was a significant correlation (r = .83) between FFR F0 amplitude in the nonimplanted ear and bimodal benefit. Other measures of auditory function of the nonimplanted ear were not significantly correlated with bimodal benefit. The FFR holds potential as an objective tool that may allow data-driven counseling regarding expected benefit from the nonimplanted ear. It is possible that this information may eventually be used for clinical decision-making, particularly in difficult-to-test populations such as young children, regarding effectiveness of bimodal hearing versus bilateral CI candidacy.
Collapse
Affiliation(s)
- David M Kessler
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN, USA
| | | | - Spencer B Smith
- Department of Communication Sciences and Disorders, The University of Texas at Austin, TX, USA
| | - Kristen D'Onofrio
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN, USA
| | - René H Gifford
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN, USA.,Department of Otolaryngology, Vanderbilt University Medical Center, Nashville, TN, USA
| |
Collapse
|
7
|
Abstract
OBJECTIVES There is increasing interest in using the frequency following response (FFR) to describe the effects of varying different aspects of hearing aid signal processing on brainstem neural representation of speech. To this end, recent studies have examined the effects of filtering on brainstem neural representation of the speech fundamental frequency (f0) in listeners with normal hearing sensitivity by measuring FFRs to low- and high-pass filtered signals. However, the stimuli used in these studies do not reflect the entire range of typical cutoff frequencies used in frequency-specific gain adjustments during hearing aid fitting. Further, there has been limited discussion on the effect of filtering on brainstem neural representation of formant-related harmonics. Here, the effects of filtering on brainstem neural representation of speech fundamental frequency (f0) and harmonics related to first formant frequency (F1) were assessed by recording envelope and spectral FFRs to a vowel low-, high-, and band-pass filtered at cutoff frequencies ranging from 0.125 to 8 kHz. DESIGN FFRs were measured to a synthetically generated vowel stimulus /u/ presented in a full bandwidth and low-pass (experiment 1), high-pass (experiment 2), and band-pass (experiment 3) filtered conditions. In experiment 1, FFRs were measured to a synthetically generated vowel stimulus /u/ presented in a full bandwidth condition as well as 11 low-pass filtered conditions (low-pass cutoff frequencies: 0.125, 0.25, 0.5, 0.75, 1, 1.5, 2, 3, 4, 6, and 8 kHz) in 19 adult listeners with normal hearing sensitivity. In experiment 2, FFRs were measured to the same synthetically generated vowel stimulus /u/ presented in a full bandwidth condition as well as 10 high-pass filtered conditions (high-pass cutoff frequencies: 0.125, 0.25, 0.5, 0.75, 1, 1.5, 2, 3, 4, and 6 kHz) in 7 adult listeners with normal hearing sensitivity. In experiment 3, in addition to the full bandwidth condition, FFRs were measured to vowel /u/ low-pass filtered at 2 kHz, band-pass filtered between 2-4 kHz and 4-6 kHz in 10 adult listeners with normal hearing sensitivity. A Fast Fourier Transform analysis was conducted to measure the strength of f0 and the F1-related harmonic relative to the noise floor in the brainstem neural responses obtained to the full bandwidth and filtered stimulus conditions. RESULTS Brainstem neural representation of f0 was reduced when the low-pass filter cutoff frequency was between 0.25 and 0.5 kHz; no differences in f0 strength were noted between conditions when the low-pass filter cutoff condition was at or greater than 0.75 kHz. While envelope FFR f0 strength was reduced when the stimulus was high-pass filtered at 6 kHz, there was no effect of high-pass filtering on brainstem neural representation of f0 when the high-pass filter cutoff frequency ranged from 0.125 to 4 kHz. There was a weakly significant global effect of band-pass filtering on brainstem neural phase-locking to f0. A trends analysis indicated that mean f0 magnitude in the brainstem neural response was greater when the stimulus was band-pass filtered between 2 and 4 kHz as compared to when the stimulus was band-pass filtered between 4 and 6 kHz, low-pass filtered at 2 kHz or presented in the full bandwidth condition. Last, neural phase-locking to f0 was reduced or absent in envelope FFRs measured to filtered stimuli that lacked spectral energy above 0.125 kHz or below 6 kHz. Similarly, little to no energy was seen at F1 in spectral FFRs obtained to low-, high-, or band-pass filtered stimuli that did not contain energy in the F1 region. For stimulus conditions that contained energy at F1, the strength of the peak at F1 in the spectral FFR varied little with low-, high-, or band-pass filtering. CONCLUSIONS Energy at f0 in envelope FFRs may arise due to neural phase-locking to low-, mid-, or high-frequency stimulus components, provided the stimulus envelope is modulated by at least two interacting harmonics. Stronger neural responses at f0 are measured when filtering results in stimulus bandwidths that preserve stimulus energy at F1 and F2. In addition, results suggest that unresolved harmonics may favorably influence f0 strength in the neural response. Lastly, brainstem neural representation of the F1-related harmonic measured in spectral FFRs obtained to filtered stimuli is related to the presence or absence of stimulus energy at F1. These findings add to the existing literature exploring the viability of the FFR as an objective technique to evaluate hearing aid fitting where stimulus bandwidth is altered by design due to frequency-specific gain applied by amplification algorithms.
Collapse
|
8
|
Brainstem correlates of cochlear nonlinearity measured via the scalp-recorded frequency-following response. Neuroreport 2020; 31:702-707. [PMID: 32453027 DOI: 10.1097/wnr.0000000000001452] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
The frequency-following response (FFR) is an EEG-based potential used to characterize the brainstem encoding of complex sounds. Adopting techniques from auditory signal processing, we assessed the degree to which FFRs encode important properties of cochlear processing (e.g. nonlinearities) and their relation to speech-in-noise (SIN) listening skills. Based on the premise that normal cochlear transduction is characterized by rectification and compression, we reasoned these nonlinearities would create measurable harmonic distortion in FFRs in response to even pure tone input. We recorded FFRs to nonspeech (pure- and amplitude-modulated-tones) stimuli in normal-hearing individuals. We then compared conventional indices of cochlear nonlinearity, via distortion product otoacoustic emission (DPOAE) I/O functions, to total harmonic distortion measured from neural FFRs (FFRTHD). Analysis of DPOAE growth and the FFRTHD revealed listeners with higher cochlear compression thresholds had lower neural FFRTHD distortion (i.e. more linear FFRs), thus linking cochlear and brainstem correlates of auditory nonlinearity. Importantly, FFRTHD was also negatively correlated with SIN perception whereby listeners with higher FFRTHD (i.e. more nonlinear responses) showed better performance on the QuickSIN. We infer individual differences in SIN perception and FFR nonlinearity even in normal-hearing individuals may reflect subtle differences in auditory health and suprathreshold hearing skills not captured by normal audiometric evaluation. Future studies in hearing-impaired individuals and animal models are necessary to confirm the diagnostic utility of FFRTHD and its relation to cochlear hearing loss or peripheral neurodegeneration in humans.
Collapse
|
9
|
Effects of Directional Microphone and Noise Reduction on Subcortical and Cortical Auditory-Evoked Potentials in Older Listeners With Hearing Loss. Ear Hear 2020; 41:1282-1293. [PMID: 32058351 DOI: 10.1097/aud.0000000000000847] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Understanding how signal processing influences neural activity in the brain with hearing loss is relevant to the design and evaluation of features intended to alleviate speech-in-noise deficits faced by many hearing aid wearers. Here, we examine whether hearing aid processing schemes that are designed to improve speech-in-noise intelligibility (i.e., directional microphone and noise reduction) also improve electrophysiological indices of speech processing in older listeners with hearing loss. DESIGN The study followed a double-blind within-subjects design. A sample of 19 older adults (8 females; mean age = 73.6 years, range = 56-86 years; 17 experienced hearing aid users) with a moderate to severe sensorineural hearing impairment participated in the experiment. Auditory-evoked potentials associated with processing in cortex (P1-N1-P2) and subcortex (frequency-following response) were measured over the course of two 2-hour visits. Listeners were presented with sequences of the consonant-vowel syllable /da/ in continuous speech-shaped noise at signal to noise ratios (SNRs) of 0, +5, and +10 dB. Speech and noise stimuli were pre-recorded using a Knowles Electronics Manikin for Acoustic Research (KEMAR) head and torso simulator outfitted with hearing aids programmed for each listener's loss. The study aid programs were set according to 4 conditions: (1) omnidirectional microphone, (2) omnidirectional microphone with noise reduction, (3) directional microphone, and (4) directional microphone with noise reduction. For each hearing aid condition, speech was presented from a loudspeaker located at 1 m directly in front of KEMAR (i.e., 0° in the azimuth) at 75 dB SPL and noise was presented from a matching loudspeaker located at 1 m directly behind KEMAR (i.e., 180° in the azimuth). Recorded stimulus sequences were normalized for speech level across conditions and presented to listeners over electromagnetically shielded ER-2 ear-insert transducers. Presentation levels were calibrated to match the output of listeners' study aids. RESULTS Cortical components from listeners with hearing loss were enhanced with improving SNR and with use of a directional microphone and noise reduction. On the other hand, subcortical components did not show sensitivity to SNR or microphone mode but did show enhanced encoding of temporal fine structure of speech for conditions where noise reduction was enabled. CONCLUSIONS These results suggest that auditory-evoked potentials may be useful in evaluating the benefit of different noise-mitigating hearing aid features.
Collapse
|
10
|
Speech Auditory Brainstem Responses: Effects of Background, Stimulus Duration, Consonant-Vowel, and Number of Epochs. Ear Hear 2019; 40:659-670. [PMID: 30124503 PMCID: PMC6493675 DOI: 10.1097/aud.0000000000000648] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/03/2022]
Abstract
Supplemental Digital Content is available in the text. Objectives: The aims of this study were to systematically explore the effects of stimulus duration, background (quiet versus noise), and three consonant–vowels on speech-auditory brainstem responses (ABRs). Additionally, the minimum number of epochs required to record speech-ABRs with clearly identifiable waveform components was assessed. The purpose was to evaluate whether shorter duration stimuli could be reliably used to record speech-ABRs both in quiet and in background noise to the three consonant–vowels, as opposed to longer duration stimuli that are commonly used in the literature. Shorter duration stimuli and a smaller number of epochs would require shorter test sessions and thus encourage the transition of the speech-ABR from research to clinical practice. Design: Speech-ABRs in response to 40 msec [da], 50 msec [ba] [da] [ga], and 170 msec [ba] [da] [ga] stimuli were collected from 12 normal-hearing adults with confirmed normal click-ABRs. Monaural (right-ear) speech-ABRs were recorded to all stimuli in quiet and to 40 msec [da], 50 msec [ba] [da] [ga], and 170 msec [da] in a background of two-talker babble at +10 dB signal to noise ratio using a 2-channel electrode montage (Cz-Active, A1 and A2-reference, Fz-ground). Twelve thousand epochs (6000 per polarity) were collected for each stimulus and background from all participants. Latencies and amplitudes of speech-ABR peaks (V, A, D, E, F, O) were compared across backgrounds (quiet and noise) for all stimulus durations, across stimulus durations (50 and 170 msec) and across consonant–vowels ([ba], [da], and [ga]). Additionally, degree of phase locking to the stimulus fundamental frequency (in quiet versus noise) was evaluated for the frequency following response in speech-ABRs to the 170 msec [da]. Finally, the number of epochs required for a robust response was evaluated using Fsp statistic and bootstrap analysis at different epoch iterations. Results: Background effect: the addition of background noise resulted in speech-ABRs with longer peak latencies and smaller peak amplitudes compared with speech-ABRs in quiet, irrespective of stimulus duration. However, there was no effect of background noise on the degree of phase locking of the frequency following response to the stimulus fundamental frequency in speech-ABRs to the 170 msec [da]. Duration effect: speech-ABR peak latencies and amplitudes did not differ in response to the 50 and 170 msec stimuli. Consonant–vowel effect: different consonant–vowels did not have an effect on speech-ABR peak latencies regardless of stimulus duration. Number of epochs: a larger number of epochs was required to record speech-ABRs in noise compared with in quiet, and a smaller number of epochs was required to record speech-ABRs to the 40 msec [da] compared with the 170 msec [da]. Conclusions: This is the first study that systematically investigated the clinical feasibility of speech-ABRs in terms of stimulus duration, background noise, and number of epochs. Speech-ABRs can be reliably recorded to the 40 msec [da] without compromising response quality even when presented in background noise. Because fewer epochs were needed for the 40 msec [da], this would be the optimal stimulus for clinical use. Finally, given that there was no effect of consonant–vowel on speech-ABR peak latencies, there is no evidence that speech-ABRs are suitable for assessing auditory discrimination of the stimuli used.
Collapse
|
11
|
Gnanateja GN, Maruthy S. Dichotic phase effects on frequency following responses reveal phase variant and invariant harmonic distortion products. Hear Res 2019; 380:84-99. [PMID: 31212114 DOI: 10.1016/j.heares.2019.04.008] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/07/2018] [Revised: 08/28/2018] [Accepted: 04/15/2019] [Indexed: 01/24/2023]
Abstract
The dichotic frequency following responses (FFR) have been used in studies to infer about dichotic auditory processing. In the present study, we hypothesize that the proximity of the binaural neural generators of the FFR would result in interference of the volume-conducted electrical fields. This might lead to contamination of the scalp-recorded dichotic FFRs due to which it might be difficult to infer about true dichotic processing in the putative neural generators. We investigated this by recording FFRs to binaurally presented 200 Hz pure tone with graded dichotic phase offsets (0°, 90°, 180° and 270°) in normal hearing young adults. Spectral analysis of the FFRs was performed for the estimation of the magnitude and phase at the component frequencies. FFR spectra were compared using non-parametric paired randomizations within the subjects. We found that the brainstem responses to a 200 Hz pure tone consisted of prominent peaks at 200 Hz, and at frequencies corresponding to the harmonics of 200 Hz. The FFR spectral magnitude at 200 Hz diminished with a phase offset of 180°. Phase offsets of 90° and 270° showed reduced spectral magnitudes at 200 Hz than those in the 0° condition. Our findings, in line with the hypothesis, show that the dichotic FFRs do not reflect true dichotic processing and that they are contaminated during volume conduction. Additionally, we found harmonic distortion products (HDP) in the FFRs. We found that the response at 200 Hz and the 3rd HDP systematically varied with a change in phase of the stimulus, while the even HDPs (2nd and 4th) were phase-invariant. Based on our findings, and modeling FFRs using auditory models, we propose a rectification process as the contributors for the generation of HDPs. We also discuss the implications of this HDP generating mechanism in understanding the pitch represented in FFRs.
Collapse
Affiliation(s)
- G Nike Gnanateja
- Department of Communication Sciences and Disorders, School of Health and Rehabilitation Sciences, University of Pittsburgh, Forbes Tower, Pittsburgh, PA, 15260, USA.
| | - Sandeep Maruthy
- Department of Audiology, All India Institute of Speech and Hearing, Mysuru, Karnataka, 570006, India.
| |
Collapse
|
12
|
Bidelman GM, Price CN, Shen D, Arnott SR, Alain C. Afferent-efferent connectivity between auditory brainstem and cortex accounts for poorer speech-in-noise comprehension in older adults. Hear Res 2019; 382:107795. [PMID: 31479953 DOI: 10.1016/j.heares.2019.107795] [Citation(s) in RCA: 30] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/08/2019] [Revised: 08/14/2019] [Accepted: 08/22/2019] [Indexed: 12/19/2022]
Abstract
Speech-in-noise (SIN) comprehension deficits in older adults have been linked to changes in both subcortical and cortical auditory evoked responses. However, older adults' difficulty understanding SIN may also be related to an imbalance in signal transmission (i.e., functional connectivity) between brainstem and auditory cortices. By modeling high-density scalp recordings of speech-evoked responses with sources in brainstem (BS) and bilateral primary auditory cortices (PAC), we show that beyond attenuating neural activity, hearing loss in older adults compromises the transmission of speech information between subcortical and early cortical hubs of the speech network. We found that the strength of afferent BS→PAC neural signaling (but not the reverse efferent flow; PAC→BS) varied with mild declines in hearing acuity and this "bottom-up" functional connectivity robustly predicted older adults' performance in a SIN identification task. Connectivity was also a better predictor of SIN processing than unitary subcortical or cortical responses alone. Our neuroimaging findings suggest that in older adults (i) mild hearing loss differentially reduces neural output at several stages of auditory processing (PAC > BS), (ii) subcortical-cortical connectivity is more sensitive to peripheral hearing loss than top-down (cortical-subcortical) control, and (iii) reduced functional connectivity in afferent auditory pathways plays a significant role in SIN comprehension problems.
Collapse
Affiliation(s)
- Gavin M Bidelman
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA; University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA.
| | - Caitlin N Price
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA
| | - Dawei Shen
- Rotman Research Institute-Baycrest Centre for Geriatric Care, Toronto, Ontario, Canada
| | - Stephen R Arnott
- Rotman Research Institute-Baycrest Centre for Geriatric Care, Toronto, Ontario, Canada
| | - Claude Alain
- Rotman Research Institute-Baycrest Centre for Geriatric Care, Toronto, Ontario, Canada; University of Toronto, Department of Psychology, Toronto, Ontario, Canada; University of Toronto, Institute of Medical Sciences, Toronto, Ontario, Canada
| |
Collapse
|
13
|
Yellamsetty A, Bidelman GM. Brainstem correlates of concurrent speech identification in adverse listening conditions. Brain Res 2019; 1714:182-192. [PMID: 30796895 DOI: 10.1016/j.brainres.2019.02.025] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2018] [Revised: 01/07/2019] [Accepted: 02/19/2019] [Indexed: 01/20/2023]
Abstract
When two voices compete, listeners can segregate and identify concurrent speech sounds using pitch (fundamental frequency, F0) and timbre (harmonic) cues. Speech perception is also hindered by the signal-to-noise ratio (SNR). How clear and degraded concurrent speech sounds are represented at early, pre-attentive stages of the auditory system is not well understood. To this end, we measured scalp-recorded frequency-following responses (FFR) from the EEG while human listeners heard two concurrently presented, steady-state (time-invariant) vowels whose F0 differed by zero or four semitones (ST) presented diotically in either clean (no noise) or noise-degraded (+5dB SNR) conditions. Listeners also performed a speeded double vowel identification task in which they were required to identify both vowels correctly. Behavioral results showed that speech identification accuracy increased with F0 differences between vowels, and this perceptual F0 benefit was larger for clean compared to noise degraded (+5dB SNR) stimuli. Neurophysiological data demonstrated more robust FFR F0 amplitudes for single compared to double vowels and considerably weaker responses in noise. F0 amplitudes showed speech-on-speech masking effects, along with a non-linear constructive interference at 0ST, and suppression effects at 4ST. Correlations showed that FFR F0 amplitudes failed to predict listeners' identification accuracy. In contrast, FFR F1 amplitudes were associated with faster reaction times, although this correlation was limited to noise conditions. The limited number of brain-behavior associations suggests subcortical activity mainly reflects exogenous processing rather than perceptual correlates of concurrent speech perception. Collectively, our results demonstrate that FFRs reflect pre-attentive coding of concurrent auditory stimuli that only weakly predict the success of identifying concurrent speech.
Collapse
Affiliation(s)
- Anusha Yellamsetty
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Department of Communication Sciences & Disorders, University of South Florida, USA.
| | - Gavin M Bidelman
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA; University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA.
| |
Collapse
|
14
|
Peng F, McKay CM, Mao D, Hou W, Innes-Brown H. Auditory Brainstem Representation of the Voice Pitch Contours in the Resolved and Unresolved Components of Mandarin Tones. Front Neurosci 2018; 12:820. [PMID: 30505262 PMCID: PMC6250765 DOI: 10.3389/fnins.2018.00820] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2018] [Accepted: 10/22/2018] [Indexed: 11/24/2022] Open
Abstract
Accurate perception of voice pitch plays a vital role in speech understanding, especially for tonal languages such as Mandarin. Lexical tones are primarily distinguished by the fundamental frequency (F0) contour of the acoustic waveform. It has been shown that the auditory system could extract the F0 from the resolved and unresolved harmonics, and the tone identification performance of resolved harmonics was better than unresolved harmonics. To evaluate the neural response to the resolved and unresolved components of Mandarin tones in quiet and in speech-shaped noise, we recorded the frequency-following response. In this study, four types of stimuli were used: speech with either only-resolved harmonics or only-unresolved harmonics, both in quiet and in speech-shaped noise. Frequency-following responses (FFRs) were recorded to alternating-polarity stimuli and were added or subtracted to enhance the neural response to the envelope (FFRENV) or fine structure (FFRTFS), respectively. The neural representation of the F0 strength reflected by the FFRENV was evaluated by the peak autocorrelation value in the temporal domain and the peak phase-locking value (PLV) at F0 in the spectral domain. Both evaluation methods showed that the FFRENV F0 strength in quiet was significantly stronger than in noise for speech including unresolved harmonics, but not for speech including resolved harmonics. The neural representation of the temporal fine structure reflected by the FFRTFS was assessed by the PLV at the harmonic near to F1 (4th of F0). The PLV at harmonic near to F1 (4th of F0) of FFRTFS to resolved harmonics was significantly larger than to unresolved harmonics. Spearman's correlation showed that the FFRENV F0 strength to unresolved harmonics was correlated with tone identification performance in noise (0 dB SNR). These results showed that the FFRENV F0 strength to speech sounds with resolved harmonics was not affected by noise. In contrast, the response to speech sounds with unresolved harmonics, which were significantly smaller in noise compared to quiet. Our results suggest that coding resolved harmonics was more important than coding envelope for tone identification performance in noise.
Collapse
Affiliation(s)
- Fei Peng
- Key Laboratory of Biorheological Science and Technology, Chongqing University, Ministry of Education, Chongqing, China.,The Bionics Institute of Australia, East Melbourne, VIC, Australia.,Medical Bionics Department, University of Melbourne, Melbourne, VIC, Australia.,Collaborative Innovation Center for Brain Science, Chongqing University, Chongqing, China
| | - Colette M McKay
- The Bionics Institute of Australia, East Melbourne, VIC, Australia.,Medical Bionics Department, University of Melbourne, Melbourne, VIC, Australia
| | - Darren Mao
- The Bionics Institute of Australia, East Melbourne, VIC, Australia.,Department of Biomedical Engineering, University of Melbourne, Melbourne, VIC, Australia
| | - Wensheng Hou
- Key Laboratory of Biorheological Science and Technology, Chongqing University, Ministry of Education, Chongqing, China.,Collaborative Innovation Center for Brain Science, Chongqing University, Chongqing, China.,Chongqing Engineering Research Center of Medical Electronics Technology, Chongqing University, Chongqing, China
| | - Hamish Innes-Brown
- The Bionics Institute of Australia, East Melbourne, VIC, Australia.,Medical Bionics Department, University of Melbourne, Melbourne, VIC, Australia
| |
Collapse
|
15
|
Bidelman G, Powers L. Response properties of the human frequency-following response (FFR) to speech and non-speech sounds: level dependence, adaptation and phase-locking limits. Int J Audiol 2018; 57:665-672. [DOI: 10.1080/14992027.2018.1470338] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/16/2022]
Affiliation(s)
- Gavin Bidelman
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA
- Department of Anatomy and Neurobiology, University of Tennessee Health Sciences Center, Memphis, TN, USA
| | - Louise Powers
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA
| |
Collapse
|
16
|
Abstract
OBJECTIVES Vocoders offer an effective platform to simulate the effects of cochlear implant speech processing strategies in normal-hearing listeners. Several behavioral studies have examined the effects of varying spectral and temporal cues on vocoded speech perception; however, little is known about the neural indices of vocoded speech perception. Here, the scalp-recorded frequency following response (FFR) was used to study the effects of varying spectral and temporal cues on brainstem neural representation of specific acoustic cues, the temporal envelope periodicity related to fundamental frequency (F0) and temporal fine structure (TFS) related to formant and formant-related frequencies, as reflected in the phase-locked neural activity in response to vocoded speech. DESIGN In experiment 1, FFRs were measured in 12 normal-hearing, adult listeners in response to a steady state English back vowel /u/ presented in an unaltered, unprocessed condition and six sine-vocoder conditions with varying numbers of channels (1, 2, 4, 8, 16, and 32), while the temporal envelope cutoff frequency was fixed at 500 Hz. In experiment 2, FFRs were obtained from 14 normal-hearing, adult listeners in response to the same English vowel /u/, presented in an unprocessed condition and four vocoded conditions where both the temporal envelope cutoff frequency (50 versus 500 Hz) and carrier type (sine wave versus noise band) were varied separately with the number of channels fixed at 8. Fast Fourier Transform was applied to the time waveforms of FFR to analyze the strength of brainstem neural representation of temporal envelope periodicity (F0) and TFS-related peaks (formant structure). RESULTS Brainstem neural representation of both temporal envelope and TFS cues improved when the number of channels increased from 1 to 4, followed by a plateau with 8 and 16 channels, and a reduction in phase-locking strength with 32 channels. For the sine vocoders, peaks in the FFRTFS spectra corresponded with the low-frequency sine-wave carriers and side band frequencies in the stimulus spectra. When the temporal envelope cutoff frequency increased from 50 to 500 Hz, an improvement was observed in brainstem F0 representation with no change in brainstem representation of spectral peaks proximal to the first formant frequency (F1). There was no significant effect of carrier type (sine- versus noise-vocoder) on brainstem neural representation of F0 cues when the temporal envelope cutoff frequency was 500 Hz. CONCLUSIONS While the improvement in neural representation of temporal envelope and TFS cues with up to 4 vocoder channels is consistent with the behavioral literature, the reduced neural phase-locking strength noted with even more channels may be because of the narrow bandwidth of each channel as the number of channels increases. Stronger neural representation of temporal envelope cues with higher temporal envelope cutoff frequencies is likely a reflection of brainstem neural phase-locking to F0-related periodicity fluctuations preserved in the 500-Hz temporal envelopes, which are unavailable in the 50-Hz temporal envelopes. No effect of temporal envelope cutoff frequency was seen for neural representation of TFS cues, suggesting that spectral side band frequencies created by the 500-Hz temporal envelopes did not improve neural representation of F1 cues over the 50-Hz temporal envelopes. Finally, brainstem F0 representation was not significantly affected by carrier type with a temporal envelope cutoff frequency of 500 Hz, which is inconsistent with previous results of behavioral studies examining pitch perception of vocoded stimuli.
Collapse
|
17
|
Communicating in Challenging Environments: Noise and Reverberation. THE FREQUENCY-FOLLOWING RESPONSE 2017. [DOI: 10.1007/978-3-319-47944-6_8] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
|
18
|
Ananthakrishnan S, Krishnan A, Bartlett E. Human Frequency Following Response: Neural Representation of Envelope and Temporal Fine Structure in Listeners with Normal Hearing and Sensorineural Hearing Loss. Ear Hear 2016; 37:e91-e103. [PMID: 26583482 DOI: 10.1097/aud.0000000000000247] [Citation(s) in RCA: 42] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVE Listeners with sensorineural hearing loss (SNHL) typically experience reduced speech perception, which is not completely restored with amplification. This likely occurs because cochlear damage, in addition to elevating audiometric thresholds, alters the neural representation of speech transmitted to higher centers along the auditory neuroaxis. While the deleterious effects of SNHL on speech perception in humans have been well-documented using behavioral paradigms, our understanding of the neural correlates underlying these perceptual deficits remains limited. Using the scalp-recorded frequency following response (FFR), the authors examine the effects of SNHL and aging on subcortical neural representation of acoustic features important for pitch and speech perception, namely the periodicity envelope (F0) and temporal fine structure (TFS; formant structure), as reflected in the phase-locked neural activity generating the FFR. DESIGN FFRs were obtained from 10 listeners with normal hearing (NH) and 9 listeners with mild-moderate SNHL in response to a steady-state English back vowel /u/ presented at multiple intensity levels. Use of multiple presentation levels facilitated comparisons at equal sound pressure level (SPL) and equal sensation level. In a second follow-up experiment to address the effect of age on envelope and TFS representation, FFRs were obtained from 25 NH and 19 listeners with mild to moderately severe SNHL to the same vowel stimulus presented at 80 dB SPL. Temporal waveforms, Fast Fourier Transform and spectrograms were used to evaluate the magnitude of the phase-locked activity at F0 (periodicity envelope) and F1 (TFS). RESULTS Neural representation of both envelope (F0) and TFS (F1) at equal SPLs was stronger in NH listeners compared with listeners with SNHL. Also, comparison of neural representation of F0 and F1 across stimulus levels expressed in SPL and sensation level (accounting for audibility) revealed that level-related changes in F0 and F1 magnitude were different for listeners with SNHL compared with listeners with NH. Furthermore, the degradation in subcortical neural representation was observed to persist in listeners with SNHL even when the effects of age were controlled for. CONCLUSIONS Overall, our results suggest a relatively greater degradation in the neural representation of TFS compared with periodicity envelope in individuals with SNHL. This degraded neural representation of TFS in SNHL, as reflected in the brainstem FFR, may reflect a disruption in the temporal pattern of phase-locked neural activity arising from altered tonotopic maps and/or wider filters causing poor frequency selectivity in these listeners. Finally, while preliminary results indicate that the deleterious effects of SNHL may be greater than age-related degradation in subcortical neural representation, the lack of a balanced age-matched control group in this study does not permit us to completely rule out the effects of age on subcortical neural representation.
Collapse
Affiliation(s)
- Saradha Ananthakrishnan
- 1Department of Speech Language Hearing Sciences, Purdue University, West Lafayette, Indiana, USA; 2Department of Audiology, Speech-Language Pathology and Deaf studies, Towson University, Towson, Maryland, USA; 3Department of Biomedical Engineering, Purdue University, West Lafayette, Indiana, USA; and 4Department of Biological Sciences, Purdue University, West Lafayette, Indiana, USA
| | | | | |
Collapse
|
19
|
Schvartz-Leyzac KC, Chatterjee M. Fundamental-frequency discrimination using noise-band-vocoded harmonic complexes in older listeners with normal hearing. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 138:1687-1695. [PMID: 26428806 PMCID: PMC4592424 DOI: 10.1121/1.4929938] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/17/2014] [Revised: 08/06/2015] [Accepted: 08/20/2015] [Indexed: 06/04/2023]
Abstract
Voice-pitch cues provide detailed information about a talker that help a listener to understand speech in complex environments. Temporal-envelope based voice-pitch coding is important for listeners with hearing impairment, especially listeners with cochlear implants, as spectral resolution is not sufficient to provide a spectrally based voice-pitch cue. The effect of aging on the ability to glean voice-pitch information using temporal envelope cues is not completely understood. The current study measured fundamental frequency (f0) discrimination limens in normal-hearing younger and older adults while listening to noise-band vocoded harmonic complexes with varying numbers of spectral channels. Age-related disparities in performance were apparent across all conditions, independent of spectral degradation and/or fundamental frequency. The findings have important implications for older listeners with normal hearing and hearing loss, who may be inherently limited in their ability to perceive f0 cues due to senescent decline in auditory function.
Collapse
Affiliation(s)
- Kara C Schvartz-Leyzac
- Department of Hearing and Speech Sciences, University of Maryland, 0100 LeFrak Hall, College Park, Maryland 20742, USA
| | - Monita Chatterjee
- Department of Hearing and Speech Sciences, University of Maryland, 0100 LeFrak Hall, College Park, Maryland 20742, USA
| |
Collapse
|
20
|
Losing the music: aging affects the perception and subcortical neural representation of musical harmony. J Neurosci 2015; 35:4071-80. [PMID: 25740534 DOI: 10.1523/jneurosci.3214-14.2015] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
When two musical notes with simple frequency ratios are played simultaneously, the resulting musical chord is pleasing and evokes a sense of resolution or "consonance". Complex frequency ratios, on the other hand, evoke feelings of tension or "dissonance". Consonance and dissonance form the basis of harmony, a central component of Western music. In earlier work, we provided evidence that consonance perception is based on neural temporal coding in the brainstem (Bones et al., 2014). Here, we show that for listeners with clinically normal hearing, aging is associated with a decline in both the perceptual distinction and the distinctiveness of the neural representations of different categories of two-note chords. Compared with younger listeners, older listeners rated consonant chords as less pleasant and dissonant chords as more pleasant. Older listeners also had less distinct neural representations of consonant and dissonant chords as measured using a Neural Consonance Index derived from the electrophysiological "frequency-following response." The results withstood a control for the effect of age on general affect, suggesting that different mechanisms are responsible for the perceived pleasantness of musical chords and affective voices and that, for listeners with clinically normal hearing, age-related differences in consonance perception are likely to be related to differences in neural temporal coding.
Collapse
|
21
|
Xu Q, Ye D. Evaluation of a posteriori Wiener filtering applied to frequency-following response extraction in the auditory brainstem. Biomed Signal Process Control 2014. [DOI: 10.1016/j.bspc.2014.08.003] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
22
|
Xu Q, Gong Q. Frequency difference beyond behavioral limen reflected by frequency following response of human auditory Brainstem. Biomed Eng Online 2014; 13:114. [PMID: 25108552 PMCID: PMC4132204 DOI: 10.1186/1475-925x-13-114] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2014] [Accepted: 08/04/2014] [Indexed: 11/10/2022] Open
Abstract
Background The present study investigated whether the frequency-following response (FFR) of the auditory brainstem can represent individual frequency-discrimination ability. Method We measured behavioral frequency-difference limens (FDLs) in normal hearing young adults. Then FFRs were evoked by two pure tones, whose frequency difference was no larger than behavioral FDL. Discrimination of FFRs to individual frequencies was conducted as the neural representation of stimulus frequency difference. Participants were 15 Chinese college students (ages 19–25; 3 males, 12 females) with normal hearing characteristics. Results According to discriminative neural representations of individual frequencies, FFRs accurately reflected individual FDLs and detected stimulus-frequency differences smaller than behavioral threshold (e.g., 75% of FDL). Conclusions These results suggest that when a frequency difference cannot be behaviorally distinguished, there is still a possibility of it being detected physiologically.
Collapse
Affiliation(s)
| | - Qin Gong
- Postal address: Department of Biomedical Engineering, Medical School, Tsinghua University, Beijing 100084, China.
| |
Collapse
|
23
|
Abstract
To enhance weak sounds while compressing the dynamic intensity range, auditory sensory cells amplify sound-induced vibrations in a nonlinear, intensity-dependent manner. In the course of this process, instantaneous waveform distortion is produced, with two conspicuous kinds of interwoven consequences, the introduction of new sound frequencies absent from the original stimuli, which are audible and detectable in the ear canal as otoacoustic emissions, and the possibility for an interfering sound to suppress the response to a probe tone, thereby enhancing contrast among frequency components. We review how the diverse manifestations of auditory nonlinearity originate in the gating principle of their mechanoelectrical transduction channels; how they depend on the coordinated opening of these ion channels ensured by connecting elements; and their links to the dynamic behavior of auditory sensory cells. This paper also reviews how the complex properties of waves traveling through the cochlea shape the manifestations of auditory nonlinearity. Examination methods based on the detection of distortions open noninvasive windows on the modes of activity of mechanosensitive structures in auditory sensory cells and on the distribution of sites of nonlinearity along the cochlear tonotopic axis, helpful for deciphering cochlear molecular physiology in hearing-impaired animal models. Otoacoustic emissions enable fast tests of peripheral sound processing in patients. The study of auditory distortions also contributes to the understanding of the perception of complex sounds.
Collapse
Affiliation(s)
- Paul Avan
- Laboratory of Neurosensory Biophysics, University of Auvergne, School of Medicine, Clermont-Ferrand, France; Institut National de la Santé et de la Recherche Médicale (INSERM), UMR 1107, Clermont-Ferrand, France; Centre Jean Perrin, Clermont-Ferrand, France; Department of Otolaryngology, County Hospital, Krems an der Donau, Austria; Laboratory of Genetics and Physiology of Hearing, Department of Neuroscience, Institut Pasteur, Paris, France; Collège de France, Genetics and Cell Physiology, Paris, France
| | - Béla Büki
- Laboratory of Neurosensory Biophysics, University of Auvergne, School of Medicine, Clermont-Ferrand, France; Institut National de la Santé et de la Recherche Médicale (INSERM), UMR 1107, Clermont-Ferrand, France; Centre Jean Perrin, Clermont-Ferrand, France; Department of Otolaryngology, County Hospital, Krems an der Donau, Austria; Laboratory of Genetics and Physiology of Hearing, Department of Neuroscience, Institut Pasteur, Paris, France; Collège de France, Genetics and Cell Physiology, Paris, France
| | - Christine Petit
- Laboratory of Neurosensory Biophysics, University of Auvergne, School of Medicine, Clermont-Ferrand, France; Institut National de la Santé et de la Recherche Médicale (INSERM), UMR 1107, Clermont-Ferrand, France; Centre Jean Perrin, Clermont-Ferrand, France; Department of Otolaryngology, County Hospital, Krems an der Donau, Austria; Laboratory of Genetics and Physiology of Hearing, Department of Neuroscience, Institut Pasteur, Paris, France; Collège de France, Genetics and Cell Physiology, Paris, France
| |
Collapse
|
24
|
Marmel F, Linley D, Carlyon RP, Gockel HE, Hopkins K, Plack CJ. Subcortical neural synchrony and absolute thresholds predict frequency discrimination independently. J Assoc Res Otolaryngol 2013; 14:757-66. [PMID: 23760984 PMCID: PMC3767871 DOI: 10.1007/s10162-013-0402-3] [Citation(s) in RCA: 54] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/03/2012] [Accepted: 05/20/2013] [Indexed: 11/25/2022] Open
Abstract
The neural mechanisms of pitch coding have been debated for more than a century. The two main mechanisms are coding based on the profiles of neural firing rates across auditory nerve fibers with different characteristic frequencies (place-rate coding), and coding based on the phase-locked temporal pattern of neural firing (temporal coding). Phase locking precision can be partly assessed by recording the frequency-following response (FFR), a scalp-recorded electrophysiological response that reflects synchronous activity in subcortical neurons. Although features of the FFR have been widely used as indices of pitch coding acuity, only a handful of studies have directly investigated the relation between the FFR and behavioral pitch judgments. Furthermore, the contribution of degraded neural synchrony (as indexed by the FFR) to the pitch perception impairments of older listeners and those with hearing loss is not well known. Here, the relation between the FFR and pure-tone frequency discrimination was investigated in listeners with a wide range of ages and absolute thresholds, to assess the respective contributions of subcortical neural synchrony and other age-related and hearing loss-related mechanisms to frequency discrimination performance. FFR measures of neural synchrony and absolute thresholds independently contributed to frequency discrimination performance. Age alone, i.e., once the effect of subcortical neural synchrony measures or absolute thresholds had been partialed out, did not contribute to frequency discrimination. Overall, the results suggest that frequency discrimination of pure tones may depend both on phase locking precision and on separate mechanisms affected in hearing loss.
Collapse
Affiliation(s)
- F. Marmel
- />School of Psychological Sciences, The University of Manchester, Manchester, M13 9PL UK
| | - D. Linley
- />School of Psychological Sciences, The University of Manchester, Manchester, M13 9PL UK
| | - R. P. Carlyon
- />MRC Cognition and Brain Sciences Unit, Cambridge, CB2 7EF UK
| | - H. E. Gockel
- />MRC Cognition and Brain Sciences Unit, Cambridge, CB2 7EF UK
| | - K. Hopkins
- />School of Psychological Sciences, The University of Manchester, Manchester, M13 9PL UK
| | - C. J. Plack
- />School of Psychological Sciences, The University of Manchester, Manchester, M13 9PL UK
| |
Collapse
|