1
|
Joris PX, Verschooten E, Mc Laughlin M, Versteegh C, van der Heijden M. Frequency selectivity in monkey auditory nerve studied with suprathreshold multicomponent stimuli. Hear Res 2024; 443:108964. [PMID: 38277882 DOI: 10.1016/j.heares.2024.108964] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/17/2023] [Revised: 01/15/2024] [Accepted: 01/20/2024] [Indexed: 01/28/2024]
Abstract
Data from non-human primates can help extend observations from non-primate species to humans. Here we report measurements on the auditory nerve of macaque monkeys in the context of a controversial topic important to human hearing. A range of techniques have been used to examine the claim, which is not generally accepted, that human frequency tuning is sharper than traditionally thought, and sharper than in commonly used animal models. Data from single auditory-nerve fibers occupy a pivotal position to examine this claim, but are not available for humans. A previous study reported sharper tuning in auditory-nerve fibers of macaque relative to the cat. A limitation of these and other single-fiber data is that frequency selectivity was measured with tonal threshold-tuning curves, which do not directly assess spectral filtering and whose shape is sharpened by cochlear nonlinearity. Our aim was to measure spectral filtering with wideband suprathreshold stimuli in the macaque auditory nerve. We obtained responses of single nerve fibers of anesthetized macaque monkeys and cats to a suprathreshold, wideband, multicomponent stimulus designed to allow characterization of spectral filtering at any cochlear locus. Quantitatively the differences between the two species are smaller than in previous studies, but consistent with these studies the filters obtained show a trend of sharper tuning in macaque, relative to the cat, for fibers in the basal half of the cochlea. We also examined differences in group delay measured on the phase data near the characteristic frequency versus in the low-frequency tail. The phase data are consistent with the interpretation of sharper frequency tuning in monkey in the basal half of the cochlea. We conclude that use of suprathreshold, wide-band stimuli supports the interpretation of sharper frequency selectivity in macaque nerve fibers relative to the cat, although the difference is less marked than apparent from the assessment with tonal threshold-based data.
Collapse
Affiliation(s)
- P X Joris
- Lab of Auditory Neurophysiology, KU Leuven, O&N2 KU Leuven, Herestraat 49 bus 1021, Leuven B-3000, Belgium.
| | - E Verschooten
- Lab of Auditory Neurophysiology, KU Leuven, O&N2 KU Leuven, Herestraat 49 bus 1021, Leuven B-3000, Belgium
| | - M Mc Laughlin
- Lab of Auditory Neurophysiology, KU Leuven, O&N2 KU Leuven, Herestraat 49 bus 1021, Leuven B-3000, Belgium
| | - Cpc Versteegh
- Department of Neuroscience, Erasmus MC, Rotterdam, the Netherlands
| | | |
Collapse
|
2
|
Michalsky J. Questioning questions - the perception of f0 scaling in German questions between categorical function and continuous attitude. PHONETICA 2023; 80:357-392. [PMID: 37534609 DOI: 10.1515/phon-2023-2002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/04/2023]
Abstract
In previous studies comparing the intonation of questions and statements in German, greater f0 excursions of phrase-final rises have been associated with questions in both read speech and spontaneous speech. This holds for production studies as well as perception studies. However, a major question remains whether these differences are perceived categorically or continuously. Furthermore, we ask whether the differences in f0 scaling correspond to categorical linguistic functions or rather an attitudinal continuum. We conducted three different perception experiments: a classical categorical perception task, an imitation task, and a semantic evaluation task. The results suggest that f0 scaling in phrase-final rises is perceived as a phonetic continuum rather than in phonological categories. Furthermore, the gradual increase of the final rise is associated with a gradual increase in perceived questioning. Lastly, the phonetic cues to this degree of questioning are distinct from those to the other investigated meanings surprise and uncertainty. Accordingly, this study supports the assumption that questioning constitutes an attitudinal meaning in its own right.
Collapse
|
3
|
Nisha KV, Uppunda AK, Kumar RT. Spatial rehabilitation using virtual auditory space training paradigm in individuals with sensorineural hearing impairment. Front Neurosci 2023; 16:1080398. [PMID: 36733923 PMCID: PMC9887142 DOI: 10.3389/fnins.2022.1080398] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2022] [Accepted: 12/20/2022] [Indexed: 01/18/2023] Open
Abstract
Purpose The present study aimed to quantify the effects of spatial training using virtual sources on a battery of spatial acuity measures in listeners with sensorineural hearing impairment (SNHI). Methods An intervention-based time-series comparison design involving 82 participants divided into three groups was adopted. Group I (n = 27, SNHI-spatially trained) and group II (n = 25, SNHI-untrained) consisted of SNHI listeners, while group III (n = 30) had listeners with normal hearing (NH). The study was conducted in three phases. In the pre-training phase, all the participants underwent a comprehensive assessment of their spatial processing abilities using a battery of tests including spatial acuity in free-field and closed-field scenarios, tests for binaural processing abilities (interaural time threshold [ITD] and level difference threshold [ILD]), and subjective ratings. While spatial acuity in the free field was assessed using a loudspeaker-based localization test, the closed-field source identification test was performed using virtual stimuli delivered through headphones. The ITD and ILD thresholds were obtained using a MATLAB psychoacoustic toolbox, while the participant ratings on the spatial subsection of speech, spatial, and qualities questionnaire in Kannada were used for the subjective ratings. Group I listeners underwent virtual auditory spatial training (VAST), following pre-evaluation assessments. All tests were re-administered on the group I listeners halfway through training (mid-training evaluation phase) and after training completion (post-training evaluation phase), whereas group II underwent these tests without any training at the same time intervals. Results and discussion Statistical analysis showed the main effect of groups in all tests at the pre-training evaluation phase, with post hoc comparisons that revealed group equivalency in spatial performance of both SNHI groups (groups I and II). The effect of VAST in group I was evident on all the tests, with the localization test showing the highest predictive power for capturing VAST-related changes on Fischer discriminant analysis (FDA). In contrast, group II demonstrated no changes in spatial acuity across timelines of measurements. FDA revealed increased errors in the categorization of NH as SNHI-trained at post-training evaluation compared to pre-training evaluation, as the spatial performance of the latter improved with VAST in the post-training phase. Conclusion The study demonstrated positive outcomes of spatial training using VAST in listeners with SNHI. The utility of this training program can be extended to other clinical population with spatial auditory processing deficits such as auditory neuropathy spectrum disorder, cochlear implants, central auditory processing disorders etc.
Collapse
|
4
|
Oxenham AJ. Questions and controversies surrounding the perception and neural coding of pitch. Front Neurosci 2023; 16:1074752. [PMID: 36699531 PMCID: PMC9868815 DOI: 10.3389/fnins.2022.1074752] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2022] [Accepted: 12/16/2022] [Indexed: 01/12/2023] Open
Abstract
Pitch is a fundamental aspect of auditory perception that plays an important role in our ability to understand speech, appreciate music, and attend to one sound while ignoring others. The questions surrounding how pitch is represented in the auditory system, and how our percept relates to the underlying acoustic waveform, have been a topic of inquiry and debate for well over a century. New findings and technological innovations have led to challenges of some long-standing assumptions and have raised new questions. This article reviews some recent developments in the study of pitch coding and perception and focuses on the topic of how pitch information is extracted from peripheral representations based on frequency-to-place mapping (tonotopy), stimulus-driven auditory-nerve spike timing (phase locking), or a combination of both. Although a definitive resolution has proved elusive, the answers to these questions have potentially important implications for mitigating the effects of hearing loss via devices such as cochlear implants.
Collapse
Affiliation(s)
- Andrew J. Oxenham
- Center for Applied and Translational Sensory Science, University of Minnesota Twin Cities, Minneapolis, MN, United States
- Department of Psychology, University of Minnesota Twin Cities, Minneapolis, MN, United States
| |
Collapse
|
5
|
Wang H, Chen R, Yan Y, McGettigan C, Rosen S, Adank P. Perceptual Learning of Noise-Vocoded Speech Under Divided Attention. Trends Hear 2023; 27:23312165231192297. [PMID: 37547940 PMCID: PMC10408355 DOI: 10.1177/23312165231192297] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2023] [Revised: 07/13/2023] [Accepted: 07/19/2023] [Indexed: 08/08/2023] Open
Abstract
Speech perception performance for degraded speech can improve with practice or exposure. Such perceptual learning is thought to be reliant on attention and theoretical accounts like the predictive coding framework suggest a key role for attention in supporting learning. However, it is unclear whether speech perceptual learning requires undivided attention. We evaluated the role of divided attention in speech perceptual learning in two online experiments (N = 336). Experiment 1 tested the reliance of perceptual learning on undivided attention. Participants completed a speech recognition task where they repeated forty noise-vocoded sentences in a between-group design. Participants performed the speech task alone or concurrently with a domain-general visual task (dual task) at one of three difficulty levels. We observed perceptual learning under divided attention for all four groups, moderated by dual-task difficulty. Listeners in easy and intermediate visual conditions improved as much as the single-task group. Those who completed the most challenging visual task showed faster learning and achieved similar ending performance compared to the single-task group. Experiment 2 tested whether learning relies on domain-specific or domain-general processes. Participants completed a single speech task or performed this task together with a dual task aiming to recruit domain-specific (lexical or phonological), or domain-general (visual) processes. All secondary task conditions produced patterns and amount of learning comparable to the single speech task. Our results demonstrate that the impact of divided attention on perceptual learning is not strictly dependent on domain-general or domain-specific processes and speech perceptual learning persists under divided attention.
Collapse
Affiliation(s)
- Han Wang
- Department of Speech, Hearing and Phonetic Sciences, University College London, London, UK
| | - Rongru Chen
- Department of Speech, Hearing and Phonetic Sciences, University College London, London, UK
| | - Yu Yan
- Department of Speech, Hearing and Phonetic Sciences, University College London, London, UK
| | - Carolyn McGettigan
- Department of Speech, Hearing and Phonetic Sciences, University College London, London, UK
| | - Stuart Rosen
- Department of Speech, Hearing and Phonetic Sciences, University College London, London, UK
| | - Patti Adank
- Department of Speech, Hearing and Phonetic Sciences, University College London, London, UK
| |
Collapse
|
6
|
Kawar K, Kishon-Rabin L, Segal O. Identification and Comprehension of Narrow Focus by Arabic-Speaking Adolescents With Moderate-to-Profound Hearing Loss. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:2029-2046. [PMID: 35472256 DOI: 10.1044/2022_jslhr-21-00296] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
PURPOSE Processing narrow focus (NF), the stressed word in the sentence, includes both the perceptual ability to identify the stressed word in the sentence and the pragmatic-semantic ability to comprehend the nonexplicit linguistic message. NF and its underlying meaning can be conveyed only via the auditory modality. Therefore, NF can be considered as a measure for assessing the efficacy of the hearing aid (HA) and cochlear implants (CIs) for acquiring nonexplicit language skills. The purpose of this study was to assess identification and comprehension of NF by HA and CI users who are native speakers of Arabic and to associate NF outcomes with speech perception and cognitive and linguistic abilities. METHOD A total of 46 adolescents (age range: 11;2-18;8) participated: 18 with moderate-to-severe hearing loss who used HAs, 10 with severe-to-profound hearing loss who used CIs, and 18 with typical hearing (TH). Test materials included the Arabic Narrow Focus Test (ANFT), which includes three subtests assessing identification (ANFT1), comprehension of NF in simple four-word sentences (ANFT2), and longer sentences with a construction list at the clause or noun phrase level (ANFT3). In addition, speech perception, vocabulary, and working memory were assessed. RESULTS All the participants successfully identified the word carrying NF, with no significant difference between the groups. Comprehension of NF in ANFT2 and ANFT3 was reduced for HA and CI users compared with TH peers, and speech perception, hearing status, and memory for digits predicted the variability in the overall results of ANFT1, ANFT2, and ANFT3, respectively. CONCLUSIONS Arabic speakers who used HAs or CIs were able to identify NF successfully, suggesting that the acoustic cues were perceptually available to them. However, HA and CI users had considerable difficulty in understanding NF. Different factors may contribute to this difficulty, including the memory load during the task as well as pragmatic-linguistic knowledge on the possible meanings of NF.
Collapse
Affiliation(s)
- Khaloob Kawar
- Department of Special Education, Beit Berl College, Kfar Saba, Israel
- Department of Communication Disorders, Steyer School of Health Professions, Sackler Faculty of Medicine, Tel Aviv University, Israel
| | - Liat Kishon-Rabin
- Department of Communication Disorders, Steyer School of Health Professions, Sackler Faculty of Medicine, Tel Aviv University, Israel
| | - Osnat Segal
- Department of Communication Disorders, Steyer School of Health Professions, Sackler Faculty of Medicine, Tel Aviv University, Israel
| |
Collapse
|
7
|
Abstract
OBJECTIVES The purpose of the present study was to determine whether age and hearing ability influence selective attention during childhood. Specifically, we hypothesized that immaturity and disrupted auditory experience impede selective attention during childhood. DESIGN Seventy-seven school-age children (5 to 12 years of age) participated in this study: 61 children with normal hearing and 16 children with bilateral hearing loss who use hearing aids and/or cochlear implants. Children performed selective attention-based behavioral change detection tasks comprised of target and distractor streams in the auditory and visual modalities. In the auditory modality, children were presented with two streams of single-syllable words spoken by a male and female talker. In the visual modality, children were presented with two streams of grayscale images. In each task, children were instructed to selectively attend to the target stream, inhibit attention to the distractor stream, and press a key as quickly as possible when they detected a frequency (auditory modality) or color (visual modality) deviant stimulus in the target, but not distractor, stream. Performance on the auditory and visual change detection tasks was quantified by response sensitivity, which reflects children's ability to selectively attend to deviants in the target stream and inhibit attention to those in the distractor stream. Children also completed a standardized measure of attention and inhibitory control. RESULTS Younger children and children with hearing loss demonstrated lower response sensitivity, and therefore poorer selective attention, than older children and children with normal hearing, respectively. The effect of hearing ability on selective attention was observed across the auditory and visual modalities, although the extent of this group difference was greater in the auditory modality than the visual modality due to differences in children's response patterns. Additionally, children's performance on a standardized measure of attention and inhibitory control related to their performance during the auditory and visual change detection tasks. CONCLUSIONS Overall, the findings from the present study suggest that age and hearing ability influence children's ability to selectively attend to a target stream in both the auditory and visual modalities. The observed differences in response patterns across modalities, however, reveal a complex interplay between hearing ability, task modality, and selective attention during childhood. While the effect of age on selective attention is expected to reflect the immaturity of cognitive and linguistic processes, the effect of hearing ability may reflect altered development of selective attention due to disrupted auditory experience early in life and/or a differential allocation of attentional resources to meet task demands.
Collapse
|
8
|
Saddler MR, Gonzalez R, McDermott JH. Deep neural network models reveal interplay of peripheral coding and stimulus statistics in pitch perception. Nat Commun 2021; 12:7278. [PMID: 34907158 PMCID: PMC8671597 DOI: 10.1038/s41467-021-27366-6] [Citation(s) in RCA: 16] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2020] [Accepted: 11/12/2021] [Indexed: 11/15/2022] Open
Abstract
Perception is thought to be shaped by the environments for which organisms are optimized. These influences are difficult to test in biological organisms but may be revealed by machine perceptual systems optimized under different conditions. We investigated environmental and physiological influences on pitch perception, whose properties are commonly linked to peripheral neural coding limits. We first trained artificial neural networks to estimate fundamental frequency from biologically faithful cochlear representations of natural sounds. The best-performing networks replicated many characteristics of human pitch judgments. To probe the origins of these characteristics, we then optimized networks given altered cochleae or sound statistics. Human-like behavior emerged only when cochleae had high temporal fidelity and when models were optimized for naturalistic sounds. The results suggest pitch perception is critically shaped by the constraints of natural environments in addition to those of the cochlea, illustrating the use of artificial neural networks to reveal underpinnings of behavior.
Collapse
Affiliation(s)
- Mark R Saddler
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA.
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA.
- Center for Brains, Minds and Machines, MIT, Cambridge, MA, USA.
| | - Ray Gonzalez
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Center for Brains, Minds and Machines, MIT, Cambridge, MA, USA
| | - Josh H McDermott
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA.
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA.
- Center for Brains, Minds and Machines, MIT, Cambridge, MA, USA.
- Program in Speech and Hearing Biosciences and Technology, Harvard University, Cambridge, MA, USA.
| |
Collapse
|
9
|
Effects of Hearing Loss on School-Aged Children's Ability to Benefit From F0 Differences Between Target and Masker Speech. Ear Hear 2021; 42:1084-1096. [PMID: 33538428 PMCID: PMC8222052 DOI: 10.1097/aud.0000000000000979] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES The objectives of the study were to (1) evaluate the impact of hearing loss on children's ability to benefit from F0 differences between target/masker speech in the context of aided speech-in-speech recognition and (2) to determine whether compromised F0 discrimination associated with hearing loss predicts F0 benefit in individual children. We hypothesized that children wearing appropriately fitted amplification would benefit from F0 differences, but they would not show the same magnitude of benefit as children with normal hearing. Reduced audibility and poor suprathreshold encoding that degrades frequency discrimination were expected to impair children's ability to segregate talkers based on F0. DESIGN Listeners were 9 to 17 year olds with bilateral, symmetrical, sensorineural hearing loss ranging in degree from mild to severe. A four-alternative, forced-choice procedure was used to estimate thresholds for disyllabic word recognition in a 60-dB-SPL two-talker masker. The same male talker produced target and masker speech. Target words had either the same mean F0 as the masker or were digitally shifted higher than the masker by three, six, or nine semitones. The F0 benefit was defined as the difference in thresholds between the shifted-F0 conditions and the unshifted-F0 condition. Thresholds for discriminating F0 were also measured, using a three-alternative, three-interval forced choice procedure, to determine whether compromised sensitivity to F0 differences due to hearing loss would predict children's ability to benefit from F0. Testing was performed in the sound field, and all children wore their personal hearing aids at user settings. RESULTS Children with hearing loss benefited from an F0 difference of nine semitones between target words and masker speech, with older children generally benefitting more than younger children. Some children benefitted from an F0 difference of six semitones, but this was not consistent across listeners. Thresholds for discriminating F0 improved with increasing age and predicted F0 benefit in the nine-semitone condition. An exploratory analysis indicated that F0 benefit was not significantly correlated with the four-frequency pure-tone average (0.5, 1, 2, and 4 kHz), aided audibility, or consistency of daily hearing aid use, although there was a trend for an association with the low-frequency pure-tone average (0.25 and 0.5 kHz). Comparisons of the present data to our previous study of children with normal hearing demonstrated that children with hearing loss benefitted less than children with normal hearing for the F0 differences tested. CONCLUSIONS The results demonstrate that children with mild-to-severe hearing loss who wear hearing aids benefit from relatively large F0 differences between target and masker speech during aided speech-in-speech recognition. The size of the benefit increases with increasing age, consistent with previously reported age effects for children with normal hearing. However, hearing loss reduces children's ability to capitalize on F0 differences between talkers. Audibility alone does not appear to be responsible for this effect; aided audibility and degree of loss were not primary predictors of performance. The ability to benefit from F0 differences may be limited by immature central processing or aspects of peripheral encoding that are not characterized in standard clinical assessments.
Collapse
|
10
|
No interaction between fundamental-frequency differences and spectral region when perceiving speech in a speech background. PLoS One 2021; 16:e0249654. [PMID: 33826663 PMCID: PMC8026035 DOI: 10.1371/journal.pone.0249654] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2020] [Accepted: 03/22/2021] [Indexed: 02/06/2023] Open
Abstract
Differences in fundamental frequency (F0) or pitch between competing voices facilitate our ability to segregate a target voice from interferers, thereby enhancing speech intelligibility. Although lower-numbered harmonics elicit a stronger and more accurate pitch sensation than higher-numbered harmonics, it is unclear whether the stronger pitch leads to an increased benefit of pitch differences when segregating competing talkers. To answer this question, sentence recognition was tested in young normal-hearing listeners in the presence of a single competing talker. The stimuli were presented in a broadband condition or were highpass or lowpass filtered to manipulate the pitch accuracy of the voicing, while maintaining roughly equal speech intelligibility in the highpass and lowpass regions. Performance was measured with average F0 differences (ΔF0) between the target and single-talker masker of 0, 2, and 4 semitones. Pitch discrimination abilities were also measured to confirm that the lowpass-filtered stimuli elicited greater pitch accuracy than the highpass-filtered stimuli. No interaction was found between filter type and ΔF0 in the sentence recognition task, suggesting little or no effect of harmonic rank or pitch accuracy on the ability to use F0 to segregate natural voices, even when the average ΔF0 is relatively small. The results suggest that listeners are able to obtain some benefit of pitch differences between competing voices, even when pitch salience and accuracy is low.
Collapse
|
11
|
Shen J. Older Listeners' Perception of Speech With Strengthened and Weakened Dynamic Pitch Cues in Background Noise. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2021; 64:348-358. [PMID: 33439741 PMCID: PMC8632513 DOI: 10.1044/2020_jslhr-20-00116] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/15/2020] [Revised: 07/28/2020] [Accepted: 09/21/2020] [Indexed: 06/12/2023]
Abstract
Purpose Dynamic pitch, which is defined as the variation in fundamental frequency, is an acoustic cue that aids speech perception in noise. This study examined the effects of strengthened and weakened dynamic pitch cues on older listeners' speech perception in noise, as well as how these effects were modulated by individual factors including spectral perception ability. Method The experiment measured speech reception thresholds in noise in both younger listeners with normal hearing and older listeners whose hearing status ranged from near-normal hearing to mild-to-moderate sensorineural hearing loss. The pitch contours of the target speech were manipulated to create four levels of dynamic pitch strength: weakened, original, mildly strengthened, and strengthened. Listeners' spectral perception ability was measured using tests of spectral ripple and frequency modulation discrimination. Results Both younger and older listeners performed worse with manipulated dynamic pitch cues than with original dynamic pitch. The effects of dynamic pitch on older listeners' speech recognition were associated with their age but not with their perception of spectral information. Those older listeners who were relatively younger were more negatively affected by dynamic pitch manipulations. Conclusions The findings suggest the current pitch manipulation strategy is detrimental for older listeners to perceive speech in noise, as compared to original dynamic pitch. While the influence of age on the effects of dynamic pitch is likely due to age-related declines in pitch perception, the spectral measures used in this study were not strong predictors for dynamic pitch effects. Taken together, these results indicate next steps in this line of work should be focused on how to manipulate acoustic cues in speech in order to improve speech perception in noise for older listeners.
Collapse
Affiliation(s)
- Jing Shen
- Department of Speech, Language and Hearing Sciences, Western Michigan University, Kalamazoo
| |
Collapse
|
12
|
Liu Y, Xu R, Gong Q. Human Auditory-Frequency Tuning Is Sensitive to Tonal Language Experience. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:4277-4288. [PMID: 33151817 DOI: 10.1044/2020_jslhr-20-00152] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Purpose The aim of this study is to investigate whether human auditory frequency tuning can be influenced by tonal language experience. Method Perceptual tuning measured via psychophysical tuning curves and cochlear tuning derived via stimulus-frequency otoacoustic emission suppression tuning curves in 14 native speakers of a tonal language (Mandarin) were compared to those of 14 native speakers of a nontonal language (English) at 1 and 4 kHz. Results Group comparisons of both psychophysical tuning curves (p = .046) and stimulus-frequency otoacoustic emission suppression tuning curves (p = .007) in the 4-kHz region indicated sharper frequency tuning in the Mandarin-speaking group relative to the English-speaking group. The auditory tuning was better at the higher (4 kHz) than the lower (1 kHz) probe frequencies (p < .001). Conclusions The sharper auditory tuning in the 4-kHz cochlear region is associated with long-term tonal language (i.e., Mandarin) experience. Experience-dependent plasticity of tonal language may occur before the sound signal reaches central neural stages, as peripheral as the cochlea.
Collapse
Affiliation(s)
- Yin Liu
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China
| | - Runyi Xu
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China
| | - Qin Gong
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China
- School of Medicine, Shanghai University, China
| |
Collapse
|
13
|
Burton JA, Mackey CA, MacDonald KS, Hackett TA, Ramachandran R. Changes in audiometric threshold and frequency selectivity correlate with cochlear histopathology in macaque monkeys with permanent noise-induced hearing loss. Hear Res 2020; 398:108082. [PMID: 33045479 PMCID: PMC7769151 DOI: 10.1016/j.heares.2020.108082] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/14/2020] [Revised: 09/12/2020] [Accepted: 09/20/2020] [Indexed: 02/07/2023]
Abstract
Exposure to loud noise causes damage to the inner ear, including but not limited to outer and inner hair cells (OHCs and IHCs) and IHC ribbon synapses. This cochlear damage impairs auditory processing and increases audiometric thresholds (noise-induced hearing loss, NIHL). However, the exact relationship between the perceptual consequences of NIHL and its underlying cochlear pathology are poorly understood. This study used a nonhuman primate model of NIHL to relate changes in frequency selectivity and audiometric thresholds to indices of cochlear histopathology. Three macaques (one Macaca mulatta and two Macaca radiata) were trained to detect tones in quiet and in noises that were spectrally notched around the tone frequency. Audiograms were derived from tone thresholds in quiet; perceptual auditory filters were derived from tone thresholds in notched-noise maskers using the rounded-exponential fit. Data were obtained before and after a four-hour exposure to a 50-Hz noise centered at 2 kHz at 141 or 146 dB SPL. Noise exposure caused permanent audiometric threshold shifts and broadening of auditory filters at and above 2 kHz, with greater changes observed for the 146-dB-exposed monkeys. The normalized bandwidth of the perceptual auditory filters was strongly correlated with audiometric threshold at each tone frequency. While changes in audiometric threshold and perceptual auditory filter widths were primarily determined by the extent of OHC survival, additional variability was explained by including interactions among OHC, IHC, and ribbon synapse survival. This is the first study to provide within-subject comparisons of auditory filter bandwidths in an animal model of NIHL and correlate these NIHL-related perceptual changes with cochlear histopathology. These results expand the foundations for ongoing investigations of the neural correlates of NIHL-related perceptual changes.
Collapse
Affiliation(s)
- Jane A Burton
- Neuroscience Graduate Program, Vanderbilt University, Nashville, TN 37235, United States.
| | - Chase A Mackey
- Neuroscience Graduate Program, Vanderbilt University, Nashville, TN 37235, United States.
| | - Kaitlyn S MacDonald
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN 37232, United States.
| | - Troy A Hackett
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN 37232, United States.
| | - Ramnarayan Ramachandran
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN 37232, United States.
| |
Collapse
|
14
|
Adaptation to pitch-altered feedback is independent of one's own voice pitch sensitivity. Sci Rep 2020; 10:16860. [PMID: 33033324 PMCID: PMC7544828 DOI: 10.1038/s41598-020-73932-1] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2020] [Accepted: 09/23/2020] [Indexed: 01/17/2023] Open
Abstract
Monitoring voice pitch is a fine-tuned process in daily conversations as conveying accurately the linguistic and affective cues in a given utterance depends on the precise control of phonation and intonation. This monitoring is thought to depend on whether the error is treated as self-generated or externally-generated, resulting in either a correction or inflation of errors. The present study reports on two separate paradigms of adaptation to altered feedback to explore whether participants could behave in a more cohesive manner once the error is of comparable size perceptually. The vocal behavior of normal-hearing and fluent speakers was recorded in response to a personalized size of pitch shift versus a non-specific size, one semitone. The personalized size of shift was determined based on the just-noticeable difference in fundamental frequency (F0) of each participant’s voice. Here we show that both tasks successfully demonstrated opposing responses to a constant and predictable F0 perturbation (on from the production onset) but these effects barely carried over once the feedback was back to normal, depicting a pattern that bears some resemblance to compensatory responses. Experiencing a F0 shift that is perceived as self-generated (because it was precisely just-noticeable) is not enough to force speakers to behave more consistently and more homogeneously in an opposing manner. On the contrary, our results suggest that the type of the response as well as the magnitude of the response do not depend in any trivial way on the sensitivity of participants to their own voice pitch. Based on this finding, we speculate that error correction could possibly occur even with a bionic ear, typically even when F0 cues are too subtle for cochlear implant users to detect accurately.
Collapse
|
15
|
Ahrens A, Joshi SN, Epp B. Perceptual Weighting of Binaural Lateralization Cues across Frequency Bands. J Assoc Res Otolaryngol 2020; 21:485-496. [PMID: 32915339 PMCID: PMC7644719 DOI: 10.1007/s10162-020-00770-3] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2019] [Accepted: 08/04/2020] [Indexed: 11/25/2022] Open
Abstract
The auditory system uses interaural time and level differences (ITD and ILD) as cues to localize and lateralize sounds. The availability of ITDs and ILDs in the auditory system is limited by neural phase-locking and by the head size, respectively. Although the frequency-specific limitations are well known, the relative contribution of ITDs and ILDs in individual frequency bands in broadband stimuli is unknown. To determine these relative contributions, or spectral weights, listeners were asked to lateralize stimuli consisting of eleven simultaneously presented 1-ERB-wide noise bands centered between 442 and 5544 Hz and separated by 1-ERB-wide gaps. Either ITDs or ILDs were varied independently across each noise band, while fixing the other interaural disparity to either 0 dB or 0 μs. The weights were obtained using a multiple linear regression analysis. In a second experiment, the effect of auditory enhancement on the spectral weights was investigated. The enhancement of single noise bands was realized by presenting ten of the noise bands as preceding and following sounds (pre- and post-cursors, respectively). Listeners were asked to lateralize the stimuli as in the first experiment. Results show that in the absence of pre- and post-cursors, only the lowest or highest frequency band received highest weight for ITD and ILD, respectively. Auditory enhancement led to significantly enhanced weights given to the band without the pre- and post-cursor. The weight enhancement could only be observed at low frequencies, when determined with ITD cues and for low and high frequencies for ILDs. Hence, the auditory system seems to be able to change the spectral weighting of binaural information depending on the information content.
Collapse
Affiliation(s)
- Axel Ahrens
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Suyash Narendra Joshi
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Kgs. Lyngby, Denmark
- Present Address: Brain & Sound Lab, Department of Biomedicine, Basel University, Basel, Switzerland
| | - Bastian Epp
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Kgs. Lyngby, Denmark
| |
Collapse
|
16
|
Mehta AH, Oxenham AJ. Effect of lowest harmonic rank on fundamental-frequency difference limens varies with fundamental frequency. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:2314. [PMID: 32359332 PMCID: PMC7166120 DOI: 10.1121/10.0001092] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/19/2019] [Revised: 03/25/2020] [Accepted: 03/27/2020] [Indexed: 06/11/2023]
Abstract
This study investigated the relationship between fundamental frequency difference limens (F0DLs) and the lowest harmonic number present over a wide range of F0s (30-2000 Hz) for 12-component harmonic complex tones that were presented in either sine or random phase. For fundamental frequencies (F0s) between 100 and 400 Hz, a transition from low (∼1%) to high (∼5%) F0DLs occurred as the lowest harmonic number increased from about seven to ten, in line with earlier studies. At lower and higher F0s, the transition between low and high F0DLs occurred at lower harmonic numbers. The worsening performance at low F0s was reasonably well predicted by the expected decrease in spectral resolution below about 500 Hz. At higher F0s, the degradation in performance at lower harmonic numbers could not be predicted by changes in spectral resolution but remained relatively good (<2%-3%) in some conditions, even when all harmonics were above 8 kHz, confirming that F0 can be extracted from harmonics even when temporal envelope or fine-structure cues are weak or absent.
Collapse
Affiliation(s)
- Anahita H Mehta
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| |
Collapse
|
17
|
Mehta AH, Lu H, Oxenham AJ. The Perception of Multiple Simultaneous Pitches as a Function of Number of Spectral Channels and Spectral Spread in a Noise-Excited Envelope Vocoder. J Assoc Res Otolaryngol 2020; 21:61-72. [PMID: 32048077 DOI: 10.1007/s10162-019-00738-y] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2018] [Accepted: 10/30/2019] [Indexed: 01/06/2023] Open
Abstract
Cochlear implant (CI) listeners typically perform poorly on tasks involving the pitch of complex tones. This limitation in performance is thought to be mainly due to the restricted number of active channels and the broad current spread that leads to channel interactions and subsequent loss of precise spectral information, with temporal information limited primarily to temporal-envelope cues. Little is known about the degree of spectral resolution required to perceive combinations of multiple pitches, or a single pitch in the presence of other interfering tones in the same spectral region. This study used noise-excited envelope vocoders that simulate the limited resolution of CIs to explore the perception of multiple pitches presented simultaneously. The results show that the resolution required for perceiving multiple complex pitches is comparable to that found in a previous study using single complex tones. Although relatively high performance can be achieved with 48 channels, performance remained near chance when even limited spectral spread (with filter slopes as steep as 144 dB/octave) was introduced to the simulations. Overall, these tight constraints suggest that current CI technology will not be able to convey the pitches of combinations of spectrally overlapping complex tones.
Collapse
Affiliation(s)
- Anahita H Mehta
- Department of Psychology, University of Minnesota, N218 Elliott Hall, 75 East River Parkway, Minneapolis, MN, 55455, USA.
| | - Hao Lu
- Department of Psychology, University of Minnesota, N218 Elliott Hall, 75 East River Parkway, Minneapolis, MN, 55455, USA
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, N218 Elliott Hall, 75 East River Parkway, Minneapolis, MN, 55455, USA
| |
Collapse
|
18
|
Halliday LF, Rosen S, Tuomainen O, Calcus A. Impaired frequency selectivity and sensitivity to temporal fine structure, but not envelope cues, in children with mild-to-moderate sensorineural hearing loss. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:4299. [PMID: 31893709 DOI: 10.1121/1.5134059] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/01/2019] [Accepted: 10/24/2019] [Indexed: 06/10/2023]
Abstract
Psychophysical thresholds were measured for 8-16 year-old children with mild-to-moderate sensorineural hearing loss (MMHL; N = 46) on a battery of auditory processing tasks that included measures designed to be dependent upon frequency selectivity and sensitivity to temporal fine structure (TFS) or envelope cues. Children with MMHL who wore hearing aids were tested in both unaided and aided conditions, and all were compared to a group of normally hearing (NH) age-matched controls. Children with MMHL performed more poorly than NH controls on tasks considered to be dependent upon frequency selectivity, sensitivity to TFS, and speech discrimination (/bɑ/-/dɑ/), but not on tasks measuring sensitivity to envelope cues. Auditory processing deficits remained regardless of age, were observed in both unaided and aided conditions, and could not be attributed to differences in nonverbal IQ or attention between groups. However, better auditory processing in children with MMHL was predicted by better audiometric thresholds and, for aided tasks only, higher levels of maternal education. These results suggest that, as for adults with MMHL, children with MMHL may show deficits in frequency selectivity and sensitivity to TFS, but sensitivity to the envelope may remain intact.
Collapse
Affiliation(s)
- Lorna F Halliday
- Speech, Hearing, and Phonetic Sciences, University College London, Chandler House, 2 Wakefield Street, London WC1N 1PF, United Kingdom
| | - Stuart Rosen
- Speech, Hearing, and Phonetic Sciences, University College London, Chandler House, 2 Wakefield Street, London WC1N 1PF, United Kingdom
| | - Outi Tuomainen
- Speech, Hearing, and Phonetic Sciences, University College London, Chandler House, 2 Wakefield Street, London WC1N 1PF, United Kingdom
| | - Axelle Calcus
- Speech, Hearing, and Phonetic Sciences, University College London, Chandler House, 2 Wakefield Street, London WC1N 1PF, United Kingdom
| |
Collapse
|
19
|
Bowen Z, Winkowski DE, Seshadri S, Plenz D, Kanold PO. Neuronal Avalanches in Input and Associative Layers of Auditory Cortex. Front Syst Neurosci 2019; 13:45. [PMID: 31551721 PMCID: PMC6737089 DOI: 10.3389/fnsys.2019.00045] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2019] [Accepted: 08/16/2019] [Indexed: 12/26/2022] Open
Abstract
The primary auditory cortex processes acoustic sequences for the perception of behaviorally meaningful sounds such as speech. Sound information arrives at its input layer four from where activity propagates to associative layer 2/3. It is currently not known whether there is a characteristic organization of neuronal population activity across layers and sound levels during sound processing. Here, we identify neuronal avalanches, which in theory and experiments have been shown to maximize dynamic range and optimize information transfer within and across networks, in primary auditory cortex. We used in vivo 2-photon imaging of pyramidal neurons in cortical layers L4 and L2/3 of mouse A1 to characterize the populations of neurons that were active spontaneously, i.e., in the absence of a sound stimulus, and those recruited by single-frequency tonal stimuli at different sound levels. Single-frequency sounds recruited neurons of widely ranging frequency selectivity in both layers. We defined neuronal ensembles as neurons being active within or during successive temporal windows at the temporal resolution of our imaging. For both layers, neuronal ensembles were highly variable in size during spontaneous activity as well as during sound presentation. Ensemble sizes distributed according to power laws, the hallmark of neuronal avalanches, and were similar across sound levels. Avalanches activated by sound were composed of neurons with diverse tuning preference, yet with selectivity independent of avalanche size. Our results suggest that optimization principles identified for avalanches guide population activity in L4 and L2/3 of auditory cortex during and in-between stimulus processing.
Collapse
Affiliation(s)
- Zac Bowen
- Department of Biology, University of Maryland, College Park, College Park, MD, United States
| | - Daniel E Winkowski
- Department of Biology, University of Maryland, College Park, College Park, MD, United States.,Institute for Systems Research, University of Maryland, College Park, College Park, MD, United States
| | - Saurav Seshadri
- Section on Critical Brain Dynamics, National Institute of Mental Health, Bethesda, MD, United States
| | - Dietmar Plenz
- Section on Critical Brain Dynamics, National Institute of Mental Health, Bethesda, MD, United States
| | - Patrick O Kanold
- Department of Biology, University of Maryland, College Park, College Park, MD, United States.,Institute for Systems Research, University of Maryland, College Park, College Park, MD, United States
| |
Collapse
|
20
|
Bianchi F, Carney LH, Dau T, Santurette S. Effects of Musical Training and Hearing Loss on Fundamental Frequency Discrimination and Temporal Fine Structure Processing: Psychophysics and Modeling. J Assoc Res Otolaryngol 2019; 20:263-277. [PMID: 30693416 PMCID: PMC6513935 DOI: 10.1007/s10162-018-00710-2] [Citation(s) in RCA: 20] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2018] [Accepted: 12/19/2018] [Indexed: 11/01/2022] Open
Abstract
Several studies have shown that musical training leads to improved fundamental frequency (F0) discrimination for young listeners with normal hearing (NH). It is unclear whether a comparable effect of musical training occurs for listeners whose sensory encoding of F0 is degraded. To address this question, the effect of musical training was investigated for three groups of listeners (young NH, older NH, and older listeners with hearing impairment, HI). In a first experiment, F0 discrimination was investigated using complex tones that differed in harmonic content and phase configuration (sine, positive, or negative Schroeder). Musical training was associated with significantly better F0 discrimination of complex tones containing low-numbered harmonics for all groups of listeners. Part of this effect was caused by the fact that musicians were more robust than non-musicians to harmonic roving. Despite the benefit relative to their non-musicians counterparts, the older musicians, with or without HI, performed worse than the young musicians. In a second experiment, binaural sensitivity to temporal fine structure (TFS) cues was assessed for the same listeners by estimating the highest frequency at which an interaural phase difference was perceived. Performance was better for musicians for all groups of listeners and the use of TFS cues was degraded for the two older groups of listeners. These findings suggest that musical training is associated with an enhancement of both TFS cues encoding and F0 discrimination in young and older listeners with or without HI, although the musicians' benefit decreased with increasing hearing loss. Additionally, models of the auditory periphery and midbrain were used to examine the effect of HI on F0 encoding. The model predictions reflected the worsening in F0 discrimination with increasing HI and accounted for up to 80 % of the variance in the data.
Collapse
Affiliation(s)
- Federica Bianchi
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Ørsteds Plads, Building 352, 2800, Lyngby, Denmark.
- Current Affiliation: Oticon Medical, Kongebakken 9, Smørum, Denmark.
| | - Laurel H Carney
- Departments of Biomedical Engineering and Neuroscience, University of Rochester, Rochester, NY, USA
| | - Torsten Dau
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Ørsteds Plads, Building 352, 2800, Lyngby, Denmark
| | - Sébastien Santurette
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Ørsteds Plads, Building 352, 2800, Lyngby, Denmark
- Department of Otorhinolaryngology, Head and Neck Surgery & Audiology, Rigshospitalet, 2100, Copenhagen, Denmark
| |
Collapse
|
21
|
Qi B, Liu P, Gu X, Dong R, Liu B. Categorical perception of lexical tones in native Mandarin-speaking listeners with sensorineural hearing loss. Acta Otolaryngol 2018; 138:801-806. [PMID: 29764263 DOI: 10.1080/00016489.2018.1467040] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/16/2022]
Abstract
BACKGROUND Categorical perception (CP) of lexical tones was examined in normal hearing (NH) people, but it was unclear whether lexical tones can be perceived categorically in sensorineural hearing loss (SNHL) people. OBJECTIVES To explore the characteristic of lexical tone perception in native Mandarin speakers with SNHL. MATERIALS AND METHODS Three types of continuum (Tone1/Tone2, Tone1/Tone4 and Tone2/Tone3) were constructed and each of them includes 15 stimuli which were resynthesized by applying the pitch-synchronous overlap and add (PSOLA) method implemented in Praat to the same Mandarin syllable, /a/, with a high-level tone produced by a female speaker. Forty native Mandarin NH speakers and 23 native Mandarin speakers with mild to moderate SNHL were recruited. A two alternative-forced-choice identification task was used to acquire the tonal perceptual data. RESULTS All tone perception curves owns the characteristic of CP in SNHL subjects. All tone perception curves were S-shape in SNHL subjects same as those in NH subjects. No significant difference of each continuum was observed between SNHL and NH. CONCLUSIONS CP of lexical tone perception could be observed in native Mandarin speakers with mild to moderate SNHL. The slight damage in the peripheral auditory system did not change characteristic of lexical tone perception.
Collapse
Affiliation(s)
- Beier Qi
- Department of Otolaryngology Head and Neck Surgery, Beijing TongRen Hospital, Capital Medical University, Beijing, PR China
- Key Laboratory of Otolaryngology Head and Neck Surgery (Ministry of Education), Beijing Institute of Otolaryngology, Beijing, PR China
| | - Peng Liu
- Capital Institute of Paediatrics, Beijing, PR China
| | - Xin Gu
- Department of Otolaryngology Head and Neck Surgery, Beijing TongRen Hospital, Capital Medical University, Beijing, PR China
- Key Laboratory of Otolaryngology Head and Neck Surgery (Ministry of Education), Beijing Institute of Otolaryngology, Beijing, PR China
| | - Ruijuan Dong
- Department of Otolaryngology Head and Neck Surgery, Beijing TongRen Hospital, Capital Medical University, Beijing, PR China
- Key Laboratory of Otolaryngology Head and Neck Surgery (Ministry of Education), Beijing Institute of Otolaryngology, Beijing, PR China
| | - Bo Liu
- Department of Otolaryngology Head and Neck Surgery, Beijing TongRen Hospital, Capital Medical University, Beijing, PR China
- Key Laboratory of Otolaryngology Head and Neck Surgery (Ministry of Education), Beijing Institute of Otolaryngology, Beijing, PR China
| |
Collapse
|
22
|
Madsen SMK, Dau T, Moore BCJ. Effect of harmonic rank on sequential sound segregation. Hear Res 2018; 367:161-168. [PMID: 30006111 DOI: 10.1016/j.heares.2018.06.002] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/24/2018] [Revised: 06/01/2018] [Accepted: 06/08/2018] [Indexed: 11/26/2022]
Abstract
The ability to segregate sounds from different sound sources is thought to depend on the perceptual salience of differences between the sounds, such as differences in frequency or fundamental frequency (F0). F0 discrimination of complex tones is better for tones with low harmonics than for tones that only contain high harmonics, suggesting greater pitch salience for the former. This leads to the expectation that the sequential stream segregation (streaming) of complex tones should be better for tones with low harmonics than for tones with only high harmonics. However, the results of previous studies are conflicting about whether this is the case. The goals of this study were to determine the effect of harmonic rank on streaming and to establish whether streaming is related to F0 discrimination. Thirteen young normal-hearing participants were tested. Streaming was assessed for pure tones and complex tones containing harmonics with various ranks using sequences of ABA triplets, where A and B differed in frequency or in F0. The participants were asked to try to hear two streams and to indicate when they heard one and when they heard two streams. F0 discrimination was measured for the same tones that were used as A tones in the streaming experiment. Both streaming and F0 discrimination worsened significantly with increasing harmonic rank. There was a significant relationship between streaming and F0 discrimination, indicating that good F0 discrimination is associated with good streaming. This supports the idea that the extent of stream segregation depends on the salience of the perceptual difference between successive sounds.
Collapse
Affiliation(s)
- Sara M K Madsen
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, DK-2800, Lyngby, Denmark.
| | - Torsten Dau
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, DK-2800, Lyngby, Denmark
| | - Brian C J Moore
- Department of Psychology, University of Cambridge, Cambridge, UK
| |
Collapse
|
23
|
McPherson MJ, McDermott JH. Diversity in pitch perception revealed by task dependence. Nat Hum Behav 2018; 2:52-66. [PMID: 30221202 PMCID: PMC6136452 DOI: 10.1038/s41562-017-0261-8] [Citation(s) in RCA: 33] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2017] [Accepted: 11/08/2017] [Indexed: 01/12/2023]
Abstract
Pitch conveys critical information in speech, music, and other natural sounds, and is conventionally defined as the perceptual correlate of a sound's fundamental frequency (F0). Although pitch is widely assumed to be subserved by a single F0 estimation process, real-world pitch tasks vary enormously, raising the possibility of underlying mechanistic diversity. To probe pitch mechanisms we conducted a battery of pitch-related music and speech tasks using conventional harmonic sounds and inharmonic sounds whose frequencies lack a common F0. Some pitch-related abilities - those relying on musical interval or voice recognition - were strongly impaired by inharmonicity, suggesting a reliance on F0. However, other tasks, including those dependent on pitch contours in speech and music, were unaffected by inharmonicity, suggesting a mechanism that tracks the frequency spectrum rather than the F0. The results suggest that pitch perception is mediated by several different mechanisms, only some of which conform to traditional notions of pitch.
Collapse
Affiliation(s)
- Malinda J McPherson
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, USA.
- Program in Speech and Hearing Bioscience and Technology, Harvard University, Cambridge, MA, USA.
| | - Josh H McDermott
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, USA
- Program in Speech and Hearing Bioscience and Technology, Harvard University, Cambridge, MA, USA
| |
Collapse
|
24
|
Segal O, Kishon-Rabin L. Recognition and Comprehension of "Narrow Focus" by Young Adults With Prelingual Hearing Loss Using Hearing Aids or Cochlear Implants. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2017; 60:3609-3624. [PMID: 29121171 DOI: 10.1044/2017_jslhr-h-16-0342] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/01/2016] [Accepted: 06/02/2017] [Indexed: 06/07/2023]
Abstract
PURPOSE The stressed word in a sentence (narrow focus [NF]) conveys information about the intent of the speaker and is therefore important for processing spoken language and in social interactions. The ability of participants with severe-to-profound prelingual hearing loss to comprehend NF has rarely been investigated. The purpose of this study was to assess the recognition and comprehension of NF by young adults with prelingual hearing loss compared with those of participants with normal hearing (NH). METHOD The participants included young adults with hearing aids (HA; n = 10), cochlear implants (CI; n = 12), and NH (n = 18). The test material included the Hebrew Narrow Focus Test (Segal, Kaplan, Patael, & Kishon-Rabin, in press), with 3 subtests, which was used to assess the recognition and comprehension of NF in different contexts. RESULTS The following results were obtained: (a) CI and HA users successfully recognized the stressed word, with the worst performance for CI; (b) HA and CI comprehended NF less well than NH; and (c) the comprehension of NF was associated with verbal working memory and expressive vocabulary in CI users. CONCLUSIONS Most CI and HA users were able to recognize the stressed word in a sentence but had considerable difficulty understanding it. Different factors may contribute to this difficulty, including the memory load during the task itself and linguistic and pragmatic abilities. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.5572792.
Collapse
Affiliation(s)
- Osnat Segal
- Department of Communication Disorders, The Stanley Steyer School of Health Professions, Sackler Faculty of Medicine, Tel-Aviv University, Israel
| | - Liat Kishon-Rabin
- Department of Communication Disorders, The Stanley Steyer School of Health Professions, Sackler Faculty of Medicine, Tel-Aviv University, Israel
| |
Collapse
|
25
|
Bianchi F, Hjortkjær J, Santurette S, Zatorre RJ, Siebner HR, Dau T. Subcortical and cortical correlates of pitch discrimination: Evidence for two levels of neuroplasticity in musicians. Neuroimage 2017; 163:398-412. [DOI: 10.1016/j.neuroimage.2017.07.057] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2017] [Revised: 07/11/2017] [Accepted: 07/27/2017] [Indexed: 10/19/2022] Open
|
26
|
Abstract
Auditory perception is our main gateway to communication with others via speech and music, and it also plays an important role in alerting and orienting us to new events. This review provides an overview of selected topics pertaining to the perception and neural coding of sound, starting with the first stage of filtering in the cochlea and its profound impact on perception. The next topic, pitch, has been debated for millennia, but recent technical and theoretical developments continue to provide us with new insights. Cochlear filtering and pitch both play key roles in our ability to parse the auditory scene, enabling us to attend to one auditory object or stream while ignoring others. An improved understanding of the basic mechanisms of auditory perception will aid us in the quest to tackle the increasingly important problem of hearing loss in our aging population.
Collapse
Affiliation(s)
- Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455;
| |
Collapse
|
27
|
Vocoder Simulations Explain Complex Pitch Perception Limitations Experienced by Cochlear Implant Users. J Assoc Res Otolaryngol 2017; 18:789-802. [PMID: 28733803 DOI: 10.1007/s10162-017-0632-x] [Citation(s) in RCA: 31] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2017] [Accepted: 06/26/2017] [Indexed: 01/06/2023] Open
Abstract
Pitch plays a crucial role in speech and music, but is highly degraded for people with cochlear implants, leading to severe communication challenges in noisy environments. Pitch is determined primarily by the first few spectrally resolved harmonics of a tone. In implants, access to this pitch is limited by poor spectral resolution, due to the limited number of channels and interactions between adjacent channels. Here we used noise-vocoder simulations to explore how many channels, and how little channel interaction, are required to elicit pitch. Results suggest that two to four times the number of channels are needed, along with interactions reduced by an order of magnitude, than available in current devices. These new constraints not only provide insights into the basic mechanisms of pitch coding in normal hearing but also suggest that spectrally based complex pitch is unlikely to be generated in implant users without significant changes in the method or site of stimulation.
Collapse
|
28
|
Borucki E, Berg BG. Isolating spectral cues in amplitude and quasi-frequency modulation discrimination by reducing stimulus duration. Hear Res 2017; 348:129-133. [PMID: 28284735 DOI: 10.1016/j.heares.2017.03.004] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/22/2016] [Revised: 03/01/2017] [Accepted: 03/02/2017] [Indexed: 11/30/2022]
Abstract
This study investigated the psychophysical effects of distortion products in a listening task traditionally used to estimate the bandwidth of phase sensitivity. For a 2000 Hz carrier, estimates of modulation depth necessary to discriminate amplitude modulated (AM) tones and quasi-frequency modulated (QFM) were measured in a two interval forced choice task as a function modulation frequency. Temporal modulation transfer functions were often non-monotonic at modulation frequencies above 300 Hz. This was likely to be due to a spectral cue arising from the interaction of auditory distortion products and the lower sideband of the stimulus complex. When the stimulus duration was decreased from 200 ms to 20 ms, thresholds for low-frequency modulators rose to near-chance levels, whereas thresholds in the region of non-monotonicities were less affected. The decrease in stimulus duration appears to hinder the listener's ability to use temporal cues in order to discriminate between AM and QFM, whereas spectral information derived from distortion product cues appears more resilient.
Collapse
Affiliation(s)
- Ewa Borucki
- Department of Cognitive Sciences, 2201 Social & Behavioral Sciences Gateway Building (SBSG), University of California, Irvine, CA 92697-5100, USA.
| | - Bruce G Berg
- Department of Cognitive Sciences, 2201 Social & Behavioral Sciences Gateway Building (SBSG), University of California, Irvine, CA 92697-5100, USA
| |
Collapse
|
29
|
A Bayesian Account of Vocal Adaptation to Pitch-Shifted Auditory Feedback. PLoS One 2017; 12:e0169795. [PMID: 28135267 PMCID: PMC5279726 DOI: 10.1371/journal.pone.0169795] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2016] [Accepted: 12/21/2016] [Indexed: 11/19/2022] Open
Abstract
Motor systems are highly adaptive. Both birds and humans compensate for synthetically induced shifts in the pitch (fundamental frequency) of auditory feedback stemming from their vocalizations. Pitch-shift compensation is partial in the sense that large shifts lead to smaller relative compensatory adjustments of vocal pitch than small shifts. Also, compensation is larger in subjects with high motor variability. To formulate a mechanistic description of these findings, we adapt a Bayesian model of error relevance. We assume that vocal-auditory feedback loops in the brain cope optimally with known sensory and motor variability. Based on measurements of motor variability, optimal compensatory responses in our model provide accurate fits to published experimental data. Optimal compensation correctly predicts sensory acuity, which has been estimated in psychophysical experiments as just-noticeable pitch differences. Our model extends the utility of Bayesian approaches to adaptive vocal behaviors.
Collapse
|
30
|
Bianchi F, Fereczkowski M, Zaar J, Santurette S, Dau T. Complex-Tone Pitch Discrimination in Listeners With Sensorineural Hearing Loss. Trends Hear 2016; 20:20/0/2331216516655793. [PMID: 27604780 PMCID: PMC5017569 DOI: 10.1177/2331216516655793] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
Physiological studies have shown that noise-induced sensorineural hearing loss (SNHL) enhances the amplitude of envelope coding in auditory-nerve fibers. As pitch coding of unresolved complex tones is assumed to rely on temporal envelope coding mechanisms, this study investigated pitch-discrimination performance in listeners with SNHL. Pitch-discrimination thresholds were obtained for 14 normal-hearing (NH) and 10 hearing-impaired (HI) listeners for sine-phase (SP) and random-phase (RP) complex tones. When all harmonics were unresolved, the HI listeners performed, on average, worse than NH listeners in the RP condition but similarly to NH listeners in the SP condition. The increase in pitch-discrimination performance for the SP relative to the RP condition (F0DL ratio) was significantly larger in the HI as compared with the NH listeners. Cochlear compression and auditory-filter bandwidths were estimated in the same listeners. The estimated reduction of cochlear compression was significantly correlated with the increase in the F0DL ratio, while no correlation was found with filter bandwidth. The effects of degraded frequency selectivity and loss of compression were considered in a simplified peripheral model as potential factors in envelope enhancement. The model revealed that reducing cochlear compression significantly enhanced the envelope of an unresolved SP complex tone, while not affecting the envelope of a RP complex tone. This envelope enhancement in the SP condition was significantly correlated with the increased pitch-discrimination performance for the SP relative to the RP condition in the HI listeners.
Collapse
Affiliation(s)
- Federica Bianchi
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Michal Fereczkowski
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Johannes Zaar
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Sébastien Santurette
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Torsten Dau
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Kgs. Lyngby, Denmark
| |
Collapse
|
31
|
Bianchi F, Santurette S, Wendt D, Dau T. Pitch Discrimination in Musicians and Non-Musicians: Effects of Harmonic Resolvability and Processing Effort. J Assoc Res Otolaryngol 2015; 17:69-79. [PMID: 26637239 PMCID: PMC4722018 DOI: 10.1007/s10162-015-0548-2] [Citation(s) in RCA: 33] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2015] [Accepted: 11/11/2015] [Indexed: 11/30/2022] Open
Abstract
Musicians typically show enhanced pitch discrimination abilities compared to non-musicians. The present study investigated this perceptual enhancement behaviorally and objectively for resolved and unresolved complex tones to clarify whether the enhanced performance in musicians can be ascribed to increased peripheral frequency selectivity and/or to a different processing effort in performing the task. In a first experiment, pitch discrimination thresholds were obtained for harmonic complex tones with fundamental frequencies (F0s) between 100 and 500 Hz, filtered in either a low- or a high-frequency region, leading to variations in the resolvability of audible harmonics. The results showed that pitch discrimination performance in musicians was enhanced for resolved and unresolved complexes to a similar extent. Additionally, the harmonics became resolved at a similar F0 in musicians and non-musicians, suggesting similar peripheral frequency selectivity in the two groups of listeners. In a follow-up experiment, listeners' pupil dilations were measured as an indicator of the required effort in performing the same pitch discrimination task for conditions of varying resolvability and task difficulty. Pupillometry responses indicated a lower processing effort in the musicians versus the non-musicians, although the processing demand imposed by the pitch discrimination task was individually adjusted according to the behavioral thresholds. Overall, these findings indicate that the enhanced pitch discrimination abilities in musicians are unlikely to be related to higher peripheral frequency selectivity and may suggest an enhanced pitch representation at more central stages of the auditory system in musically trained listeners.
Collapse
Affiliation(s)
- Federica Bianchi
- Hearing Systems group, Technical University of Denmark, Ørsteds Plads Building 352, 2800, Kongens Lyngby, Denmark.
| | - Sébastien Santurette
- Hearing Systems group, Technical University of Denmark, Ørsteds Plads Building 352, 2800, Kongens Lyngby, Denmark
| | - Dorothea Wendt
- Hearing Systems group, Technical University of Denmark, Ørsteds Plads Building 352, 2800, Kongens Lyngby, Denmark
| | - Torsten Dau
- Hearing Systems group, Technical University of Denmark, Ørsteds Plads Building 352, 2800, Kongens Lyngby, Denmark
| |
Collapse
|
32
|
Wilsch A, Obleser J. What works in auditory working memory? A neural oscillations perspective. Brain Res 2015; 1640:193-207. [PMID: 26556773 DOI: 10.1016/j.brainres.2015.10.054] [Citation(s) in RCA: 35] [Impact Index Per Article: 3.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2015] [Revised: 10/28/2015] [Accepted: 10/30/2015] [Indexed: 11/16/2022]
Abstract
Working memory is a limited resource: brains can only maintain small amounts of sensory input (memory load) over a brief period of time (memory decay). The dynamics of slow neural oscillations as recorded using magneto- and electroencephalography (M/EEG) provide a window into the neural mechanics of these limitations. Especially oscillations in the alpha range (8-13Hz) are a sensitive marker for memory load. Moreover, according to current models, the resultant working memory load is determined by the relative noise in the neural representation of maintained information. The auditory domain allows memory researchers to apply and test the concept of noise quite literally: Employing degraded stimulus acoustics increases memory load and, at the same time, allows assessing the cognitive resources required to process speech in noise in an ecologically valid and clinically relevant way. The present review first summarizes recent findings on neural oscillations, especially alpha power, and how they reflect memory load and memory decay in auditory working memory. The focus is specifically on memory load resulting from acoustic degradation. These findings are then contrasted with contextual factors that benefit neural as well as behavioral markers of memory performance, by reducing representational noise. We end on discussing the functional role of alpha power in auditory working memory and suggest extensions of the current methodological toolkit. This article is part of a Special Issue entitled SI: Auditory working memory.
Collapse
Affiliation(s)
- Anna Wilsch
- Max Planck Research Group "Auditory Cognition", Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.
| | - Jonas Obleser
- Max Planck Research Group "Auditory Cognition", Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany; Department of Psychology, University of Lübeck, Ratzeburger Allee 160, 23562 Lübeck, Germany.
| |
Collapse
|
33
|
Wang S, Dong R, Liu D, Wang Y, Mao Y, Zhang H, Zhang L, Xu L. Perceptual separation of sensorineural hearing loss and auditory neuropathy spectrum disorder. Laryngoscope 2015; 126:1420-5. [PMID: 26344821 DOI: 10.1002/lary.25595] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2015] [Accepted: 07/24/2015] [Indexed: 11/12/2022]
Abstract
OBJECTIVES/HYPOTHESIS The present study aimed to examine whether the response patterns to the chimeric lexical tone tokens, combined with their pure tone audiometry (PTA) results, could separate listeners with sensorineural hearing loss (SNHL) from listeners with auditory neuropathy spectrum disorder (ANSD). STUDY DESIGN Case-control study. METHODS Forty-three SNHL subjects and 46 ANSD subjects participated in a Mandarin lexical tone perception test using original and chimeric tone tokens. Ten sets of monosyllables, with four tone patterns for each, were processed through a 16-channel chimeric synthesizer in which a temporal envelope (E) from a monosyllabic word of one tone was paired with a fine structure (FS) from the same monosyllable of other tones. RESULTS Significantly negative correlations were present between tone perception scores and PTA0.5-4 kHz for both SNHL (P < 0.001) and ANSD (P < 0.001) subjects. Overall, 72.4%, 66.4%, and 46.3% of the tone responses were consistent with FS for the SNHL subjects with mild, moderate, and severe degree of hearing loss, respectively; and 28.4%, 23.1%, and 22.7% were consistent with FS for the ANSD subjects, with the equivalent degree of hearing loss. Similarly, 17.6%, 24.2%, and 37.7% of the tone responses were consistent with E for the SNHL subjects with mild, moderate, and severe degree of hearing loss, respectively; and 45.5%, 44.3%, and 36.5% were consistent with E for the ANSD subjects. CONCLUSION Subjects with SNHL and ANSD may be separated by representing their FS- and E-consistent tone responses as a function of their pure-tone hearing thresholds. LEVEL OF EVIDENCE 3b. Laryngoscope, 126:1420-1425, 2016.
Collapse
Affiliation(s)
- Shuo Wang
- Otolaryngology-Head and Neck Surgery, Beijing Tongren Hospital, Beijing Institute of Otolaryngology, Capital Medical University, Beijing, People's Republic of China
| | - Ruijuan Dong
- Otolaryngology-Head and Neck Surgery, Beijing Tongren Hospital, Beijing Institute of Otolaryngology, Capital Medical University, Beijing, People's Republic of China
| | - Dongxin Liu
- Otolaryngology-Head and Neck Surgery, Beijing Tongren Hospital, Beijing Institute of Otolaryngology, Capital Medical University, Beijing, People's Republic of China
| | - Yuan Wang
- Otolaryngology-Head and Neck Surgery, Beijing Tongren Hospital, Beijing Institute of Otolaryngology, Capital Medical University, Beijing, People's Republic of China
| | - Yitao Mao
- Communication Sciences and Disorders, Ohio University, Athens, Ohio, U.S.A
| | - Hua Zhang
- Otolaryngology-Head and Neck Surgery, Beijing Tongren Hospital, Beijing Institute of Otolaryngology, Capital Medical University, Beijing, People's Republic of China
| | - Luo Zhang
- Otolaryngology-Head and Neck Surgery, Beijing Tongren Hospital, Beijing Institute of Otolaryngology, Capital Medical University, Beijing, People's Republic of China
| | - Li Xu
- Communication Sciences and Disorders, Ohio University, Athens, Ohio, U.S.A
| |
Collapse
|
34
|
Schvartz-Leyzac KC, Chatterjee M. Fundamental-frequency discrimination using noise-band-vocoded harmonic complexes in older listeners with normal hearing. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 138:1687-1695. [PMID: 26428806 PMCID: PMC4592424 DOI: 10.1121/1.4929938] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/17/2014] [Revised: 08/06/2015] [Accepted: 08/20/2015] [Indexed: 06/04/2023]
Abstract
Voice-pitch cues provide detailed information about a talker that help a listener to understand speech in complex environments. Temporal-envelope based voice-pitch coding is important for listeners with hearing impairment, especially listeners with cochlear implants, as spectral resolution is not sufficient to provide a spectrally based voice-pitch cue. The effect of aging on the ability to glean voice-pitch information using temporal envelope cues is not completely understood. The current study measured fundamental frequency (f0) discrimination limens in normal-hearing younger and older adults while listening to noise-band vocoded harmonic complexes with varying numbers of spectral channels. Age-related disparities in performance were apparent across all conditions, independent of spectral degradation and/or fundamental frequency. The findings have important implications for older listeners with normal hearing and hearing loss, who may be inherently limited in their ability to perceive f0 cues due to senescent decline in auditory function.
Collapse
Affiliation(s)
- Kara C Schvartz-Leyzac
- Department of Hearing and Speech Sciences, University of Maryland, 0100 LeFrak Hall, College Park, Maryland 20742, USA
| | - Monita Chatterjee
- Department of Hearing and Speech Sciences, University of Maryland, 0100 LeFrak Hall, College Park, Maryland 20742, USA
| |
Collapse
|
35
|
Pham CQ, Bremen P, Shen W, Yang SM, Middlebrooks JC, Zeng FG, Mc Laughlin M. Central Auditory Processing of Temporal and Spectral-Variance Cues in Cochlear Implant Listeners. PLoS One 2015; 10:e0132423. [PMID: 26176553 PMCID: PMC4503639 DOI: 10.1371/journal.pone.0132423] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2014] [Accepted: 06/13/2015] [Indexed: 11/25/2022] Open
Abstract
Cochlear implant (CI) listeners have difficulty understanding speech in complex listening environments. This deficit is thought to be largely due to peripheral encoding problems arising from current spread, which results in wide peripheral filters. In normal hearing (NH) listeners, central processing contributes to segregation of speech from competing sounds. We tested the hypothesis that basic central processing abilities are retained in post-lingually deaf CI listeners, but processing is hampered by degraded input from the periphery. In eight CI listeners, we measured auditory nerve compound action potentials to characterize peripheral filters. Then, we measured psychophysical detection thresholds in the presence of multi-electrode maskers placed either inside (peripheral masking) or outside (central masking) the peripheral filter. This was intended to distinguish peripheral from central contributions to signal detection. Introduction of temporal asynchrony between the signal and masker improved signal detection in both peripheral and central masking conditions for all CI listeners. Randomly varying components of the masker created spectral-variance cues, which seemed to benefit only two out of eight CI listeners. Contrastingly, the spectral-variance cues improved signal detection in all five NH listeners who listened to our CI simulation. Together these results indicate that widened peripheral filters significantly hamper central processing of spectral-variance cues but not of temporal cues in post-lingually deaf CI listeners. As indicated by two CI listeners in our study, however, post-lingually deaf CI listeners may retain some central processing abilities similar to NH listeners.
Collapse
Affiliation(s)
- Carol Q. Pham
- Center for Hearing Research, University of California Irvine, Irvine, California, United States of America
- Department of Anatomy and Neurobiology, University of California Irvine, Irvine, California, United States of America
- * E-mail:
| | - Peter Bremen
- Center for Hearing Research, University of California Irvine, Irvine, California, United States of America
- Department of Otolaryngology- Head and Neck Surgery, University of California Irvine, Irvine, California, United States of America
| | - Weidong Shen
- Institute of Otolaryngology, Chinese PLA Genera Hospital, Beijing, China
| | - Shi-Ming Yang
- Institute of Otolaryngology, Chinese PLA Genera Hospital, Beijing, China
| | - John C. Middlebrooks
- Center for Hearing Research, University of California Irvine, Irvine, California, United States of America
- Department of Otolaryngology- Head and Neck Surgery, University of California Irvine, Irvine, California, United States of America
- Department of Neurobiology and Behavior, University of California Irvine, Irvine, California, United States of America
- Department of Biomedical Engineering, University of California Irvine, Irvine, California, United States of America
- Department of Cognitive Sciences, University of California Irvine, Irvine, California, United States of America
| | - Fan-Gang Zeng
- Center for Hearing Research, University of California Irvine, Irvine, California, United States of America
- Department of Anatomy and Neurobiology, University of California Irvine, Irvine, California, United States of America
- Department of Otolaryngology- Head and Neck Surgery, University of California Irvine, Irvine, California, United States of America
- Department of Biomedical Engineering, University of California Irvine, Irvine, California, United States of America
- Department of Cognitive Sciences, University of California Irvine, Irvine, California, United States of America
| | - Myles Mc Laughlin
- Center for Hearing Research, University of California Irvine, Irvine, California, United States of America
- Department of Otolaryngology- Head and Neck Surgery, University of California Irvine, Irvine, California, United States of America
| |
Collapse
|
36
|
The Role of Temporal Envelope and Fine Structure in Mandarin Lexical Tone Perception in Auditory Neuropathy Spectrum Disorder. PLoS One 2015; 10:e0129710. [PMID: 26052707 PMCID: PMC4459992 DOI: 10.1371/journal.pone.0129710] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2015] [Accepted: 05/12/2015] [Indexed: 11/19/2022] Open
Abstract
Temporal information in a signal can be partitioned into temporal envelope (E) and fine structure (FS). Fine structure is important for lexical tone perception for normal-hearing (NH) listeners, and listeners with sensorineural hearing loss (SNHL) have an impaired ability to use FS in lexical tone perception due to the reduced frequency resolution. The present study was aimed to assess which of the acoustic aspects (E or FS) played a more important role in lexical tone perception in subjects with auditory neuropathy spectrum disorder (ANSD) and to determine whether it was the deficit in temporal resolution or frequency resolution that might lead to more detrimental effects on FS processing in pitch perception. Fifty-eight native Mandarin Chinese-speaking subjects (27 with ANSD, 16 with SNHL, and 15 with NH) were assessed for (1) their ability to recognize lexical tones using acoustic E or FS cues with the “auditory chimera” technique, (2) temporal resolution as measured with temporal gap detection (TGD) threshold, and (3) frequency resolution as measured with the Q10dB values of the psychophysical tuning curves. Overall, 26.5%, 60.2%, and 92.1% of lexical tone responses were consistent with FS cues for tone perception for listeners with ANSD, SNHL, and NH, respectively. The mean TGD threshold was significantly higher for listeners with ANSD (11.9 ms) than for SNHL (4.0 ms; p < 0.001) and NH (3.9 ms; p < 0.001) listeners, with no significant difference between SNHL and NH listeners. In contrast, the mean Q10dB for listeners with SNHL (1.8±0.4) was significantly lower than that for ANSD (3.5±1.0; p < 0.001) and NH (3.4±0.9; p < 0.001) listeners, with no significant difference between ANSD and NH listeners. These results suggest that reduced temporal resolution, as opposed to reduced frequency selectivity, in ANSD subjects leads to greater degradation of FS processing for pitch perception.
Collapse
|
37
|
Woods DL, Arbogast T, Doss Z, Younus M, Herron TJ, Yund EW. Aided and unaided speech perception by older hearing impaired listeners. PLoS One 2015; 10:e0114922. [PMID: 25730423 PMCID: PMC4346396 DOI: 10.1371/journal.pone.0114922] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2014] [Accepted: 11/14/2014] [Indexed: 12/11/2022] Open
Abstract
The most common complaint of older hearing impaired (OHI) listeners is difficulty understanding speech in the presence of noise. However, tests of consonant-identification and sentence reception threshold (SeRT) provide different perspectives on the magnitude of impairment. Here we quantified speech perception difficulties in 24 OHI listeners in unaided and aided conditions by analyzing (1) consonant-identification thresholds and consonant confusions for 20 onset and 20 coda consonants in consonant-vowel-consonant (CVC) syllables presented at consonant-specific signal-to-noise (SNR) levels, and (2) SeRTs obtained with the Quick Speech in Noise Test (QSIN) and the Hearing in Noise Test (HINT). Compared to older normal hearing (ONH) listeners, nearly all unaided OHI listeners showed abnormal consonant-identification thresholds, abnormal consonant confusions, and reduced psychometric function slopes. Average elevations in consonant-identification thresholds exceeded 35 dB, correlated strongly with impairments in mid-frequency hearing, and were greater for hard-to-identify consonants. Advanced digital hearing aids (HAs) improved average consonant-identification thresholds by more than 17 dB, with significant HA benefit seen in 83% of OHI listeners. HAs partially normalized consonant-identification thresholds, reduced abnormal consonant confusions, and increased the slope of psychometric functions. Unaided OHI listeners showed much smaller elevations in SeRTs (mean 6.9 dB) than in consonant-identification thresholds and SeRTs in unaided listening conditions correlated strongly (r = 0.91) with identification thresholds of easily identified consonants. HAs produced minimal SeRT benefit (2.0 dB), with only 38% of OHI listeners showing significant improvement. HA benefit on SeRTs was accurately predicted (r = 0.86) by HA benefit on easily identified consonants. Consonant-identification tests can accurately predict sentence processing deficits and HA benefit in OHI listeners.
Collapse
Affiliation(s)
- David L. Woods
- Human Cognitive Neurophysiology Laboratory, VANCHCS, 150 Muir Rd., Martinez, California, 95553, United States of America
- UC Davis Department of Neurology, 4860 Y St., Suite 3700, Sacramento, California, 95817, United States of America
- Center for Neurosciences, UC Davis, 1544 Newton Ct., Davis, California, 95616, United States of America
- UC Davis Center for Mind and Brain, 202 Cousteau Place, Suite 201, Davis, California, 95616, United States of America
- * E-mail:
| | - Tanya Arbogast
- Human Cognitive Neurophysiology Laboratory, VANCHCS, 150 Muir Rd., Martinez, California, 95553, United States of America
| | - Zoe Doss
- Human Cognitive Neurophysiology Laboratory, VANCHCS, 150 Muir Rd., Martinez, California, 95553, United States of America
| | - Masood Younus
- Human Cognitive Neurophysiology Laboratory, VANCHCS, 150 Muir Rd., Martinez, California, 95553, United States of America
| | - Timothy J. Herron
- Human Cognitive Neurophysiology Laboratory, VANCHCS, 150 Muir Rd., Martinez, California, 95553, United States of America
| | - E. William Yund
- Human Cognitive Neurophysiology Laboratory, VANCHCS, 150 Muir Rd., Martinez, California, 95553, United States of America
| |
Collapse
|
38
|
Abstract
Sensorineural hearing loss is the most common type of hearing impairment worldwide. It arises as a consequence of damage to the cochlea or auditory nerve, and several structures are often affected simultaneously. There are many causes, including genetic mutations affecting the structures of the inner ear, and environmental insults such as noise, ototoxic substances, and hypoxia. The prevalence increases dramatically with age. Clinical diagnosis is most commonly accomplished by measuring detection thresholds and comparing these to normative values to determine the degree of hearing loss. In addition to causing insensitivity to weak sounds, sensorineural hearing loss has a number of adverse perceptual consequences, including loudness recruitment, poor perception of pitch and auditory space, and difficulty understanding speech, particularly in the presence of background noise. The condition is usually incurable; treatment focuses on restoring the audibility of sounds made inaudible by hearing loss using either hearing aids or cochlear implants.
Collapse
Affiliation(s)
- Kathryn Hopkins
- School of Psychological Sciences, University of Manchester, Manchester, UK.
| |
Collapse
|
39
|
Cousineau M, Oxenham AJ, Peretz I. Congenital amusia: a cognitive disorder limited to resolved harmonics and with no peripheral basis. Neuropsychologia 2015; 66:293-301. [PMID: 25433224 PMCID: PMC4300951 DOI: 10.1016/j.neuropsychologia.2014.11.031] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2014] [Revised: 11/01/2014] [Accepted: 11/25/2014] [Indexed: 11/17/2022]
Abstract
Pitch plays a fundamental role in audition, from speech and music perception to auditory scene analysis. Congenital amusia is a neurogenetic disorder that appears to affect primarily pitch and melody perception. Pitch is normally conveyed by the spectro-temporal fine structure of low harmonics, but some pitch information is available in the temporal envelope produced by the interactions of higher harmonics. Using 10 amusic subjects and 10 matched controls, we tested the hypothesis that amusics suffer exclusively from impaired processing of spectro-temporal fine structure. We also tested whether the inability of amusics to process acoustic temporal fine structure extends beyond pitch by measuring sensitivity to interaural time differences, which also rely on temporal fine structure. Further tests were carried out on basic intensity and spectral resolution. As expected, pitch perception based on spectro-temporal fine structure was impaired in amusics; however, no significant deficits were observed in amusics' ability to perceive the pitch conveyed via temporal-envelope cues. Sensitivity to interaural time differences was also not significantly different between the amusic and control groups, ruling out deficits in the peripheral coding of temporal fine structure. Finally, no significant differences in intensity or spectral resolution were found between the amusic and control groups. The results demonstrate a pitch-specific deficit in fine spectro-temporal information processing in amusia that seems unrelated to temporal or spectral coding in the auditory periphery. These results are consistent with the view that there are distinct mechanisms dedicated to processing resolved and unresolved harmonics in the general population, the former being altered in congenital amusia while the latter is spared.
Collapse
Affiliation(s)
- Marion Cousineau
- International Laboratory for Brain, Music and Sound Research (BRAMS) and Center for Research on Brain, Language and Music (CRBLM), Department of Psychology, University of Montreal, C.P. 6128, succursale Centre-ville, Montreal, QC, Canada H3C 3J7.
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, MN 55455, USA
| | - Isabelle Peretz
- International Laboratory for Brain, Music and Sound Research (BRAMS) and Center for Research on Brain, Language and Music (CRBLM), Department of Psychology, University of Montreal, C.P. 6128, succursale Centre-ville, Montreal, QC, Canada H3C 3J7
| |
Collapse
|
40
|
Heinrich A, de la Rosa S, Schneider BA. The role of stimulus complexity, spectral overlap, and pitch for gap-detection thresholds in young and old listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2014; 136:1797-1807. [PMID: 25324081 DOI: 10.1121/1.4894788] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
Thresholds for detecting a gap between two complex tones were determined for young listeners with normal hearing and old listeners with mild age-related hearing loss. The leading tonal marker was always a 20-ms, 250-Hz complex tone with energy at 250, 500, 750, and 1000 Hz. The lagging marker, also tonal, could differ from the leading marker with respect to fundamental frequency (f0), the presence versus absence of energy at f0, and the degree to which it overlapped spectrally with the leading marker. All stimuli were presented with steeper (1 ms) and less steep (4 ms) envelope rise and fall times. F0 differences, decreases in the degree of spectral overlap between the markers, and shallower envelope shape all contributed to increases in gap-detection thresholds. Age differences for gap detection of complex sounds were generally small and constant when gap-detection thresholds were measured on a log scale. When comparing the results for complex sounds to thresholds obtained for pure-tones in a previous study by Heinrich and Schneider [(2006). J. Acoust. Soc. Am. 119, 2316-2326], thresholds increased in an orderly fashion from markers with identical (within-channel) pure tones to different (between-channel) pure tones to complex sounds. This pattern of results was true for listeners of both ages although younger listeners had smaller thresholds overall.
Collapse
Affiliation(s)
- A Heinrich
- Human Communication Laboratory, Department of Psychology, University of Toronto Mississauga, 3359 Mississauga Road North, Mississauga, Ontario L5L 1C6, Canada
| | - S de la Rosa
- Human Communication Laboratory, Department of Psychology, University of Toronto Mississauga, 3359 Mississauga Road North, Mississauga, Ontario L5L 1C6, Canada
| | - B A Schneider
- Human Communication Laboratory, Department of Psychology, University of Toronto Mississauga, 3359 Mississauga Road North, Mississauga, Ontario L5L 1C6, Canada
| |
Collapse
|
41
|
Bidelman GM, Syed Khaja A. Spectrotemporal resolution tradeoff in auditory processing as revealed by human auditory brainstem responses and psychophysical indices. Neurosci Lett 2014; 572:53-7. [PMID: 24793771 DOI: 10.1016/j.neulet.2014.04.037] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2014] [Revised: 04/20/2014] [Accepted: 04/22/2014] [Indexed: 11/15/2022]
Abstract
Auditory filter theory dictates a physiological compromise between frequency and temporal resolution of cochlear signal processing. We examined neurophysiological correlates of these spectrotemporal tradeoffs in the human auditory system using auditory evoked brain potentials and psychophysical responses. Temporal resolution was assessed using scalp-recorded auditory brainstem responses (ABRs) elicited by paired clicks. The inter-click interval (ICI) between successive pulses was parameterized from 0.7 to 25 ms to map ABR amplitude recovery as a function of stimulus spacing. Behavioral frequency difference limens (FDLs) and auditory filter selectivity (Q10 of psychophysical tuning curves) were obtained to assess relations between behavioral spectral acuity and electrophysiological estimates of temporal resolvability. Neural responses increased monotonically in amplitude with increasing ICI, ranging from total suppression (0.7 ms) to full recovery (25 ms) with a temporal resolution of ∼3-4 ms. ABR temporal thresholds were correlated with behavioral Q10 (frequency selectivity) but not FDLs (frequency discrimination); no correspondence was observed between Q10 and FDLs. Results suggest that finer frequency selectivity, but not discrimination, is associated with poorer temporal resolution. The inverse relation between ABR recovery and perceptual frequency tuning demonstrates a time-frequency tradeoff between the temporal and spectral resolving power of the human auditory system.
Collapse
Affiliation(s)
- Gavin M Bidelman
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA; School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA.
| | - Ameenuddin Syed Khaja
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA
| |
Collapse
|
42
|
Winn MB, Chatterjee M, Idsardi WJ. Roles of voice onset time and F0 in stop consonant voicing perception: effects of masking noise and low-pass filtering. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2013; 56:1097-107. [PMID: 23785185 PMCID: PMC3755127 DOI: 10.1044/1092-4388(2012/12-0086)] [Citation(s) in RCA: 31] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
PURPOSE The contributions of voice onset time (VOT) and fundamental frequency (F0) were evaluated for the perception of voicing in syllable-initial stop consonants in words that were low-pass filtered and/or masked by speech-shaped noise. It was expected that listeners would rely less on VOT and more on F0 in these degraded conditions. METHOD Twenty young listeners with normal hearing identified modified natural speech tokens that varied by VOT and F0 in several conditions of low-pass filtering and masking noise. Stimuli included /b/-/p/ and /d/-/t/ continua that were presented in separate blocks. Identification results were modeled using mixed-effects logistic regression. RESULTS When speech was filtered and/or masked by noise, listeners' voicing perceptions were driven less by VOT and more by F0. Speech-shaped masking noise exerted greater effects on the /b/-/p/ contrast, while low-pass filtering exerted greater effects on the /d/-/t/ contrast, consistent with the acoustics of these contrasts. CONCLUSION Listeners can adjust their use of acoustic-phonetic cues in a dynamic way that is appropriate for challenging listening conditions; cues that are less influential in ideal conditions can gain priority in challenging conditions.
Collapse
|
43
|
Meister H, Schreitmüller S, Grugel L, Ortmann M, Beutner D, Walger M, Meister I. Cognitive resources related to speech recognition with a competing talker in young and older listeners. Neuroscience 2013; 232:74-82. [DOI: 10.1016/j.neuroscience.2012.12.006] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2012] [Revised: 11/28/2012] [Accepted: 12/05/2012] [Indexed: 10/27/2022]
|
44
|
Impaired Timing and Frequency Discrimination in High-functioning Autism Spectrum Disorders. J Autism Dev Disord 2013; 43:2312-28. [DOI: 10.1007/s10803-013-1778-y] [Citation(s) in RCA: 27] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/08/2023]
|
45
|
Abstract
Pitch is one of the primary auditory sensations and plays a defining role in music, speech, and auditory scene analysis. Although the main physical correlate of pitch is acoustic periodicity, or repetition rate, there are many interactions that complicate the relationship between the physical stimulus and the perception of pitch. In particular, the effects of other acoustic parameters on pitch judgments, and the complex interactions between perceptual organization and pitch, have uncovered interesting perceptual phenomena that should help to reveal the underlying neural mechanisms.
Collapse
|
46
|
Tabuchi H, Borucki E, Berg BG. Effects of randomizing phase on the discrimination between amplitude-modulated and quasi-frequency-modulated tones. Hear Res 2012; 290:72-82. [PMID: 22609773 DOI: 10.1016/j.heares.2012.04.021] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/26/2011] [Revised: 04/10/2012] [Accepted: 04/23/2012] [Indexed: 10/28/2022]
Abstract
This study investigated the bandwidth of phase sensitivity. Subjects discriminated amplitude-modulated tones (AM), and quasi-frequency-modulated tones (QFM) in a two-interval, forced-choice task. An adaptive threshold procedure was used to estimate the modulation depth needed to discriminate the stimuli as a function of carrier and modulation frequency. Non-monotonicities in threshold-bandwidth functions were often observed at higher modulation frequencies. The results are discussed in terms of two potential cues: (1) waveform envelope, (2) cubic distortion products. In order to degrade the information obtained from auditory distortions, the phase for the carrier frequency was randomly sampled from a uniform distribution, which diminished the non-monotonicities with minimal effect at lower modulation frequencies. Model simulations demonstrated that phase randomization degrades distortion product cues with only a modest effect on temporal cues. Final results show that maximum bandwidths for phase sensitivity (BW(max)) were not proportional to carrier frequencies.
Collapse
Affiliation(s)
- Hisaaki Tabuchi
- Department of Cognitive Sciences, University of California, 2201 Social & Behavioral Sciences, Gateway Building, Irvine, CA 92697-5100, USA
| | | | | |
Collapse
|
47
|
Desloge JG, Reed CM, Braida LD, Perez ZD, Delhorne LA. Auditory-filter characteristics for listeners with real and simulated hearing impairment. Trends Amplif 2012; 16:19-39. [PMID: 22593204 PMCID: PMC4040846 DOI: 10.1177/1084713812445510] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Functional simulation of sensorineural hearing impairment is an important research tool that can elucidate the nature of hearing impairments and suggest or eliminate compensatory signal-processing schemes. The objective of the current study was to evaluate the capability of an audibility-based functional simulation of hearing loss to reproduce the auditory-filter characteristics of listeners with sensorineural hearing loss. The hearing-loss simulation used either threshold-elevating noise alone or a combination of threshold-elevating noise and multiband expansion to reproduce the audibility-based characteristics of the loss (including detection thresholds, dynamic range, and loudness recruitment). The hearing losses of 10 listeners with bilateral, mild-to-severe hearing loss were simulated in 10 corresponding groups of 3 age-matched normal-hearing listeners. Frequency selectivity was measured using a notched-noise masking paradigm at five probe frequencies in the range of 250 to 4000 Hz with a fixed probe level of either 70 dB SPL or 8 dB SL (whichever was greater) and probe duration of 200 ms. The hearing-loss simulation reproduced the absolute thresholds of individual hearing-impaired listeners with an average root-mean-squared (RMS) difference of 2.2 dB and the notched-noise masked thresholds with an RMS difference of 5.6 dB. A rounded-exponential model of the notched-noise data was used to estimate equivalent rectangular bandwidths and slopes of the auditory filters. For some subjects and probe frequencies, the simulations were accurate in reproducing the auditory-filter characteristics of the hearing-impaired listeners. In other cases, however, the simulations underestimated the magnitude of the auditory bandwidths for the hearing-impaired listeners, which suggests the possibility of suprathreshold deficits.
Collapse
Affiliation(s)
- Joseph G Desloge
- Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, MA 02139, USA.
| | | | | | | | | |
Collapse
|
48
|
Moore BCJ, Glasberg BR. The effect of hearing loss on the resolution of partials and fundamental frequency discrimination. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2011; 130:2891-2901. [PMID: 22087918 DOI: 10.1121/1.3640852] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
The relationship between the ability to hear out partials in complex tones, discrimination of the fundamental frequency (F0) of complex tones, and frequency selectivity was examined for subjects with mild-to-moderate cochlear hearing loss. The ability to hear out partials was measured using a two-interval task. Each interval included a sinusoid followed by a complex tone; one complex contained a partial with the same frequency as the sinusoid, whereas in the other complex that partial was missing. Subjects had to indicate the interval in which the partial was present in the complex. The components in the complex were uniformly spaced on the ERB(N)-number scale. Performance was generally good for the two "edge" partials, but poorer for the inner partials. Performance for the latter improved with increasing spacing. F0 discrimination was measured for a bandpass-filtered complex tone containing low harmonics. The equivalent rectangular bandwidth (ERB) of the auditory filter was estimated using the notched-noise method for center frequencies of 0.5, 1, and 2 kHz. Significant correlations were found between the ability to hear out inner partials, F0 discrimination, and the ERB. The results support the idea that F0 discrimination of tones with low harmonics depends on the ability to resolve the harmonics.
Collapse
Affiliation(s)
- Brian C J Moore
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, United Kingdom.
| | | |
Collapse
|
49
|
Wang S, Xu L, Mannell R. Relative contributions of temporal envelope and fine structure cues to lexical tone recognition in hearing-impaired listeners. J Assoc Res Otolaryngol 2011; 12:783-94. [PMID: 21833816 DOI: 10.1007/s10162-011-0285-0] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2010] [Accepted: 07/25/2011] [Indexed: 11/24/2022] Open
Abstract
It has been reported that normal-hearing Chinese speakers base their lexical tone recognition on fine structure regardless of temporal envelope cues. However, a few psychoacoustic and perceptual studies have demonstrated that listeners with sensorineural hearing impairment may have an impaired ability to use fine structure information, whereas their ability to use temporal envelope information is close to normal. The purpose of this study is to investigate the relative contributions of temporal envelope and fine structure cues to lexical tone recognition in normal-hearing and hearing-impaired native Mandarin Chinese speakers. Twenty-two normal-hearing subjects and 31 subjects with various degrees of sensorineural hearing loss participated in the study. Sixteen sets of Mandarin monosyllables with four tone patterns for each were processed through a "chimeric synthesizer" in which temporal envelope from a monosyllabic word of one tone was paired with fine structure from the same monosyllable of other tones. The chimeric tokens were generated in the three channel conditions (4, 8, and 16 channels). Results showed that differences in tone responses among the three channel conditions were minor. On average, 90.9%, 70.9%, 57.5%, and 38.2% of tone responses were consistent with fine structure for normal-hearing, moderate, moderate to severe, and severely hearing-impaired groups respectively, whereas 6.8%, 21.1%, 31.4%, and 44.7% of tone responses were consistent with temporal envelope cues for the above-mentioned groups. Tone responses that were consistent neither with temporal envelope nor fine structure had averages of 2.3%, 8.0%, 11.1%, and 17.1% for the above-mentioned groups of subjects. Pure-tone average thresholds were negatively correlated with tone responses that were consistent with fine structure, but were positively correlated with tone responses that were based on the temporal envelope cues. Consistent with the idea that the spectral resolvability is responsible for fine structure coding, these results demonstrated that, as hearing loss becomes more severe, lexical tone recognition relies increasingly on temporal envelope rather than fine structure cues due to the widened auditory filters.
Collapse
Affiliation(s)
- Shuo Wang
- Beijing Institute of Otolaryngology, Key Laboratory of Otolaryngology Head and Neck Surgery (Capital Medical University), Ministry of Education, Beijing Tongren Hospital, Capital Medical University, Beijing, People's Republic of China
| | | | | |
Collapse
|
50
|
Mackersie CL, Dewey J, Guthrie LA. Effects of fundamental frequency and vocal-tract length cues on sentence segregation by listeners with hearing loss. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2011; 130:1006-19. [PMID: 21877813 PMCID: PMC3190663 DOI: 10.1121/1.3605548] [Citation(s) in RCA: 30] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/15/2010] [Revised: 05/22/2011] [Accepted: 06/01/2011] [Indexed: 05/20/2023]
Abstract
The purpose was to determine the effect of hearing loss on the ability to separate competing talkers using talker differences in fundamental frequency (F0) and apparent vocal-tract length (VTL). Performance of 13 adults with hearing loss and 6 adults with normal hearing was measured using the Coordinate Response Measure. For listeners with hearing loss, the speech was amplified and filtered according to the NAL-RP hearing aid prescription. Target-to-competition ratios varied from 0 to 9 dB. The target sentence was randomly assigned to the higher or lower values of F0 or VTL on each trial. Performance improved for F0 differences up to 9 and 6 semitones for people with normal hearing and hearing loss, respectively, but only when the target talker had the higher F0. Recognition for the lower F0 target improved when trial-to-trial uncertainty was removed (9-semitone condition). Scores improved with increasing differences in VTL for the normal-hearing group. On average, hearing-impaired listeners did not benefit from VTL cues, but substantial inter-subject variability was observed. The amount of benefit from VTL cues was related to the average hearing loss in the 1-3-kHz region when the target talker had the shorter VTL.
Collapse
|