1
|
Cychosz M, Edwards JR, Munson B, Romeo R, Kosie J, Newman RS. The everyday speech environments of preschoolers with and without cochlear implants. JOURNAL OF CHILD LANGUAGE 2025; 52:377-398. [PMID: 38362892 PMCID: PMC11327381 DOI: 10.1017/s0305000924000023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/17/2024]
Abstract
Children who receive cochlear implants develop spoken language on a protracted timescale. The home environment facilitates speech-language development, yet it is relatively unknown how the environment differs between children with cochlear implants and typical hearing. We matched eighteen preschoolers with implants (31-65 months) to two groups of children with typical hearing: by chronological age and hearing age. Each child completed a long-form, naturalistic audio recording of their home environment (appx. 16 hours/child; >730 hours of observation) to measure adult speech input, child vocal productivity, and caregiver-child interaction. Results showed that children with cochlear implants and typical hearing were exposed to and engaged in similar amounts of spoken language with caregivers. However, the home environment did not reflect developmental stages as closely for children with implants, or predict their speech outcomes as strongly. Home-based speech-language interventions should focus on the unique input-outcome relationships for this group of children with hearing loss.
Collapse
|
2
|
Meijer A, Benard MR, Woonink A, Başkent D, Dirks E. The Auditory Environment at Early Intervention Groups for Young Children With Hearing Loss: Signal to Noise Ratio, Background Noise, and Reverberation. Ear Hear 2025:00003446-990000000-00385. [PMID: 39789707 DOI: 10.1097/aud.0000000000001627] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/12/2025]
Abstract
OBJECTIVES One important aspect in facilitating language access for children with hearing loss (HL) is the auditory environment. An optimal auditory environment is characterized by high signal to noise ratios (SNRs), low background noise levels, and low reverberation times. In this study, the authors describe the auditory environment of early intervention groups specifically equipped for young children with HL. DESIGN Seven early intervention groups for children with HL were included in the study. A total of 26 young children (22 to 46 months) visiting those groups participated. Language Environmental Analysis recorders were used to record all sounds around a child during one group visit. The recordings were analyzed to estimate SNR levels and background noise levels during the intervention groups. The unoccupied noise levels and reverberation times were measured in the unoccupied room either directly before or after the group visit. RESULTS The average SNR encountered by the children in the intervention groups was +13 dB SNR. The detected speech of the attending professionals achieved the +15 dB SNR recommended by the American Speech-Language-Hearing Association in approximately 42% of the time. The unoccupied noise levels were between 29 and 39 dBA, complying with acoustic norms for classroom environments (≤35 dBA, by ANSI/ASA 12.60-2010 Part 1) for six out of seven groups. Reverberation time was between 0.3 and 0.6 sec for all groups, which complies to the acoustic norms for classroom environments for children without HL (0.6 or 0.7 sec, depending on the room size), while only one group complied to the stricter norm for children with HL (0.3 sec). CONCLUSIONS The current findings show characteristics of the auditory environment of a setting that is specifically equipped and designed for groups of children with HL. Maintaining favorable SNRs seems to be the largest challenge to achieve within the constraints of an environment where young children gather, play, and learn. The results underscore the importance of staying attentive to keep spoken language accessible for children with HL in a group setting.
Collapse
Affiliation(s)
| | | | | | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, The Netherlands
| | - Evelien Dirks
- Dutch Foundation of the Deaf and Hard of Hearing Child (NSDSK), Amsterdam, The Netherlands
- Department Tranzo, Tilburg University, the Netherlands
| |
Collapse
|
3
|
Heffner CC, Al-Thowaini BM, Rytting CA. Arabic consonant length perception depends on the relative speech rate of the distal context. PLoS One 2024; 19:e0312249. [PMID: 39436869 PMCID: PMC11495589 DOI: 10.1371/journal.pone.0312249] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2023] [Accepted: 10/03/2024] [Indexed: 10/25/2024] Open
Abstract
Listeners adapt to the speech rate of talkers. Many studies of speech rate adaptation have focused on the influence of rate information on the perception of word segmentation or segmental perception in English. The effects of immediately adjacent (i.e., proximal) information are generally strong on the perception of both segments and segmentation, but the effects of rate information temporally remote from (i.e., distal to) ambiguous speech signals are less clear, especially for segments. The present study examines the influence of distal rate adaptation on the perception of geminate consonants in Arabic that straddle a morpheme boundary (i.e., heteromorphemic geminates). Participants heard sentences that at one point were ambiguous to the presence of the Arabic definite clitic al, which, under certain circumstances, can be realized as gemination of the subsequent word-initial consonant. The sentences were either recorded with or without the clitic and with three possible distal speech rates in the context of the clitic. They transcribed the sentences and reported what they heard, and those transcriptions were analyzed for the contribution of the original recording and the distal speech rate on the perception of al. It was found that the perception of geminates in Arabic is rate dependent. This extends the knowledge of the effects of distal rate cues to Arabic, showing that Arabic geminate consonants are perceived relative to the rate of the distal context.
Collapse
Affiliation(s)
- Christopher C. Heffner
- Program in Neuroscience and Cognitive Science, University of Maryland, College Park, Maryland, United States of America
- Department of Linguistics, University of Maryland, College Park, Maryland, United States of America
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland, United States of America
| | - Buthainah M. Al-Thowaini
- Department of English, College of Language Sciences, King Saud University, Riyadh, Kingdom of Saudi Arabia
| | - C. Anton Rytting
- Center for Advanced Study of Language, University of Maryland, College Park, Maryland, United States of America
| |
Collapse
|
4
|
Arras T, Rachman L, van Wieringen A, Başkent D. Perception of voice cues and speech-in-speech by children with prelingual single-sided deafness and a cochlear implant. Hear Res 2024; 454:109133. [PMID: 39546877 DOI: 10.1016/j.heares.2024.109133] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/16/2024] [Revised: 10/10/2024] [Accepted: 10/14/2024] [Indexed: 11/17/2024]
Abstract
Voice cues, such as fundamental frequency (F0) and vocal tract length (VTL), help listeners identify the speaker's gender, perceive the linguistic and emotional prosody, and segregate competing talkers. Postlingually implanted adult cochlear implant (CI) users seem to have difficulty in perceiving and making use of voice cues, especially of VTL. Early implanted child CI users, in contrast, perceive and make use of both voice cues better than CI adults, and in patterns similar to their peers with normal hearing (NH). In our study, we investigated the perception and use of voice cues in children with single-sided deafness (SSD) who received their CI at an early age (SSD+CI), in an attempt to bridge the gap between these two groups. The SSD+CI children have access to bilateral auditory information and often receive their CI at an early age, similar to CI children. They may also have dominant acoustic representations, similar to CI adults who acquired hearing loss at a later age. As such, the current study aimed to investigate the perception and use of voice cues by a group of nine early-implanted children with prelingual SSD. The study consisted of three experiments: F0 and VTL discrimination, voice gender categorization, and speech-in-speech perception. In each experiment, the results of the SSD group are compared to children and adults with CIs (for their CI ear) and with typical hearing (for their NH ear). Overall, the SSD+CI children had poorer VTL detection thresholds with their CI compared to their NH ear, while their F0 perception was similar across ears. Detection thresholds for both F0 and VTL with their CI ear was comparable to those of bilaterally implanted CI children, suggesting that SSD+CI children do not only rely on their NH ear, but actually make use of their CI. SSD+CI children relied more heavily on F0 cues than on VTL cues for voice gender categorization, with cue weighting patterns comparable to those of CI adults. In contrast to CI children, the SSD+CI children showed limited speech perception benefit based on F0 and VTL differences between the target and masker speaker, which again corresponded to the results of CI adults. Altogether, the SSD+CI children make good use of their CI, despite a good-hearing ear, however, the perceptual patterns seem to fall in-between those of CI children and CI adults. Perhaps a combination of childhood neuroplasticity, limited experience with relying only on the CI, and a dominant acoustic representation of voice gender explain these results.
Collapse
Affiliation(s)
- Tine Arras
- ExpORL, Dept. Neurosciences, KU Leuven, Belgium; Cochlear Technology Centre, Belgium
| | - Laura Rachman
- Dept. of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, The Netherlands; Research School of Behavioral and Cognitive Neuroscience, Graduate School of Medical Sciences, University of Groningen, The Netherlands; W.J. Kolff Institute for Biomedical Engineering and Materials Science, Graduate School of Medical Sciences, University of Groningen, The Netherlands
| | - Astrid van Wieringen
- ExpORL, Dept. Neurosciences, KU Leuven, Belgium; Dept. of Special Needs Education, University of Oslo, Norway
| | - Deniz Başkent
- Dept. of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, The Netherlands; Research School of Behavioral and Cognitive Neuroscience, Graduate School of Medical Sciences, University of Groningen, The Netherlands; W.J. Kolff Institute for Biomedical Engineering and Materials Science, Graduate School of Medical Sciences, University of Groningen, The Netherlands.
| |
Collapse
|
5
|
Isnard V, Chastres V, Andéol G. French version of the coordinate response measure corpus and its validation on a speech-on-speech task. JASA EXPRESS LETTERS 2024; 4:075203. [PMID: 39051871 DOI: 10.1121/10.0028059] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/01/2024] [Accepted: 07/08/2024] [Indexed: 07/27/2024]
Abstract
Since its creation, the coordinate response measure (CRM) corpus has been applied in hundreds of studies to explore the mechanisms of informational masking in multi-talker situations, but also in speech-in-noise or auditory attentional tasks. Here, we present its French version, with equivalent content to the original version in English. Furthermore, an evaluation of speech-on-speech intelligibility in French shows informational masking with similar result patterns to the original data in English. This validation of the French CRM corpus allows to propose the use of the CRM for intelligibility tests in French, and for comparisons with a foreign language under masking conditions.
Collapse
Affiliation(s)
- Vincent Isnard
- Institut de Recherche Biomédicale des Armées, Brétigny-sur-Orge, , ,
| | | | - Guillaume Andéol
- Institut de Recherche Biomédicale des Armées, Brétigny-sur-Orge, , ,
| |
Collapse
|
6
|
Nagels L, Gaudrain E, Vickers D, Hendriks P, Başkent D. Prelingually Deaf Children With Cochlear Implants Show Better Perception of Voice Cues and Speech in Competing Speech Than Postlingually Deaf Adults With Cochlear Implants. Ear Hear 2024; 45:952-968. [PMID: 38616318 PMCID: PMC11175806 DOI: 10.1097/aud.0000000000001489] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/29/2022] [Accepted: 01/10/2024] [Indexed: 04/16/2024]
Abstract
OBJECTIVES Postlingually deaf adults with cochlear implants (CIs) have difficulties with perceiving differences in speakers' voice characteristics and benefit little from voice differences for the perception of speech in competing speech. However, not much is known yet about the perception and use of voice characteristics in prelingually deaf implanted children with CIs. Unlike CI adults, most CI children became deaf during the acquisition of language. Extensive neuroplastic changes during childhood could make CI children better at using the available acoustic cues than CI adults, or the lack of exposure to a normal acoustic speech signal could make it more difficult for them to learn which acoustic cues they should attend to. This study aimed to examine to what degree CI children can perceive voice cues and benefit from voice differences for perceiving speech in competing speech, comparing their abilities to those of normal-hearing (NH) children and CI adults. DESIGN CI children's voice cue discrimination (experiment 1), voice gender categorization (experiment 2), and benefit from target-masker voice differences for perceiving speech in competing speech (experiment 3) were examined in three experiments. The main focus was on the perception of mean fundamental frequency (F0) and vocal-tract length (VTL), the primary acoustic cues related to speakers' anatomy and perceived voice characteristics, such as voice gender. RESULTS CI children's F0 and VTL discrimination thresholds indicated lower sensitivity to differences compared with their NH-age-equivalent peers, but their mean discrimination thresholds of 5.92 semitones (st) for F0 and 4.10 st for VTL indicated higher sensitivity than postlingually deaf CI adults with mean thresholds of 9.19 st for F0 and 7.19 st for VTL. Furthermore, CI children's perceptual weighting of F0 and VTL cues for voice gender categorization closely resembled that of their NH-age-equivalent peers, in contrast with CI adults. Finally, CI children had more difficulties in perceiving speech in competing speech than their NH-age-equivalent peers, but they performed better than CI adults. Unlike CI adults, CI children showed a benefit from target-masker voice differences in F0 and VTL, similar to NH children. CONCLUSION Although CI children's F0 and VTL voice discrimination scores were overall lower than those of NH children, their weighting of F0 and VTL cues for voice gender categorization and their benefit from target-masker differences in F0 and VTL resembled that of NH children. Together, these results suggest that prelingually deaf implanted CI children can effectively utilize spectrotemporally degraded F0 and VTL cues for voice and speech perception, generally outperforming postlingually deaf CI adults in comparable tasks. These findings underscore the presence of F0 and VTL cues in the CI signal to a certain degree and suggest other factors contributing to the perception challenges faced by CI adults.
Collapse
Affiliation(s)
- Leanne Nagels
- Center for Language and Cognition Groningen (CLCG), University of Groningen, Groningen, The Netherlands
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Groningen, The Netherlands
| | - Etienne Gaudrain
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Groningen, The Netherlands
- CNRS UMR 5292, Lyon Neuroscience Research Center, Auditory Cognition and Psychoacoustics, Inserm UMRS 1028, Université Claude Bernard Lyon 1, Université de Lyon, Lyon, France
| | - Deborah Vickers
- Cambridge Hearing Group, Sound Lab, Clinical Neurosciences Department, University of Cambridge, Cambridge, United Kingdom
| | - Petra Hendriks
- Center for Language and Cognition Groningen (CLCG), University of Groningen, Groningen, The Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Groningen, The Netherlands
| | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Groningen, The Netherlands
- W.J. Kolff Institute for Biomedical Engineering and Materials Science, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
| |
Collapse
|
7
|
Fontan L, Desreumaux J. Developmental Effects in the "Vocale Rapide dans le Bruit" Speech-in-Noise Identification Test: Reference Performances of Normal-Hearing Children and Adolescents. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:1624-1634. [PMID: 38564490 DOI: 10.1044/2024_jslhr-23-00467] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/04/2024]
Abstract
PURPOSE The main objective of this study was to assess the existence of developmental effects on the performance of the Vocale Rapide dans le Bruit (VRB) speech-in-noise (SIN) identification test that was recently developed for the French language and to collect reference scores for children and adolescents. METHOD Seventy-two native French speakers, aged 10-20 years, participated in the study. Each participant listened and repeated four lists of eight sentences, each containing three key words to be scored. The sentences were presented in free field at different signal-to-noise ratios (SNRs) using a four-talker babble noise. The SNR yielding 50% of correct repetitions of key words (SNR50) was recorded for each list. RESULTS A strong relationship between age and SNR50 was found, better performance occurring with increasing age (average drop in SNR50 per year: 0.34 dB). Large differences (Cohen's d ≥ 1.2) were observed between the SNR50 achieved by 10- to 13-year-old participants and those of adults. For participants aged 14-15 years, the difference fell just above the 5% level of significance. No effects of hearing thresholds or level of education were observed. CONCLUSIONS The study confirms the existence of developmental effects on SIN identification performance as measured using the VRB test and provides reference data for taking into account these effects during clinical practice. Explanations as to why age effects perdure during adolescence are discussed.
Collapse
|
8
|
Gransier R, Carlyon RP, Richardson ML, Middlebrooks JC, Wouters J. Artifact removal by template subtraction enables recordings of the frequency following response in cochlear-implant users. Sci Rep 2024; 14:6158. [PMID: 38486005 PMCID: PMC10940306 DOI: 10.1038/s41598-024-56047-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2023] [Accepted: 03/01/2024] [Indexed: 03/18/2024] Open
Abstract
Electrically evoked frequency-following responses (eFFRs) provide insight in the phase-locking ability of brainstem of cochlear-implant (CI) users. eFFRs can potentially be used to gain insight in the individual differences in the biological limitation on temporal encoding of the electrically stimulated auditory pathway, which can be inherent to the electrical stimulation itself and/or the degenerative processes associated with hearing loss. One of the major challenge of measuring eFFRs in CI users is the process of isolating the stimulation artifact from the neural response, as both the response and the artifact overlap in time and have similar frequency characteristics. Here we introduce a new artifact removal method based on template subtraction that successfully removes the stimulation artifacts from the recordings when CI users are stimulated with pulse trains from 128 to 300 pulses per second in a monopolar configuration. Our results show that, although artifact removal was successful in all CI users, the phase-locking ability of the brainstem to the different pulse rates, as assessed with the eFFR differed substantially across participants. These results show that the eFFR can be measured, free from artifacts, in CI users and that they can be used to gain insight in individual differences in temporal processing of the electrically stimulated auditory pathway.
Collapse
Affiliation(s)
- Robin Gransier
- ExpORL, Department of Neurosciences, Leuven Brain Institute, KU Leuven, Leuven, Belgium
| | - Robert P Carlyon
- Cambridge Hearing Group, MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
| | - Matthew L Richardson
- Department of Otolaryngology, University of California at Irvine, Irvine, CA, USA
- Center for Hearing Research, University of California at Irvine, Irvine, CA, USA
| | - John C Middlebrooks
- Department of Otolaryngology, University of California at Irvine, Irvine, CA, USA
- Center for Hearing Research, University of California at Irvine, Irvine, CA, USA
- Departments of Neurobiology and Behavior, Biomedical Engineering, Cognitive Sciences, University of California at Irvine, Irvine, CA, USA
| | - Jan Wouters
- ExpORL, Department of Neurosciences, Leuven Brain Institute, KU Leuven, Leuven, Belgium.
| |
Collapse
|
9
|
Meyer L, Araiza-Illan G, Rachman L, Gaudrain E, Başkent D. Evaluating speech-in-speech perception via a humanoid robot. Front Neurosci 2024; 18:1293120. [PMID: 38406584 PMCID: PMC10884269 DOI: 10.3389/fnins.2024.1293120] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2023] [Accepted: 01/15/2024] [Indexed: 02/27/2024] Open
Abstract
Introduction Underlying mechanisms of speech perception masked by background speakers, a common daily listening condition, are often investigated using various and lengthy psychophysical tests. The presence of a social agent, such as an interactive humanoid NAO robot, may help maintain engagement and attention. However, such robots potentially have limited sound quality or processing speed. Methods As a first step toward the use of NAO in psychophysical testing of speech- in-speech perception, we compared normal-hearing young adults' performance when using the standard computer interface to that when using a NAO robot to introduce the test and present all corresponding stimuli. Target sentences were presented with colour and number keywords in the presence of competing masker speech at varying target-to-masker ratios. Sentences were produced by the same speaker, but voice differences between the target and masker were introduced using speech synthesis methods. To assess test performance, speech intelligibility and data collection duration were compared between the computer and NAO setups. Human-robot interaction was assessed using the Negative Attitude Toward Robot Scale (NARS) and quantification of behavioural cues (backchannels). Results Speech intelligibility results showed functional similarity between the computer and NAO setups. Data collection durations were longer when using NAO. NARS results showed participants had a relatively positive attitude toward "situations of interactions" with robots prior to the experiment, but otherwise showed neutral attitudes toward the "social influence" of and "emotions in interaction" with robots. The presence of more positive backchannels when using NAO suggest higher engagement with the robot in comparison to the computer. Discussion Overall, the study presents the potential of the NAO for presenting speech materials and collecting psychophysical measurements for speech-in-speech perception.
Collapse
Affiliation(s)
- Luke Meyer
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, Netherlands
- University Medical Center Groningen, W.J. Kolff Institute for Biomedical Engineering and Materials Science, University of Groningen, Groningen, Netherlands
| | - Gloria Araiza-Illan
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, Netherlands
- University Medical Center Groningen, W.J. Kolff Institute for Biomedical Engineering and Materials Science, University of Groningen, Groningen, Netherlands
| | - Laura Rachman
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, Netherlands
- University Medical Center Groningen, W.J. Kolff Institute for Biomedical Engineering and Materials Science, University of Groningen, Groningen, Netherlands
- Pento Audiology Centre, Zwolle, Netherlands
| | - Etienne Gaudrain
- Lyon Neuroscience Research Center, CNRS UMR 5292, INSERM UMRS 1028, Université Claude Bernard Lyon 1, Université de Lyon, Lyon, France
| | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, Netherlands
- University Medical Center Groningen, W.J. Kolff Institute for Biomedical Engineering and Materials Science, University of Groningen, Groningen, Netherlands
| |
Collapse
|
10
|
Everhardt MK, Jung DE, Stiensma B, Lowie W, Başkent D, Sarampalis A. Foreign Language Acquisition in Adolescent Cochlear Implant Users. Ear Hear 2024; 45:174-185. [PMID: 37747307 PMCID: PMC10718217 DOI: 10.1097/aud.0000000000001410] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2022] [Accepted: 06/20/2023] [Indexed: 09/26/2023]
Abstract
OBJECTIVES This study explores to what degree adolescent cochlear implant (CI) users can learn a foreign language in a school setting similar to their normal-hearing (NH) peers despite the degraded auditory input. DESIGN A group of native Dutch adolescent CI users (age range 13 to 17 years) learning English as a foreign language at secondary school and a group of NH controls (age range 12 to 15 years) were assessed on their Dutch and English language skills using various language tasks that either relied on the processing of auditory information (i.e., listening task) or on the processing of orthographic information (i.e., reading and/or gap-fill task). The test battery also included various auditory and cognitive tasks to assess whether the auditory and cognitive functioning of the learners could explain the potential variation in language skills. RESULTS Results showed that adolescent CI users can learn English as a foreign language, as the English language skills of the CI users and their NH peers were comparable when assessed with reading or gap-fill tasks. However, the performance of the adolescent CI users was lower for English listening tasks. This discrepancy between task performance was not observed in their native language Dutch. The auditory tasks confirmed that the adolescent CI users had coarser temporal and spectral resolution than their NH peers, supporting the notion that the difference in foreign language listening skills may be due to a difference in auditory functioning. No differences in the cognitive functioning of the CI users and their NH peers were found that could explain the variation in the foreign language listening tasks. CONCLUSIONS In short, acquiring a foreign language with degraded auditory input appears to affect foreign language listening skills, yet does not appear to impact foreign language skills when assessed with tasks that rely on the processing of orthographic information. CI users could take advantage of orthographic information to facilitate foreign language acquisition and potentially support the development of listening-based foreign language skills.
Collapse
Affiliation(s)
- Marita K. Everhardt
- Center for Language and Cognition Groningen, University of Groningen, Netherlands
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Netherlands
| | - Dorit Enja Jung
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Netherlands
- Department of Psychology, University of Groningen, Netherlands
| | - Berrit Stiensma
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Netherlands
| | - Wander Lowie
- Center for Language and Cognition Groningen, University of Groningen, Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Netherlands
| | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Netherlands
- W.J. Kolff Institute for Biomedical Engineering and Materials Science, University Medical Center Groningen, University of Groningen, Netherlands
| | - Anastasios Sarampalis
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Netherlands
- Department of Psychology, University of Groningen, Netherlands
| |
Collapse
|
11
|
Nittrouer S, Lowenstein JH. Recognition of Sentences With Complex Syntax in Speech Babble by Adolescents With Normal Hearing or Cochlear Implants. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2023; 66:1110-1135. [PMID: 36758200 PMCID: PMC10205108 DOI: 10.1044/2022_jslhr-22-00407] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/13/2022] [Revised: 10/17/2022] [Accepted: 11/22/2022] [Indexed: 05/25/2023]
Abstract
PURPOSE General language abilities of children with cochlear implants have been thoroughly investigated, especially at young ages, but far less is known about how well they process language in real-world settings, especially in higher grades. This study addressed this gap in knowledge by examining recognition of sentences with complex syntactic structures in backgrounds of speech babble by adolescents with cochlear implants, and peers with normal hearing. DESIGN Two experiments were conducted. First, new materials were developed using young adults with normal hearing as the normative sample, creating a corpus of sentences with controlled, but complex syntactic structures presented in three kinds of babble that varied in voice gender and number of talkers. Second, recognition by adolescents with normal hearing or cochlear implants was examined for these new materials and for sentence materials used with these adolescents at younger ages. Analyses addressed three objectives: (1) to assess the stability of speech recognition across a multiyear age range, (2) to evaluate speech recognition of sentences with complex syntax in babble, and (3) to explore how bottom-up and top-down mechanisms account for performance under these conditions. RESULTS Results showed: (1) Recognition was stable across the ages of 10-14 years for both groups. (2) Adolescents with normal hearing performed similarly to young adults with normal hearing, showing effects of syntactic complexity and background babble; adolescents with cochlear implants showed poorer recognition overall, and diminished effects of both factors. (3) Top-down language and working memory primarily explained recognition for adolescents with normal hearing, but the bottom-up process of perceptual organization primarily explained recognition for adolescents with cochlear implants. CONCLUSIONS Comprehension of language in real-world settings relies on different mechanisms for adolescents with cochlear implants than for adolescents with normal hearing. A novel finding was that perceptual organization is a critical factor. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.21965228.
Collapse
Affiliation(s)
- Susan Nittrouer
- Department of Speech, Language, and Hearing Sciences, University of Florida, Gainesville
| | - Joanna H. Lowenstein
- Department of Speech, Language, and Hearing Sciences, University of Florida, Gainesville
| |
Collapse
|
12
|
Flaherty MM, Buss E, Libert K. Effects of Target and Masker Fundamental Frequency Contour Depth on School-Age Children's Speech Recognition in a Two-Talker Masker. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2023; 66:400-414. [PMID: 36580582 DOI: 10.1044/2022_jslhr-22-00207] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/17/2023]
Abstract
PURPOSE Maturation of the ability to recognize target speech in the presence of a two-talker speech masker extends into early adolescence. This study evaluated whether children benefit from differences in fundamental frequency (f o) contour depth between the target and masker speech, a cue that has been shown to improve recognition in adults. METHOD Speech stimuli were recorded from talkers using three speaking styles, with f o contour depths that were Flat, Normal, or Exaggerated. Targets were open-set, declarative sentences produced by a female talker, and maskers were two streams of concatenated sentences produced by a second female talker. Listeners were children (ages 5-17 years) and adults (ages 18-24 years) with normal hearing. Each listener was tested in one of the three masker styles paired with all three target styles. Speech recognition thresholds (SRTs) corresponding to 50% correct were estimated by fitting psychometric functions to adaptive track data. RESULTS For adults, performance did not differ significantly across conditions with matched speaking styles. A mismatch benefit was observed when combining Flat targets with the Exaggerated masker and Exaggerated targets with the Flat masker, and for both Flat and Exaggerated targets paired with the Normal masker. For children, there was a significant effect of age in all conditions. Flat targets in the Flat masker were associated with lower SRTs than the other two matched conditions, and a mismatch benefit was observed for young children only when the target f o contour was less variable than the masker f o contour. CONCLUSIONS Whereas child-directed speech often has exaggerated pitch contours, young children were better able to recognize speech with less variable f o. Age effects were observed in the benefit of mismatched speaking styles for some conditions, which could be related to differences in baseline SRTs rather than differences in segregation abilities.
Collapse
Affiliation(s)
- Mary M Flaherty
- Department of Speech and Hearing Science, University of Illinois at Urbana-Champaign, Champaign
| | - Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, The University of North Carolina at Chapel Hill
| | - Kelsey Libert
- Department of Speech and Hearing Science, University of Illinois at Urbana-Champaign, Champaign
| |
Collapse
|
13
|
Colby S, Orena AJ. Recognizing Voices Through a Cochlear Implant: A Systematic Review of Voice Perception, Talker Discrimination, and Talker Identification. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:3165-3194. [PMID: 35926089 PMCID: PMC9911123 DOI: 10.1044/2022_jslhr-21-00209] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/12/2021] [Revised: 02/02/2022] [Accepted: 05/03/2022] [Indexed: 06/15/2023]
Abstract
OBJECTIVE Some cochlear implant (CI) users report having difficulty accessing indexical information in the speech signal, presumably due to limitations in the transmission of fine spectrotemporal cues. The purpose of this review article was to systematically review and evaluate the existing research on talker processing in CI users. Specifically, we reviewed the performance of CI users in three types of talker- and voice-related tasks. We also examined the different factors (such as participant, hearing, and device characteristics) that might influence performance in these specific tasks. DESIGN We completed a systematic search of the literature with select key words using citation aggregation software to search Google Scholar. We included primary reports that tested (a) talker discrimination, (b) voice perception, and (c) talker identification. Each report must have had at least one group of participants with CIs. Each included study was also evaluated for quality of evidence. RESULTS The searches resulted in 1,561 references, which were first screened for inclusion and then evaluated in full. Forty-three studies examining talker discrimination, voice perception, and talker identification were included in the final review. Most studies were focused on postlingually deafened and implanted adult CI users, with fewer studies focused on prelingual implant users. In general, CI users performed above chance in these tasks. When there was a difference between groups, CI users performed less accurately than their normal-hearing (NH) peers. A subset of CI users reached the same level of performance as NH participants exposed to noise-vocoded stimuli. Some studies found that CI users and NH participants relied on different cues for talker perception. Within groups of CI users, there is moderate evidence for a bimodal benefit for talker processing, and there are mixed findings about the effects of hearing experience. CONCLUSIONS The current review highlights the challenges faced by CI users in tracking and recognizing voices and how they adapt to it. Although large variability exists, there is evidence that CI users can process indexical information from speech, though with less accuracy than their NH peers. Recent work has described some of the factors that might ease the challenges of talker processing in CI users. We conclude by suggesting some future avenues of research to optimize real-world speech outcomes.
Collapse
Affiliation(s)
- Sarah Colby
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City
| | - Adriel John Orena
- Department of Psychology, University of British Columbia, Vancouver, Canada
| |
Collapse
|
14
|
Huet MP, Micheyl C, Gaudrain E, Parizet E. Vocal and semantic cues for the segregation of long concurrent speech stimuli in diotic and dichotic listening-The Long-SWoRD test. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 151:1557. [PMID: 35364949 DOI: 10.1121/10.0007225] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/04/2020] [Accepted: 10/25/2021] [Indexed: 06/14/2023]
Abstract
It is not always easy to follow a conversation in a noisy environment. To distinguish between two speakers, a listener must mobilize many perceptual and cognitive processes to maintain attention on a target voice and avoid shifting attention to the background noise. The development of an intelligibility task with long stimuli-the Long-SWoRD test-is introduced. This protocol allows participants to fully benefit from the cognitive resources, such as semantic knowledge, to separate two talkers in a realistic listening environment. Moreover, this task also provides the experimenters with a means to infer fluctuations in auditory selective attention. Two experiments document the performance of normal-hearing listeners in situations where the perceptual separability of the competing voices ranges from easy to hard using a combination of voice and binaural cues. The results show a strong effect of voice differences when the voices are presented diotically. In addition, analyzing the influence of the semantic context on the pattern of responses indicates that the semantic information induces a response bias in situations where the competing voices are distinguishable and indistinguishable from one another.
Collapse
Affiliation(s)
- Moïra-Phoebé Huet
- Laboratory of Vibration and Acoustics, National Institute of Applied Sciences, University of Lyon, 20 Avenue Albert Einstein, Villeurbanne, 69100, France
| | | | - Etienne Gaudrain
- Lyon Neuroscience Research Center, Auditory Cognition and Psychoacoustics, Centre National de la Recerche Scientifique UMR5292, Institut National de la Santé et de la Recherche Médicale U1028, Université Claude Bernard Lyon 1, Université de Lyon, Centre Hospitalier Le Vinatier, Neurocampus, 95 boulevard Pinel, Bron Cedex, 69675, France
| | - Etienne Parizet
- Laboratory of Vibration and Acoustics, National Institute of Applied Sciences, University of Lyon, 20 Avenue Albert Einstein, Villeurbanne, 69100, France
| |
Collapse
|