1
|
Levi SV, Cheng HS, O'Brien G, Harel D. A Picture May Be Worth 1,000 Words, but Is It Worth a Letter? Examining Whether the Choice of Label Affects the Perception of Speech Sounds. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:2115-2127. [PMID: 38754023 DOI: 10.1044/2024_jslhr-23-00616] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/18/2024]
Abstract
PURPOSE Researchers often use identification or goodness rating tasks to assess speech perception for different populations. These tasks provide useful information about a listener's willingness to accept a range of acoustically variable stimuli as belonging to the same category and also about assessing how stimuli that are labeled the same may not be perceived as equally good versions of a particular speech sound. Many methodological aspects of these simple tasks have been tested, but one aspect that has not is the choice of label. In this study, we examine response patterns to images versus letters, as studies with different populations (children vs. adults) or different methods (typical behavioral study vs. visual world paradigm) may vary in the type of label used. METHOD Eighty-one adult listeners completed phoneme identification and goodness ratings tasks with either images of response options (a picture of a bear and a picture of a pear) or with letter labels (a capital B and P). RESULTS The results suggest that choice of label does not alter performance within the tasks studied here. In addition, the results did show the expected finding that the slope of the response curve is steeper in an identification task than in a goodness rating task. CONCLUSION These results suggest that it is possible to compare across studies that use different response options, a benefit to research and practice because letter labels can be used for nonimageable words and nonwords, whereas images may be best used for participants who are younger or have poorer reading skills.
Collapse
Affiliation(s)
- Susannah V Levi
- Department of Communicative Science and Disorders, New York University, NY
| | - Hung-Shao Cheng
- Department of Communicative Science and Disorders, New York University, NY
| | | | - Daphna Harel
- Department of Applied Statistics, Social Sciences, and Humanities, New York University, NY
| |
Collapse
|
2
|
Kronenberger WG, Castellanos I, Pisoni DB. Association of domain-general speed of information processing with spoken language outcomes in prelingually-deaf children with cochlear implants. Hear Res 2024; 450:109069. [PMID: 38889562 DOI: 10.1016/j.heares.2024.109069] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/19/2023] [Revised: 05/24/2024] [Accepted: 06/11/2024] [Indexed: 06/20/2024]
Abstract
Spoken language development after pediatric cochlear implantation requires rapid and efficient processing of novel, degraded auditory signals and linguistic information. These demands for rapid adaptation tax the information processing speed ability of children who receive cochlear implants. This study investigated the association of speed of information processing ability with spoken language outcomes after cochlear implantation in prelingually deaf children aged 4-6 years. Two domain-general (visual, non-linguistic) speed of information processing measures were administered to 21 preschool-aged children with cochlear implants and 23 normal-hearing peers. Measures of speech recognition, language (vocabulary and comprehension), nonverbal intelligence, and executive functioning skills were also obtained from each participant. Speed of information processing was positively associated with speech recognition and language skills in preschool-aged children with cochlear implants but not in normal-hearing peers. This association remained significant after controlling for hearing group, age, nonverbal intelligence, and executive functioning skills. These findings are consistent with models suggesting that domain-general, fast-efficient information processing speed underlies adaptation to speech perception and language learning following implantation. Assessment and intervention strategies targeting speed of information processing may provide better understanding and development of speech-language skills after cochlear implantation.
Collapse
Affiliation(s)
- William G Kronenberger
- Department of Otolaryngology - Head and Neck Surgery, Indiana University School of Medicine, Indianapolis, IN 46202, USA; Department of Psychiatry, Indiana University School of Medicine, Indianapolis, IN 46202, USA.
| | - Irina Castellanos
- Department of Otolaryngology - Head and Neck Surgery, Indiana University School of Medicine, Indianapolis, IN 46202, USA; Department of Psychiatry, Indiana University School of Medicine, Indianapolis, IN 46202, USA
| | - David B Pisoni
- Department of Otolaryngology - Head and Neck Surgery, Indiana University School of Medicine, Indianapolis, IN 46202, USA; Department of Psychiatry, Indiana University School of Medicine, Indianapolis, IN 46202, USA; Department of Psychological and Brain Sciences, Indiana University, Bloomington, IN 47405, USA
| |
Collapse
|
3
|
Hendrickson K, Bay K, Combiths P, Foody M, Walker E. Speech Sound Categories Affect Lexical Competition: Implications for Analytic Auditory Training. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:1281-1289. [PMID: 38517230 PMCID: PMC11005953 DOI: 10.1044/2024_jslhr-23-00307] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/16/2023] [Revised: 12/13/2023] [Accepted: 02/01/2024] [Indexed: 03/23/2024]
Abstract
OBJECTIVES We provide a novel application of psycholinguistic theories and methods to the field of auditory training to provide preliminary data regarding which minimal pair contrasts are more difficult for listeners with typical hearing to distinguish in real-time. DESIGN Using eye-tracking, participants heard a word and selected the corresponding image from a display of four: the target word, two unrelated words, and a word from one of four contrast categories (i.e., voiced-initial [e.g., peach-beach], voiced-final [e.g., back-bag], manner-initial [e.g., talk-sock], and manner-final [e.g., bat-bass]). RESULTS Fixations were monitored to measure how strongly words compete for recognition depending on the contrast type (voicing, manner) and location (word-initial or final). Manner contrasts competed more for recognition than did voicing contrasts, and contrasts that occurred in word-final position were harder to distinguish than word-initial position. CONCLUSION These results are an important initial step toward creating an evidence-based hierarchy for auditory training for individuals who use cochlear implants.
Collapse
Affiliation(s)
- Kristi Hendrickson
- Department of Communication Sciences and Disorder, The University of Iowa, Iowa City
| | - Katlyn Bay
- Department of Communication Sciences and Disorder, The University of Iowa, Iowa City
| | - Philip Combiths
- Department of Communication Sciences and Disorder, The University of Iowa, Iowa City
| | - Meaghan Foody
- Department of Communication Sciences and Disorder, The University of Iowa, Iowa City
| | - Elizabeth Walker
- Department of Communication Sciences and Disorder, The University of Iowa, Iowa City
| |
Collapse
|
4
|
Colby S, Seedorff M, McMurray B. Audiological and Demographic Factors that Impact the Precision of Speech Categorization in Cochlear Implant Users. Ear Hear 2023; 44:572-587. [PMID: 36542839 PMCID: PMC10101869 DOI: 10.1097/aud.0000000000001307] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
Abstract
OBJECTIVES The ability to adapt to subtle variations in acoustic input is a necessary skill for successful speech perception. Cochlear implant (CI) users tend to show speech perception benefits from the maintenance of their residual acoustic hearing. However, previous studies often compare CI users in different listening conditions within-subjects (i.e., in their typical Acoustic + Electric configuration compared with Acoustic-only or Electric-only configurations) and comparisons among different groups of CI users do not always reflect an Acoustic + Electric benefit. Existing work suggests that CI users with residual acoustic hearing perform similarly to Electric-only listeners on phonetic voicing contrasts and unexpectedly poorer with fricative contrasts which have little energy in the range of the Acoustic + Electric listeners' acoustic hearing. To further investigate how residual acoustic hearing impacts sensitivity to phonetic ambiguity, we examined whether device configuration, age, and device experience influenced phonetic categorization in a large individual differences study. DESIGN CI users with various device configurations (Electric-only N = 41; Acoustic + Electric N = 95) categorized tokens from five /b-p/ and five /s-ʃ/ minimal pair continua (e.g., bet-pet; sock-shock). We investigated age, device experience, and when applicable, residual acoustic hearing (pure tone hearing thresholds) as predictors of categorization. We also examined the relationship between phonetic categorization and clinical outcomes (CNC, AzBio) in a subset of our sample. RESULTS Acoustic + Electric CI users were better able to categorize along the voicing contrast (steeper categorization slope) compared with Electric-only users, but there was no group-level difference for fricatives. There were differences within the subgroups for fricatives: bilateral users showed better categorization than unilateral users and bimodal users had better categorization than hybrid users. Age was a significant factor for voicing, while device experience was significant for fricatives. Critically, within the Acoustic + Electric group, hybrid CI users had shallower slopes than bimodal CI users. CONCLUSIONS Our findings suggest residual acoustic hearing is beneficial for categorizing stop voicing, but not frication. Age impacts the categorization of voicing, while device experience matters for fricatives. For CI users with ipsilateral residual acoustic hearing, those with better hearing thresholds may be over-relying on their acoustic hearing rather than extracting as much information as possible from their CI, and thus have shallower fricative categorization.
Collapse
Affiliation(s)
- Sarah Colby
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City, Iowa, USA
- Department of Otolaryngology, University of Iowa, Iowa City, Iowa, USA
| | - Michael Seedorff
- Department of Biostatistics, University of Iowa, Iowa City, Iowa, USA
- Alphabet, Inc., Mountain View, California, USA
| | - Bob McMurray
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City, Iowa, USA
- Department of Otolaryngology, University of Iowa, Iowa City, Iowa, USA
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City, Iowa, USA
- Department of Linguistics, University of Iowa, Iowa City, Iowa, USA
| |
Collapse
|
5
|
Buz E, Dwyer NC, Lai W, Watson DG, Gifford RH. Integration of fundamental frequency and voice-onset-time to voicing categorization: Listeners with normal hearing and bimodal hearing configurations. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 153:1580. [PMID: 37002096 PMCID: PMC9995168 DOI: 10.1121/10.0017429] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/08/2022] [Revised: 02/13/2023] [Accepted: 02/13/2023] [Indexed: 05/18/2023]
Abstract
This study investigates the integration of word-initial fundamental frequency (F0) and voice-onset-time (VOT) in stop voicing categorization for adult listeners with normal hearing (NH) and unilateral cochlear implant (CI) recipients utilizing a bimodal hearing configuration [CI + contralateral hearing aid (HA)]. Categorization was assessed for ten adults with NH and ten adult bimodal listeners, using synthesized consonant stimuli interpolating between /ba/ and /pa/ exemplars with five-step VOT and F0 conditions. All participants demonstrated the expected categorization pattern by reporting /ba/ for shorter VOTs and /pa/ for longer VOTs, with NH listeners showing more use of VOT as a voicing cue than CI listeners in general. When VOT becomes ambiguous between voiced and voiceless stops, NH users make more use of F0 as a cue to voicing than CI listeners, and CI listeners showed greater utilization of initial F0 during voicing identification in their bimodal (CI + HA) condition than in the CI-alone condition. The results demonstrate the adjunctive benefit of acoustic hearing from the non-implanted ear for listening conditions involving spectrotemporally complex stimuli. This finding may lead to the development of a clinically feasible perceptual weighting task that could inform clinicians about bimodal efficacy and the risk-benefit profile associated with bilateral CI recommendation.
Collapse
Affiliation(s)
- Esteban Buz
- Department of Psychology and Human Development, Vanderbilt University, Nashville, Tennessee 37203, USA
| | - Nichole C Dwyer
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida 33620, USA
| | - Wei Lai
- Department of Psychology and Human Development, Vanderbilt University, Nashville, Tennessee 37203, USA
| | - Duane G Watson
- Department of Psychology and Human Development, Vanderbilt University, Nashville, Tennessee 37203, USA
| | - René H Gifford
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, Tennessee 37203, USA
| |
Collapse
|
6
|
Bochner J, Samar V, Prud'hommeaux E, Huenerfauth M. Phoneme Categorization in Prelingually Deaf Adult Cochlear Implant Users. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:4429-4453. [PMID: 36279201 DOI: 10.1044/2022_jslhr-22-00038] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
PURPOSE Phoneme categorization (PC) for voice onset time and second formant transition was studied in adult cochlear implant (CI) users with early-onset deafness and hearing controls. METHOD Identification and discrimination tasks were administered to 30 participants implanted before 4 years of age, 21 participants implanted after 7 years of age, and 21 hearing individuals. RESULTS Distinctive identification and discrimination functions confirmed PC within all groups. Compared to hearing participants, the CI groups generally displayed longer/higher category boundaries, shallower identification function slopes, reduced identification consistency, and reduced discrimination performance. A principal component analysis revealed that identification consistency, discrimination accuracy, and identification function slope, but not boundary location, loaded on a single factor, reflecting general PC performance. Earlier implantation was associated with better PC performance within the early CI group, but not the late CI group. Within the early CI group, earlier implantation age but not PC performance was associated with better speech recognition. Conversely, within the late CI group, better PC performance but not earlier implantation age was associated with better speech recognition. CONCLUSIONS Results suggest that implantation timing within the sensitive period before 4 years of age partly determines the level of PC performance. They also suggest that early implantation may promote development of higher level processes that can compensate for relatively poor PC performance, as can occur in challenging listening conditions.
Collapse
Affiliation(s)
- Joseph Bochner
- National Technical Institute for the Deaf, Rochester Institute of Technology, NY
| | - Vincent Samar
- National Technical Institute for the Deaf, Rochester Institute of Technology, NY
| | | | - Matt Huenerfauth
- Golisano College of Computing and Information Sciences, Rochester Institute of Technology, NY
| |
Collapse
|
7
|
Lexical Access Changes Based on Listener Needs: Real-Time Word Recognition in Continuous Speech in Cochlear Implant Users. Ear Hear 2022; 43:1487-1501. [PMID: 35067570 PMCID: PMC9300769 DOI: 10.1097/aud.0000000000001203] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/23/2023]
Abstract
OBJECTIVES A key challenge in word recognition is the temporary ambiguity created by the fact that speech unfolds over time. In normal hearing (NH) listeners, this temporary ambiguity is resolved through incremental processing and competition among lexical candidates. Post-lingually deafened cochlear implant (CI) users show similar incremental processing and competition but with slight delays. However, even brief delays could lead to drastic changes when compounded across multiple words in a phrase. This study asks whether words presented in non-informative continuous speech (a carrier phrase) are processed differently than in isolation and whether NH listeners and CI users exhibit different effects of a carrier phrase. DESIGN In a Visual World Paradigm experiment, listeners heard words either in isolation or in non-informative carrier phrases (e.g., "click on the…" ). Listeners selected the picture corresponding to the target word from among four items including the target word (e.g., mustard ), a cohort competitor (e.g., mustache ), a rhyme competitor (e.g., custard ), and an unrelated item (e.g., penguin ). Eye movements were tracked as an index of the relative activation of each lexical candidate as competition unfolds over the course of word recognition. Participants included 21 post-lingually deafened cochlear implant users and 21 NH controls. A replication experiment presented in the Supplemental Digital Content, http://links.lww.com/EANDH/A999 included an additional 22 post-lingually deafened CI users and 18 NH controls. RESULTS Both CI users and the NH controls were accurate at recognizing the words both in continuous speech and in isolation. The time course of lexical activation (indexed by the fixations) differed substantially between groups. CI users were delayed in fixating the target relative to NH controls. Additionally, CI users showed less competition from cohorts than NH controls (even as previous studies have often report increased competition). However, CI users took longer to suppress the cohort and suppressed it less fully than the NH controls. For both CI users and NH controls, embedding words in carrier phrases led to more immediacy in lexical access as observed by increases in cohort competition relative to when words were presented in isolation. However, CI users were not differentially affected by the carriers. CONCLUSIONS Unlike prior work, CI users appeared to exhibit "wait-and-see" profile, in which lexical access is delayed minimizing early competition. However, CI users simultaneously sustained competitor activation late in the trial, possibly to preserve flexibility. This hybrid profile has not been observed previously. When target words are heard in continuous speech, both CI users and NH controls more heavily weight early information. However, CI users (but not NH listeners) also commit less fully to the target, potentially keeping options open if they need to recover from a misperception. This mix of patterns reflects a lexical system that is extremely flexible and adapts to fit the needs of a listener.
Collapse
|
8
|
Abstract
OBJECTIVES Whispered speech offers a unique set of challenges to speech perception and word recognition. The goals of the present study were twofold: First, to determine how listeners recognize whispered speech. Second, to inform major theories of spoken word recognition by considering how recognition changes when major cues to phoneme identity are reduced or largely absent compared with normal voiced speech. DESIGN Using eye tracking in the Visual World Paradigm, we examined how listeners recognize whispered speech. After hearing a target word (normal or whispered), participants selected the corresponding image from a display of four-a target (e.g., money), a word that shares sounds with the target at the beginning (cohort competitor, e.g., mother), a word that shares sounds with the target at the end (rhyme competitor, e.g., honey), and a phonologically unrelated word (e.g., whistle). Eye movements to each object were monitored to measure (1) how fast listeners process whispered speech, and (2) how strongly they consider lexical competitors (cohorts and rhymes) as the speech signal unfolds. RESULTS Listeners were slower to recognize whispered words. Compared with normal speech, listeners displayed slower reaction times to click the target image, were slower to fixate the target, and fixated the target less overall. Further, we found clear evidence that the dynamics of lexical competition are altered during whispered speech recognition. Relative to normal speech, words that overlapped with the target at the beginning (cohorts) displayed slower, reduced, and delayed activation, whereas words that overlapped with the target at the end (rhymes) exhibited faster, more robust, and longer lasting activation. CONCLUSION When listeners are confronted with whispered speech, they engage in a "wait-and-see" approach. Listeners delay lexical access, and by the time they begin to consider what word they are hearing, the beginning of the word has largely come and gone, and activation for cohorts is reduced. However, delays in lexical access actually increase consideration of rhyme competitors; the delay pushes lexical activation to a point later in processing, and the recognition system puts more weight on the word-final overlap between the target and the rhyme.
Collapse
|
9
|
Galle ME, Klein-Packard J, Schreiber K, McMurray B. What Are You Waiting For? Real-Time Integration of Cues for Fricatives Suggests Encapsulated Auditory Memory. Cogn Sci 2020; 43. [PMID: 30648798 DOI: 10.1111/cogs.12700] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2017] [Revised: 10/15/2018] [Accepted: 10/25/2018] [Indexed: 11/30/2022]
Abstract
Speech unfolds over time, and the cues for even a single phoneme are rarely available simultaneously. Consequently, to recognize a single phoneme, listeners must integrate material over several hundred milliseconds. Prior work contrasts two accounts: (a) a memory buffer account in which listeners accumulate auditory information in memory and only access higher level representations (i.e., lexical representations) when sufficient information has arrived; and (b) an immediate integration scheme in which lexical representations can be partially activated on the basis of early cues and then updated when more information arises. These studies have uniformly shown evidence for immediate integration for a variety of phonetic distinctions. We attempted to extend this to fricatives, a class of speech sounds which requires not only temporal integration of asynchronous cues (the frication, followed by the formant transitions 150-350 ms later), but also integration across different frequency bands and compensation for contextual factors like coarticulation. Eye movements in the visual world paradigm showed clear evidence for a memory buffer. Results were replicated in five experiments, ruling out methodological factors and tying the release of the buffer to the onset of the vowel. These findings support a general auditory account for speech by suggesting that the acoustic nature of particular speech sounds may have large effects on how they are processed. It also has major implications for theories of auditory and speech perception by raising the possibility of an encapsulated memory buffer in early auditory processing.
Collapse
Affiliation(s)
- Marcus E Galle
- Department of Psychological and Brain Sciences, University of Iowa
| | | | | | - Bob McMurray
- Department of Psychological and Brain Sciences, University of Iowa.,Department of Communication Sciences and Disorders, University of Iowa.,Department of Linguistics, University of Iowa.,Department of Otolaryngology, University of Iowa
| |
Collapse
|
10
|
Balkenhol T, Wallhäusser-Franke E, Rotter N, Servais JJ. Changes in Speech-Related Brain Activity During Adaptation to Electro-Acoustic Hearing. Front Neurol 2020; 11:161. [PMID: 32300327 PMCID: PMC7145411 DOI: 10.3389/fneur.2020.00161] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2019] [Accepted: 02/19/2020] [Indexed: 12/17/2022] Open
Abstract
Objectives: Hearing improves significantly with bimodal provision, i.e., a cochlear implant (CI) at one ear and a hearing aid (HA) at the other, but performance shows a high degree of variability resulting in substantial uncertainty about the performance that can be expected by the individual CI user. The objective of this study was to explore how auditory event-related potentials (AERPs) of bimodal listeners in response to spoken words approximate the electrophysiological response of normal hearing (NH) listeners. Study Design: Explorative prospective analysis during the first 6 months of bimodal listening using a within-subject repeated measures design. Setting: Academic tertiary care center. Participants: Twenty-seven adult participants with bilateral sensorineural hearing loss who received a HiRes 90K CI and continued use of a HA at the non-implanted ear. Age-matched NH listeners served as controls. Intervention: Cochlear implantation. Main Outcome Measures: Obligatory auditory evoked potentials N1 and P2, and the event-related N2 potential in response to monosyllabic words and their reversed sound traces before, as well as 3 and 6 months post-implantation. The task required word/non-word classification. Stimuli were presented within speech-modulated noise. Loudness of word/non-word signals was adjusted individually to achieve the same intelligibility across groups and assessments. Results: Intelligibility improved significantly with bimodal hearing, and the N1-P2 response approximated the morphology seen in NH with enhanced and earlier responses to the words compared to their reversals. For bimodal listeners, a prominent negative deflection was present between 370 and 570 ms post stimulus onset (N2), irrespective of stimulus type. This was absent for NH controls; hence, this response did not approximate the NH response during the study interval. N2 source localization evidenced extended activation of general cognitive areas in frontal and prefrontal brain areas in the CI group. Conclusions: Prolonged and spatially extended processing in bimodal CI users suggests employment of additional auditory-cognitive mechanisms during speech processing. This does not reduce within 6 months of bimodal experience and may be a correlate of the enhanced listening effort described by CI listeners.
Collapse
|
11
|
Davies B, Xu Rattanasone N, Davis A, Demuth K. The Acquisition of Productive Plural Morphology by Children With Hearing Loss. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:552-568. [PMID: 32004109 DOI: 10.1044/2019_jslhr-19-00208] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Purpose Normal-hearing (NH) children acquire plural morphemes at different rates, with the segmental allomorphs /-s, -z/ (e.g., cat-s) being acquired before the syllabic allomorph /-əz/ (e.g., bus-es). Children with hearing loss (HL) have been reported to show delays in the production of plural morphology, raising the possibility that this might be due to challenges acquiring different types of lexical/morphological representations. This study therefore examined the comprehension of plural morphology by 3- to 7-year-olds with HL and compared this with performance by their NH peers. We also investigated comprehension as a function of wearing hearing aids (HAs) versus cochlear implants (CIs). Method Participants included 129 NH children aged 3-5 years and 25 children with HL aged 3-7 years (13 with HAs, 12 with CIs). All participated in a novel word two-alternative forced-choice task presented on an iPad. The task tested comprehension of the segmental (e.g., teps, mubz) and syllabic (e.g., kosses) plural, as well as their singular counterparts (e.g., tep, mub, koss). Results While the children with NH were above chance for all conditions, those with HL performed at chance. As a group, the performance of the children with HL did not improve with age. However, results suggest possible differences between children with HAs and those with CIs, where those with HAs appeared to be in the process of developing representations of consonant-vowel-consonant singulars. Conclusions Results suggest that preschoolers with HL do not yet have a robust representation of plural morphology for words they have not heard before. However, those with HAs are beginning to access the singular/plural system as they get older.
Collapse
Affiliation(s)
- Benjamin Davies
- Department of Linguistics, Macquarie University, Sydney, New South Wales, Australia
- ARC Centre of Excellence in Cognition and Its Disorders, Macquarie University, Sydney, New South Wales, Australia
- The HEARing Cooperative Research Centre, Melbourne, Victoria, Australia
| | - Nan Xu Rattanasone
- Department of Linguistics, Macquarie University, Sydney, New South Wales, Australia
- ARC Centre of Excellence in Cognition and Its Disorders, Macquarie University, Sydney, New South Wales, Australia
- The HEARing Cooperative Research Centre, Melbourne, Victoria, Australia
| | - Aleisha Davis
- The HEARing Cooperative Research Centre, Melbourne, Victoria, Australia
- The Shepherd Centre, Sydney, New South Wales, Australia
| | - Katherine Demuth
- Department of Linguistics, Macquarie University, Sydney, New South Wales, Australia
- ARC Centre of Excellence in Cognition and Its Disorders, Macquarie University, Sydney, New South Wales, Australia
- The HEARing Cooperative Research Centre, Melbourne, Victoria, Australia
| |
Collapse
|
12
|
Hendrickson K, Spinelli J, Walker E. Cognitive processes underlying spoken word recognition during soft speech. Cognition 2020; 198:104196. [PMID: 32004934 DOI: 10.1016/j.cognition.2020.104196] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2019] [Revised: 01/06/2020] [Accepted: 01/18/2020] [Indexed: 11/25/2022]
Abstract
In two eye-tracking experiments using the Visual World Paradigm, we examined how listeners recognize words when faced with speech at lower intensities (40, 50, and 65 dBA). After hearing the target word, participants (n = 32) clicked the corresponding picture from a display of four images - a target (e.g., money), a cohort competitor (e.g., mother), a rhyme competitor (e.g., honey) and an unrelated item (e.g., whistle) - while their eye-movements were tracked. For slightly soft speech (50 dBA), listeners demonstrated an increase in cohort activation, whereas for rhyme competitors, activation started later and was sustained longer in processing. For very soft speech (40 dBA), listeners waited until later in processing to activate potential words, as illustrated by a decrease in activation for cohorts, and an increase in activation for rhymes. Further, the extent to which words were considered depended on word length (mono- vs. bi-syllabic words), and speech-extrinsic factors such as the surrounding listening environment. These results advance current theories of spoken word recognition by considering a range of speech levels more typical of everyday listening environments. From an applied perspective, these results motivate models of how individuals who are hard of hearing approach the task of recognizing spoken words.
Collapse
Affiliation(s)
- Kristi Hendrickson
- Department of Communication Sciences & Disorders, University of Iowa, 250 Hawkins Drive, 52242 Iowa City, IA, United States of America; Department of Psychological & Brain Sciences, University of Iowa, 250 Hawkins Drive, 52242 Iowa City, IA, United States of America.
| | - Jessica Spinelli
- Department of Communication Sciences & Disorders, University of Iowa, 250 Hawkins Drive, 52242 Iowa City, IA, United States of America.
| | - Elizabeth Walker
- Department of Communication Sciences & Disorders, University of Iowa, 250 Hawkins Drive, 52242 Iowa City, IA, United States of America.
| |
Collapse
|
13
|
Nagels L, Bastiaanse R, Başkent D, Wagner A. Individual Differences in Lexical Access Among Cochlear Implant Users. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:286-304. [PMID: 31855606 DOI: 10.1044/2019_jslhr-19-00192] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Purpose The current study investigates how individual differences in cochlear implant (CI) users' sensitivity to word-nonword differences, reflecting lexical uncertainty, relate to their reliance on sentential context for lexical access in processing continuous speech. Method Fifteen CI users and 14 normal-hearing (NH) controls participated in an auditory lexical decision task (Experiment 1) and a visual-world paradigm task (Experiment 2). Experiment 1 tested participants' reliance on lexical statistics, and Experiment 2 studied how sentential context affects the time course and patterns of lexical competition leading to lexical access. Results In Experiment 1, CI users had lower accuracy scores and longer reaction times than NH listeners, particularly for nonwords. In Experiment 2, CI users' lexical competition patterns were, on average, similar to those of NH listeners, but the patterns of individual CI users varied greatly. Individual CI users' word-nonword sensitivity (Experiment 1) explained differences in the reliance on sentential context to resolve lexical competition, whereas clinical speech perception scores explained competition with phonologically related words. Conclusions The general analysis of CI users' lexical competition patterns showed merely quantitative differences with NH listeners in the time course of lexical competition, but our additional analysis revealed more qualitative differences in CI users' strategies to process speech. Individuals' word-nonword sensitivity explained different parts of individual variability than clinical speech perception scores. These results stress, particularly for heterogeneous clinical populations such as CI users, the importance of investigating individual differences in addition to group averages, as they can be informative for clinical rehabilitation. Supplemental Material https://doi.org/10.23641/asha.11368106.
Collapse
Affiliation(s)
- Leanne Nagels
- Department of Otorhinolaryngology-Head & Neck Surgery, University Medical Center Groningen, University of Groningen, the Netherlands
- Center for Language and Cognition Groningen, University of Groningen, the Netherlands
| | - Roelien Bastiaanse
- Center for Language and Cognition Groningen, University of Groningen, the Netherlands
- National Research University Higher School of Economics, Moscow, Russia
| | - Deniz Başkent
- Department of Otorhinolaryngology-Head & Neck Surgery, University Medical Center Groningen, University of Groningen, the Netherlands
- Research School of Behavioural and Cognitive Neurosciences, Graduate School of Medical Sciences, University of Groningen, the Netherlands
| | - Anita Wagner
- Department of Otorhinolaryngology-Head & Neck Surgery, University Medical Center Groningen, University of Groningen, the Netherlands
- Research School of Behavioural and Cognitive Neurosciences, Graduate School of Medical Sciences, University of Groningen, the Netherlands
| |
Collapse
|
14
|
McMurray B, Ellis TP, Apfelbaum KS. How Do You Deal With Uncertainty? Cochlear Implant Users Differ in the Dynamics of Lexical Processing of Noncanonical Inputs. Ear Hear 2020; 40:961-980. [PMID: 30531260 PMCID: PMC6551335 DOI: 10.1097/aud.0000000000000681] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES Work in normal-hearing (NH) adults suggests that spoken language processing involves coping with ambiguity. Even a clearly spoken word contains brief periods of ambiguity as it unfolds over time, and early portions will not be sufficient to uniquely identify the word. However, beyond this temporary ambiguity, NH listeners must also cope with the loss of information due to reduced forms, dialect, and other factors. A recent study suggests that NH listeners may adapt to increased ambiguity by changing the dynamics of how they commit to candidates at a lexical level. Cochlear implant (CI) users must also frequently deal with highly degraded input, in which there is less information available in the input to recover a target word. The authors asked here whether their frequent experience with this leads to lexical dynamics that are better suited for coping with uncertainty. DESIGN Listeners heard words either correctly pronounced (dog) or mispronounced at onset (gog) or offset (dob). Listeners selected the corresponding picture from a screen containing pictures of the target and three unrelated items. While they did this, fixations to each object were tracked as a measure of the time course of identifying the target. The authors tested 44 postlingually deafened adult CI users in 2 groups (23 used standard electric only configurations, and 21 supplemented the CI with a hearing aid), along with 28 age-matched age-typical hearing (ATH) controls. RESULTS All three groups recognized the target word accurately, though each showed a small decrement for mispronounced forms (larger in both types of CI users). Analysis of fixations showed a close time locking to the timing of the mispronunciation. Onset mispronunciations delayed initial fixations to the target, but fixations to the target showed partial recovery by the end of the trial. Offset mispronunciations showed no effect early, but suppressed looking later. This pattern was attested in all three groups, though both types of CI users were slower and did not commit fully to the target. When the authors quantified the degree of disruption (by the mispronounced forms), they found that both groups of CI users showed less disruption than ATH listeners during the first 900 msec of processing. Finally, an individual differences analysis showed that within the CI users, the dynamics of fixations predicted speech perception outcomes over and above accuracy in this task and that CI users with the more rapid fixation patterns of ATH listeners showed better outcomes. CONCLUSIONS Postlingually deafened CI users process speech incrementally (as do ATH listeners), though they commit more slowly and less strongly to a single item than do ATH listeners. This may allow them to cope more flexible with mispronunciations.
Collapse
Affiliation(s)
- Bob McMurray
- Departments of Psychological and Brain Sciences, Communication Sciences and Disorders, Otolaryngology, University of Iowa, Iowa City, Iowa, USA
| | - Tyler P Ellis
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City, Iowa, USA
| | - Keith S Apfelbaum
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City, Iowa, USA
- Foundations in Learning, Inc., Coralville, Iowa, USA
| |
Collapse
|
15
|
Amichetti NM, Atagi E, Kong YY, Wingfield A. Linguistic Context Versus Semantic Competition in Word Recognition by Younger and Older Adults With Cochlear Implants. Ear Hear 2019; 39:101-109. [PMID: 28700448 PMCID: PMC5741484 DOI: 10.1097/aud.0000000000000469] [Citation(s) in RCA: 17] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
Abstract
OBJECTIVES The increasing numbers of older adults now receiving cochlear implants raises the question of how the novel signal produced by cochlear implants may interact with cognitive aging in the recognition of words heard spoken within a linguistic context. The objective of this study was to pit the facilitative effects of a constraining linguistic context against a potential age-sensitive negative effect of response competition on effectiveness of word recognition. DESIGN Younger (n = 8; mean age = 22.5 years) and older (n = 8; mean age = 67.5 years) adult implant recipients heard 20 target words as the final words in sentences that manipulated the target word's probability of occurrence within the sentence context. Data from published norms were also used to measure response entropy, calculated as the total number of different responses and the probability distribution of the responses suggested by the sentence context. Sentence-final words were presented to participants using a word-onset gating paradigm, in which a target word was presented with increasing amounts of its onset duration in 50 msec increments until the word was correctly identified. RESULTS Results showed that for both younger and older adult implant users, the amount of word-onset information needed for correct recognition of sentence-final words was inversely proportional to their likelihood of occurrence within the sentence context, with older adults gaining differential advantage from the contextual constraints offered by a sentence context. On the negative side, older adults' word recognition was differentially hampered by high response entropy, with this effect being driven primarily by the number of competing responses that might also fit the sentence context. CONCLUSIONS Consistent with previous research with normal-hearing younger and older adults, the present results showed older adult implant users' recognition of spoken words to be highly sensitive to linguistic context. This sensitivity, however, also resulted in a greater degree of interference from other words that might also be activated by the context, with negative effects on ease of word recognition. These results are consistent with an age-related inhibition deficit extending to the domain of semantic constraints on word recognition.
Collapse
Affiliation(s)
- Nicole M. Amichetti
- Volen National Center for Complex Systems, Brandeis University, Waltham, MA, USA
| | - Eriko Atagi
- Volen National Center for Complex Systems, Brandeis University, Waltham, MA, USA
- Department of Communication Sciences and Disorders, Northeastern University, Boston, MA, USA
| | - Ying-Yee Kong
- Department of Communication Sciences and Disorders, Northeastern University, Boston, MA, USA
| | - Arthur Wingfield
- Volen National Center for Complex Systems, Brandeis University, Waltham, MA, USA
| |
Collapse
|
16
|
Kaandorp MW, Smits C, Merkus P, Festen JM, Goverts ST. Lexical-Access Ability and Cognitive Predictors of Speech Recognition in Noise in Adult Cochlear Implant Users. Trends Hear 2019; 21:2331216517743887. [PMID: 29205095 PMCID: PMC5721962 DOI: 10.1177/2331216517743887] [Citation(s) in RCA: 16] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/10/2023] Open
Abstract
Not all of the variance in speech-recognition performance of cochlear implant (CI) users can be explained by biographic and auditory factors. In normal-hearing listeners, linguistic and cognitive factors determine most of speech-in-noise performance. The current study explored specifically the influence of visually measured lexical-access ability compared with other cognitive factors on speech recognition of 24 postlingually deafened CI users. Speech-recognition performance was measured with monosyllables in quiet (consonant-vowel-consonant [CVC]), sentences-in-noise (SIN), and digit-triplets in noise (DIN). In addition to a composite variable of lexical-access ability (LA), measured with a lexical-decision test (LDT) and word-naming task, vocabulary size, working-memory capacity (Reading Span test [RSpan]), and a visual analogue of the SIN test (text reception threshold test) were measured. The DIN test was used to correct for auditory factors in SIN thresholds by taking the difference between SIN and DIN: SRTdiff. Correlation analyses revealed that duration of hearing loss (dHL) was related to SIN thresholds. Better working-memory capacity was related to SIN and SRTdiff scores. LDT reaction time was positively correlated with SRTdiff scores. No significant relationships were found for CVC or DIN scores with the predictor variables. Regression analyses showed that together with dHL, RSpan explained 55% of the variance in SIN thresholds. When controlling for auditory performance, LA, LDT, and RSpan separately explained, together with dHL, respectively 37%, 36%, and 46% of the variance in SRTdiff outcome. The results suggest that poor verbal working-memory capacity and to a lesser extent poor lexical-access ability limit speech-recognition ability in listeners with a CI.
Collapse
Affiliation(s)
- Marre W Kaandorp
- 1 Department of Otolaryngology-Head and Neck Surgery, Section Ear & Hearing and EMGO Institute for Health and Care Research, VU University Medical Center, Amsterdam, The Netherlands
| | - Cas Smits
- 1 Department of Otolaryngology-Head and Neck Surgery, Section Ear & Hearing and EMGO Institute for Health and Care Research, VU University Medical Center, Amsterdam, The Netherlands
| | - Paul Merkus
- 1 Department of Otolaryngology-Head and Neck Surgery, Section Ear & Hearing and EMGO Institute for Health and Care Research, VU University Medical Center, Amsterdam, The Netherlands
| | - Joost M Festen
- 1 Department of Otolaryngology-Head and Neck Surgery, Section Ear & Hearing and EMGO Institute for Health and Care Research, VU University Medical Center, Amsterdam, The Netherlands
| | - S Theo Goverts
- 1 Department of Otolaryngology-Head and Neck Surgery, Section Ear & Hearing and EMGO Institute for Health and Care Research, VU University Medical Center, Amsterdam, The Netherlands
| |
Collapse
|
17
|
McMurray B, Danelz A, Rigler H, Seedorff M. Speech categorization develops slowly through adolescence. Dev Psychol 2018; 54:1472-1491. [PMID: 29952600 PMCID: PMC6062449 DOI: 10.1037/dev0000542] [Citation(s) in RCA: 35] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
The development of the ability to categorize speech sounds is often viewed as occurring primarily during infancy via perceptual learning mechanisms. However, a number of studies suggest that even after infancy, children's categories become more categorical and well defined through about age 12. We investigated the cognitive changes that may be responsible for such development using a visual world paradigm experiment based on (McMurray, Tanenhaus, & Aslin, 2002). Children from 3 age groups (7-8, 12-13, and 17-18 years) heard a token from either a b/p or s/∫ continua spanning 2 words (beach/peach, ship/sip) and selected its referent from a screen containing 4 pictures of potential lexical candidates. Eye movements to each object were monitored as a measure of how strongly children were committing to each candidate as perception unfolds in real-time. Results showed an ongoing sharpening of speech categories through 18, which was particularly apparent during the early stages of real-time perception. When analysis targeted to specifically within-category sensitivity to continuous detail, children exhibited increasingly gradient categories over development, suggesting that increasing sensitivity to fine-grained detail in the signal enables these more discrete categorizations. Together these suggest that speech development is a protracted process in which children's increasing sensitivity to within-category detail in the signal enables increasingly sharp phonetic categories. (PsycINFO Database Record
Collapse
Affiliation(s)
- Bob McMurray
- Department of Psychological and Brain Sciences, University of Iowa
| | - Ani Danelz
- Department of Communication Sciences and Disorders, University of Iowa
| | | | | |
Collapse
|
18
|
Van Engen KJ, McLaughlin DJ. Eyes and ears: Using eye tracking and pupillometry to understand challenges to speech recognition. Hear Res 2018; 369:56-66. [PMID: 29801981 DOI: 10.1016/j.heares.2018.04.013] [Citation(s) in RCA: 20] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/03/2017] [Revised: 04/12/2018] [Accepted: 04/25/2018] [Indexed: 11/16/2022]
Abstract
Although human speech recognition is often experienced as relatively effortless, a number of common challenges can render the task more difficult. Such challenges may originate in talkers (e.g., unfamiliar accents, varying speech styles), the environment (e.g. noise), or in listeners themselves (e.g., hearing loss, aging, different native language backgrounds). Each of these challenges can reduce the intelligibility of spoken language, but even when intelligibility remains high, they can place greater processing demands on listeners. Noisy conditions, for example, can lead to poorer recall for speech, even when it has been correctly understood. Speech intelligibility measures, memory tasks, and subjective reports of listener difficulty all provide critical information about the effects of such challenges on speech recognition. Eye tracking and pupillometry complement these methods by providing objective physiological measures of online cognitive processing during listening. Eye tracking records the moment-to-moment direction of listeners' visual attention, which is closely time-locked to unfolding speech signals, and pupillometry measures the moment-to-moment size of listeners' pupils, which dilate in response to increased cognitive load. In this paper, we review the uses of these two methods for studying challenges to speech recognition.
Collapse
|
19
|
Simeon KM, Bicknell K, Grieco-Calub TM. Belief Shift or Only Facilitation: How Semantic Expectancy Affects Processing of Speech Degraded by Background Noise. Front Psychol 2018; 9:116. [PMID: 29472883 PMCID: PMC5809983 DOI: 10.3389/fpsyg.2018.00116] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2017] [Accepted: 01/24/2018] [Indexed: 11/13/2022] Open
Abstract
Individuals use semantic expectancy - applying conceptual and linguistic knowledge to speech input - to improve the accuracy and speed of language comprehension. This study tested how adults use semantic expectancy in quiet and in the presence of speech-shaped broadband noise at -7 and -12 dB signal-to-noise ratio. Twenty-four adults (22.1 ± 3.6 years, mean ±SD) were tested on a four-alternative-forced-choice task whereby they listened to sentences and were instructed to select an image matching the sentence-final word. The semantic expectancy of the sentences was unrelated to (neutral), congruent with, or conflicting with the acoustic target. Congruent expectancy improved accuracy and conflicting expectancy decreased accuracy relative to neutral, consistent with a theory where expectancy shifts beliefs toward likely words and away from unlikely words. Additionally, there were no significant interactions of expectancy and noise level when analyzed in log-odds, supporting the predictions of ideal observer models of speech perception.
Collapse
Affiliation(s)
- Katherine M. Simeon
- The Roxelyn and Richard Pepper Department of Communication Sciences and Disorders, Northwestern University, Evanston, IL, United States
| | - Klinton Bicknell
- Department of Linguistics, Northwestern University, Evanston, IL, United States
| | - Tina M. Grieco-Calub
- The Roxelyn and Richard Pepper Department of Communication Sciences and Disorders, Northwestern University, Evanston, IL, United States
- Hugh Knowles Hearing Center, Northwestern University, Evanston, IL, United States
| |
Collapse
|
20
|
McMurray B, Farris-Trimble A, Rigler H. Waiting for lexical access: Cochlear implants or severely degraded input lead listeners to process speech less incrementally. Cognition 2017; 169:147-164. [PMID: 28917133 DOI: 10.1016/j.cognition.2017.08.013] [Citation(s) in RCA: 50] [Impact Index Per Article: 7.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2016] [Revised: 08/27/2017] [Accepted: 08/31/2017] [Indexed: 11/30/2022]
Abstract
Spoken language unfolds over time. Consequently, there are brief periods of ambiguity, when incomplete input can match many possible words. Typical listeners solve this problem by immediately activating multiple candidates which compete for recognition. In two experiments using the visual world paradigm, we examined real-time lexical competition in prelingually deaf cochlear implant (CI) users, and normal hearing (NH) adults listening to severely degraded speech. In Experiment 1, adolescent CI users and NH controls matched spoken words to arrays of pictures including pictures of the target word and phonological competitors. Eye-movements to each referent were monitored asa measure of how strongly that candidate was considered over time. Relative to NH controls, CI users showed a large delay in fixating any object, less competition from onset competitors (e.g., sandwich after hearing sandal), and increased competition from rhyme competitors (e.g., candle after hearing sandal). Experiment 2 observed the same pattern with NH listeners hearing highly degraded speech. These studies suggests that in contrast to all prior studies of word recognition in typical listeners, listeners recognizing words in severely degraded conditions can exhibit a substantively different pattern of dynamics, waiting to begin lexical access until substantial information has accumulated.
Collapse
Affiliation(s)
- Bob McMurray
- Dept. of Psychological and Brain Sciences, University of Iowa, United States; Dept. of Communication Sciences and Disorders, University of Iowa, United States; Dept. of Otolaryngology, University of Iowa, United States; DeLTA Center, University of Iowa, United States.
| | | | - Hannah Rigler
- Dept. of Psychological and Brain Sciences, University of Iowa, United States
| |
Collapse
|