1
|
Feng Y, Peng G. Development of categorical speech perception in Mandarin-speaking children and adolescents. Child Dev 2023; 94:28-43. [PMID: 35920586 PMCID: PMC10087708 DOI: 10.1111/cdev.13837] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
Abstract
Although children develop categorical speech perception at a very young age, the maturation process remains unclear. A cross-sectional study in Mandarin-speaking 4-, 6-, and 10-year-old children, 14-year-old adolescents, and adults (n = 104, 56 males, all Asians from mainland China) was conducted to investigate the development of categorical perception of four Mandarin phonemic contrasts: lexical tone contrast Tone 1-2, vowel contrast /u/-/i/, consonant aspiration contrast /p/-/ph /, and consonant formant transition contrast /p/-/t/. The results indicated that different types of phonemic contrasts, and even the identification and discrimination of the same phonemic contrast, matured asynchronously. The observation that tone and vowel perception are achieved earlier than consonant perception supports the phonological saliency hypothesis.
Collapse
Affiliation(s)
- Yan Feng
- School of Foreign Studies, Nanjing University of Science and Technology, Nanjing, China.,Department of Chinese and Bilingual Studies, Research Centre for Language, Cognition, and Neuroscience, The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Gang Peng
- Department of Chinese and Bilingual Studies, Research Centre for Language, Cognition, and Neuroscience, The Hong Kong Polytechnic University, Hong Kong SAR, China.,Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| |
Collapse
|
2
|
McMurray B. The acquisition of speech categories: Beyond perceptual narrowing, beyond unsupervised learning and beyond infancy. LANGUAGE, COGNITION AND NEUROSCIENCE 2022; 38:419-445. [PMID: 38425732 PMCID: PMC10904032 DOI: 10.1080/23273798.2022.2105367] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/24/2022] [Accepted: 07/01/2022] [Indexed: 03/02/2024]
Abstract
An early achievement in language is carving a variable acoustic space into categories. The canonical story is that infants accomplish this by the second year, when only unsupervised learning is plausible. I challenge this view, synthesizing five lines of developmental, phonetic and computational work. First, unsupervised learning may be insufficient given the statistics of speech (including infant-directed). Second, evidence that infants "have" speech categories rests on tenuous methodological assumptions. Third, the fact that the ecology of the learning environment is unsupervised does not rule out more powerful error driven learning mechanisms. Fourth, several implicit supervisory signals are available to older infants. Finally, development is protracted through adolescence, enabling richer avenues for development. Infancy may be a time of organizing the auditory space, but true categorization only arises via complex developmental cascades later in life. This has implications for critical periods, second language acquisition, and our basic framing of speech perception.
Collapse
Affiliation(s)
- Bob McMurray
- Dept. of Psychological and Brain Sciences, Dept. of Communication Sciences and Disorders, Dept. of Linguistics, University of Iowa and Haskins Laboratories
| |
Collapse
|
3
|
Feldman NH, Goldwater S, Dupoux E, Schatz T. Do Infants Really Learn Phonetic Categories? OPEN MIND 2022; 5:113-131. [PMID: 35024527 PMCID: PMC8746127 DOI: 10.1162/opmi_a_00046] [Citation(s) in RCA: 14] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2020] [Accepted: 08/06/2021] [Indexed: 11/16/2022] Open
Abstract
Early changes in infants’ ability to perceive native and nonnative speech sound contrasts are typically attributed to their developing knowledge of phonetic categories. We critically examine this hypothesis and argue that there is little direct evidence of category knowledge in infancy. We then propose an alternative account in which infants’ perception changes because they are learning a perceptual space that is appropriate to represent speech, without yet carving up that space into phonetic categories. If correct, this new account has substantial implications for understanding early language development.
Collapse
Affiliation(s)
- Naomi H Feldman
- Department of Linguistics and UMIACS, University of Maryland, College Park, MD, USA
| | | | - Emmanuel Dupoux
- Cognitive Machine Learning (ENS - EHESS - PSL Research University - CNRS - INRIA), Paris, France
| | - Thomas Schatz
- Department of Linguistics and UMIACS, University of Maryland, College Park, MD, USA
| |
Collapse
|
4
|
Voice Emotion Recognition by Mandarin-Speaking Children with Cochlear Implants. Ear Hear 2021; 43:165-180. [PMID: 34288631 DOI: 10.1097/aud.0000000000001085] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
Objectives Emotional expressions are very important in social interactions. Children with cochlear implants can have voice emotion recognition deficits due to device limitations. Mandarin-speaking children with cochlear implants may face greater challenges than those speaking nontonal languages; the pitch information is not well preserved in cochlear implants, and such children could benefit from child-directed speech, which carries more exaggerated distinctive acoustic cues for different emotions. This study investigated voice emotion recognition, using both adult-directed and child-directed materials, in Mandarin-speaking children with cochlear implants compared with normal hearing peers. The authors hypothesized that both the children with cochlear implants and those with normal hearing would perform better with child-directed materials than with adult-directed materials. Design Thirty children (7.17-17 years of age) with cochlear implants and 27 children with normal hearing (6.92-17.08 years of age) were recruited in this study. Participants completed a nonverbal reasoning test, speech recognition tests, and a voice emotion recognition task. Children with cochlear implants over the age of 10 years also completed the Chinese version of the Nijmegen Cochlear Implant Questionnaire to evaluate the health-related quality of life. The voice emotion recognition task was a five-alternative, forced-choice paradigm, which contains sentences spoken with five emotions (happy, angry, sad, scared, and neutral) in a child-directed or adult-directed manner. Results Acoustic analyses showed substantial variations across emotions in all materials, mainly on measures of mean fundamental frequency and fundamental frequency range. Mandarin-speaking children with cochlear implants displayed a significantly poorer performance than normal hearing peers in voice emotion perception tasks, regardless of whether the performance is measured in accuracy scores, Hu value, or reaction time. Children with cochlear implants and children with normal hearing were mainly affected by the mean fundamental frequency in speech emotion recognition tasks. Chronological age had a significant effect on speech emotion recognition in children with normal hearing; however, there was no significant correlation between chronological age and accuracy scores in speech emotion recognition in children with implants. Significant effects of specific emotion and test materials (better performance with child-directed materials) in both groups of children were observed. Among the children with cochlear implants, age at implantation, percentage scores of nonverbal intelligence quotient test, and sentence recognition threshold in quiet could predict recognition performance in both accuracy scores and Hu values. Time wearing cochlear implant could predict reaction time in emotion perception tasks among children with cochlear implants. No correlation was observed between the accuracy score in voice emotion perception and the self-reported scores of health-related quality of life; however, the latter were significantly correlated with speech recognition skills among Mandarin-speaking children with cochlear implants. Conclusions Mandarin-speaking children with cochlear implants could have significant deficits in voice emotion recognition tasks compared with their normally hearing peers and can benefit from the exaggerated prosody of child-directed speech. The effects of age at cochlear implantation, speech and language development, and cognition could play an important role in voice emotion perception by Mandarin-speaking children with cochlear implants.
Collapse
|
5
|
Ma J, Zhu J, Yang Y, Chen F. The Development of Categorical Perception of Segments and Suprasegments in Mandarin-Speaking Preschoolers. Front Psychol 2021; 12:693366. [PMID: 34354636 PMCID: PMC8329735 DOI: 10.3389/fpsyg.2021.693366] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2021] [Accepted: 05/27/2021] [Indexed: 11/13/2022] Open
Abstract
This study investigated the developmental trajectories of categorical perception (CP) of segments (i.e., stops) and suprasegments (i.e., lexical tones) in an attempt to examine the perceptual development of phonological categories and whether CP of suprasegments develops in parallel with that of segments. Forty-seven Mandarin-speaking monolingual preschoolers aged four to six years old, and fourteen adults completed both identification and discrimination tasks of the Tone 1-2 continuum and the /pa/-/pha/ continuum. Results revealed that children could perceive both lexical tones and aspiration of stops in a categorical manner by age four. The boundary position did not depend on age, with children having similar positions to adults regardless of speech continuum types. The boundary width, on the other hand, reached the adult-like level at age six for lexical tones, but not for stops. In addition, the within-category discrimination score did not differ significantly between children and adults for both continua. The between-category discrimination score improved with age and achieved the adult-like level at age five for lexical tones, but still not for stops even at age six. It suggests that the fine-grained perception of phonological categories is a protracted process, and the improvement and varying timeline of the development of segments and suprasegments are discussed in relation to statistical learning of the regularities of speech sounds in ambient language, ongoing maturation of perceptual systems, the memory mechanism underlying perceptual learning, and the intrinsic nature of speech elements.
Collapse
Affiliation(s)
- Junzhou Ma
- School of Foreign Languages, Taizhou University, Taizhou, China
| | - Jiaqiang Zhu
- School of Foreign Languages, Hunan University, Changsha, China
| | - Yuxiao Yang
- Foreign Studies College, Hunan Normal University, Changsha, China
| | - Fei Chen
- School of Foreign Languages, Hunan University, Changsha, China
| |
Collapse
|
6
|
Nagels L, Gaudrain E, Vickers D, Hendriks P, Başkent D. School-age children benefit from voice gender cue differences for the perception of speech in competing speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:3328. [PMID: 34241121 DOI: 10.1121/10.0004791] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/24/2020] [Accepted: 04/08/2021] [Indexed: 06/13/2023]
Abstract
Differences in speakers' voice characteristics, such as mean fundamental frequency (F0) and vocal-tract length (VTL), that primarily define speakers' so-called perceived voice gender facilitate the perception of speech in competing speech. Perceiving speech in competing speech is particularly challenging for children, which may relate to their lower sensitivity to differences in voice characteristics than adults. This study investigated the development of the benefit from F0 and VTL differences in school-age children (4-12 years) for separating two competing speakers while tasked with comprehending one of them and also the relationship between this benefit and their corresponding voice discrimination thresholds. Children benefited from differences in F0, VTL, or both cues at all ages tested. This benefit proportionally remained the same across age, although overall accuracy continued to differ from that of adults. Additionally, children's benefit from F0 and VTL differences and their overall accuracy were not related to their discrimination thresholds. Hence, although children's voice discrimination thresholds and speech in competing speech perception abilities develop throughout the school-age years, children already show a benefit from voice gender cue differences early on. Factors other than children's discrimination thresholds seem to relate more closely to their developing speech in competing speech perception abilities.
Collapse
Affiliation(s)
- Leanne Nagels
- Center for Language and Cognition Groningen (CLCG), University of Groningen, Groningen 9712EK, Netherlands
| | - Etienne Gaudrain
- CNRS UMR 5292, Lyon Neuroscience Research Center, Auditory Cognition and Psychoacoustics, Inserm UMRS 1028, Université Claude Bernard Lyon 1, Université de Lyon, Lyon, France
| | - Deborah Vickers
- Sound Lab, Cambridge Hearing Group, Clinical Neurosciences Department, University of Cambridge, Cambridge CB2 0SZ, United Kingdom
| | - Petra Hendriks
- Center for Language and Cognition Groningen (CLCG), University of Groningen, Groningen 9712EK, Netherlands
| | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen 9713GZ, Netherlands
| |
Collapse
|
7
|
Aberrant auditory system and its developmental implications for autism. SCIENCE CHINA-LIFE SCIENCES 2021; 64:861-878. [DOI: 10.1007/s11427-020-1863-6] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/20/2020] [Accepted: 12/06/2020] [Indexed: 12/26/2022]
|
8
|
Flaherty MM, Buss E, Leibold LJ. Independent and Combined Effects of Fundamental Frequency and Vocal Tract Length Differences for School-Age Children's Sentence Recognition in a Two-Talker Masker. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2021; 64:206-217. [PMID: 33375828 PMCID: PMC8610228 DOI: 10.1044/2020_jslhr-20-00327] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/08/2020] [Revised: 09/08/2020] [Accepted: 09/29/2020] [Indexed: 06/12/2023]
Abstract
Purpose The purpose of this study was to examine the independent and combined contributions of fundamental frequency (F0) and vocal tract length (VTL) differences on children's speech-in-speech recognition in the presence of a competing two-talker masker. Method Participants were 64 children (5-17 years old) and 25 adults (18-39 years old). Sentence recognition thresholds were measured in a two-talker masker. Target sentences had either the same mean F0 and VTL of the masker or were digitally altered so that the target and masker differed in F0 (Experiment 1), differed in VTL (Experiment 2), or differed in both F0 and VTL (Experiment 3). To determine the benefit, masking release was computed by subtracting thresholds in each shifted condition from the threshold in the unshifted condition. Results Results demonstrate that children's ability to benefit from either F0 or VTL differences (Experiments 1 and 2) depended on listener age, with younger children showing less improvement in speech reception thresholds compared to older children and adults. Age effects were also evident in the combined-cue conditions (Experiment 3), but children showed greater improvements compared to F0-only or VTL-only manipulations. Conclusions There was a prolonged pattern of development in children's ability to benefit from F0 or VTL differences between target and masker speech. Young children failed to capitalize on F0 and VTL differences to the same extent as older children and adults but did show a robust benefit when the cues were combined, supporting the hypothesis that younger children rely more heavily on redundant cues compared to older children and adults.
Collapse
Affiliation(s)
- Mary M. Flaherty
- Department of Speech and Hearing Science, University of Illinois at Urbana-Champaign
| | - Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, School of Medicine, University of North Carolina at Chapel Hill
| | - Lori J. Leibold
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, NE
| |
Collapse
|
9
|
Cummings AE, Ogiela DA, Wu YC. Evidence for [Coronal] Underspecification in Typical and Atypical Phonological Development. Front Hum Neurosci 2021; 14:580697. [PMID: 33414710 PMCID: PMC7782969 DOI: 10.3389/fnhum.2020.580697] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2020] [Accepted: 11/30/2020] [Indexed: 11/13/2022] Open
Abstract
The Featurally Underspecified Lexicon (FUL) theory predicts that [coronal] is the language universal default place of articulation for phonemes. This assumption has been consistently supported with adult behavioral and event-related potential (ERP) data; however, this underspecification claim has not been tested in developmental populations. The purpose of this study was to determine whether children demonstrate [coronal] underspecification patterns similar to those of adults. Two English consonants differing in place of articulation, [labial] /b/ and [coronal] /d/, were presented to 24 children (ages 4-6 years) characterized by either a typically developing phonological system (TD) or a phonological disorder (PD). Two syllables, /bɑ/ and /dɑ/, were presented in an ERP oddball paradigm where both syllables served as the standard and deviant stimulus in opposite stimulus sets. Underspecification was examined with three analyses: traditional mean amplitude measurements, cluster-based permutation tests, and single-trial general linear model (GLM) analyses of single-subject data. Contrary to previous adult findings, children with PD demonstrated a large positive mismatch response (PMR) to /bɑ/ while the children with TD exhibited a negative mismatch response (MMN); significant group differences were not observed in the /dɑ/ responses. Moreover, the /bɑ/ deviant ERP response was significantly larger in the TD children than in the children with PD. At the single-subject level, more children demonstrated mismatch responses to /dɑ/ than to /bɑ/, though some children had a /bɑ/ mismatch response and no /dɑ/ mismatch response. While both groups of children demonstrated similar responses to the underspecified /dɑ/, their neural responses to the more specified /bɑ/ varied. These findings are interpreted within a proposed developmental model of phonological underspecification, wherein children with PD are functioning at a developmentally less mature stage of phonological acquisition than their same-aged TD peers. Thus, phonological underspecification is a phenomenon that likely develops over time with experience and exposure to language.
Collapse
Affiliation(s)
- Alycia E Cummings
- Department of Communication Sciences and Disorders, Idaho State University, Meridian, ID, United States
| | - Diane A Ogiela
- Department of Communication Sciences and Disorders, Idaho State University, Meridian, ID, United States
| | - Ying C Wu
- Swartz Center for Computational Neuroscience, University of California, San Diego, San Diego, CA, United States
| |
Collapse
|
10
|
Hitchcock ER, Cabbage KL, Swartz MT, Carrell TD. Measuring Speech Perception Using the Wide-Range Acoustic Accuracy Scale: Preliminary Findings. ACTA ACUST UNITED AC 2020. [DOI: 10.1044/2020_persp-20-00037] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Purpose
Speech perception requires individuals to hear and differentiate acoustic signals integral to effective communication. Measuring speech perception in children is challenging because speech perception methodology typically requires lengthy experiments that may fatigue children, resulting in limited knowledge of developmental perceptual skills. In this study, we used an adaptive tracking measure along with a wide range of acoustic stimuli to explore how adults, typically developing (TD) children, and children with speech sound disorder (SSD) perceive small acoustic differences in synthetic speech stimuli.
Method
Twenty-four adults, 15 TD children, and 15 children with SSD between the ages of 7;0–14;0 (years;months) were administered a newly developed perceptual assessment, the Wide-Range Acoustic Accuracy Scale, to determine the just-noticeable difference in discrimination for three separate syllable contrasts. Each syllable contrast varied along a single acoustic parameter: formant transition duration for /bɑ/–/wɑ/, F3 onset frequency for /dɑ/–/ɡɑ/, and F3–F2 distance for /rɑ/–/wɑ/.
Results
Findings revealed that adults and TD children did not differ in their discrimination of any syllable contrast, but adults significantly differed from SSD children on all syllable contrasts. TD children and children with SSD differed only on the /rɑ/–/wɑ/ contrast.
Conclusions
Children with SSD demonstrate less accurate and more variable perception skills relative to adults and TD children for /bɑ/–/wɑ/, /dɑ/–/ɡɑ/, and /rɑ/–/wɑ/ syllable contrasts. Clinical implications of the utility of the Wide-Range Acoustic Accuracy Scale are discussed.
Collapse
Affiliation(s)
- Elaine R. Hitchcock
- Department of Communication Science and Disorders, Montclair State University, Bloomfield, NJ
| | - Kathryn L. Cabbage
- Department of Communication Disorders, Brigham Young University, Provo, UT
| | - Michelle T. Swartz
- Department of Communication Science and Disorders, Montclair State University, Bloomfield, NJ
| | - Thomas D. Carrell
- Deparment of Communication Sciences and Disorders, University of Nebraska–Lincoln
| |
Collapse
|
11
|
DiNino M, Arenberg JG, Duchen ALR, Winn MB. Effects of Age and Cochlear Implantation on Spectrally Cued Speech Categorization. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:2425-2440. [PMID: 32552327 PMCID: PMC7838840 DOI: 10.1044/2020_jslhr-19-00127] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/06/2019] [Revised: 08/12/2019] [Accepted: 03/30/2020] [Indexed: 06/11/2023]
Abstract
Purpose Weighting of acoustic cues for perceiving place-of-articulation speech contrasts was measured to determine the separate and interactive effects of age and use of cochlear implants (CIs). It has been found that adults with normal hearing (NH) show reliance on fine-grained spectral information (e.g., formants), whereas adults with CIs show reliance on broad spectral shape (e.g., spectral tilt). In question was whether children with NH and CIs would demonstrate the same patterns as adults, or show differences based on ongoing maturation of hearing and phonetic skills. Method Children and adults with NH and with CIs categorized a /b/-/d/ speech contrast based on two orthogonal spectral cues. Among CI users, phonetic cue weights were compared to vowel identification scores and Spectral-Temporally Modulated Ripple Test thresholds. Results NH children and adults both relied relatively more on the fine-grained formant cue and less on the broad spectral tilt cue compared to participants with CIs. However, early-implanted children with CIs better utilized the formant cue compared to adult CI users. Formant cue weights correlated with CI participants' vowel recognition and in children, also related to Spectral-Temporally Modulated Ripple Test thresholds. Adults and child CI users with very poor phonetic perception showed additive use of the two cues, whereas those with better and/or more mature cue usage showed a prioritized trading relationship, akin to NH listeners. Conclusions Age group and hearing modality can influence phonetic cue-weighting patterns. Results suggest that simple nonlexical categorization tests correlate with more general speech recognition skills of children and adults with CIs.
Collapse
Affiliation(s)
- Mishaela DiNino
- Department of Psychology, Carnegie Mellon University, Pittsburgh, PA
| | - Julie G. Arenberg
- Massachusetts Eye and Ear, Harvard Medical School Department of Otolaryngology, Boston
| | | | - Matthew B. Winn
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis
| |
Collapse
|
12
|
Yu YH, Tessel C, Han X, Campanelli L, Vidal N, Gerometta J, Garrido-Nag K, Datta H, Shafer VL. Neural Indices of Vowel Discrimination in Monolingual and Bilingual Infants and Children. Ear Hear 2020; 40:1376-1390. [PMID: 31033699 PMCID: PMC6814506 DOI: 10.1097/aud.0000000000000726] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES To examine maturation of neural discriminative responses to an English vowel contrast from infancy to 4 years of age and to determine how biological factors (age and sex) and an experiential factor (amount of Spanish versus English input) modulate neural discrimination of speech. DESIGN Event-related potential (ERP) mismatch responses (MMRs) were used as indices of discrimination of the American English vowels [ε] versus [I] in infants and children between 3 months and 47 months of age. A total of 168 longitudinal and cross-sectional data sets were collected from 98 children (Bilingual Spanish-English: 47 male and 31 female sessions; Monolingual English: 48 male and 42 female sessions). Language exposure and other language measures were collected. ERP responses were examined in an early time window (160 to 360 msec, early MMR [eMMR]) and late time window (400 to 600 msec, late MMR). RESULTS The eMMR became more negative with increasing age. Language experience and sex also influenced the amplitude of the eMMR. Specifically, bilingual children, especially bilingual females, showed more negative eMMR compared with monolingual children and with males. However, the subset of bilingual children with more exposure to English than Spanish compared with those with more exposure to Spanish than English (as reported by caretakers) showed similar amplitude of the eMMR to their monolingual peers. Age was the only factor that influenced the amplitude of the late MMR. More negative late MMR was observed in older children with no difference found between bilingual and monolingual groups. CONCLUSIONS Consistent with previous studies, our findings revealed that biological factors (age and sex) and language experience modulated the amplitude of the eMMR in young children. The early negative MMR is likely to be the mismatch negativity found in older children and adults. In contrast, the late MMR amplitude was influenced only by age and may be equivalent to the Nc in infants and to the late negativity observed in some auditory passive oddball designs.
Collapse
Affiliation(s)
- Yan H. Yu
- Department of Communication Sciences and Disorders, St.
John’s University, Queens, NY, USA
| | - Carol Tessel
- Department of Communication Sciences and Disorders, Florida
Atlantic University, Boca Raton, FL, USA
| | - Xiaoxu Han
- Department of Computer and Information Science, Fordham
University, New York, NY, USA
| | - Luca Campanelli
- Ph.D. Program in Speech-Language-Hearing Sciences, The
Graduate Center, City University of New York, New York, NY, USA
| | - Nancy Vidal
- Speech Communication Studies, Iona College, New Rochelle,
NY, USA
| | | | - Karen Garrido-Nag
- Hearing, Speech, Language Sciences, Gallaudet University,
Washington DC, USA
| | - Hia Datta
- Speech-Language Pathology, Molloy College, Rockville
Centre, NY, USA
| | - Valerie L. Shafer
- Ph.D. Program in Speech-Language-Hearing Sciences, The
Graduate Center, City University of New York, New York, NY, USA
| |
Collapse
|
13
|
Nagels L, Gaudrain E, Vickers D, Matos Lopes M, Hendriks P, Başkent D. Development of vocal emotion recognition in school-age children: The EmoHI test for hearing-impaired populations. PeerJ 2020; 8:e8773. [PMID: 32274264 PMCID: PMC7130108 DOI: 10.7717/peerj.8773] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/09/2019] [Accepted: 02/21/2020] [Indexed: 11/20/2022] Open
Abstract
Traditionally, emotion recognition research has primarily used pictures and videos, while audio test materials are not always readily available or are not of good quality, which may be particularly important for studies with hearing-impaired listeners. Here we present a vocal emotion recognition test with pseudospeech productions from multiple speakers expressing three core emotions (happy, angry, and sad): the EmoHI test. The high sound quality recordings make the test suitable for use with populations of children and adults with normal or impaired hearing. Here we present normative data for vocal emotion recognition development in normal-hearing (NH) school-age children using the EmoHI test. Furthermore, we investigated cross-language effects by testing NH Dutch and English children, and the suitability of the EmoHI test for hearing-impaired populations, specifically for prelingually deaf Dutch children with cochlear implants (CIs). Our results show that NH children's performance improved significantly with age from the youngest age group onwards (4-6 years: 48.9%, on average). However, NH children's performance did not reach adult-like values (adults: 94.1%) even for the oldest age group tested (10-12 years: 81.1%). Additionally, the effect of age on NH children's development did not differ across languages. All except one CI child performed at or above chance-level showing the suitability of the EmoHI test. In addition, seven out of 14 CI children performed within the NH age-appropriate range, and nine out of 14 CI children did so when performance was adjusted for hearing age, measured from their age at CI implantation. However, CI children showed great variability in their performance, ranging from ceiling (97.2%) to below chance-level performance (27.8%), which could not be explained by chronological age alone. The strong and consistent development in performance with age, the lack of significant differences across the tested languages for NH children, and the above-chance performance of most CI children affirm the usability and versatility of the EmoHI test.
Collapse
Affiliation(s)
- Leanne Nagels
- Center for Language and Cognition Groningen (CLCG), University of Groningen, Groningen, The Netherlands.,Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, Groningen, The Netherlands
| | - Etienne Gaudrain
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, Groningen, The Netherlands.,CNRS, Lyon Neuroscience Research Center, Université de Lyon, Lyon, France
| | - Deborah Vickers
- Cambridge Hearing Group, Clinical Neurosciences Department, University of Cambridge, Cambridge, United Kingdom
| | - Marta Matos Lopes
- Hearbase Ltd, The Hearing Specialists, Kent, United Kingdom.,The Ear Institute, University College London, London, United Kingdom
| | - Petra Hendriks
- Center for Language and Cognition Groningen (CLCG), University of Groningen, Groningen, The Netherlands
| | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, Groningen, The Netherlands
| |
Collapse
|
14
|
Nagels L, Gaudrain E, Vickers D, Hendriks P, Başkent D. Development of voice perception is dissociated across gender cues in school-age children. Sci Rep 2020; 10:5074. [PMID: 32193411 PMCID: PMC7081243 DOI: 10.1038/s41598-020-61732-6] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2019] [Accepted: 02/27/2020] [Indexed: 11/11/2022] Open
Abstract
Children's ability to distinguish speakers' voices continues to develop throughout childhood, yet it remains unclear how children's sensitivity to voice cues, such as differences in speakers' gender, develops over time. This so-called voice gender is primarily characterized by speakers' mean fundamental frequency (F0), related to glottal pulse rate, and vocal-tract length (VTL), related to speakers' size. Here we show that children's acquisition of adult-like performance for discrimination, a lower-order perceptual task, and categorization, a higher-order cognitive task, differs across voice gender cues. Children's discrimination was adult-like around the age of 8 for VTL but still differed from adults at the age of 12 for F0. Children's perceptual weight attributed to F0 for gender categorization was adult-like around the age of 6 but around the age of 10 for VTL. Children's discrimination and weighting of F0 and VTL were only correlated for 4- to 6-year-olds. Hence, children's development of discrimination and weighting of voice gender cues are dissociated, i.e., adult-like performance for F0 and VTL is acquired at different rates and does not seem to be closely related. The different developmental patterns for auditory discrimination and categorization highlight the complexity of the relationship between perceptual and cognitive mechanisms of voice perception.
Collapse
Affiliation(s)
- Leanne Nagels
- Center for Language and Cognition Groningen (CLCG), University of Groningen, Groningen, The Netherlands.
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands.
- Research School of Behavioral and Cognitive Neuroscience, Graduate School of Medical Sciences, University of Groningen, Groningen, The Netherlands.
| | - Etienne Gaudrain
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
- Research School of Behavioral and Cognitive Neuroscience, Graduate School of Medical Sciences, University of Groningen, Groningen, The Netherlands
- CNRS UMR 5292, Lyon Neuroscience Research Center, Auditory Cognition and Psychoacoustics, Université de Lyon, Lyon, France
| | - Deborah Vickers
- Cambridge Hearing Group, Clinical Neurosciences Department, University of Cambridge, Cambridge, United Kingdom
| | - Petra Hendriks
- Center for Language and Cognition Groningen (CLCG), University of Groningen, Groningen, The Netherlands
- Research School of Behavioral and Cognitive Neuroscience, Graduate School of Medical Sciences, University of Groningen, Groningen, The Netherlands
| | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
- Research School of Behavioral and Cognitive Neuroscience, Graduate School of Medical Sciences, University of Groningen, Groningen, The Netherlands
| |
Collapse
|
15
|
Flaherty MM, Buss E, Leibold LJ. Developmental Effects in Children's Ability to Benefit From F0 Differences Between Target and Masker Speech. Ear Hear 2020; 40:927-937. [PMID: 30334835 PMCID: PMC6467703 DOI: 10.1097/aud.0000000000000673] [Citation(s) in RCA: 27] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES The objectives of this study were to (1) evaluate the extent to which school-age children benefit from fundamental frequency (F0) differences between target words and competing two-talker speech, and (2) assess whether this benefit changes with age. It was predicted that while children would be more susceptible to speech-in-speech masking compared to adults, they would benefit from differences in F0 between target and masker speech. A second experiment was conducted to evaluate the relationship between frequency discrimination thresholds and the ability to benefit from target/masker differences in F0. DESIGN Listeners were children (5 to 15 years) and adults (20 to 36 years) with normal hearing. In the first experiment, speech reception thresholds (SRTs) for disyllabic words were measured in a continuous, 60-dB SPL two-talker speech masker. The same male talker produced both the target and masker speech (average F0 = 120 Hz). The level of the target words was adaptively varied to estimate the level associated with 71% correct identification. The procedure was a four-alternative forced-choice with a picture-pointing response. Target words either had the same mean F0 as the masker or it was shifted up by 3, 6, or 9 semitones. To determine the benefit of target/masker F0 separation on word recognition, masking release was computed by subtracting thresholds in each shifted-F0 condition from the threshold in the unshifted-F0 condition. In the second experiment, frequency discrimination thresholds were collected for a subset of listeners to determine whether sensitivity to F0 differences would be predictive of SRTs. The standard was the syllable /ba/ with an F0 of 250 Hz; the target stimuli had a higher F0. Discrimination thresholds were measured using a three-alternative, three-interval forced choice procedure. RESULTS Younger children (5 to 12 years) had significantly poorer SRTs than older children (13 to 15 years) and adults in the unshifted-F0 condition. The benefit of F0 separations generally increased with increasing child age and magnitude of target/masker F0 separation. For 5- to 7-year-olds, there was a small benefit of F0 separation in the 9-semitone condition only. For 8- to 12-year-olds, there was a benefit from both 6- and 9-semitone separations, but to a lesser degree than what was observed for older children (13 to 15 years) and adults, who showed a substantial benefit in the 6- and 9-semitone conditions. Examination of individual data found that children younger than 7 years of age did not benefit from any of the F0 separations tested. Results for the frequency discrimination task indicated that, while there was a trend for improved thresholds with increasing age, these thresholds were not predictive of the ability to use F0 differences in the speech-in-speech recognition task after controlling for age. CONCLUSIONS The overall pattern of results suggests that children's ability to benefit from F0 differences in speech-in-speech recognition follows a prolonged developmental trajectory. Younger children are less able to capitalize on differences in F0 between target and masker speech. The extent to which individual children benefitted from target/masker F0 differences was not associated with their frequency discrimination thresholds.
Collapse
Affiliation(s)
- Mary M. Flaherty
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, Nebraska, USA
| | - Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, School of Medicine, University of North Carolina, Chapel Hill, North Carolina, USA
| | - Lori J. Leibold
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, Nebraska, USA
| |
Collapse
|
16
|
Jacewicz E, Fox RA. Perception of local and non-local vowels by adults and children in the South. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:627. [PMID: 32006983 PMCID: PMC7043861 DOI: 10.1121/10.0000542] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/23/2019] [Revised: 07/23/2019] [Accepted: 09/05/2019] [Indexed: 06/10/2023]
Abstract
This study assessed the ability of Southern listeners to accommodate extensive talker variability in identifying vowels in their local Appalachian community in the context of sound change. Building on prior work, the current experiment targeted a subset of spectrally overlapping vowels in local and two non-local varieties to establish whether adult and child listeners will demonstrate the local dialect advantage. Listeners responded to isolated target words, which minimized the interaction of multiple linguistic and dialect-specific features. For most vowel categories, the local dialect advantage was not demonstrated. However, adult listeners showed sensitivity to generational changes, indicating their familiarity with the local norms. A differential response pattern in children suggests that children perceived the vowels through the lens of their own experience with vowel production, representing a sound change in the community. Compared with the adults, children also relied more on stress cues, with increased confusions when the vowels were unstressed. The study provides evidence that identification accuracy is dependent upon the robustness of cues in individual vowel categories-whether local or non-local-and suggests that the bottom-up processes underlying phonetic vowel categorization in isolated monosyllables can interact with the top-down processing of dialect- and talker-specific information.
Collapse
Affiliation(s)
- Ewa Jacewicz
- Department of Speech and Hearing Science, The Ohio State University, 1070 Carmack Road, Columbus, Ohio 43210, USA
| | - Robert Allen Fox
- Department of Speech and Hearing Science, The Ohio State University, 1070 Carmack Road, Columbus, Ohio 43210, USA
| |
Collapse
|
17
|
Toscano JC, Lansing CR. Age-Related Changes in Temporal and Spectral Cue Weights in Speech. LANGUAGE AND SPEECH 2019; 62:61-79. [PMID: 29103359 DOI: 10.1177/0023830917737112] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Listeners weight acoustic cues in speech according to their reliability, but few studies have examined how cue weights change across the lifespan. Previous work has suggested that older adults have deficits in auditory temporal discrimination, which could affect the reliability of temporal phonetic cues, such as voice onset time (VOT), and in turn, impact speech perception in real-world listening environments. We addressed this by examining younger and older adults' use of VOT and onset F0 (a secondary phonetic cue) for voicing judgments (e.g., /b/ vs. /p/), using both synthetic and naturally produced speech. We found age-related differences in listeners' use of the two voicing cues, such that older adults relied more heavily on onset F0 than younger adults, even though this cue is less reliable in American English. These results suggest that phonetic cue weights continue to change across the lifespan.
Collapse
|
18
|
McMurray B, Danelz A, Rigler H, Seedorff M. Speech categorization develops slowly through adolescence. Dev Psychol 2018; 54:1472-1491. [PMID: 29952600 PMCID: PMC6062449 DOI: 10.1037/dev0000542] [Citation(s) in RCA: 35] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
The development of the ability to categorize speech sounds is often viewed as occurring primarily during infancy via perceptual learning mechanisms. However, a number of studies suggest that even after infancy, children's categories become more categorical and well defined through about age 12. We investigated the cognitive changes that may be responsible for such development using a visual world paradigm experiment based on (McMurray, Tanenhaus, & Aslin, 2002). Children from 3 age groups (7-8, 12-13, and 17-18 years) heard a token from either a b/p or s/∫ continua spanning 2 words (beach/peach, ship/sip) and selected its referent from a screen containing 4 pictures of potential lexical candidates. Eye movements to each object were monitored as a measure of how strongly children were committing to each candidate as perception unfolds in real-time. Results showed an ongoing sharpening of speech categories through 18, which was particularly apparent during the early stages of real-time perception. When analysis targeted to specifically within-category sensitivity to continuous detail, children exhibited increasingly gradient categories over development, suggesting that increasing sensitivity to fine-grained detail in the signal enables these more discrete categorizations. Together these suggest that speech development is a protracted process in which children's increasing sensitivity to within-category detail in the signal enables increasingly sharp phonetic categories. (PsycINFO Database Record
Collapse
Affiliation(s)
- Bob McMurray
- Department of Psychological and Brain Sciences, University of Iowa
| | - Ani Danelz
- Department of Communication Sciences and Disorders, University of Iowa
| | | | | |
Collapse
|
19
|
Spratford M, McLean HH, McCreery R. Relationship of Grammatical Context on Children's Recognition of s/z-Inflected Words. J Am Acad Audiol 2018; 28:799-809. [PMID: 28972469 DOI: 10.3766/jaaa.16151] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
BACKGROUND Access to aided high-frequency speech information is currently assessed behaviorally using recognition of plural monosyllabic words. Because of semantic and grammatical cues that support word+morpheme recognition in sentence materials, the contribution of high-frequency audibility to sentence recognition is less than that for isolated words. However, young children may not yet have the linguistic competence to take advantage of these cues. A low-predictability sentence recognition task that controls for language ability could be used to assess the impact of high-frequency audibility in a context that more closely represents how children learn language. PURPOSE To determine if differences exist in recognition of s/z-inflected monosyllabic words for children with normal hearing (CNH) and children who are hard of hearing (CHH) across stimuli context (presented in isolation versus embedded medially within a sentence that has low semantic and syntactic predictability) and varying levels of high-frequency audibility (4- and 8-kHz low-pass filtered for CNH and 8-kHz low-pass filtered for CHH). RESEARCH DESIGN A prospective, cross-sectional design was used to analyze word+morpheme recognition in noise for stimuli varying in grammatical context and high-frequency audibility. Low-predictability sentence stimuli were created so that the target word+morpheme could not be predicted by semantic or syntactic cues. Electroacoustic measures of aided access to high-frequency speech sounds were used to predict individual differences in recognition for CHH. STUDY SAMPLE Thirty-five children, aged 5-12 yrs, were recruited to participate in the study; 24 CNH and 11 CHH (bilateral mild to severe hearing loss) who wore hearing aids (HAs). All children were native speakers of English. DATA COLLECTION AND ANALYSIS Monosyllabic word+morpheme recognition was measured in isolated and sentence-embedded conditions at a +10 dB signal-to-noise ratio using steady state, speech-shaped noise. Real-ear probe microphone measures of HAs were obtained for CHH. To assess the effects of high-frequency audibility on word+morpheme recognition for CNH, a repeated-measures ANOVA was used with bandwidth (8 kHz, 4 kHz) and context (isolated, sentence embedded) as within-subjects factors. To compare recognition between CNH and CHH, a mixed-model ANOVA was completed with context (isolated, sentence-embedded) as a within-subjects factor and hearing status as a between-subjects factor. Bivariate correlations between word+morpheme recognition scores and electroacoustic measures of high-frequency audibility were used to assess which measures might be sensitive to differences in perception for CHH. RESULTS When high-frequency audibility was maximized, CNH and CHH had better word+morpheme recognition in the isolated condition compared with sentence-embedded. When high-frequency audibility was limited, CNH had better word+morpheme recognition in the sentence-embedded condition compared with the isolated condition. CHH whose HAs had greater high-frequency speech bandwidth, as measured by the maximum audible frequency, had better word+morpheme recognition in sentences. CONCLUSIONS High-frequency audibility supports word+morpheme recognition within low-predictability sentences for both CNH and CHH. Maximum audible frequency can be used to estimate word+morpheme recognition for CHH. Low-predictability sentences that do not contain semantic or grammatical context may be of clinical use in estimating children's use of high-frequency audibility in a manner that approximates how they learn language.
Collapse
Affiliation(s)
- Meredith Spratford
- Audibility, Perception and Cognition Laboratory, Boys Town National Research Hospital, Omaha, NE
| | | | - Ryan McCreery
- Audibility, Perception and Cognition Laboratory, Boys Town National Research Hospital, Omaha, NE
| |
Collapse
|
20
|
Dugan SH, Silbert N, McAllister T, Preston JL, Sotto C, Boyce SE. Modelling category goodness judgments in children with residual sound errors. CLINICAL LINGUISTICS & PHONETICS 2018; 33:295-315. [PMID: 29792525 PMCID: PMC6733520 DOI: 10.1080/02699206.2018.1477834] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/23/2018] [Revised: 05/12/2018] [Accepted: 05/14/2018] [Indexed: 05/29/2023]
Abstract
This study investigates category goodness judgments of /r/ in adults and children with and without residual speech errors (RSEs) using natural speech stimuli. Thirty adults, 38 children with RSE (ages 7-16) and 35 age-matched typically developing (TD) children provided category goodness judgments on whole words, recorded from 27 child speakers, with /r/ in various phonetic environments. The salient acoustic property of /r/ - the lowered third formant (F3) - was normalized in two ways. A logistic mixed-effect model quantified the relationships between listeners' responses and the third formant frequency, vowel context and clinical group status. Goodness judgments from the adult group showed a statistically significant interaction with the F3 parameter when compared to both child groups (p < 0.001) using both normalization methods. The RSE group did not differ significantly from the TD group in judgments of /r/. All listeners were significantly more likely to judge /r/ as correct in a front-vowel context. Our results suggest that normalized /r/ F3 is a statistically significant predictor of category goodness judgments for both adults and children, but children do not appear to make adult-like judgments. Category goodness judgments do not have a clear relationship with /r/ production abilities in children with RSE. These findings may have implications for clinical activities that include category goodness judgments in natural speech, especially for recorded productions.
Collapse
|
21
|
Peter V, Kalashnikova M, Burnham D. Weighting of Amplitude and Formant Rise Time Cues by School-Aged Children: A Mismatch Negativity Study. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2018; 61:1322-1333. [PMID: 29800360 DOI: 10.1044/2018_jslhr-h-17-0334] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/05/2017] [Accepted: 02/07/2018] [Indexed: 06/08/2023]
Abstract
PURPOSE An important skill in the development of speech perception is to apply optimal weights to acoustic cues so that phonemic information is recovered from speech with minimum effort. Here, we investigated the development of acoustic cue weighting of amplitude rise time (ART) and formant rise time (FRT) cues in children as measured by mismatch negativity (MMN). METHOD Twelve adults and 36 children aged 6-12 years listened to a /ba/-/wa/ contrast in an oddball paradigm in which the standard stimulus had the ART and FRT cues of /ba/. In different blocks, the deviant stimulus had either the ART or FRT cues of /wa/. RESULTS The results revealed that children younger than 10 years were sensitive to both ART and FRT cues whereas 10- to 12-year-old children and adults were sensitive only to FRT cues. Moreover, children younger than 10 years generated a positive mismatch response, whereas older children and adults generated MMN. CONCLUSION These results suggest that preattentive adultlike weighting of ART and FRT cues is attained only by 10 years of age and accompanies the change from mismatch response to the more mature MMN response. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.6207608.
Collapse
Affiliation(s)
- Varghese Peter
- MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, New South Wales, Australia
| | - Marina Kalashnikova
- MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, New South Wales, Australia
| | - Denis Burnham
- MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, New South Wales, Australia
| |
Collapse
|
22
|
Abstract
OBJECTIVES Formant rise time (FRT) and amplitude rise time (ART) are acoustic cues that inform phonetic identity. FRT represents the rate of transition of the formant(s) to a steady state, while ART represents the rate at which the sound reaches its peak amplitude. Normal-hearing (NH) native English speakers weight FRT more than ART during the perceptual labeling of the /ba/-/wa/ contrast. This weighting strategy is reflected neurophysiologically in the magnitude of the mismatch negativity (MMN)-MMN is larger during the FRT than the ART distinction. The present study examined the neurophysiological basis of acoustic cue weighting in adult cochlear implant (CI) listeners using the MMN design. It was hypothesized that individuals with CIs who weight ART more in behavioral labeling (ART users) would show larger MMNs during the ART than the FRT contrast, and the opposite would be seen for FRT users. DESIGN Electroencephalography was recorded while 20 adults with CIs listened passively to combinations of 3 synthetic speech stimuli: a /ba/ with /ba/-like FRT and ART; a /wa/ with /wa/-like FRT and ART; and a /ba/ stimulus with /ba/-like FRT and /wa/-like ART. The MMN response was elicited during the FRT contrast by having participants passively listen to a train of /wa/ stimuli interrupted occasionally by /ba/ stimuli, and vice versa. For the ART contrast, the same procedure was implemented using the /ba/ and /ba/ stimuli. RESULTS Both ART and FRT users with CIs elicited MMNs that were equal in magnitudes during FRT and ART contrasts, with the exception that FRT users exhibited MMNs for ART and FRT contrasts that were temporally segregated. That is, their MMNs occurred significantly earlier during the ART contrast (~100 msec following sound onset) than during the FRT contrast (~200 msec). In contrast, the MMNs for ART users of both contrasts occurred later and were not significantly separable in time (~230 msec). Interestingly, this temporal segregation observed in FRT users is consistent with the MMN behavior in NH listeners. CONCLUSIONS Results suggest that listeners with CIs who learn to classify phonemes based on formant dynamics, consistent with NH listeners, develop a strategy similar to NH listeners, in which the organization of the amplitude and spectral representations of phonemes in auditory memory are temporally segregated.
Collapse
|
23
|
Heald SLM, Van Hedger SC, Nusbaum HC. Perceptual Plasticity for Auditory Object Recognition. Front Psychol 2017; 8:781. [PMID: 28588524 PMCID: PMC5440584 DOI: 10.3389/fpsyg.2017.00781] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2016] [Accepted: 04/26/2017] [Indexed: 01/25/2023] Open
Abstract
In our auditory environment, we rarely experience the exact acoustic waveform twice. This is especially true for communicative signals that have meaning for listeners. In speech and music, the acoustic signal changes as a function of the talker (or instrument), speaking (or playing) rate, and room acoustics, to name a few factors. Yet, despite this acoustic variability, we are able to recognize a sentence or melody as the same across various kinds of acoustic inputs and determine meaning based on listening goals, expectations, context, and experience. The recognition process relates acoustic signals to prior experience despite variability in signal-relevant and signal-irrelevant acoustic properties, some of which could be considered as "noise" in service of a recognition goal. However, some acoustic variability, if systematic, is lawful and can be exploited by listeners to aid in recognition. Perceivable changes in systematic variability can herald a need for listeners to reorganize perception and reorient their attention to more immediately signal-relevant cues. This view is not incorporated currently in many extant theories of auditory perception, which traditionally reduce psychological or neural representations of perceptual objects and the processes that act on them to static entities. While this reduction is likely done for the sake of empirical tractability, such a reduction may seriously distort the perceptual process to be modeled. We argue that perceptual representations, as well as the processes underlying perception, are dynamically determined by an interaction between the uncertainty of the auditory signal and constraints of context. This suggests that the process of auditory recognition is highly context-dependent in that the identity of a given auditory object may be intrinsically tied to its preceding context. To argue for the flexible neural and psychological updating of sound-to-meaning mappings across speech and music, we draw upon examples of perceptual categories that are thought to be highly stable. This framework suggests that the process of auditory recognition cannot be divorced from the short-term context in which an auditory object is presented. Implications for auditory category acquisition and extant models of auditory perception, both cognitive and neural, are discussed.
Collapse
|
24
|
Richardson K, Sussman JE. Discrimination and Identification of a Third Formant Frequency Cue to Place of Articulation by Young Children and Adults. LANGUAGE AND SPEECH 2017; 60:27-47. [PMID: 28326988 DOI: 10.1177/0023830915625680] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Typically-developing children, 4 to 6 years of age, and adults participated in discrimination and identification speech perception tasks using a synthetic consonant-vowel continuum ranging from /da/ to /ga/. The seven-step synthetic /da/-/ga/ continuum was created by adjusting the first 40 ms of the third formant frequency transition. For the discrimination task, listeners participated in a Change/No-Change paradigm with four different stimuli compared to the endpoint-1 /da/ token. For the identification task, listeners labeled each token along the /da/-/ga/ continuum as either "DA" or "GA." Results of the discrimination experiment showed that sensitivity to the third-formant transition cue improved for the adult listeners as the stimulus contrast increased, whereas the performance of the children remained poor across all stimulus comparisons. Results of the identification experiment support previous hypotheses of age-related differences in phonetic categorization. Results have implications for normative data on identification and discrimination tasks. These norms provide a metric against which children with auditory-based speech sound disorders can be compared. Furthermore, the results provide some insight into the developmental nature of categorical and non-categorical speech perception.
Collapse
|
25
|
Law JM, Vandermosten M, Ghesquière P, Wouters J. Predicting Future Reading Problems Based on Pre-reading Auditory Measures: A Longitudinal Study of Children with a Familial Risk of Dyslexia. Front Psychol 2017; 8:124. [PMID: 28223953 PMCID: PMC5293743 DOI: 10.3389/fpsyg.2017.00124] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2016] [Accepted: 01/17/2017] [Indexed: 11/13/2022] Open
Abstract
Purpose: This longitudinal study examines measures of temporal auditory processing in pre-reading children with a family risk of dyslexia. Specifically, it attempts to ascertain whether pre-reading auditory processing, speech perception, and phonological awareness (PA) reliably predict later literacy achievement. Additionally, this study retrospectively examines the presence of pre-reading auditory processing, speech perception, and PA impairments in children later found to be literacy impaired. Method: Forty-four pre-reading children with and without a family risk of dyslexia were assessed at three time points (kindergarten, first, and second grade). Auditory processing measures of rise time (RT) discrimination and frequency modulation (FM) along with speech perception, PA, and various literacy tasks were assessed. Results: Kindergarten RT uniquely contributed to growth in literacy in grades one and two, even after controlling for letter knowledge and PA. Highly significant concurrent and predictive correlations were observed with kindergarten RT significantly predicting first grade PA. Retrospective analysis demonstrated atypical performance in RT and PA at all three time points in children who later developed literacy impairments. Conclusions: Although significant, kindergarten auditory processing contributions to later literacy growth lack the power to be considered as a single-cause predictor; thus results support temporal processing deficits' contribution within a multiple deficit model of dyslexia.
Collapse
Affiliation(s)
- Jeremy M Law
- Parenting and Special Education Research Unit, Faculty of Psychology and Educational Sciences, KU Leuven Leuven, Belgium
| | - Maaike Vandermosten
- Parenting and Special Education Research Unit, Faculty of Psychology and Educational Sciences, KU LeuvenLeuven, Belgium; Laboratory for Experimental ORL, Department of Neuroscience, KU LeuvenLeuven, Belgium
| | - Pol Ghesquière
- Parenting and Special Education Research Unit, Faculty of Psychology and Educational Sciences, KU Leuven Leuven, Belgium
| | - Jan Wouters
- Laboratory for Experimental ORL, Department of Neuroscience, KU Leuven Leuven, Belgium
| |
Collapse
|
26
|
Kong EJ, Edwards J. Individual differences in categorical perception of speech: Cue weighting and executive function. JOURNAL OF PHONETICS 2016; 59:40-57. [PMID: 28503007 PMCID: PMC5423668 DOI: 10.1016/j.wocn.2016.08.006] [Citation(s) in RCA: 34] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2023]
Abstract
This study examined individual differences in categorical perception and the use of multiple acoustic cues in the perception of the stop voicing contrast. Goals were to investigate whether gradiency of speech perception was related to listeners' differential sensitivity to acoustic cues and to individual differences in executive function. The experiment included two speech perception tasks (visual analogue scaling [VAS] and anticipatory eye movement [AEM]) administered to 30 English-speaking adults in two separate experimental sessions. Stimuli were a /ta/ to /da/ continuum that systematically varied VOT and f0. Findings were that some listeners had a more gradient pattern of responses on the VAS task; the listeners who had a gradient response pattern on the VAS task also showed more sensitivity to f0 on the AEM task. The patterns were consistent across individuals tested on two separate occasions. These results suggest that variability in how categorically listeners perceive speech sounds is consistent and systematic within individuals.
Collapse
Affiliation(s)
- Eun Jong Kong
- Korea Aerospace University, 100, Hanggongdae gil, Hwajeon-dong, Deogyang-gu, Goyang-city, Gyeonggi-do, South Korea 412-791
| | - Jan Edwards
- University of Wisconsin-Madison, 301 Goodnight Hall, 1975 Willow Dr., Madison, WI 53706 USA,
| |
Collapse
|
27
|
Word Recognition Variability With Cochlear Implants: "Perceptual Attention" Versus "Auditory Sensitivity". Ear Hear 2016; 37:14-26. [PMID: 26301844 DOI: 10.1097/aud.0000000000000204] [Citation(s) in RCA: 26] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/21/2022]
Abstract
OBJECTIVES Cochlear implantation does not automatically result in robust spoken language understanding for postlingually deafened adults. Enormous outcome variability exists, related to the complexity of understanding spoken language through cochlear implants (CIs), which deliver degraded speech representations. This investigation examined variability in word recognition as explained by "perceptual attention" and "auditory sensitivity" to acoustic cues underlying speech perception. DESIGN Thirty postlingually deafened adults with CIs and 20 age-matched controls with normal hearing (NH) were tested. Participants underwent assessment of word recognition in quiet and perceptual attention (cue-weighting strategies) based on labeling tasks for two phonemic contrasts: (1) "cop"-"cob," based on a duration cue (easily accessible through CIs) or a dynamic spectral cue (less accessible through CIs), and (2) "sa"-"sha," based on static or dynamic spectral cues (both potentially poorly accessible through CIs). Participants were also assessed for auditory sensitivity to the speech cues underlying those labeling decisions. RESULTS Word recognition varied widely among CI users (20 to 96%), but it was generally poorer than for NH participants. Implant users and NH controls showed similar perceptual attention and auditory sensitivity to the duration cue, while CI users showed poorer attention and sensitivity to all spectral cues. Both attention and sensitivity to spectral cues predicted variability in word recognition. CONCLUSIONS For CI users, both perceptual attention and auditory sensitivity are important in word recognition. Efforts should be made to better represent spectral cues through implants, while also facilitating attention to these cues through auditory training.
Collapse
|
28
|
Giezen MR, Escudero P, Baker AE. Rapid learning of minimally different words in five- to six-year-old children: effects of acoustic salience and hearing impairment. JOURNAL OF CHILD LANGUAGE 2016; 43:310-337. [PMID: 25994361 DOI: 10.1017/s0305000915000197] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
This study investigates the role of acoustic salience and hearing impairment in learning phonologically minimal pairs. Picture-matching and object-matching tasks were used to investigate the learning of consonant and vowel minimal pairs in five- to six-year-old deaf children with a cochlear implant (CI), and children of the same age with normal hearing (NH). In both tasks, the CI children showed clear difficulties with learning minimal pairs. The NH children also showed some difficulties, however, particularly in the picture-matching task. Vowel minimal pairs were learned more successfully than consonant minimal pairs, particularly in the object-matching task. These results suggest that the ability to encode phonetic detail in novel words is not fully developed at age six and is affected by task demands and acoustic salience. CI children experience persistent difficulties with accurately mapping sound contrasts to novel meanings, but seem to benefit from the relative acoustic salience of vowel sounds.
Collapse
Affiliation(s)
- Marcel R Giezen
- Laboratory for Language and Cognitive Neuroscience,San Diego State University
| | | | - Anne E Baker
- University of Amsterdam,Department of Linguistics
| |
Collapse
|
29
|
Tuomainen O, Stuart NJ, van der Lely HKJ. Phonetic categorisation and cue weighting in adolescents with Specific Language Impairment (SLI). CLINICAL LINGUISTICS & PHONETICS 2015; 29:557-572. [PMID: 25970138 DOI: 10.3109/02699206.2015.1036464] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
This study investigates phonetic categorisation and cue weighting in adolescents and young adults with Specific Language Impairment (SLI). We manipulated two acoustic cues, vowel duration and F1 offset frequency, that signal word-final stop consonant voicing ([t] and [d]) in English. Ten individuals with SLI (14.0-21.4 years), 10 age-matched controls (CA; 14.6-21.9 years) and 10 non-matched adult controls (23.3-36.0 years) labelled synthetic CVC non-words in an identification task. The results showed that the adolescents and young adults with SLI were less consistent than controls in the identification of the good category representatives. The group with SLI also assigned less weight to vowel duration than the adult controls. However, no direct relationship between phonetic categorisation, cue weighting and language skills was found. These findings indicate that some individuals with SLI have speech perception deficits but they are not necessarily associated with oral language skills.
Collapse
Affiliation(s)
- Outi Tuomainen
- Department of Speech, Hearing and Phonetic Sciences, University College London , UK
| | | | | |
Collapse
|
30
|
Effects of frequency compression and frequency transposition on fricative and affricate perception in listeners with normal hearing and mild to moderate hearing loss. Ear Hear 2015; 35:519-32. [PMID: 24699702 DOI: 10.1097/aud.0000000000000040] [Citation(s) in RCA: 39] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES The authors have demonstrated that the limited bandwidth associated with conventional hearing aid amplification prevents useful high-frequency speech information from being transmitted. The purpose of this study was to examine the efficacy of two popular frequency-lowering algorithms and one novel algorithm (spectral envelope decimation) in adults with mild to moderate sensorineural hearing loss and in normal-hearing controls. DESIGN Participants listened monaurally through headphones to recordings of nine fricatives and affricates spoken by three women in a vowel-consonant context. Stimuli were mixed with speech-shaped noise at 10 dB SNR and recorded through a Widex Inteo IN-9 and a Phonak Naída UP V behind-the-ear (BTE) hearing aid. Frequency transposition (FT) is used in the Inteo and nonlinear frequency compression (NFC) used in the Naída. Both devices were programmed to lower frequencies above 4 kHz, but neither device could lower frequencies above 6 to 7 kHz. Each device was tested under four conditions: frequency lowering deactivated (FT-off and NFC-off), frequency lowering activated (FT and NFC), wideband (WB), and a fourth condition unique to each hearing aid. The WB condition was constructed by mixing recordings from the first condition with high-pass filtered versions of the source stimuli. For the Inteo, the fourth condition consisted of recordings made with the same settings as the first, but with the noise-reduction feature activated (FT-off). For the Naída, the fourth condition was the same as the first condition except that source stimuli were preprocessed by a novel frequency compression algorithm, spectral envelope decimation (SED), designed in MATLAB, which allowed for a more complete lowering of the 4 to 10 kHz input band. A follow-up experiment with NFC used Phonak's Naída SP V BTE, which could also lower a greater range of input frequencies. RESULTS For normal-hearing and hearing-impaired listeners, performance with FT was significantly worse compared with that in the other conditions. Consistent with previous findings, performance for the hearing-impaired listeners in the WB condition was significantly better than in the FT-off condition. In addition, performance in the SED and WB conditions were both significantly better than in the NFC-off condition and the NFC condition with 6 kHz input bandwidth. There were no significant differences between SED and WB, indicating that improvements in fricative identification obtained by increasing bandwidth can also be obtained using this form of frequency compression. Significant differences between most conditions could be largely attributed to an increase or decrease in confusions for the phonemes /s/ and /z/. In the follow-up experiment, performance in the NFC condition with 10 kHz input bandwidth was significantly better than NFC-off, replicating the results obtained with SED. Furthermore, listeners who performed poorly with NFC-off tended to show the most improvement with NFC. CONCLUSIONS Improvements in the identification of stimuli chosen to be sensitive to the effects of frequency lowering have been demonstrated using two forms of frequency compression (NFC and SED) in individuals with mild to moderate high-frequency sensorineural hearing loss. However, negative results caution against using FT for this population. Results also indicate that the advantage of an extended bandwidth as reported here and elsewhere applies to the input bandwidth for frequency compression (NFC/SED) when the start frequency is ≥4 kHz.
Collapse
|
31
|
The development of voicing categories: a quantitative review of over 40 years of infant speech perception research. Psychon Bull Rev 2015; 21:884-906. [PMID: 24550074 DOI: 10.3758/s13423-013-0569-y] [Citation(s) in RCA: 18] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Most research on infant speech categories has relied on measures of discrimination. Such work often employs categorical perception as a linking hypothesis to enable inferences about categorization on the basis of discrimination measures. However, a large number of studies with adults challenge the utility of categorical perception in describing adult speech perception, and this in turn calls into question how to interpret measures of infant speech discrimination. We propose here a parallel channels model of discrimination (built on Pisoni and Tash Perception & Psychophysics, 15(2), 285-290, 1974), which posits that both a noncategorical or veridical encoding of speech cues and category representations can simultaneously contribute to discrimination. This can thus produce categorical perception effects without positing any warping of the acoustic signal, but it also reframes how we think about infant discrimination and development. We test this model by conducting a quantitative review of 20 studies examining infants' discrimination of voice onset time contrasts. This review suggests that within-category discrimination is surprisingly prevalent even in classic studies and that, averaging across studies, discrimination is related to continuous acoustic distance. It also identifies several methodological factors that may mask our ability to see this. Finally, it suggests that infant discrimination may improve over development, contrary to commonly held notion of perceptual narrowing. These results are discussed in terms of theories of speech development that may require such continuous sensitivity.
Collapse
|
32
|
Nittrouer S, Lowenstein JH. Dynamic spectral structure specifies vowels for adults and children. LANGUAGE AND SPEECH 2014; 57:487-512. [PMID: 25536845 PMCID: PMC4315365 DOI: 10.1177/0023830913508075] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
The dynamic specification account of vowel recognition suggests that formant movement between vowel targets and consonant margins is used by listeners to recognize vowels. This study tested that account by measuring contributions to vowel recognition of dynamic (i.e., time-varying) spectral structure and coarticulatory effects on stationary structure. Adults and children (four- and seven-year-olds) were tested with three kinds of consonant-vowel-consonant syllables: (I) unprocessed; (2) sine waves that preserved both stationary coarticulated and dynamic spectral structure; and (3) vocoded signals that primarily preserved that stationary, but not dynamic structure. Sections of two lengths were removed from syllable middles: (I) half the vocalic portion; and (2) all but the first and last three pitch periods. Adults performed accurately with unprocessed and sine-wave signals, as long as half the syllable remained; their recognition was poorer for vocoded signals, but above chance. Seven-year-olds performed more poorly than adults with both sorts of processed signals, but disproportionately worse with vocoded than sine-wave signals. Most four-year-olds were unable to recognize vowels at all with vocoded signals. Conclusions were that both dynamic and stationary coarticulated structures support vowel recognition for adults, but children attend to dynamic spectral structure more strongly because early phonological organization favors whole words.
Collapse
|
33
|
Abstract
The end-result of perceptual reorganization in infancy is currently viewed as a reconfigured perceptual space, "warped" around native-language phonetic categories, which then acts as a direct perceptual filter on any non-native sounds: naïve-listener discrimination of non-native-sounds is determined by their mapping onto native-language phonetic categories that are acoustically/articulatorily most similar. We report results that suggest another factor in non-native speech perception: some perceptual sensitivities cannot be attributed to listeners' warped perceptual space alone, but rather to enhanced general sensitivity along phonetic dimensions that the listeners' native language employs to distinguish between categories. Specifically, we show that the knowledge of a language with short and long vowel categories leads to enhanced discrimination of non-native consonant length contrasts. We argue that these results support a view of perceptual reorganization as the consequence of learners' hierarchical inductive inferences about the structure of the language's sound system: infants not only acquire the specific phonetic category inventory, but also draw higher-order generalizations over the set of those categories, such as the overall informativity of phonetic dimensions for sound categorization. Non-native sound perception is then also determined by sensitivities that emerge from these generalizations, rather than only by mappings of non-native sounds onto native-language phonetic categories.
Collapse
Affiliation(s)
- Bozena Pajak
- University of California, San Diego, 9500 Gilman Drive, La Jolla, California 92093-0108, USA
| | - Roger Levy
- University of California, San Diego, 9500 Gilman Drive, La Jolla, California 92093-0108, USA
| |
Collapse
|
34
|
Lewis DE, Manninen CM, Valente DL, Smith NA. Children's understanding of instructions presented in noise and reverberation. Am J Audiol 2014; 23:326-36. [PMID: 25036922 DOI: 10.1044/2014_aja-14-0020] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2014] [Accepted: 06/27/2014] [Indexed: 11/09/2022] Open
Abstract
PURPOSE This study examined children's ability to follow audio-visual instructions presented in noise and reverberation. METHOD Children (8-12 years of age) with normal hearing followed instructions in noise or noise plus reverberation. Performance was compared for a single talker (ST), multiple talkers speaking one at a time (MT), and multiple talkers with competing comments from other talkers (MTC). Working memory was assessed using measures of digit span. RESULTS Performance was better for children in noise than for those in noise plus reverberation. In noise, performance for ST was better than for either MT or MTC, and performance for MT was better than for MTC. In noise plus reverberation, performance for ST and MT was better than for MTC, but there were no differences between ST and MT. Digit span did not account for significant variance in the task. CONCLUSIONS Overall, children performed better in noise than in noise plus reverberation. However, differing patterns across conditions for the 2 environments suggested that the addition of reverberation may have affected performance in a way that was not apparent in noise alone. Continued research is needed to examine the differing effects of noise and reverberation on children's speech understanding.
Collapse
Affiliation(s)
| | - Crystal M. Manninen
- Boys Town National Research Hospital, Omaha, NE
- University of Nebraska–Lincoln
| | | | | |
Collapse
|
35
|
McCarthy KM, Mahon M, Rosen S, Evans BG. Speech perception and production by sequential bilingual children: a longitudinal study of voice onset time acquisition. Child Dev 2014; 85:1965-80. [PMID: 25123987 PMCID: PMC4282029 DOI: 10.1111/cdev.12275] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
Abstract
The majority of bilingual speech research has focused on simultaneous bilinguals. Yet, in immigrant communities, children are often initially exposed to their family language (L1), before becoming gradually immersed in the host country's language (L2). This is typically referred to as sequential bilingualism. Using a longitudinal design, this study explored the perception and production of the English voicing contrast in 55 children (40 Sylheti-English sequential bilinguals and 15 English monolinguals). Children were tested twice: when they were in nursery (52-month-olds) and 1 year later. Sequential bilinguals' perception and production of English plosives were initially driven by their experience with their L1, but after starting school, changed to match that of their monolingual peers.
Collapse
|
36
|
Moberly AC, Lowenstein JH, Tarr E, Caldwell-Tarr A, Welling DB, Shahin AJ, Nittrouer S. Do adults with cochlear implants rely on different acoustic cues for phoneme perception than adults with normal hearing? JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2014; 57:566-82. [PMID: 24686722 PMCID: PMC4008700 DOI: 10.1044/2014_jslhr-h-12-0323] [Citation(s) in RCA: 25] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
PURPOSE Several acoustic cues specify any single phonemic contrast. Nonetheless, adult, native speakers of a language share weighting strategies, showing preferential attention to some properties over others. Cochlear implant (CI) signal processing disrupts the salience of some cues: In general, amplitude structure remains readily available, but spectral structure less so. This study asked how well speech recognition is supported if CI users shift attention to salient cues not weighted strongly by native speakers. METHOD Twenty adults with CIs participated. The /bɑ/-/wɑ/ contrast was used because spectral and amplitude structure varies in correlated fashion for this contrast. Adults with normal hearing weight the spectral cue strongly but the amplitude cue negligibly. Three measurements were made: labeling decisions, spectral and amplitude discrimination, and word recognition. RESULTS Outcomes varied across listeners: Some weighted the spectral cue strongly, some weighted the amplitude cue, and some weighted neither. Spectral discrimination predicted spectral weighting. Spectral weighting explained the most variance in word recognition. Age of onset of hearing loss predicted spectral weighting but not unique variance in word recognition. CONCLUSION The weighting strategies of listeners with normal hearing likely support speech recognition best, so efforts in implant design, fitting, and training should focus on developing those strategies.
Collapse
|
37
|
Heald SLM, Nusbaum HC. Speech perception as an active cognitive process. Front Syst Neurosci 2014; 8:35. [PMID: 24672438 PMCID: PMC3956139 DOI: 10.3389/fnsys.2014.00035] [Citation(s) in RCA: 98] [Impact Index Per Article: 9.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2013] [Accepted: 02/25/2014] [Indexed: 11/13/2022] Open
Abstract
One view of speech perception is that acoustic signals are transformed into representations for pattern matching to determine linguistic structure. This process can be taken as a statistical pattern-matching problem, assuming realtively stable linguistic categories are characterized by neural representations related to auditory properties of speech that can be compared to speech input. This kind of pattern matching can be termed a passive process which implies rigidity of processing with few demands on cognitive processing. An alternative view is that speech recognition, even in early stages, is an active process in which speech analysis is attentionally guided. Note that this does not mean consciously guided but that information-contingent changes in early auditory encoding can occur as a function of context and experience. Active processing assumes that attention, plasticity, and listening goals are important in considering how listeners cope with adverse circumstances that impair hearing by masking noise in the environment or hearing loss. Although theories of speech perception have begun to incorporate some active processing, they seldom treat early speech encoding as plastic and attentionally guided. Recent research has suggested that speech perception is the product of both feedforward and feedback interactions between a number of brain regions that include descending projections perhaps as far downstream as the cochlea. It is important to understand how the ambiguity of the speech signal and constraints of context dynamically determine cognitive resources recruited during perception including focused attention, learning, and working memory. Theories of speech perception need to go beyond the current corticocentric approach in order to account for the intrinsic dynamics of the auditory encoding of speech. In doing so, this may provide new insights into ways in which hearing disorders and loss may be treated either through augementation or therapy.
Collapse
|
38
|
Moberly AC, Bhat J, Welling DB, Shahin AJ. Neurophysiology of spectrotemporal cue organization of spoken language in auditory memory. BRAIN AND LANGUAGE 2014; 130:42-49. [PMID: 24576808 PMCID: PMC3989417 DOI: 10.1016/j.bandl.2014.01.007] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/08/2013] [Revised: 01/07/2014] [Accepted: 01/23/2014] [Indexed: 06/03/2023]
Abstract
Listeners assign different weights to spectral dynamics, such as formant rise time (FRT), and temporal dynamics, such as amplitude rise time (ART), during phonetic judgments. We examined the neurophysiological basis of FRT and ART weighting in the /ba/-/wa/ contrast. Electroencephalography was recorded for thirteen adult English speakers during a mismatch negativity (MMN) design using synthetic stimuli: a /ba/ with /ba/-like FRT and ART; a /wa/ with /wa/-like FRT and ART; and a /ba/(wa) with /ba/-like FRT and /wa/-like ART. We hypothesized that because of stronger reliance on FRT, subjects would encode a stronger memory trace and exhibit larger MMN during the FRT than the ART contrast. Results supported this hypothesis. The effect was most robust in the later portion of MMN. Findings suggest that MMN is generated by multiple sources, differentially reflecting acoustic change detection (earlier MMN, bottom-up process) and perceptual weighting of ART and FRT (later MMN, top-down process).
Collapse
Affiliation(s)
- Aaron C Moberly
- The Ohio State University, Wexner Medical Center, Department of Otolaryngology - Head & Neck Surgery, United States.
| | - Jyoti Bhat
- The Ohio State University, Wexner Medical Center, Department of Otolaryngology - Head & Neck Surgery, United States
| | - D Bradley Welling
- The Ohio State University, Wexner Medical Center, Department of Otolaryngology - Head & Neck Surgery, United States
| | - Antoine J Shahin
- The Ohio State University, Wexner Medical Center, Department of Otolaryngology - Head & Neck Surgery, United States
| |
Collapse
|
39
|
Nittrouer S, Lowenstein JH. SEPARATING THE EFFECTS OF ACOUSTIC AND PHONETIC FACTORS IN LINGUISTIC PROCESSING WITH IMPOVERISHED SIGNALS BY ADULTS AND CHILDREN. APPLIED PSYCHOLINGUISTICS 2014; 35:333-370. [PMID: 24729642 PMCID: PMC3981461 DOI: 10.1017/s0142716412000410] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
Cochlear implants allow many individuals with profound hearing loss to understand spoken language, even though the impoverished signals provided by these devices poorly preserve acoustic attributes long believed to support recovery of phonetic structure. Consequently questions may be raised regarding whether traditional psycholinguistic theories rely too heavily on phonetic segments to explain linguistic processing while ignoring potential roles of other forms of acoustic structure. This study tested that possibility. Adults and children (8 years old) performed two tasks: one involving explicit segmentation, phonemic awareness, and one involving a linguistic task thought to operate more efficiently with well-defined phonetic segments, short-term memory. Stimuli were unprocessed signals (UP), amplitude envelopes (AE) analogous to implant signals, and unprocessed signals in noise (NOI) which provided a degraded signal for comparison. Adults' results for short-term recall were similar for UP and NOI, but worse for AE stimuli. The phonemic awareness task revealed the opposite pattern across AE and NOI. Children's results for short-term recall showed similar decrements in performance for AE and NOI compared to UP, even though only NOI stimuli showed diminished results for segmentation. Conclusions were that perhaps traditional accounts are too focused on phonetic segments, something implant designers and clinicians need to consider.
Collapse
|
40
|
Macrae T, Tyler AA, Lewis KE. Lexical and phonological variability in preschool children with speech sound disorder. AMERICAN JOURNAL OF SPEECH-LANGUAGE PATHOLOGY 2014; 23:27-35. [PMID: 23813198 DOI: 10.1044/1058-0360(2013/12-0037)] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
PURPOSE The authors of this study examined relationships between measures of word and speech error variability and between these and other speech and language measures in preschool children with speech sound disorder (SSD). METHOD In this correlational study, 18 preschool children with SSD, age-appropriate receptive vocabulary, and normal oral motor functioning and hearing were assessed across 2 sessions. Experimental measures included word and speech error variability, receptive vocabulary, nonword repetition (NWR), and expressive language. Pearson product–moment correlation coefficients were calculated among the experimental measures. RESULTS The correlation between word and speech error variability was slight and nonsignificant. The correlation between word variability and receptive vocabulary was moderate and negative, although nonsignificant. High word variability was associated with small receptive vocabularies. The correlations between speech error variability and NWR and between speech error variability and the mean length of children's utterances were moderate and negative, although both were nonsignificant. High speech error variability was associated with poor NWR and language scores. CONCLUSION High word variability may reflect unstable lexical representations, whereas high speech error variability may reflect indistinct phonological representations. Preschool children with SSD who show abnormally high levels of different types of speech variability may require slightly different approaches to intervention.
Collapse
|
41
|
Kong EJ, Yoon IH. L2 Proficiency Effect on the Acoustic Cue-Weighting Pattern by Korean L2 Learners of English: Production and Perception of English Stops. ACTA ACUST UNITED AC 2013. [DOI: 10.13064/ksss.2013.5.4.081] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
|
42
|
Spectral information in nonspeech contexts influences children's categorization of ambiguous speech sounds. J Exp Child Psychol 2013; 116:728-37. [PMID: 23827642 DOI: 10.1016/j.jecp.2013.05.008] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2012] [Revised: 05/20/2013] [Accepted: 05/26/2013] [Indexed: 11/20/2022]
Abstract
For both adults and children, acoustic context plays an important role in speech perception. For adults, both speech and nonspeech acoustic contexts influence perception of subsequent speech items, consistent with the argument that effects of context are due to domain-general auditory processes. However, prior research examining the effects of context on children's speech perception have focused on speech contexts; nonspeech contexts have not been explored previously. To better understand the developmental progression of children's use of contexts in speech perception and the mechanisms underlying that development, we created a novel experimental paradigm testing 5-year-old children's speech perception in several acoustic contexts. The results demonstrated that nonspeech context influences children's speech perception, consistent with claims that context effects arise from general auditory system properties rather than speech-specific mechanisms. This supports theoretical accounts of language development suggesting that domain-general processes play a role across the lifespan.
Collapse
|
43
|
Idemaru K, Holt LL. The developmental trajectory of children's perception and production of English /r/-/l/. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 133:4232-46. [PMID: 23742374 PMCID: PMC3689790 DOI: 10.1121/1.4802905] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/09/2011] [Revised: 02/07/2013] [Accepted: 04/05/2013] [Indexed: 05/23/2023]
Abstract
The English /l-r/ distinction is difficult to learn for some second language learners as well as for native-speaking children. This study examines the use of the second (F2) and third (F3) formants in the production and perception of /l/ and /r/ sounds in 4-, 4.5-, 5.5-, and 8.5-yr-old English-speaking children. The children were tested with elicitation and repetition tasks as well as word recognition tasks. The results indicate that whereas young children's /l/ and /r/ in both production and perception show fairly high accuracy and were well defined along the primary acoustic parameter that differentiates them, F3 frequency, these children were still developing in regard to the integration of the secondary cue, F2 frequency. The pattern of development is consistent with the distribution of these features in the ambient input relative to the /l/ and /r/ category distinction: F3 is robust and reliable, whereas F2 is less reliable in distinguishing /l/ and /r/. With delayed development of F2, cue weighting of F3 and F2 for the English /l-r/ categorization seems to continue to develop beyond 8 or 9 yr of age. These data are consistent with a rather long trajectory of phonetic development whereby native categories are refined and tuned well into childhood.
Collapse
Affiliation(s)
- Kaori Idemaru
- Department of East Asian Languages and Literatures, University of Oregon, Eugene, Oregon 97405, USA.
| | | |
Collapse
|
44
|
Strait DL, O'Connell S, Parbery-Clark A, Kraus N. Musicians' enhanced neural differentiation of speech sounds arises early in life: developmental evidence from ages 3 to 30. Cereb Cortex 2013; 24:2512-21. [PMID: 23599166 DOI: 10.1093/cercor/bht103] [Citation(s) in RCA: 69] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/14/2022] Open
Abstract
The perception and neural representation of acoustically similar speech sounds underlie language development. Music training hones the perception of minute acoustic differences that distinguish sounds; this training may generalize to speech processing given that adult musicians have enhanced neural differentiation of similar speech syllables compared with nonmusicians. Here, we asked whether this neural advantage in musicians is present early in life by assessing musically trained and untrained children as young as age 3. We assessed auditory brainstem responses to the speech syllables /ba/ and /ga/ as well as auditory and visual cognitive abilities in musicians and nonmusicians across 3 developmental time-points: preschoolers, school-aged children, and adults. Cross-phase analyses objectively measured the degree to which subcortical responses differed to these speech syllables in musicians and nonmusicians for each age group. Results reveal that musicians exhibit enhanced neural differentiation of stop consonants early in life and with as little as a few years of training. Furthermore, the extent of subcortical stop consonant distinction correlates with auditory-specific cognitive abilities (i.e., auditory working memory and attention). Results are interpreted according to a corticofugal framework for auditory learning in which subcortical processing enhancements are engendered by strengthened cognitive control over auditory function in musicians.
Collapse
Affiliation(s)
- Dana L Strait
- Institute for Neuroscience, Auditory Neuroscience Laboratory
| | | | | | - Nina Kraus
- Institute for Neuroscience, Auditory Neuroscience Laboratory, Department of Communication Sciences, Department of Neurobiology and Physiology and Department of Otolaryngology, Northwestern University, Evanston, IL, USA
| |
Collapse
|
45
|
Nittrouer S, Lowenstein JH, Tarr E. Amplitude rise time does not cue the /ba/-/wa/ contrast for adults or children. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2013; 56:427-440. [PMID: 22992704 PMCID: PMC3810943 DOI: 10.1044/1092-4388(2012/12-0075)] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
PURPOSE Previous research has demonstrated that children weight the acoustic cues to many phonemic decisions differently than do adults and gradually shift those strategies as they gain language experience. However, that research has focused on spectral and duration cues rather than on amplitude cues. In the current study, the authors examined amplitude rise time (ART; an amplitude cue) and formant rise time (FRT; a spectral cue) in the /b/-/w/ manner contrast for adults and children, and related those speech decisions to outcomes of nonspeech discrimination tasks. METHOD Twenty adults and 30 children (ages 4-5 years) labeled natural and synthetic speech stimuli manipulated to vary ARTs and FRTs, and discriminated nonspeech analogs that varied only by ART in an AX paradigm. RESULTS Three primary results were obtained. First, listeners in both age groups based speech labeling judgments on FRT, not on ART. Second, the fundamental frequency of the natural speech samples did not influence labeling judgments. Third, discrimination performance for the nonspeech stimuli did not predict how listeners would perform with the speech stimuli. CONCLUSION Even though both adults and children are sensitive to ART, it was not weighted in phonemic judgments by these typical listeners.
Collapse
|
46
|
Storkel HL, Maekawa J, Aschenbrenner AJ. The effect of homonymy on learning correctly articulated versus misarticulated words. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2013; 56:694-707. [PMID: 23275395 PMCID: PMC3615102 DOI: 10.1044/1092-4388(2012/12-0122)] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/27/2023]
Abstract
PURPOSE The goal of the current study was to examine the effect of homonymy (learning a second meaning for a known word form vs. learning a novel meaning and novel word form) and articulation accuracy (IN vs. OUT sounds) on word learning by preschool children. An added goal was to determine whether word frequency altered the effect of homonymy on word learning. METHOD Twenty-nine 3- to 4-year-old children were taught homonyms and novel words. Stimuli further varied in whether homonymy was present in both the adult input and the child's output (as for IN sounds) versus present only in the child's output (as for OUT sounds). RESULTS For IN sounds, children learned homonyms more rapidly than novel words. Moreover, the homonym advantage was modulated by word frequency, such that children learned a new meaning for a high-frequency word more accurately than they learned a new meaning for a low-frequency word. In contrast, for OUT sounds, there was no evidence that homonymy influenced learning. CONCLUSIONS Homonymy in the adult input facilitates word learning by preschool children, whereas homonymy in the child's output alone does not. This effect is captured in a usage-based model of phonology and the lexicon.
Collapse
|
47
|
Holt RF, Lalonde K. Assessing toddlers' speech-sound discrimination. Int J Pediatr Otorhinolaryngol 2012; 76:680-92. [PMID: 22402014 PMCID: PMC3335986 DOI: 10.1016/j.ijporl.2012.02.020] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/27/2011] [Revised: 02/03/2012] [Accepted: 02/05/2012] [Indexed: 11/26/2022]
Abstract
OBJECTIVE Valid and reliable methods for assessing speech perception in toddlers are lacking in the field, leading to conspicuous gaps in understanding how speech perception develops and limited clinical tools for assessing sensory aid benefit in toddlers. The objective of this investigation was to evaluate speech-sound discrimination in toddlers using modifications to the Change/No-Change procedure [1]. METHODS Normal-hearing 2- and 3-year-olds' discrimination of acoustically dissimilar ("easy") and similar ("hard") speech-sound contrasts were evaluated in a combined repeated measures and factorial design. Performance was measured in d'. Effects of contrast difficulty and age were examined, as was test-retest reliability, using repeated measures ANOVAs, planned post hoc tests, and correlation analyses. RESULTS The easy contrast (M=2.53) was discriminated better than the hard contrast (M=1.72) across all ages (p<.0001). The oldest group of children (M=3.13) discriminated the contrasts better than youngest (M=1.04; p<.0001) and the mid-age children (M=2.20; p=.037), who in turn discriminated the contrasts better than the youngest children (p=.010). Test-retest reliability was excellent (r=.886, p<.0001). Almost 90% of the children met the teaching criterion. The vast majority demonstrated the ability to be tested with the modified procedure and discriminated the contrasts. The few who did not were 2.5 years of age and younger. CONCLUSIONS The modifications implemented resulted, at least preliminarily, in a procedure that is reliable and sensitive to contrast difficulty and age in this young group of children, suggesting that these modifications are appropriate for this age group. With further development, the procedure holds promise for use in clinical populations who are believed to have core deficits in rapid phonological encoding, such as children with hearing loss or specific language impairment, children who are struggling to read, and second-language learners.
Collapse
Affiliation(s)
- Rachael Frush Holt
- Department of Speech and Hearing Sciences, Indiana University, 200 South Jordan Avenue, Bloomington, IN 47405, USA.
| | | |
Collapse
|
48
|
Li F. Language-specific developmental differences in speech production: a cross-language acoustic study. Child Dev 2012; 83:1303-15. [PMID: 22540834 DOI: 10.1111/j.1467-8624.2012.01773.x] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
Abstract
Speech productions of 40 English- and 40 Japanese-speaking children (aged 2-5) were examined and compared with the speech produced by 20 adult speakers (10 speakers per language). Participants were recorded while repeating words that began with "s" and "sh" sounds. Clear language-specific patterns in adults' speech were found, with English speakers differentiating "s" and "sh" in 1 acoustic dimension (i.e., spectral mean) and Japanese speakers differentiating the 2 categories in 3 acoustic dimensions (i.e., spectral mean, standard deviation, and onset F2 frequency). For both language groups, children's speech exhibited a gradual change from an early undifferentiated form to later differentiated categories. The separation processes, however, only occur in those acoustic dimensions used by adults in the corresponding languages.
Collapse
Affiliation(s)
- Fangfang Li
- Department of Psychology, University of Lethbridge, 4401 University Drive, Lethbridge, Alberta, Canada.
| |
Collapse
|
49
|
Shafer VL, Yu YH, Datta H. The Development of English Vowel Perception in Monolingual and Bilingual Infants: Neurophysiological Correlates. JOURNAL OF PHONETICS 2011; 39:527-545. [PMID: 22046059 PMCID: PMC3201800 DOI: 10.1016/j.wocn.2010.11.010] [Citation(s) in RCA: 37] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/06/2023]
Abstract
The goal of this paper was to examine intrinsic and extrinsic factors contributing to the development of speech perception in monolingual and bilingual infants and toddlers. A substantial number of behavioral studies have characterized when infants show changes in behavior towards speech sounds in relation to amount of experience with these sounds. However, these studies cannot explain to what extent the developmental timeline is influenced by experience with the language versus constraints imposed by cortical maturation. Studies using electrophysiological measures to examine the development of auditory and speech processing have shown great differences in infant and adult electrophysiological correlates of processing. Many of these differences are a function of immature cortex in the infant. In this paper, we examined the maturation of infant and child event-related-potential (ERP) electrophysiological components in processing an English vowel contrast and explored to what extent these components are influenced by intrinsic (e.g., sex) versus extrinsic factors, such as language experience (monolingual vs. bilingual). Our findings demonstrate differences in the pattern of ERP responses related to age and sex, as well as language experience. These differences make it clear that general maturational factors need to be taken into consideration in examining the effect of language experience on the neurodevelopment of speech perception.
Collapse
|
50
|
Holt RF. Enhancing speech discrimination through stimulus repetition. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2011; 54:1431-1447. [PMID: 21646420 DOI: 10.1044/1092-4388(2011/09-0242)] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
PURPOSE To evaluate the effects of sequential and alternating repetition on speech-sound discrimination. METHOD Typically hearing adults' discrimination of 3 pairs of speech-sound contrasts was assessed at 3 signal-to-noise ratios using the change/no-change procedure. On change trials, the standard and comparison stimuli differ; on no-change trials, they are identical. Listeners were presented with 5 repetition conditions: 2 and 4 sequential repetitions of the standard followed by sequential repetitions of the comparison; 2 and 4 alternating presentations of the standard and comparison; and 1 repetition of the standard and comparison. RESULTS Both sequential and alternating repetition improved discrimination of the fricative and liquid contrasts, but neither was clearly superior to the other across the conditions. CONCLUSIONS The results support previous findings that increasing the number of fricative and liquid stimulus presentations improves discriminability and extends the findings to natural speech stimuli. Further, the effect of repetition is robust: Both sequential and alternating repetitions improve speech-sound discrimination, and few differences emerge between the two types of stimulus repetitions. The results have implications for evaluating the strength of the internal representation of speech stimuli in clinical populations believed to have a core deficit in phonological encoding, such as children with hearing loss.
Collapse
|