1
|
Werker JF. Phonetic perceptual reorganization across the first year of life: Looking back. Infant Behav Dev 2024; 75:101935. [PMID: 38569416 DOI: 10.1016/j.infbeh.2024.101935] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2023] [Revised: 02/26/2024] [Accepted: 02/27/2024] [Indexed: 04/05/2024]
Abstract
This paper provides a selective overview of some of the research that has followed from the publication of Werker and Tees (1984a) "Cross-language speech perception: Evidence for Perceptual Reorganization During the First Year of Life." Specifically, I briefly present the original finding, our interpretation of its meaning, and some key replications and extensions. I then review some of the work that has followed, including work with different kinds of populations, different kinds of speech sound contrasts, as well as attunement (perceptual reorganization) to additional properties of language beyond phonetic contrasts. Included is the body of work that queries whether perceptual attunement is a critical period phenomenon. Potential learning mechanisms for how experience functions to guide phonetic perceptual development are also presented, as is work on the relation between speech perception and word learning.
Collapse
Affiliation(s)
- Janet F Werker
- Department of Psychology, The University of British Columbia, Canada.
| |
Collapse
|
2
|
Ren J, Wang M. Contribution of statistical learning in learning to read across languages. PLoS One 2024; 19:e0298670. [PMID: 38527080 PMCID: PMC10962809 DOI: 10.1371/journal.pone.0298670] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2023] [Accepted: 01/26/2024] [Indexed: 03/27/2024] Open
Abstract
Statistical Learning (SL) refers to human's ability to detect regularities from environment Kirkham, N. Z. (2002) & Saffran, J. R. (1996). There has been a growing interest in understanding how sensitivity to statistical regularities influences learning to read. The current study systematically examined whether and how non-linguistic SL, Chinese SL, and English SL contribute to Chinese and English word reading among native Chinese-speaking 4th, 6th and 8th graders who learn English as a second language (L2). Children showed above-chance learning across all SL tasks and across all grades. In addition, developmental improvements were shown across at least two of the three grade ranges on all SL tasks. In terms of the contribution of SL to reading, non-linguistic auditory SL (ASL), English visual SL (VSL), and Chinese ASL accounted for a significant amount of variance in English L2 word reading. Non-linguistic ASL, Chinese VSL, English VSL, and English ASL accounted for a significant amount of variance in Chinese word reading. Our results provide clear and novel evidence for cross-linguistic contribution from Chinese SL to English reading, and from English SL to Chinese reading, highlighting a bi-directional relationship between SL in one language and reading in another language.
Collapse
Affiliation(s)
- Jinglei Ren
- Department of Human Development and Quantitative Methodology, University of Maryland, College Park, Maryland, United States of America
| | - Min Wang
- Department of Human Development and Quantitative Methodology, University of Maryland, College Park, Maryland, United States of America
| |
Collapse
|
3
|
Ko ES, Jun J. Phonological Variation in Child-Directed Speech is Modulated by Lexical Frequency. JOURNAL OF CHILD LANGUAGE 2024; 51:288-313. [PMID: 37737231 DOI: 10.1017/s0305000923000466] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/23/2023]
Abstract
We investigate whether child-directed speech (CDS) contains a higher proportion of canonical pronunciations compared to adult-directed speech (ADS), focusing on Korean noun stem-final obstruent variation. In a word-teaching task, we observed that mothers use a higher rate of canonical pronunciation when addressing infants than when addressing adults. In a follow-up experiment, adults exhibited a higher rate of canonical pronunciation for high- than low-frequency words. Additional analyses conducted with only the high-frequency monosyllabic words from the two experiments found no evidence for simplified phonology in CDS when lexical frequency was controlled for. Our findings suggest that the higher rate of canonical forms in CDS, with respect to Korean morphophonological rules, is mediated by the frequency of word usage. Thus, the didactic function of CDS phonology appears to be a byproduct of mothers using familiar words with children. These results highlight the importance of considering word usage in investigating the nature of CDS.
Collapse
Affiliation(s)
- Eon-Suk Ko
- Department of English Language and Literature, Chosun University, Gwangju, Korea
| | - Jongho Jun
- Department of Linguistics, Seoul National University, Seoul, Korea
| |
Collapse
|
4
|
Zhou X, Wang L, Hong X, Wong PCM. Infant-directed speech facilitates word learning through attentional mechanisms: An fNIRS study of toddlers. Dev Sci 2024; 27:e13424. [PMID: 37322865 DOI: 10.1111/desc.13424] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/10/2022] [Revised: 05/27/2023] [Accepted: 05/30/2023] [Indexed: 06/17/2023]
Abstract
The speech register that adults especially caregivers use when interacting with infants and toddlers, that is, infant-directed speech (IDS) or baby talk, has been reported to facilitate language development throughout the early years. However, the neural mechanisms as well as why IDS results in such a developmental faciliatory effect remain to be investigated. The current study uses functional near-infrared spectroscopy (fNIRS) to evaluate two alternative hypotheses of such a facilitative effect, that IDS serves to enhance linguistic contrastiveness or to attract the child's attention. Behavioral and fNIRS data were acquired from twenty-seven Cantonese-learning toddlers 15-20 months of age when their parents spoke to them in either an IDS or adult-directed speech (ADS) register in a naturalistic task in which the child learned four disyllabic pseudowords. fNIRS results showed significantly greater neural responses to IDS than ADS register in the left dorsolateral prefrontal cortex (L-dlPFC), but opposite response patterns in the bilateral inferior frontal gyrus (IFG). The differences in fNIRS responses to IDS and to ADS in the L-dlPFC and the left parietal cortex (L-PC) showed significantly positive correlations with the differences in the behavioral word-learning performance of toddlers. The same fNIRS measures in the L-dlPFC and right PC (R-PC) of toddlers were significantly correlated with pitch range differences of parents between the two speech conditions. Together, our results suggest that the dynamic prosody in IDS increased toddlers' attention through greater involvement of the left frontoparietal network that facilitated word learning, compared to ADS. RESEARCH HIGHLIGHTS: This study for the first time examined the neural mechanisms of how infant-directed speech (IDS) facilitates word learning in toddlers. Using fNIRS, we identified the cortical regions that were directly involved in IDS processing. Our results suggest that IDS facilitates word learning by engaging a right-lateralized prosody processing and top-down attentional mechanisms in the left frontoparietal networks. The language network including the inferior frontal gyrus and temporal cortex was not directly involved in IDS processing to support word learning.
Collapse
Affiliation(s)
- Xin Zhou
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong SAR, China
- Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong SAR, China
| | - Luchang Wang
- Department of Applied Linguistics, Xi'an Jiaotong-Liverpool University, Suzhou, China
| | - Xuancu Hong
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong SAR, China
| | - Patrick C M Wong
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong SAR, China
- Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong SAR, China
| |
Collapse
|
5
|
Geangu E, Smith WAP, Mason HT, Martinez-Cedillo AP, Hunter D, Knight MI, Liang H, del Carmen Garcia de Soria Bazan M, Tse ZTH, Rowland T, Corpuz D, Hunter J, Singh N, Vuong QC, Abdelgayed MRS, Mullineaux DR, Smith S, Muller BR. EgoActive: Integrated Wireless Wearable Sensors for Capturing Infant Egocentric Auditory-Visual Statistics and Autonomic Nervous System Function 'in the Wild'. SENSORS (BASEL, SWITZERLAND) 2023; 23:7930. [PMID: 37765987 PMCID: PMC10534696 DOI: 10.3390/s23187930] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/28/2023] [Revised: 08/25/2023] [Accepted: 09/11/2023] [Indexed: 09/29/2023]
Abstract
There have been sustained efforts toward using naturalistic methods in developmental science to measure infant behaviors in the real world from an egocentric perspective because statistical regularities in the environment can shape and be shaped by the developing infant. However, there is no user-friendly and unobtrusive technology to densely and reliably sample life in the wild. To address this gap, we present the design, implementation and validation of the EgoActive platform, which addresses limitations of existing wearable technologies for developmental research. EgoActive records the active infants' egocentric perspective of the world via a miniature wireless head-mounted camera concurrently with their physiological responses to this input via a lightweight, wireless ECG/acceleration sensor. We also provide software tools to facilitate data analyses. Our validation studies showed that the cameras and body sensors performed well. Families also reported that the platform was comfortable, easy to use and operate, and did not interfere with daily activities. The synchronized multimodal data from the EgoActive platform can help tease apart complex processes that are important for child development to further our understanding of areas ranging from executive function to emotion processing and social learning.
Collapse
Affiliation(s)
- Elena Geangu
- Psychology Department, University of York, York YO10 5DD, UK; (A.P.M.-C.); (M.d.C.G.d.S.B.)
| | - William A. P. Smith
- Department of Computer Science, University of York, York YO10 5DD, UK; (W.A.P.S.); (J.H.); (M.R.S.A.); (B.R.M.)
| | - Harry T. Mason
- School of Physics, Engineering and Technology, University of York, York YO10 5DD, UK; (H.T.M.); (D.H.); (N.S.); (S.S.)
| | | | - David Hunter
- School of Physics, Engineering and Technology, University of York, York YO10 5DD, UK; (H.T.M.); (D.H.); (N.S.); (S.S.)
| | - Marina I. Knight
- Department of Mathematics, University of York, York YO10 5DD, UK; (M.I.K.); (D.R.M.)
| | - Haipeng Liang
- School of Engineering and Materials Science, Queen Mary University of London, London E1 2AT, UK; (H.L.); (Z.T.H.T.)
| | | | - Zion Tsz Ho Tse
- School of Engineering and Materials Science, Queen Mary University of London, London E1 2AT, UK; (H.L.); (Z.T.H.T.)
| | - Thomas Rowland
- Protolabs, Halesfield 8, Telford TF7 4QN, UK; (T.R.); (D.C.)
| | - Dom Corpuz
- Protolabs, Halesfield 8, Telford TF7 4QN, UK; (T.R.); (D.C.)
| | - Josh Hunter
- Department of Computer Science, University of York, York YO10 5DD, UK; (W.A.P.S.); (J.H.); (M.R.S.A.); (B.R.M.)
| | - Nishant Singh
- School of Physics, Engineering and Technology, University of York, York YO10 5DD, UK; (H.T.M.); (D.H.); (N.S.); (S.S.)
| | - Quoc C. Vuong
- Biosciences Institute, Newcastle University, Newcastle upon Tyne NE1 7RU, UK;
| | - Mona Ragab Sayed Abdelgayed
- Department of Computer Science, University of York, York YO10 5DD, UK; (W.A.P.S.); (J.H.); (M.R.S.A.); (B.R.M.)
| | - David R. Mullineaux
- Department of Mathematics, University of York, York YO10 5DD, UK; (M.I.K.); (D.R.M.)
| | - Stephen Smith
- School of Physics, Engineering and Technology, University of York, York YO10 5DD, UK; (H.T.M.); (D.H.); (N.S.); (S.S.)
| | - Bruce R. Muller
- Department of Computer Science, University of York, York YO10 5DD, UK; (W.A.P.S.); (J.H.); (M.R.S.A.); (B.R.M.)
| |
Collapse
|
6
|
Bundgaard-Nielsen RL, O'Shannessy C, Wang Y, Nelson A, Bartlett J, Davis V. Two-part vowel modifications in Child Directed Speech in Warlpiri may enhance child attention to speech and scaffold noun acquisition. PHONETICA 2023; 0:phon-2022-0039. [PMID: 37314963 DOI: 10.1515/phon-2022-0039] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/28/2022] [Accepted: 05/08/2023] [Indexed: 06/16/2023]
Abstract
Study 1 compared vowels in Child Directed Speech (CDS; child ages 25-46 months) to vowels in Adult Directed Speech (ADS) in natural conversation in the Australian Indigenous language Warlpiri, which has three vowels (/i/, /a/, /u). Study 2 compared the vowels of the child interlocutors from Study 1 to caregiver ADS and CDS. Study 1 indicates that Warlpiri CDS vowels are characterised by fronting, /a/-lowering, f o -raising, and increased duration, but not vowel space expansion. Vowels in CDS nouns, however, show increased between-contrast differentiation and reduced within-contrast variation, similar to what has been reported for other languages. We argue that this two-part CDS modification process serves a dual purpose: Vowel space shifting induces IDS/CDS that sounds more child-like, which may enhance child attention to speech, while increased between-contrast differentiation and reduced within-contrast variation in nouns may serve didactic purposes by providing high-quality information about lexical specifications. Study 2 indicates that Warlpiri CDS vowels are more like child vowels, providing indirect evidence that aspects of CDS may serve non-linguistic purposes simultaneously with other aspects serving linguistic-didactic purposes. The studies have novel implications for the way CDS vowel modifications are considered and highlight the necessity of naturalistic data collection, novel analyses, and typological diversity.
Collapse
Affiliation(s)
- Rikke L Bundgaard-Nielsen
- MARCS Institute of Brain, Behaviour and Development, Western Sydney University, Westmead, NSW, Australia
- School of Literature, Languages and Linguistics, Australian National University, Canberra, ACT, Australia
| | - Carmel O'Shannessy
- School of Literature, Languages and Linguistics, Australian National University, Canberra, ACT, Australia
| | - Yizhou Wang
- School of Literature, Languages and Linguistics, Australian National University, Canberra, ACT, Australia
| | - Alice Nelson
- Red Dust Role Models, Alice Springs, NT, Australia
| | | | - Vanessa Davis
- School of Literature, Languages and Linguistics, Australian National University, Canberra, ACT, Australia
- Tangentyere Council Research Hub, Alice Springs, NT, Australia
| |
Collapse
|
7
|
Marriott Haresign I, Phillips EAM, Whitehorn M, Lamagna F, Eliano M, Goupil L, Jones EJH, Wass SV. Gaze onsets during naturalistic infant-caregiver interaction associate with 'sender' but not 'receiver' neural responses, and do not lead to changes in inter-brain synchrony. Sci Rep 2023; 13:3555. [PMID: 36864074 PMCID: PMC9981599 DOI: 10.1038/s41598-023-28988-0] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2022] [Accepted: 01/27/2023] [Indexed: 03/04/2023] Open
Abstract
Temporal coordination during infant-caregiver social interaction is thought to be crucial for supporting early language acquisition and cognitive development. Despite a growing prevalence of theories suggesting that increased inter-brain synchrony associates with many key aspects of social interactions such as mutual gaze, little is known about how this arises during development. Here, we investigated the role of mutual gaze onsets as a potential driver of inter-brain synchrony. We extracted dual EEG activity around naturally occurring gaze onsets during infant-caregiver social interactions in N = 55 dyads (mean age 12 months). We differentiated between two types of gaze onset, depending on each partners' role. 'Sender' gaze onsets were defined at a time when either the adult or the infant made a gaze shift towards their partner at a time when their partner was either already looking at them (mutual) or not looking at them (non-mutual). 'Receiver' gaze onsets were defined at a time when their partner made a gaze shift towards them at a time when either the adult or the infant was already looking at their partner (mutual) or not (non-mutual). Contrary to our hypothesis we found that, during a naturalistic interaction, both mutual and non-mutual gaze onsets were associated with changes in the sender, but not the receiver's brain activity and were not associated with increases in inter-brain synchrony above baseline. Further, we found that mutual, compared to non-mutual gaze onsets were not associated with increased inter brain synchrony. Overall, our results suggest that the effects of mutual gaze are strongest at the intra-brain level, in the 'sender' but not the 'receiver' of the mutual gaze.
Collapse
Affiliation(s)
| | - E A M Phillips
- Department of Psychology, University of East London, London, E15 4LZ, UK
| | - M Whitehorn
- Department of Psychology, University of East London, London, E15 4LZ, UK
| | - F Lamagna
- Department of Psychology, University of East London, London, E15 4LZ, UK
| | - M Eliano
- Department of Psychology, University of East London, London, E15 4LZ, UK
| | - L Goupil
- LPNC/CNRS, Grenoble Alpes University, Grenoble, France
| | - E J H Jones
- Centre for Brain and Cognitive Development, Birkbeck College, University of London, London, UK
| | - S V Wass
- Department of Psychology, University of East London, London, E15 4LZ, UK
| |
Collapse
|
8
|
Understanding why infant-directed speech supports learning: A dynamic attention perspective. DEVELOPMENTAL REVIEW 2022. [DOI: 10.1016/j.dr.2022.101047] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
9
|
Pérez-Navarro J, Lallier M, Clark C, Flanagan S, Goswami U. Local Temporal Regularities in Child-Directed Speech in Spanish. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:3776-3788. [PMID: 36194778 DOI: 10.1044/2022_jslhr-22-00111] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
PURPOSE The purpose of this study is to characterize the local (utterance-level) temporal regularities of child-directed speech (CDS) that might facilitate phonological development in Spanish, classically termed a syllable-timed language. METHOD Eighteen female adults addressed their 4-year-old children versus other adults spontaneously and also read aloud (CDS vs. adult-directed speech [ADS]). We compared CDS and ADS speech productions using a spectrotemporal model (Leong & Goswami, 2015), obtaining three temporal metrics: (a) distribution of modulation energy, (b) temporal regularity of stressed syllables, and (c) syllable rate. RESULTS CDS was characterized by (a) significantly greater modulation energy in the lower frequencies (0.5-4 Hz), (b) more regular rhythmic occurrence of stressed syllables, and (c) a slower syllable rate than ADS, across both spontaneous and read conditions. DISCUSSION CDS is characterized by a robust local temporal organization (i.e., within utterances) with amplitude modulation bands aligning with delta and theta electrophysiological frequency bands, respectively, showing greater phase synchronization than in ADS, facilitating parsing of stress units and syllables. These temporal regularities, together with the slower rate of production of CDS, might support the automatic extraction of phonological units in speech and hence support the phonological development of children. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.21210893.
Collapse
Affiliation(s)
- Jose Pérez-Navarro
- BCBL, Basque Center on Cognition, Brain and Language, Donostia-San Sebastián, Spain
- University of the Basque Country UPV/EHU, Donostia-San Sebastián, Spain
| | - Marie Lallier
- BCBL, Basque Center on Cognition, Brain and Language, Donostia-San Sebastián, Spain
| | - Catherine Clark
- BCBL, Basque Center on Cognition, Brain and Language, Donostia-San Sebastián, Spain
- University of the Basque Country UPV/EHU, Donostia-San Sebastián, Spain
| | - Sheila Flanagan
- Centre for Neuroscience in Education, Department of Psychology, University of Cambridge, United Kingdom
| | - Usha Goswami
- Centre for Neuroscience in Education, Department of Psychology, University of Cambridge, United Kingdom
| |
Collapse
|
10
|
Zacharaki K, Sebastian-Galles N. Before perceptual narrowing: The emergence of the native sounds of language. INFANCY 2022; 27:900-915. [PMID: 35632983 DOI: 10.1111/infa.12486] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2021] [Revised: 04/15/2022] [Accepted: 04/20/2022] [Indexed: 11/28/2022]
Abstract
The present study investigates the precursors of representations of phonemes in 4.5-month-olds. The emergence of phonemes has been mainly studied within the framework of perceptual narrowing, that is, infants tuning to their native language and losing sensitivity to non-native speech. One of the mechanisms behind this phenomenon is distributional learning. In this article, we tested the preference of 4.5-month-old infants using lists of pseudowords that resemble the vowel distribution of the native or a non-native language. We found that infants prefer listening to the lists mirroring the native language. The results suggest that infants can extract vowel information from novel stimuli, and they can map it on pre-existing knowledge on vowels that leads to a preference for the native lists.
Collapse
|
11
|
Michon M, Zamorano-Abramson J, Aboitiz F. Faces and Voices Processing in Human and Primate Brains: Rhythmic and Multimodal Mechanisms Underlying the Evolution and Development of Speech. Front Psychol 2022; 13:829083. [PMID: 35432052 PMCID: PMC9007199 DOI: 10.3389/fpsyg.2022.829083] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2021] [Accepted: 03/07/2022] [Indexed: 11/24/2022] Open
Abstract
While influential works since the 1970s have widely assumed that imitation is an innate skill in both human and non-human primate neonates, recent empirical studies and meta-analyses have challenged this view, indicating other forms of reward-based learning as relevant factors in the development of social behavior. The visual input translation into matching motor output that underlies imitation abilities instead seems to develop along with social interactions and sensorimotor experience during infancy and childhood. Recently, a new visual stream has been identified in both human and non-human primate brains, updating the dual visual stream model. This third pathway is thought to be specialized for dynamics aspects of social perceptions such as eye-gaze, facial expression and crucially for audio-visual integration of speech. Here, we review empirical studies addressing an understudied but crucial aspect of speech and communication, namely the processing of visual orofacial cues (i.e., the perception of a speaker's lips and tongue movements) and its integration with vocal auditory cues. Along this review, we offer new insights from our understanding of speech as the product of evolution and development of a rhythmic and multimodal organization of sensorimotor brain networks, supporting volitional motor control of the upper vocal tract and audio-visual voices-faces integration.
Collapse
Affiliation(s)
- Maëva Michon
- Laboratory for Cognitive and Evolutionary Neuroscience, Department of Psychiatry, Faculty of Medicine, Interdisciplinary Center for Neuroscience, Pontificia Universidad Católica de Chile, Santiago, Chile
- Centro de Estudios en Neurociencia Humana y Neuropsicología, Facultad de Psicología, Universidad Diego Portales, Santiago, Chile
| | - José Zamorano-Abramson
- Centro de Investigación en Complejidad Social, Facultad de Gobierno, Universidad del Desarrollo, Santiago, Chile
| | - Francisco Aboitiz
- Laboratory for Cognitive and Evolutionary Neuroscience, Department of Psychiatry, Faculty of Medicine, Interdisciplinary Center for Neuroscience, Pontificia Universidad Católica de Chile, Santiago, Chile
| |
Collapse
|
12
|
Feldman NH, Goldwater S, Dupoux E, Schatz T. Do Infants Really Learn Phonetic Categories? OPEN MIND 2022; 5:113-131. [PMID: 35024527 PMCID: PMC8746127 DOI: 10.1162/opmi_a_00046] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2020] [Accepted: 08/06/2021] [Indexed: 11/16/2022] Open
Abstract
Early changes in infants’ ability to perceive native and nonnative speech sound contrasts are typically attributed to their developing knowledge of phonetic categories. We critically examine this hypothesis and argue that there is little direct evidence of category knowledge in infancy. We then propose an alternative account in which infants’ perception changes because they are learning a perceptual space that is appropriate to represent speech, without yet carving up that space into phonetic categories. If correct, this new account has substantial implications for understanding early language development.
Collapse
Affiliation(s)
- Naomi H Feldman
- Department of Linguistics and UMIACS, University of Maryland, College Park, MD, USA
| | | | - Emmanuel Dupoux
- Cognitive Machine Learning (ENS - EHESS - PSL Research University - CNRS - INRIA), Paris, France
| | - Thomas Schatz
- Department of Linguistics and UMIACS, University of Maryland, College Park, MD, USA
| |
Collapse
|
13
|
Smith SA, Leon Guerrero S, Surrain S, Luk G. Phonetic discrimination, phonological awareness, and pre-literacy skills in Spanish-English dual language preschoolers. JOURNAL OF CHILD LANGUAGE 2022; 49:80-113. [PMID: 33568236 DOI: 10.1017/s0305000920000768] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The current study explores variation in phonemic representation among Spanish-English dual language learners (DLLs, n = 60) who were dominant in English or in Spanish. Children were given a phonetic discrimination task with speech sounds that: 1) occur in English and Spanish, 2) are exclusive to English, and 3) are exclusive to Russian, during Fall (age m = 57 months) and Spring (age m = 62 months, n = 42). In Fall, English-dominant DLLs discriminated more accurately than Spanish-dominant DLLs between English-Spanish phones and English-exclusive phones. Both groups discriminated Russian phones at or close to chance. In Spring, however, groups no longer differed in discriminating English-exclusive phones and both groups discriminated Russian phones above chance. Additionally, joint English-Spanish and English-exclusive phonetic discrimination predicted children's phonological awareness in both groups. Results demonstrate plasticity in early childhood through diverse language exposure and suggest that phonemic representation begins to emerge driven by lexical restructuring.
Collapse
|
14
|
Wang L, Kalashnikova M, Kager R, Lai R, Wong PCM. Lexical and Prosodic Pitch Modifications in Cantonese Infant-directed Speech. JOURNAL OF CHILD LANGUAGE 2021; 48:1235-1261. [PMID: 33531090 DOI: 10.1017/s0305000920000707] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The functions of acoustic-phonetic modifications in infant-directed speech (IDS) remain a question: do they specifically serve to facilitate language learning via enhanced phonemic contrasts (the hyperarticulation hypothesis) or primarily to improve communication via prosodic exaggeration (the prosodic hypothesis)? The study of lexical tones provides a unique opportunity to shed light on this, as lexical tones are phonemically contrastive, yet their primary cue, pitch, is also a prosodic cue. This study investigated Cantonese IDS and found increased intra-talker variation of lexical tones, which more likely posed a challenge to rather than facilitated phonetic learning. Although tonal space was expanded which could facilitate phonetic learning, its expansion was a function of overall intonational modifications. Similar findings were observed in speech to pets who should not benefit from larger phonemic distinction. We conclude that lexical-tone adjustments in IDS mainly serve to broadly enhance communication rather than specifically increase phonemic contrast for learners.
Collapse
Affiliation(s)
- Luchang Wang
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong, China
- Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong, China
| | - Marina Kalashnikova
- Basque Center on Cognition, Brain and Language, San Sebastian, Spain
- IKERBASQUE, Basque Foundation for Science, Bilbao, Spain
| | - René Kager
- Utrecht Institute of Linguistics OTS, Utrecht University, Utrecht, Netherlands
| | - Regine Lai
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong, China
| | - Patrick C M Wong
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong, China
- Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong, China
| |
Collapse
|
15
|
McClay EK, Cebioglu S, Broesch T, Yeung HH. Rethinking the phonetics of baby-talk: Differences across Canada and Vanuatu in the articulation of mothers' speech to infants. Dev Sci 2021; 25:e13180. [PMID: 34633716 DOI: 10.1111/desc.13180] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2021] [Revised: 09/04/2021] [Accepted: 09/15/2021] [Indexed: 11/29/2022]
Abstract
Infant-directed speech (IDS) is phonetically distinct from adult-directed speech (ADS): It is typically considered to have special prosody-like higher pitch and slower speaking rates-as well as unique speech sound properties, for example, more breathy, hyperarticulated, and/or variable consonant and vowel articulation. These phonetic features are widely observed in the IDS of caregivers from urbanized contexts who speak a handful of very well-researched languages. Yet studies with more diverse socio-cultural and linguistic samples show that this "typical" IDS prosody is not consistently observed across cultures. We extended cross-cultural work by examining IDS speech segment articulation, which-like prosody-is also thought to be a characteristic phonetic feature of IDS that might aid speech and language development. Here we asked whether IDS vowels have different articulatory features compared to ADS vowels in two distinct linguistic and socio-cultural contexts: urban English-speaking Canadian mothers, and rural Lenakel- and Southwest Tanna-speaking ni-Vanuatu mothers (n = 57, 20-46 years of age). Replicating prior work, Canadian mothers had more variable vowels in IDS compared to ADS, but also did not show clear register differences for breathiness or hyperarticulation. Vowels spoken by ni-Vanuatu mothers showed very distinct articulatory tendencies, using less variable (and less breathy) IDS vowels. Along with other work showing diversity in IDS phonetics across populations, this paper suggests that any understanding of how IDS might aid speech and language development are best examined through a culturally- and linguistically-specific lens.
Collapse
Affiliation(s)
- Elise K McClay
- Department of Linguistics, Simon Fraser University, Burnaby, BC, Canada
| | - Senay Cebioglu
- Department of Psychology, Simon Fraser University, Burnaby, BC, Canada.,Department of Human Behavior, Ecology and Culture, Max Planck Institute for Evolutionary Anthropology, Deutscher Platz 6, Leipzig, 04103, Germany
| | - Tanya Broesch
- Department of Human Behavior, Ecology and Culture, Max Planck Institute for Evolutionary Anthropology, Deutscher Platz 6, Leipzig, 04103, Germany
| | - H Henny Yeung
- Department of Linguistics, Simon Fraser University, Burnaby, BC, Canada.,Integrative Neuroscience and Cognition Center, UMR 8002, Centre national de la recherche scientifique (CNRS) & Université de Paris, Paris, 75006, France
| |
Collapse
|
16
|
He AX. Optimal input for language development: Tailor nurture to nature. INFANT AND CHILD DEVELOPMENT 2021. [DOI: 10.1002/icd.2269] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Affiliation(s)
- Angela Xiaoxue He
- Department of English Language & Literature Hong Kong Baptist University Kowloon Tong, Hong Kong SAR China
| |
Collapse
|
17
|
|
18
|
Yasufuku K, Doyle G. Echoes of L1 Syllable Structure in L2 Phoneme Recognition. Front Psychol 2021; 12:515237. [PMID: 34354620 PMCID: PMC8329372 DOI: 10.3389/fpsyg.2021.515237] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2019] [Accepted: 03/23/2021] [Indexed: 11/13/2022] Open
Abstract
Learning to move from auditory signals to phonemic categories is a crucial component of first, second, and multilingual language acquisition. In L1 and simultaneous multilingual acquisition, learners build up phonological knowledge to structure their perception within a language. For sequential multilinguals, this knowledge may support or interfere with acquiring language-specific representations for a new phonemic categorization system. Syllable structure is a part of this phonological knowledge, and language-specific syllabification preferences influence language acquisition, including early word segmentation. As a result, we expect to see language-specific syllable structure influencing speech perception as well. Initial evidence of an effect appears in Ali et al. (2011), who argued that cross-linguistic differences in McGurk fusion within a syllable reflected listeners’ language-specific syllabification preferences. Building on a framework from Cho and McQueen (2006), we argue that this could reflect the Phonological-Superiority Hypothesis (differences in L1 syllabification preferences make some syllabic positions harder to classify than others) or the Phonetic-Superiority Hypothesis (the acoustic qualities of speech sounds in some positions make it difficult to perceive unfamiliar sounds). However, their design does not distinguish between these two hypotheses. The current research study extends the work of Ali et al. (2011) by testing Japanese, and adding audio-only and congruent audio-visual stimuli to test the effects of syllabification preferences beyond just McGurk fusion. Eighteen native English speakers and 18 native Japanese speakers were asked to transcribe nonsense words in an artificial language. English allows stop consonants in syllable codas while Japanese heavily restricts them, but both groups showed similar patterns of McGurk fusion in stop codas. This is inconsistent with the Phonological-Superiority Hypothesis. However, when visual information was added, the phonetic influences on transcription accuracy largely disappeared. This is inconsistent with the Phonetic-Superiority Hypothesis. We argue from these results that neither acoustic informativity nor interference of a listener’s phonological knowledge is superior, and sketch a cognitively inspired rational cue integration framework as a third hypothesis to explain how L1 phonological knowledge affects L2 perception.
Collapse
Affiliation(s)
- Kanako Yasufuku
- Department of Linguistics and Asian/Middle Eastern Languages, San Diego State University, San Diego, CA, United States
| | - Gabriel Doyle
- Department of Linguistics and Asian/Middle Eastern Languages, San Diego State University, San Diego, CA, United States
| |
Collapse
|
19
|
Reh RK, Hensch TK, Werker JF. Distributional learning of speech sound categories is gated by sensitive periods. Cognition 2021; 213:104653. [PMID: 33752869 DOI: 10.1016/j.cognition.2021.104653] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2020] [Revised: 02/28/2021] [Accepted: 03/01/2021] [Indexed: 01/02/2023]
Abstract
Perceptual attunement to the native phonetic repertoire occurs over the first year of life: an infant's discrimination of non-native phonetic contrasts declines while their discrimination of native phonetic contrasts improves, with the timing of change consistent with sensitive periods. The statistics of speech sound distributions is one source of input used to collapse non-native phonetic category boundaries, while sharpening native ones. Distributional learning can be a domain-general mechanism, yet given the timing of perceptual attunement, we hypothesized that this learning mechanism may be maturationally delimited in the content domain of phonetic categories. Here, we assessed whether sensitivity to the distribution of speech sounds in the environment declines as the period of perceptual attunement closes. We used electroencephalography (EEG) to investigate whether neuronal responses to native 'ra' and 'la' phones are modulated differently in older vs young infants by exposure to either a bimodal or unimodal sound distribution spanning the [r] ~ [l] phoneme space. The native contrast, ra-la, is discriminable at all three ages, ensuring that we were testing the distributional learning mechanism, rather than confounding it with a decline in discrimination to a non-native distinction. English monolingual infants (n = 131) at 5-, 9- and 12-months-old were familiarized to either a unimodal or bimodal distribution of /ra/-/la/ speech sounds. Immediately following familiarization, an ERP oddball task was used to assess discrimination. Results showed that brief exposure to a bi- vs uni-modal distribution is sufficient to alter neuronal responses to subsequent /ra/ vs /la/ speech sounds at 5-months and 9-months, but not at 12-months. These results are the first to capture a progressive decline in sensitivity to distributional statistics in the environment. A potential mechanistic explanation based on critical period biology is discussed.
Collapse
Affiliation(s)
- Rebecca K Reh
- Department of Psychology, University of British Columbia, Canada
| | - Takao K Hensch
- Department of Molecular and Cellular Biology, Harvard University, USA; International Research Center for Neurointelligence, University of Tokyo, Japan
| | - Janet F Werker
- Department of Psychology, University of British Columbia, Canada.
| |
Collapse
|
20
|
Aberrant auditory system and its developmental implications for autism. SCIENCE CHINA-LIFE SCIENCES 2021; 64:861-878. [DOI: 10.1007/s11427-020-1863-6] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/20/2020] [Accepted: 12/06/2020] [Indexed: 12/26/2022]
|
21
|
García-Sierra A, Ramírez-Esparza N, Wig N, Robertson D. Language learning as a function of infant directed speech (IDS) in Spanish: Testing neural commitment using the positive-MMR. BRAIN AND LANGUAGE 2021; 212:104890. [PMID: 33307333 DOI: 10.1016/j.bandl.2020.104890] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/03/2020] [Revised: 10/27/2020] [Accepted: 11/17/2020] [Indexed: 06/12/2023]
Abstract
Spanish-English bilingual families (N = 17) were recruited to assess the association between infant directed speech (IDS) in Spanish and their degree of neural commitment to the Spanish language. IDS was assessed by extracting the caregivers' Vowel Space Area (VSA) from recordings of a storybook reading task done at home. Infants' neural commitment was assessed by extracting the positive mismatch brain response (positive-MMR), an Event-Related Potential (ERP) thought to be indicative of higher attentional processes and early language commitment. A linear mixed model analysis demonstrated that caregivers' VSA predicted the amplitude of the positive-MMR in response to a native speech contrast (Spanish), but not to a non-native speech contrast (Chinese), even after holding other predictors constant (i.e., socioeconomic status, infants' age, and fundamental frequency). Our findings provide support to the view that quality of language exposure fosters language learning, and that this beneficial relationship expands to the bilingual population.
Collapse
Affiliation(s)
- Adrián García-Sierra
- Speech, Language, and Hearing Sciences, University of Connecticut, 2 Alethia Dr. Unit 1085, Storrs, CT 06269, USA; Connecticut Institute for the Brain and Cognitive Science, University of Connecticut, 337 Mansfield Rd Unit 1272, Storrs, CT 06269, USA.
| | - Nairán Ramírez-Esparza
- Department of Psychological Sciences, University of Connecticut, 406 Babbidge Rd, Unit 1020, Storrs, CT 06269, USA.
| | - Noelle Wig
- Speech, Language, and Hearing Sciences, University of Connecticut, 2 Alethia Dr. Unit 1085, Storrs, CT 06269, USA; Connecticut Institute for the Brain and Cognitive Science, University of Connecticut, 337 Mansfield Rd Unit 1272, Storrs, CT 06269, USA.
| | - Dylan Robertson
- Speech, Language, and Hearing Sciences, University of Connecticut, 2 Alethia Dr. Unit 1085, Storrs, CT 06269, USA.
| |
Collapse
|
22
|
Filippi P. Emotional Voice Intonation: A Communication Code at the Origins of Speech Processing and Word-Meaning Associations? JOURNAL OF NONVERBAL BEHAVIOR 2020. [DOI: 10.1007/s10919-020-00337-z] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/29/2022]
Abstract
Abstract
The aim of the present work is to investigate the facilitating effect of vocal emotional intonation on the evolution of the following processes involved in language: (a) identifying and producing phonemes, (b) processing compositional rules underlying vocal utterances, and (c) associating vocal utterances with meanings. To this end, firstly, I examine research on the presence of these abilities in animals, and the biologically ancient nature of emotional vocalizations. Secondly, I review research attesting to the facilitating effect of emotional voice intonation on these abilities in humans. Thirdly, building on these studies in animals and humans, and through taking an evolutionary perspective, I provide insights for future empirical work on the facilitating effect of emotional intonation on these three processes in animals and preverbal humans. In this work, I highlight the importance of a comparative approach to investigate language evolution empirically. This review supports Darwin’s hypothesis, according to which the ability to express emotions through voice modulation was a key step in the evolution of spoken language.
Collapse
|
23
|
Outters V, Schreiner MS, Behne T, Mani N. Maternal input and infants' response to infant-directed speech. INFANCY 2020; 25:478-499. [PMID: 32744790 DOI: 10.1111/infa.12334] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/13/2019] [Revised: 03/18/2020] [Accepted: 03/25/2020] [Indexed: 11/29/2022]
Abstract
Caregivers typically use an exaggerated speech register known as infant-directed speech (IDS) in communication with infants. Infants prefer IDS over adult-directed speech (ADS) and IDS is functionally relevant in infant-directed communication. We examined interactions among maternal IDS quality, infants' preference for IDS over ADS, and the functional relevance of IDS at 6 and 13 months. While 6-month-olds showed a preference for IDS over ADS, 13-month-olds did not. Differences in gaze following behavior triggered by speech register (IDS vs. ADS) were found in both age groups. The degree of infants' preference for IDS (relative to ADS) was linked to the quality of maternal IDS infants were exposed to. No such relationship was found between gaze following behavior and maternal IDS quality and infants' IDS preference. The results speak to a dynamic interaction between infants' preference for different kinds of social signals and the social cues available to them.
Collapse
Affiliation(s)
- Vivien Outters
- Department for Psychology of Language, University of Goettingen, Göttingen, Germany.,Leibniz ScienceCampus "Primate Cognition", Göttingen, Germany
| | - Melanie S Schreiner
- Department for Psychology of Language, University of Goettingen, Göttingen, Germany.,Leibniz ScienceCampus "Primate Cognition", Göttingen, Germany.,Clinic for Cognitive Neurology, University of Leipzig, Leipzig, Germany.,Department of Neurology, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| | - Tanya Behne
- Leibniz ScienceCampus "Primate Cognition", Göttingen, Germany.,Department of Developmental Psychology, University of Goettingen, Göttingen, Germany
| | - Nivedita Mani
- Department for Psychology of Language, University of Goettingen, Göttingen, Germany.,Leibniz ScienceCampus "Primate Cognition", Göttingen, Germany
| |
Collapse
|
24
|
Scott SK. From speech and talkers to the social world: The neural processing of human spoken language. Science 2020; 366:58-62. [PMID: 31604302 DOI: 10.1126/science.aax0288] [Citation(s) in RCA: 25] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/20/2022]
Abstract
Human speech perception is a paradigm example of the complexity of human linguistic processing; however, it is also the dominant way of expressing vocal identity and is critically important for social interactions. Here, I review the ways that the speech, the talker, and the social nature of speech interact and how this may be computed in the human brain, using models and approaches from nonhuman primate studies. I explore the extent to which domain-general approaches may be able to account for some of these neural findings. Finally, I address the importance of extending these findings into a better understanding of the social use of speech in conversations.
Collapse
Affiliation(s)
- Sophie K Scott
- Institute of Cognitive Neuroscience, University College London, London, UK
| |
Collapse
|
25
|
Abstract
Infants learn about the sounds of their language and adults process the sounds they hear, even though sound categories often overlap in their acoustics. Researchers have suggested that listeners rely on context for these tasks, and have proposed two main ways that context could be helpful: top-down information accounts, which argue that listeners use context to predict which sound will be produced, and normalization accounts, which argue that listeners compensate for the fact that the same sound is produced differently in different contexts by factoring out this systematic context-dependent variability from the acoustics. These ideas have been somewhat conflated in past research, and have rarely been tested on naturalistic speech. We implement top-down and normalization accounts separately and evaluate their relative efficacy on spontaneous speech, using the test case of Japanese vowels. We find that top-down information strategies are effective even on spontaneous speech. Surprisingly, we find that at least one common implementation of normalization is ineffective on spontaneous speech, in contrast to what has been found on lab speech. We provide analyses showing that when there are systematic regularities in which contexts different sounds occur in-which are common in naturalistic speech, but generally controlled for in lab speech-normalization can actually increase category overlap rather than decrease it. This work calls into question the usefulness of normalization in naturalistic listening tasks, and highlights the importance of applying ideas from carefully controlled lab speech to naturalistic, spontaneous speech.
Collapse
|
26
|
DE Pablo I, Murillo E, Romero A. The effect of infant-directed speech on early multimodal communicative production in Spanish and Basque. JOURNAL OF CHILD LANGUAGE 2020; 47:457-471. [PMID: 31426871 DOI: 10.1017/s0305000919000412] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
We analyzed the effect of infant-directed speech (IDS) on multimodal communicative production of children at the beginning of the second year of life in two different languages: Spanish and Basque. Twelve Spanish and twelve Basque children aged between 12 and 15 months observed two versions of an audiovisual story: one version was narrated with IDS and the other with adult-directed speech (ADS). We analyzed the use of gaze and the communicative behaviors produced by children. The time spent looking at the story increases in the IDS condition regardless of the language of the narration. Children produced more multimodal communicative behaviors while watching the IDS version both in Spanish and in Basque. These results suggest that IDS increases attention and social engagement promoting joint attention episodes.
Collapse
Affiliation(s)
- Irati DE Pablo
- Departamento de Didáctica de la Lengua y la Literatura, Universidad del País Vasco, Spain
| | - Eva Murillo
- Departamento de Psicología Básica, Universidad Autónoma de Madrid, Spain
| | - Asier Romero
- Departamento de Didáctica de la Lengua y la Literatura, Universidad del País Vasco, Spain
| |
Collapse
|
27
|
Kalashnikova M, Goswami U, Burnham D. Infant‐directed speech to infants at risk for dyslexia: A novel cross‐dyad design. INFANCY 2020. [DOI: 10.1111/infa.12329] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Affiliation(s)
- Marina Kalashnikova
- BCBL ‐ Basque Center on Cognition, Brain and Language San Sebastian Spain
- The MARCS Institute for Brain, Behaviour and Development Western Sydney University Penrith NSW Australia
| | - Usha Goswami
- Centre for Neuroscience in Education University of Cambridge Cambridge UK
| | - Denis Burnham
- The MARCS Institute for Brain, Behaviour and Development Western Sydney University Penrith NSW Australia
| |
Collapse
|
28
|
Kostilainen K, Partanen E, Mikkola K, Wikström V, Pakarinen S, Fellman V, Huotilainen M. Neural processing of changes in phonetic and emotional speech sounds and tones in preterm infants at term age. Int J Psychophysiol 2020; 148:111-118. [DOI: 10.1016/j.ijpsycho.2019.10.009] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2018] [Revised: 09/10/2019] [Accepted: 10/23/2019] [Indexed: 10/25/2022]
|
29
|
Dilley L, Gamache J, Wang Y, Houston DM, Bergeson TR. Statistical distributions of consonant variants in infant-directed speech: evidence that /t/ may be exceptional. JOURNAL OF PHONETICS 2019; 75:73-87. [PMID: 32884162 PMCID: PMC7467459 DOI: 10.1016/j.wocn.2019.05.004] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Statistical distributions of phonetic variants in spoken language influence speech perception for both language learners and mature users. We theorized that patterns of phonetic variant processing of consonants demonstrated by adults might stem in part from patterns of early exposure to statistics of phonetic variants in infant-directed (ID) speech. In particular, we hypothesized that ID speech might involve greater proportions of canonical /t/ pronunciations compared to adult-directed (AD) speech in at least some phonological contexts. This possibility was tested using a corpus of spontaneous speech of mothers speaking to other adults, or to their typically-developing infant. Tokens of word-final alveolar stops - including /t/, /d/, and the nasal stop /n/ - were examined in assimilable contexts (i.e., those followed by a word-initial labial and/or velar); these were classified as canonical, assimilated, deleted, or glottalized. Results confirmed that there were significantly more canonical pronunciations in assimilable contexts in ID compared with AD speech, an effect which was driven by the phoneme /t/. These findings suggest that at least in phonological contexts involving possible assimilation, children are exposed to more canonical /t/ variant pronunciations than adults are. This raises the possibility that perceptual processing of canonical /t/ may be partly attributable to exposure to canonical /t/ variants in ID speech. Results support the need for further research into how statistics of variant pronunciations in early language input may shape speech processing across the lifespan.
Collapse
Affiliation(s)
- Laura Dilley
- Department of Communicative Sciences and Disorders, Michigan State University
| | - Jessica Gamache
- Department of Linguistics and Germanic, Slavic, Asian and African Languages, Michigan State University
| | - Yuanyuan Wang
- Department of Otolaryngology, The Ohio State University
| | | | - Tonya R. Bergeson
- Dept. of Otolaryngology – Head & Neck Surgery, Indiana University School of Medicine
- Department of Communication Sciences and Disorders, Butler University
| |
Collapse
|
30
|
Hübscher I, Prieto P. Gestural and Prosodic Development Act as Sister Systems and Jointly Pave the Way for Children's Sociopragmatic Development. Front Psychol 2019; 10:1259. [PMID: 31244716 PMCID: PMC6581748 DOI: 10.3389/fpsyg.2019.01259] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2018] [Accepted: 05/13/2019] [Indexed: 12/04/2022] Open
Abstract
Children might combine gesture and prosody to express a pragmatic meaning such as a request, information focus, uncertainty or politeness, before they can convey these meanings in speech. However, little is known about the developmental trajectories of gestural and prosodic patterns and how they relate to a child's growing understanding and propositional use of these sociopragmatic meanings. Do gesture and prosody act as sister systems in pragmatic development? Do children acquire these components of language before they are able to express themselves through spoken language, thus acting as forerunners in children's pragmatic development? This review article assesses empirical evidence that demonstrates that gesture and prosody act as intimately related systems and, importantly, pave the way for pragmatic acquisition at different developmental stages. The review goes on to explore how the integration of gesture and prosody with semantics and syntax can impact language acquisition and how multimodal interventions can be used effectively in educational settings. Our review findings support the importance of simultaneously assessing both the prosodic and the gestural components of language in the fields of language development, language learning, and language intervention.
Collapse
Affiliation(s)
- Iris Hübscher
- URPP Language and Space, University of Zurich, Zurich, Switzerland
| | - Pilar Prieto
- Institució Catalana de Recerca i Estudis Avançats, Barcelona, Spain
- Facultat de Traducció i Ciències del Llenguatge, Universitat Pompeu Fabra, Barcelona, Spain
| |
Collapse
|
31
|
Cheng B, Zhang X, Fan S, Zhang Y. The Role of Temporal Acoustic Exaggeration in High Variability Phonetic Training: A Behavioral and ERP Study. Front Psychol 2019; 10:1178. [PMID: 31178795 PMCID: PMC6543854 DOI: 10.3389/fpsyg.2019.01178] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2019] [Accepted: 05/06/2019] [Indexed: 12/03/2022] Open
Abstract
High variability phonetic training (HVPT) has been found to be effective in helping adult learners acquire non-native phonetic contrasts. The present study investigated the role of temporal acoustic exaggeration by comparing the canonical HVPT paradigm without involving acoustic exaggeration with a modified adaptive HVPT paradigm that integrated key temporal exaggerations in infant-directed speech (IDS). Sixty native Chinese adults participated in the training of the English /i/ and /i/ vowel contrast and were randomly assigned to three subject groups. Twenty were trained with the typical HVPT paradigm (the HVPT group), twenty were trained under the modified adaptive approach with acoustic exaggeration (the HVPT-E group), and twenty were in the control group. Behavioral tasks for the pre- and post- tests used natural word identification, synthetic stimuli identification, and synthetic stimuli discrimination. Mismatch negativity (MMN) responses from the HVPT-E group were also obtained to assess the training effects in within- and across- category discrimination without requiring focused attention. Like previous studies, significant generalization effects to new talkers were found in both the HVPT group and the HVPT-E group. The HVPT-E group, by contrast, showed greater improvement as reflected in larger progress in natural word identification performance. Furthermore, the HVPT-E group exhibited more native-like categorical perception based on spectral cues after training, together with corresponding training-induced changes in the MMN responses to within- and across- category differences. These data provide the initial evidence supporting the important role of temporal acoustic exaggeration with adaptive training in facilitating phonetic learning and promoting brain plasticity at the perceptual and pre-attentive neural levels.
Collapse
Affiliation(s)
- Bing Cheng
- English Department & Language and Cognitive Neuroscience Lab, School of Foreign Studies, Xi’an Jiaotong University, Xi’an, China
| | - Xiaojuan Zhang
- English Department & Language and Cognitive Neuroscience Lab, School of Foreign Studies, Xi’an Jiaotong University, Xi’an, China
| | - Siying Fan
- English Department & Language and Cognitive Neuroscience Lab, School of Foreign Studies, Xi’an Jiaotong University, Xi’an, China
| | - Yang Zhang
- Department of Speech-Language-Hearing Sciences, Center for Neurobehavioral Development, University of Minnesota, Minneapolis, MN, United States
| |
Collapse
|
32
|
Chong AJ, Vicenik C, Sundara M. Intonation Plays a Role in Language Discrimination by Infants. INFANCY 2018. [DOI: 10.1111/infa.12257] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
Affiliation(s)
- Adam J. Chong
- Department of Linguistics Queen Mary University of London
- Department of Linguistics University of California Los Angeles
| | - Chad Vicenik
- Department of Linguistics University of California Los Angeles
| | - Megha Sundara
- Department of Linguistics University of California Los Angeles
| |
Collapse
|
33
|
Drouin JR, Theodore RM. Lexically guided perceptual learning is robust to task-based changes in listening strategy. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 144:1089. [PMID: 30180678 PMCID: PMC6117182 DOI: 10.1121/1.5047672] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/16/2018] [Revised: 06/29/2018] [Accepted: 07/02/2018] [Indexed: 05/28/2023]
Abstract
Listeners use lexical information to resolve ambiguity in the speech signal, resulting in the restructuring of speech sound categories. Recent findings suggest that lexically guided perceptual learning is attenuated when listeners use a perception-focused listening strategy (that directs attention towards surface variation) compared to when listeners use a comprehension-focused listening strategy (that directs attention towards higher-level linguistic information). However, previous investigations used the word position of the ambiguity to manipulate listening strategy, raising the possibility that attenuated learning reflected decreased strength of lexical recruitment instead of a perception-oriented listening strategy. The current work tests this hypothesis. Listeners completed an exposure phase followed by a test phase. During exposure, listeners heard an ambiguous fricative embedded in word-medial lexical contexts that supported realization of the ambiguity as /∫/. At test, listeners categorized members of an /ɑsi/-/ɑ∫i/ continuum. Listening strategy was manipulated via exposure task (experiment 1) and explicit acknowledgement of the ambiguity (experiment 2). Compared to control participants, listeners who were exposed to the ambiguity showed more /∫/ responses at the test; critically, the magnitude of learning did not differ across listening strategy conditions. These results suggest that given sufficient lexical context, lexically guided perceptual learning is robust to task-based changes in listening strategy.
Collapse
Affiliation(s)
- Julia R Drouin
- Department of Speech, Language, and Hearing Sciences, University of Connecticut, 850 Bolton Road, Unit 1085, Storrs, Connecticut 06269-1085, USA
| | - Rachel M Theodore
- Department of Speech, Language, and Hearing Sciences, University of Connecticut, 850 Bolton Road, Unit 1085, Storrs, Connecticut 06269-1085, USA
| |
Collapse
|
34
|
Adriaans F. Effects of consonantal context on the learnability of vowel categories from infant-directed speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 144:EL20. [PMID: 30075685 DOI: 10.1121/1.5045192] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Recent studies have shown that vowels in infant-directed speech (IDS) are characterized by highly variable formant distributions. The current study investigates whether vowel variability is partially due to consonantal context, and explores whether consonantal context could support the learning of vowel categories from IDS. A computational model is presented which selects contexts based on frequency in the input and generalizes across contextual categories. Improved categorization performance was found on a vowel contrast in American-English IDS. The findings support a view in which the infant's learning mechanism is anchored in context, in order to cope with acoustic variability in the input.
Collapse
Affiliation(s)
- Frans Adriaans
- Utrecht Institute of Linguistics OTS, Utrecht University, Trans 10, 3512 JK Utrecht, the Netherlands
| |
Collapse
|
35
|
Esteve-Gibert N, Guellaï B. Prosody in the Auditory and Visual Domains: A Developmental Perspective. Front Psychol 2018; 9:338. [PMID: 29615944 PMCID: PMC5868325 DOI: 10.3389/fpsyg.2018.00338] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2017] [Accepted: 02/27/2018] [Indexed: 11/13/2022] Open
Abstract
The development of body movements such as hand or head gestures, or facial expressions, seems to go hand-in-hand with the development of speech abilities. We know that very young infants rely on the movements of their caregivers' mouth to segment the speech stream, that infants' canonical babbling is temporally related to rhythmic hand movements, that narrative abilities emerge at a similar time in speech and gestures, and that children make use of both modalities to access complex pragmatic intentions. Prosody has emerged as a key linguistic component in this speech-gesture relationship, yet its exact role in the development of multimodal communication is still not well understood. For example, it is not clear what the relative weights of speech prosody and body gestures are in language acquisition, or whether both modalities develop at the same time or whether one modality needs to be in place for the other to emerge. The present paper reviews existing literature on the interactions between speech prosody and body movements from a developmental perspective in order to shed some light on these issues.
Collapse
Affiliation(s)
- Núria Esteve-Gibert
- Departament de Llengües i Literatures Modernes i d’Estudis Anglesos, Universitat de Barcelona (UB), Barcelona, Spain
| | - Bahia Guellaï
- Laboratoire Ethologie, Cognition, Développement, Université Paris Nanterre, Nanterre, France
| |
Collapse
|
36
|
Ben-Aderet T, Gallego-Abenza M, Reby D, Mathevon N. Dog-directed speech: why do we use it and do dogs pay attention to it? Proc Biol Sci 2018; 284:rspb.2016.2429. [PMID: 28077769 DOI: 10.1098/rspb.2016.2429] [Citation(s) in RCA: 29] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/04/2016] [Accepted: 12/02/2016] [Indexed: 11/12/2022] Open
Abstract
Pet-directed speech is strikingly similar to infant-directed speech, a peculiar speaking pattern with higher pitch and slower tempo known to engage infants' attention and promote language learning. Here, we report the first investigation of potential factors modulating the use of dog-directed speech, as well as its immediate impact on dogs' behaviour. We recorded adult participants speaking in front of pictures of puppies, adult and old dogs, and analysed the quality of their speech. We then performed playback experiments to assess dogs' reaction to dog-directed speech compared with normal speech. We found that human speakers used dog-directed speech with dogs of all ages and that the acoustic structure of dog-directed speech was mostly independent of dog age, except for sound pitch which was relatively higher when communicating with puppies. Playback demonstrated that, in the absence of other non-auditory cues, puppies were highly reactive to dog-directed speech, and that the pitch was a key factor modulating their behaviour, suggesting that this specific speech register has a functional value in young dogs. Conversely, older dogs did not react differentially to dog-directed speech compared with normal speech. The fact that speakers continue to use dog-directed with older dogs therefore suggests that this speech pattern may mainly be a spontaneous attempt to facilitate interactions with non-verbal listeners.
Collapse
Affiliation(s)
- Tobey Ben-Aderet
- Department of Psychology, City University of New York, Hunter College, New York, NY, USA
| | - Mario Gallego-Abenza
- Equipe Neuro-Ethologie Sensorielle, ENES/Neuro-PSI CNRS UMR9197, University of Lyon/Saint-Etienne, Saint-Etienne, France
| | - David Reby
- School of Psychology, University of Sussex, Brighton BN1 9QH, UK
| | - Nicolas Mathevon
- Department of Psychology, City University of New York, Hunter College, New York, NY, USA .,Equipe Neuro-Ethologie Sensorielle, ENES/Neuro-PSI CNRS UMR9197, University of Lyon/Saint-Etienne, Saint-Etienne, France
| |
Collapse
|
37
|
Laing CE, Vihman M, Keren-Portnoy T. How salient are onomatopoeia in the early input? A prosodic analysis of infant-directed speech. JOURNAL OF CHILD LANGUAGE 2017; 44:1117-1139. [PMID: 27670787 DOI: 10.1017/s0305000916000428] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Onomatopoeia are frequently identified amongst infants' earliest words (Menn & Vihman, 2011), yet few authors have considered why this might be, and even fewer have explored this phenomenon empirically. Here we analyze mothers' production of onomatopoeia in infant-directed speech (IDS) to provide an input-based perspective on these forms. Twelve mothers were recorded interacting with their 8-month-olds; onomatopoeic words (e.g. quack) were compared acoustically with their corresponding conventional words (duck). Onomatopoeia were more salient than conventional words across all features measured: mean pitch, pitch range, word duration, repetition, and pause length. Furthermore, a systematic pattern was observed in the production of onomatopoeia, suggesting a conventionalized approach to mothers' production of these words in IDS.
Collapse
|
38
|
Pulvermüller F. Neural reuse of action perception circuits for language, concepts and communication. Prog Neurobiol 2017; 160:1-44. [PMID: 28734837 DOI: 10.1016/j.pneurobio.2017.07.001] [Citation(s) in RCA: 96] [Impact Index Per Article: 13.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2017] [Revised: 05/12/2017] [Accepted: 07/13/2017] [Indexed: 10/19/2022]
Abstract
Neurocognitive and neurolinguistics theories make explicit statements relating specialized cognitive and linguistic processes to specific brain loci. These linking hypotheses are in need of neurobiological justification and explanation. Recent mathematical models of human language mechanisms constrained by fundamental neuroscience principles and established knowledge about comparative neuroanatomy offer explanations for where, when and how language is processed in the human brain. In these models, network structure and connectivity along with action- and perception-induced correlation of neuronal activity co-determine neurocognitive mechanisms. Language learning leads to the formation of action perception circuits (APCs) with specific distributions across cortical areas. Cognitive and linguistic processes such as speech production, comprehension, verbal working memory and prediction are modelled by activity dynamics in these APCs, and combinatorial and communicative-interactive knowledge is organized in the dynamics within, and connections between APCs. The network models and, in particular, the concept of distributionally-specific circuits, can account for some previously not well understood facts about the cortical 'hubs' for semantic processing and the motor system's role in language understanding and speech sound recognition. A review of experimental data evaluates predictions of the APC model and alternative theories, also providing detailed discussion of some seemingly contradictory findings. Throughout, recent disputes about the role of mirror neurons and grounded cognition in language and communication are assessed critically.
Collapse
Affiliation(s)
- Friedemann Pulvermüller
- Brain Language Laboratory, Department of Philosophy & Humanities, WE4, Freie Universität Berlin, 14195 Berlin, Germany; Berlin School of Mind and Brain, Humboldt Universität zu Berlin, 10099 Berlin, Germany; Einstein Center for Neurosciences, Berlin 10117 Berlin, Germany.
| |
Collapse
|
39
|
Adriaans F, Swingley D. Prosodic exaggeration within infant-directed speech: Consequences for vowel learnability. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 141:3070. [PMID: 28599541 PMCID: PMC5418129 DOI: 10.1121/1.4982246] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/19/2016] [Revised: 04/12/2017] [Accepted: 04/12/2017] [Indexed: 05/30/2023]
Abstract
Perceptual experiments with infants show that they adapt their perception of speech sounds toward the categories of the native language. How do infants learn these categories? For the most part, acoustic analyses of natural infant-directed speech have suggested that phonetic categories are not presented to learners as separable clusters of sounds in acoustic space. As a step toward explaining how infants begin to solve this problem, the current study proposes that the exaggerated prosody characteristic of infant-directed speech may highlight for infants certain speech-sound tokens that collectively form more readily identifiable categories. A database is presented, containing vowel measurements in a large sample of natural American English infant-directed speech. Analyses of the vowel space show that prosodic exaggeration in infant-directed speech has the potential to support distributional vowel learning by providing the learner with a subset of "high-quality" tokens that infants might attend to preferentially. Categorization models trained on prosodically exaggerated tokens outperformed models that were trained on tokens that were not exaggerated. Though focusing on more prominent, exaggerated tokens does not provide a solution to the categorization problem, it would make it easier to solve.
Collapse
Affiliation(s)
- Frans Adriaans
- Utrecht Institute of Linguistics OTS, Utrecht University, Trans 10, 3512 JK Utrecht, the Netherlands
| | - Daniel Swingley
- Department of Psychology, University of Pennsylvania, 425 South University Avenue, Philadelphia, Pennsylvania 19104, USA
| |
Collapse
|
40
|
Richoz AR, Quinn PC, Hillairet de Boisferon A, Berger C, Loevenbruck H, Lewkowicz DJ, Lee K, Dole M, Caldara R, Pascalis O. Audio-Visual Perception of Gender by Infants Emerges Earlier for Adult-Directed Speech. PLoS One 2017; 12:e0169325. [PMID: 28060872 PMCID: PMC5218491 DOI: 10.1371/journal.pone.0169325] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2016] [Accepted: 12/15/2016] [Indexed: 11/18/2022] Open
Abstract
Early multisensory perceptual experiences shape the abilities of infants to perform socially-relevant visual categorization, such as the extraction of gender, age, and emotion from faces. Here, we investigated whether multisensory perception of gender is influenced by infant-directed (IDS) or adult-directed (ADS) speech. Six-, 9-, and 12-month-old infants saw side-by-side silent video-clips of talking faces (a male and a female) and heard either a soundtrack of a female or a male voice telling a story in IDS or ADS. Infants participated in only one condition, either IDS or ADS. Consistent with earlier work, infants displayed advantages in matching female relative to male faces and voices. Moreover, the new finding that emerged in the current study was that extraction of gender from face and voice was stronger at 6 months with ADS than with IDS, whereas at 9 and 12 months, matching did not differ for IDS versus ADS. The results indicate that the ability to perceive gender in audiovisual speech is influenced by speech manner. Our data suggest that infants may extract multisensory gender information developmentally earlier when looking at adults engaged in conversation with other adults (i.e., ADS) than when adults are directly talking to them (i.e., IDS). Overall, our findings imply that the circumstances of social interaction may shape early multisensory abilities to perceive gender.
Collapse
Affiliation(s)
- Anne-Raphaëlle Richoz
- Eye and Brain Mapping Laboratory (iBMLab), Department of Psychology, University of Fribourg, Fribourg, Switzerland
- LPNC, University of Grenoble Alpes, Grenoble, France
- LPNC, CNRS-UMR 5105, University of Grenoble Alpes, Grenoble, France
| | - Paul C. Quinn
- Department of Psychological and Brain Sciences, University of Delaware, Newark, Delaware, United States of America
| | - Anne Hillairet de Boisferon
- LPNC, University of Grenoble Alpes, Grenoble, France
- LPNC, CNRS-UMR 5105, University of Grenoble Alpes, Grenoble, France
| | - Carole Berger
- LPNC, University of Grenoble Alpes, Grenoble, France
- LPNC, CNRS-UMR 5105, University of Grenoble Alpes, Grenoble, France
| | - Hélène Loevenbruck
- LPNC, University of Grenoble Alpes, Grenoble, France
- LPNC, CNRS-UMR 5105, University of Grenoble Alpes, Grenoble, France
| | - David J. Lewkowicz
- Department of Communication Sciences & Disorders, Northeastern University, Boston, Massachusetts, United States of America
| | - Kang Lee
- Institute of Child Study University of Toronto, Toronto, Ontario, Canada
| | - Marjorie Dole
- LPNC, University of Grenoble Alpes, Grenoble, France
- LPNC, CNRS-UMR 5105, University of Grenoble Alpes, Grenoble, France
| | - Roberto Caldara
- Eye and Brain Mapping Laboratory (iBMLab), Department of Psychology, University of Fribourg, Fribourg, Switzerland
| | - Olivier Pascalis
- LPNC, University of Grenoble Alpes, Grenoble, France
- LPNC, CNRS-UMR 5105, University of Grenoble Alpes, Grenoble, France
| |
Collapse
|
41
|
Martin A, Igarashi Y, Jincho N, Mazuka R. Utterances in infant-directed speech are shorter, not slower. Cognition 2016; 156:52-59. [DOI: 10.1016/j.cognition.2016.07.015] [Citation(s) in RCA: 57] [Impact Index Per Article: 7.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2014] [Revised: 07/16/2016] [Accepted: 07/29/2016] [Indexed: 11/26/2022]
|
42
|
Ludusan B, Cristia A, Martin A, Mazuka R, Dupoux E. Learnability of prosodic boundaries: Is infant-directed speech easier? THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 140:1239. [PMID: 27586752 DOI: 10.1121/1.4960576] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
This study explores the long-standing hypothesis that the acoustic cues to prosodic boundaries in infant-directed speech (IDS) make those boundaries easier to learn than those in adult-directed speech (ADS). Three cues (pause duration, nucleus duration, and pitch change) were investigated, by means of a systematic review of the literature, statistical analyses of a corpus of Japanese, and machine learning experiments. The review of previous work revealed that the effect of register on boundary cues is less well established than previously thought, and that results often vary across studies for certain cues. Statistical analyses run on a large database of mother-child and mother-interviewer interactions showed that the duration of a pause and the duration of the syllable nucleus preceding the boundary are two cues which are enhanced in IDS, while f0 change is actually degraded in IDS. Supervised and unsupervised machine learning techniques applied to these acoustic cues revealed that IDS boundaries were consistently better classified than ADS ones, regardless of the learning method used. The role of the cues examined in this study and the importance of these findings in the more general context of early linguistic structure acquisition is discussed.
Collapse
Affiliation(s)
- Bogdan Ludusan
- Laboratoire de Sciences Cognitives et Psycholinguistique, EHESS/École Normale Supérieure-PSL Research University/CNRS, Paris, France
| | - Alejandrina Cristia
- Laboratoire de Sciences Cognitives et Psycholinguistique, EHESS/École Normale Supérieure-PSL Research University/CNRS, Paris, France
| | - Andrew Martin
- Laboratory for Language Development, RIKEN Brain Science Institute, Wako-shi, Japan
| | - Reiko Mazuka
- Laboratory for Language Development, RIKEN Brain Science Institute, Wako-shi, Japan
| | - Emmanuel Dupoux
- Laboratoire de Sciences Cognitives et Psycholinguistique, EHESS/École Normale Supérieure-PSL Research University/CNRS, Paris, France
| |
Collapse
|
43
|
Leung KKW, Jongman A, Wang Y, Sereno JA. Acoustic characteristics of clearly spoken English tense and lax vowels. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 140:45. [PMID: 27475131 DOI: 10.1121/1.4954737] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Clearly produced vowels exhibit longer duration and more extreme spectral properties than plain, conversational vowels. These features also characterize tense relative to lax vowels. This study explored the interaction of clear-speech and tensity effects by comparing clear and plain productions of three English tense-lax vowel pairs (/i-ɪ/, /ɑ-ʌ/, /u-ʊ/ in /kVd/ words). Both temporal and spectral acoustic features were examined, including vowel duration, vowel-to-word duration ratio, formant frequency, and dynamic spectral characteristics. Results revealed that the tense-lax vowel difference was generally enhanced in clear relative to plain speech, but clear-speech modifications for tense and lax vowels showed a trade-off in the use of temporal and spectral cues. While plain-to-clear vowel lengthening was greater for tense than lax vowels, clear-speech modifications in spectral change were larger for lax than tense vowels. Moreover, peripheral tense vowels showed more consistent clear-speech modifications in the temporal than spectral domain. Presumably, articulatory constraints limit the spectral variation of these extreme vowels, so clear-speech modifications resort to temporal features and reserve the primary spectral features for tensity contrasts. These findings suggest that clear-speech and tensity interactions involve compensatory modifications in different acoustic domains.
Collapse
Affiliation(s)
- Keith K W Leung
- Department of Linguistics, Simon Fraser University, 8888 University Drive, Burnaby, British Columbia V5A 1S6, Canada
| | - Allard Jongman
- Department of Linguistics, The University of Kansas, Lawrence, Kansas 66044, USA
| | - Yue Wang
- Department of Linguistics, Simon Fraser University, 8888 University Drive, Burnaby, British Columbia V5A 1S6, Canada
| | - Joan A Sereno
- Department of Linguistics, The University of Kansas, Lawrence, Kansas 66044, USA
| |
Collapse
|
44
|
Asada M. Modeling Early Vocal Development Through Infant–Caregiver Interaction: A Review. IEEE Trans Cogn Dev Syst 2016. [DOI: 10.1109/tcds.2016.2552493] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
45
|
Thiessen ED, Girard S, Erickson LC. Statistical learning and the critical period: how a continuous learning mechanism can give rise to discontinuous learning. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2016; 7:276-88. [DOI: 10.1002/wcs.1394] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/01/2015] [Revised: 03/31/2016] [Accepted: 04/06/2016] [Indexed: 11/08/2022]
Affiliation(s)
- Erik D. Thiessen
- Department of Psychology; Carnegie Mellon University; Pittsburgh PA USA
| | - Sandrine Girard
- Department of Psychology; Carnegie Mellon University; Pittsburgh PA USA
| | - Lucy C. Erickson
- Department of Psychology; Carnegie Mellon University; Pittsburgh PA USA
| |
Collapse
|
46
|
Sandbank M, Yoder P. The Association Between Parental Mean Length of Utterance and Language Outcomes in Children With Disabilities: A Correlational Meta-Analysis. AMERICAN JOURNAL OF SPEECH-LANGUAGE PATHOLOGY 2016; 25:240-251. [PMID: 27088766 DOI: 10.1044/2015_ajslp-15-0003] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/21/2015] [Accepted: 09/01/2015] [Indexed: 06/05/2023]
Abstract
PURPOSE The purpose of this correlational meta-analysis was to examine the association between parental utterance length and language outcomes in children with disabilities and whether this association varies according to other child characteristics, such as age and disability type. This association can serve as a starting point for language intervention practices for children with disabilities. METHOD We conducted a systematic search of 42 electronic databases to identify relevant studies. Twelve studies reporting on a total of 13 populations (including 257 participants) were identified. A random-effects model was used to estimate a combined effect size across all studies as well as separate effect sizes across studies in each disability category. RESULTS The combined effect size across all studies suggests a weak positive association between parental input length and child language outcomes. However, subgroup analyses within disability categories suggest that this association may differ for children with autism. Results of 4 studies including 47 children with autism show that parental input length is strongly associated with positive language outcomes in this population. CONCLUSIONS Present evidence suggests that clinicians should reconsider intervention practices that prescribe shorter, grammatically incomplete utterances, particularly when working with children with autism.
Collapse
|
47
|
Narayan CR, McDermott LC. Speech rate and pitch characteristics of infant-directed speech: Longitudinal and cross-linguistic observations. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 139:1272-1281. [PMID: 27036263 DOI: 10.1121/1.4944634] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
The speech rate and pitch (F0) characteristics of naturalistic, longitudinally recorded infant- and adult-directed speech are reported for three, genetically diverse languages. Previous research has suggested that the prosodic characteristics of infant-directed speech are slowed speech rate, raised mean pitch, and expanded pitch range relative to adult-directed speech. Sixteen mothers (5 Sri Lankan Tamil, 5 Tagalog, 6 Korean) were recorded in their homes during natural interactions with their young infants, and adults, over the course of 12 months beginning when the infant was 4 months old. Regression models indicated that the difference between infant- and adult-directed speech rates decreased across the first year of infants' development. Models of pitch revealed predicted differences between infant- and adult-directed speech but did not provide evidence for cross-linguistic or longitudinal effects within the time period investigated for the three languages. The universality of slowed speech rate, raised pitch, and expanded pitch range is discussed in light of individuals' highly variable implementation of these prosodic features in infant-directed speech.
Collapse
Affiliation(s)
- Chandan R Narayan
- Speech and Psycholinguistics Laboratory, York University, Toronto, Ontario M3J 1P3, Canada
| | - Lily C McDermott
- Speech and Psycholinguistics Laboratory, York University, Toronto, Ontario M3J 1P3, Canada
| |
Collapse
|
48
|
Statistical Learning, Syllable Processing, and Speech Production in Healthy Hearing and Hearing-Impaired Preschool Children. Ear Hear 2016; 37:e57-71. [DOI: 10.1097/aud.0000000000000197] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
49
|
Yeung HH, Chen LM, Werker JF. Referential labeling can facilitate phonetic learning in infancy. Child Dev 2015; 85:1036-49. [PMID: 24936610 DOI: 10.1111/cdev.12185] [Citation(s) in RCA: 32] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Abstract
All languages employ certain phonetic contrasts when distinguishing words. Infant speech perception is rapidly attuned to these contrasts before many words are learned, thus phonetic attunement is thought to proceed independently of lexical and referential knowledge. Here, evidence to the contrary is provided. Ninety-eight 9-month-old English-learning infants were trained to perceive a non-native Cantonese tone contrast.Two object–tone audiovisual pairings were consistently presented, which highlighted the target contrast (Object A with Tone X; Object B with Tone Y). Tone discrimination was then assessed. Results showed improved tone discrimination if object–tone pairings were perceived as being referential word labels, although this effect was modulated by vocabulary size. Results suggest how lexical and referential knowledge could play a role in phonetic attunement.
Collapse
|
50
|
Erickson LC, Thiessen ED. Statistical learning of language: Theory, validity, and predictions of a statistical learning account of language acquisition. DEVELOPMENTAL REVIEW 2015. [DOI: 10.1016/j.dr.2015.05.002] [Citation(s) in RCA: 57] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|