1
|
Di Dona G, Scaltritti M, Sulpizio S. Formant-invariant voice and pitch representations are pre-attentively formed from constantly varying speech and non-speech stimuli. Eur J Neurosci 2022; 56:4086-4106. [PMID: 35673798 PMCID: PMC9545905 DOI: 10.1111/ejn.15730] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2022] [Revised: 05/23/2022] [Accepted: 05/24/2022] [Indexed: 11/30/2022]
Abstract
The present study investigated whether listeners can form abstract voice representations while ignoring constantly changing phonological information and if they can use the resulting information to facilitate voice change detection. Further, the study aimed at understanding whether the use of abstraction is restricted to the speech domain or can be deployed also in non‐speech contexts. We ran an electroencephalogram (EEG) experiment including one passive and one active oddball task, each featuring a speech and a rotated speech condition. In the speech condition, participants heard constantly changing vowels uttered by a male speaker (standard stimuli) which were infrequently replaced by vowels uttered by a female speaker with higher pitch (deviant stimuli). In the rotated speech condition, participants heard rotated vowels, in which the natural formant structure of speech was disrupted. In the passive task, the mismatch negativity was elicited after the presentation of the deviant voice in both conditions, indicating that listeners could successfully group together different stimuli into a formant‐invariant voice representation. In the active task, participants showed shorter reaction times (RTs), higher accuracy and a larger P3b in the speech condition with respect to the rotated speech condition. Results showed that whereas at a pre‐attentive level the cognitive system can track pitch regularities while presumably ignoring constantly changing formant information both in speech and in rotated speech, at an attentive level the use of such information is facilitated for speech. This facilitation was also testified by a stronger synchronisation in the theta band (4–7 Hz), potentially pointing towards differences in encoding/retrieval processes.
Collapse
Affiliation(s)
- Giuseppe Di Dona
- Dipartimento di Psicologia e Scienze Cognitive, Università degli Studi di Trento, Trento, Italy
| | - Michele Scaltritti
- Dipartimento di Psicologia e Scienze Cognitive, Università degli Studi di Trento, Trento, Italy
| | - Simone Sulpizio
- Dipartimento di Psicologia, Università degli Studi di Milano-Bicocca, Milano, Italy.,Milan Center for Neuroscience (NeuroMi), Università degli Studi di Milano-Bicocca, Milano, Italy
| |
Collapse
|
2
|
Ortiz-Mantilla S, Roesler CP, Realpe-Bonilla T, Benasich AA. Modulation of Theta Phase Synchrony during Syllable Processing as a Function of Interactive Acoustic Experience in Infancy. Cereb Cortex 2021; 32:919-932. [PMID: 34403462 PMCID: PMC8889996 DOI: 10.1093/cercor/bhab256] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2021] [Revised: 07/02/2021] [Accepted: 07/03/2021] [Indexed: 11/13/2022] Open
Abstract
Plasticity, a prominent characteristic of the infant brain, supports formation of cortical representations as infants begin to interact with and adapt to environmental sensory events. Enhanced acoustic processing efficiency along with improved allocation of attentional resources at 7 months and establishment of well-defined phonemic maps at 9 months have been shown to be facilitated by early interactive acoustic experience (IAE). In this study, using an oddball paradigm and measures of theta phase synchrony at source level, we examined short- and long-term effects of nonspeech IAE on syllable processing. Results demonstrated that beyond maturation alone, IAE increased the efficiency of syllabic representation and discrimination, an effect that endured well beyond the immediate training period. As compared with naive controls, the IAE-trained group at 7, 9, and 18 months showed less theta phase synchrony for the standard syllable and at 7 and 18 months for the deviant syllable. The decreased theta phase synchrony exhibited by the trained group suggests more mature, efficient, acoustic processing, and thus, better cortical representation and discrimination of syllabic content. Further, the IAE modulatory effect observed on theta phase synchrony in left auditory cortex at 7 and 9 months was differentially associated with receptive and expressive language scores at 12 and 18 months of age.
Collapse
Affiliation(s)
- Silvia Ortiz-Mantilla
- Center for Molecular and Behavioral Neuroscience, Rutgers University, Newark, NJ 07102, USA
| | - Cynthia P Roesler
- Center for Molecular and Behavioral Neuroscience, Rutgers University, Newark, NJ 07102, USA
| | - Teresa Realpe-Bonilla
- Center for Molecular and Behavioral Neuroscience, Rutgers University, Newark, NJ 07102, USA
| | - April A Benasich
- Center for Molecular and Behavioral Neuroscience, Rutgers University, Newark, NJ 07102, USA
| |
Collapse
|
3
|
The Role of Acoustic Similarity and Non-Native Categorisation in Predicting Non-Native Discrimination: Brazilian Portuguese Vowels by English vs. Spanish Listeners. LANGUAGES 2021. [DOI: 10.3390/languages6010044] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
This study tests whether Australian English (AusE) and European Spanish (ES) listeners differ in their categorisation and discrimination of Brazilian Portuguese (BP) vowels. In particular, we investigate two theoretically relevant measures of vowel category overlap (acoustic vs. perceptual categorisation) as predictors of non-native discrimination difficulty. We also investigate whether the individual listener’s own native vowel productions predict non-native vowel perception better than group averages. The results showed comparable performance for AusE and ES participants in their perception of the BP vowels. In particular, discrimination patterns were largely dependent on contrast-specific learning scenarios, which were similar across AusE and ES. We also found that acoustic similarity between individuals’ own native productions and the BP stimuli were largely consistent with the participants’ patterns of non-native categorisation. Furthermore, the results indicated that both acoustic and perceptual overlap successfully predict discrimination performance. However, accuracy in discrimination was better explained by perceptual similarity for ES listeners and by acoustic similarity for AusE listeners. Interestingly, we also found that for ES listeners, the group averages explained discrimination accuracy better than predictions based on individual production data, but that the AusE group showed no difference.
Collapse
|
4
|
Ou J, Law SP. Induced gamma oscillations index individual differences in speech sound perception and production. Neuropsychologia 2018; 121:28-36. [PMID: 30391567 DOI: 10.1016/j.neuropsychologia.2018.10.028] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/16/2018] [Revised: 10/15/2018] [Accepted: 10/29/2018] [Indexed: 11/16/2022]
Abstract
Auditory neuroscience has provided strong evidence that neural oscillations synchronize to the rhythm of speech stimuli, and oscillations at different frequencies have been linked to processing of different language structures. The present study aims to examine how these ubiquitous neurophysiological attributes may inform us about the brain processes that underpin individual differences in speech perception and production, which in turn elucidate the specific functions of neural oscillations in the domain of speech processing. To this end, we recorded electrophysiological responses to a lexical tone contrast in a passive auditory oddball paradigm from two groups of healthy tone-language speakers who were equal in perceptual discriminability but differed in response latency and production distinctiveness of the tone contrast. Time-frequency analysis was applied to the EEG data, and decomposed into theta (4-7 Hz), beta (12-30 Hz), and gamma (30-50 Hz) frequency bands. Results show that listeners with longer discrimination RT and less distinctive production showed significantly higher induced (non-phase-locked) gamma during tone processing. Moreover, among speakers with less distinctive production, individual differences in induced gamma were significantly correlated with discrimination latency and production distinction. Based on the present findings, we propose that differences in gamma oscillations reflect differential sensory/perceptual computations during acoustic encoding, impacting the quality of perceptual representations, which further mediates individual differences in speech perception and production.
Collapse
Affiliation(s)
- Jinghua Ou
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong Special Administrative Region.
| | - Sam-Po Law
- Division of Speech and Hearing Science the University of Hong Kong, Hong Kong Special Administrative Region
| |
Collapse
|
5
|
Nacar Garcia L, Guerrero-Mosquera C, Colomer M, Sebastian-Galles N. Evoked and oscillatory EEG activity differentiates language discrimination in young monolingual and bilingual infants. Sci Rep 2018; 8:2770. [PMID: 29426859 PMCID: PMC5807452 DOI: 10.1038/s41598-018-20824-0] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2017] [Accepted: 01/25/2018] [Indexed: 11/08/2022] Open
Abstract
Language discrimination is one of the core differences between bilingual and monolingual language acquisition. Here, we investigate the earliest brain specialization induced by it. Following previous research, we hypothesize that bilingual native language discrimination is a complex process involving specific processing of the prosodic properties of the speech signal. We recorded the brain activity of monolingual and bilingual 4.5-month-old infants using EEG, while listening to their native/dominant language and two foreign languages. We defined two different windows of analysis to separate discrimination and identification effects. In the early window of analysis (150-280 ms) we measured the P200 component, and in the later window of analysis we measured Theta (400-1800 ms) and Gamma (300-2800 ms) oscillations. The results point in the direction of different language discrimination strategies for bilingual and monolingual infants. While only monolingual infants show early discrimination of their native language based on familiarity, bilinguals perform a later processing which is compatible with an increase in attention to the speech signal. This is the earliest evidence found for brain specialization induced by bilingualism.
Collapse
Affiliation(s)
- Loreto Nacar Garcia
- Infant Studies Centre, University of British Columbia, 2136 West Mall, Vancouver, BC, V6T 1Z4, Canada.
- Center for Brain and Cognition, Universitat Pompeu Fabra, Ramon Trias Fargas, 25-27, 08005, Barcelona, Spain.
| | - Carlos Guerrero-Mosquera
- Center for Brain and Cognition, Universitat Pompeu Fabra, Ramon Trias Fargas, 25-27, 08005, Barcelona, Spain
| | - Marc Colomer
- Center for Brain and Cognition, Universitat Pompeu Fabra, Ramon Trias Fargas, 25-27, 08005, Barcelona, Spain
| | - Nuria Sebastian-Galles
- Center for Brain and Cognition, Universitat Pompeu Fabra, Ramon Trias Fargas, 25-27, 08005, Barcelona, Spain
| |
Collapse
|
6
|
Wang X, Wang S, Fan Y, Huang D, Zhang Y. Speech-specific categorical perception deficit in autism: An Event-Related Potential study of lexical tone processing in Mandarin-speaking children. Sci Rep 2017; 7:43254. [PMID: 28225070 PMCID: PMC5320551 DOI: 10.1038/srep43254] [Citation(s) in RCA: 36] [Impact Index Per Article: 5.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2016] [Accepted: 01/20/2017] [Indexed: 01/14/2023] Open
Abstract
Recent studies reveal that tonal language speakers with autism have enhanced neural sensitivity to pitch changes in nonspeech stimuli but not to lexical tone contrasts in their native language. The present ERP study investigated whether the distinct pitch processing pattern for speech and nonspeech stimuli in autism was due to a speech-specific deficit in categorical perception of lexical tones. A passive oddball paradigm was adopted to examine two groups (16 in the autism group and 15 in the control group) of Chinese children’s Mismatch Responses (MMRs) to equivalent pitch deviations representing within-category and between-category differences in speech and nonspeech contexts. To further examine group-level differences in the MMRs to categorical perception of speech/nonspeech stimuli or lack thereof, neural oscillatory activities at the single trial level were further calculated with the inter-trial phase coherence (ITPC) measure for the theta and beta frequency bands. The MMR and ITPC data from the children with autism showed evidence for lack of categorical perception in the lexical tone condition. In view of the important role of lexical tones in acquiring a tonal language, the results point to the necessity of early intervention for the individuals with autism who show such a speech-specific categorical perception deficit.
Collapse
Affiliation(s)
- Xiaoyue Wang
- School of Psychology, South China Normal University, Guangzhou, 510631, China
| | - Suiping Wang
- School of Psychology, South China Normal University, Guangzhou, 510631, China.,Center for Studies of Psychological Application, South China Normal University, 510631, China.,Guangdong Provincial Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, 510631, China
| | - Yuebo Fan
- Guangzhou Rehabilitation and Research Center for Children with Autism, Guangzhou Cana School, Guangzhou, 510540, China
| | - Dan Huang
- Guangzhou Rehabilitation and Research Center for Children with Autism, Guangzhou Cana School, Guangzhou, 510540, China
| | - Yang Zhang
- Department of Speech-Language-Hearing Science, University of Minnesota, Minneapolis, MN, 55455, USA.,Center for Neurobehavioral Development, University of Minnesota, Minneapolis, MN, 55455, USA
| |
Collapse
|
7
|
Oscillatory Dynamics Underlying Perceptual Narrowing of Native Phoneme Mapping from 6 to 12 Months of Age. J Neurosci 2016; 36:12095-12105. [PMID: 27903720 DOI: 10.1523/jneurosci.1162-16.2016] [Citation(s) in RCA: 64] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2016] [Revised: 09/08/2016] [Accepted: 10/07/2016] [Indexed: 11/21/2022] Open
Abstract
During the first months of life, human infants process phonemic elements from all languages similarly. However, by 12 months of age, as language-specific phonemic maps are established, infants respond preferentially to their native language. This process, known as perceptual narrowing, supports neural representation and thus efficient processing of the distinctive phonemes within the sound environment. Although oscillatory mechanisms underlying processing of native and non-native phonemic contrasts were recently delineated in 6-month-old infants, the maturational trajectory of these mechanisms remained unclear. A group of typically developing infants born into monolingual English families, were followed from 6 to 12 months and presented with English and Spanish syllable contrasts varying in voice-onset time. Brain responses were recorded with high-density electroencephalogram, and sources of event-related potential generators identified at right and left auditory cortices at 6 and 12 months and also at frontal cortex at 6 months. Time-frequency analyses conducted at source level found variations in both θ and γ ranges across age. Compared with 6-month-olds, 12-month-olds' responses to native phonemes showed smaller and faster phase synchronization and less spectral power in the θ range, and increases in left phase synchrony as well as induced high-γ activity in both frontal and left auditory sources. These results demonstrate that infants become more automatized and efficient in processing their native language as they approach 12 months of age via the interplay between θ and γ oscillations. We suggest that, while θ oscillations support syllable processing, γ oscillations underlie phonemic perceptual narrowing, progressively favoring mapping of native over non-native language across the first year of life. SIGNIFICANCE STATEMENT During early language acquisition, typically developing infants gradually construct phonemic maps of their native language in auditory cortex. It is well known that, by 12 months of age, human infants move from universal discrimination of most linguistic phonemic contrasts to phonemic expertise in their native language. This perceptual narrowing occurs at the expense of the ability to process non-native phonemes. However, the neural mechanisms underlying this process are still poorly understood. Here we demonstrate that perceptual narrowing is, at least in part, accomplished by decreasing power and phase coherence in the θ range while increasing activity in high-γ in left auditory cortex. Understanding the normative neural mechanisms that support early language acquisition is crucial to understanding and perhaps ameliorating developmental language disorders.
Collapse
|
8
|
Koerner TK, Zhang Y, Nelson PB, Wang B, Zou H. Neural indices of phonemic discrimination and sentence-level speech intelligibility in quiet and noise: A mismatch negativity study. Hear Res 2016; 339:40-9. [PMID: 27267705 DOI: 10.1016/j.heares.2016.06.001] [Citation(s) in RCA: 24] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/27/2016] [Revised: 05/16/2016] [Accepted: 06/02/2016] [Indexed: 11/17/2022]
Abstract
Successful speech communication requires the extraction of important acoustic cues from irrelevant background noise. In order to better understand this process, this study examined the effects of background noise on mismatch negativity (MMN) latency, amplitude, and spectral power measures as well as behavioral speech intelligibility tasks. Auditory event-related potentials (AERPs) were obtained from 15 normal-hearing participants to determine whether pre-attentive MMN measures recorded in response to a consonant (from /ba/ to /bu/) and vowel change (from /ba/ to /da/) in a double-oddball paradigm can predict sentence-level speech perception. The results showed that background noise increased MMN latencies and decreased MMN amplitudes with a reduction in the theta frequency band power. Differential noise-induced effects were observed for the pre-attentive processing of consonant and vowel changes due to different degrees of signal degradation by noise. Linear mixed-effects models further revealed significant correlations between the MMN measures and speech intelligibility scores across conditions and stimuli. These results confirm the utility of MMN as an objective neural marker for understanding noise-induced variations as well as individual differences in speech perception, which has important implications for potential clinical applications.
Collapse
Affiliation(s)
- Tess K Koerner
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, MN 55455, USA
| | - Yang Zhang
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, MN 55455, USA; Center for Neurobehavioral Development, University of Minnesota, Minneapolis, MN 55455, USA; Center for Applied Translational Sensory Science, University of Minnesota, Minneapolis, MN 55455, USA.
| | - Peggy B Nelson
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, MN 55455, USA; Center for Applied Translational Sensory Science, University of Minnesota, Minneapolis, MN 55455, USA
| | - Boxiang Wang
- School of Statistics, University of Minnesota, Minneapolis, MN 55455, USA
| | - Hui Zou
- School of Statistics, University of Minnesota, Minneapolis, MN 55455, USA
| |
Collapse
|
9
|
Pérez A, Carreiras M, Gillon Dowens M, Duñabeitia JA. Differential oscillatory encoding of foreign speech. BRAIN AND LANGUAGE 2015; 147:51-57. [PMID: 26070104 DOI: 10.1016/j.bandl.2015.05.008] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/06/2014] [Revised: 04/15/2015] [Accepted: 05/16/2015] [Indexed: 06/04/2023]
Abstract
Neuronal oscillations play a key role in auditory perception of verbal input, with the oscillatory rhythms of the brain showing synchronization with specific frequencies of speech. Here we investigated the neural oscillatory patterns associated with perceiving native, foreign, and unknown speech. Spectral power and phase synchronization were compared to those of a silent context. Power synchronization to native speech was found in frequency ranges corresponding to the theta band, while no synchronization patterns were found for the foreign speech context and the unknown language context. For phase synchrony, the native and unknown languages showed higher synchronization in the theta-band than the foreign language when compared to the silent condition. These results suggest that neural synchronization patterns are markedly different for native and foreign languages.
Collapse
Affiliation(s)
- Alejandro Pérez
- BCBL - Basque Center on Cognition Brain and Language, 20009 Donostia, Spain.
| | - Manuel Carreiras
- BCBL - Basque Center on Cognition Brain and Language, 20009 Donostia, Spain; Ikerbasque, Basque Foundation for Science, 48011 Bilbao, Spain; Departamento de Filología Vasca, EHU/UPV, 48015 Bilbao, Spain
| | - Margaret Gillon Dowens
- University of Nottingham Ningbo Interdisciplinary Centre on Research in Neuroscience (UNNICORN), 315100 Ningbo, China
| | | |
Collapse
|
10
|
Rocha-Muniz CN, Befi-Lopes DM, Schochat E. Mismatch negativity in children with specific language impairment and auditory processing disorder. Braz J Otorhinolaryngol 2015; 81:408-15. [PMID: 26142650 PMCID: PMC9442763 DOI: 10.1016/j.bjorl.2014.08.022] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2014] [Accepted: 08/12/2014] [Indexed: 11/27/2022] Open
Abstract
Introduction Mismatch negativity, an electrophysiological measure, evaluates the brain's capacity to discriminate sounds, regardless of attentional and behavioral capacity. Thus, this auditory event-related potential is promising in the study of the neurophysiological basis underlying auditory processing. Objective To investigate complex acoustic signals (speech) encoded in the auditory nervous system of children with specific language impairment and compare with children with auditory processing disorders and typical development through the mismatch negativity paradigm. Methods It was a prospective study. 75 children (6–12 years) participated in this study: 25 children with specific language impairment, 25 with auditory processing disorders, and 25 with typical development. Mismatch negativity was obtained by subtracting from the waves obtained by the stimuli /ga/ (frequent) and /da/ (rare). Measures of mismatch negativity latency and two amplitude measures were analyzed. Results It was possible to verify an absence of mismatch negativity in 16% children with specific language impairment and 24% children with auditory processing disorders. In the comparative analysis, auditory processing disorders and specific language impairment showed higher latency values and lower amplitude values compared to typical development. Conclusion These data demonstrate changes in the automatic discrimination of crucial acoustic components of speech sounds in children with specific language impairment and auditory processing disorders. It could indicate problems in physiological processes responsible for ensuring the discrimination of acoustic contrasts in pre-attentional and pre-conscious levels, contributing to poor perception.
Collapse
|