1
|
Bidelman GM, York A, Pearson C. Neural correlates of phonetic categorization under auditory (phoneme) and visual (grapheme) modalities. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.07.24.604940. [PMID: 39211275 PMCID: PMC11361091 DOI: 10.1101/2024.07.24.604940] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/04/2024]
Abstract
We tested whether the neural mechanisms of phonetic categorization are specific to speech sounds or generalize to graphemes (i.e., visual letters) of the same phonetic label. Given that linguistic experience shapes categorical processing, and letter-speech sound matching plays a crucial role during early reading acquisition, we hypothesized sound phoneme and visual grapheme tokens representing the same linguistic identity might recruit common neural substrates, despite originating from different sensory modalities. Behavioral and neuroelectric brain responses (ERPs) were acquired as participants categorized stimuli from sound (phoneme) and homologous letter (grapheme) continua each spanning a /da/ - /ga/ gradient. Behaviorally, listeners were faster and showed stronger categorization of phoneme compared to graphemes. At the neural level, multidimensional scaling of the EEG revealed responses self-organized in a categorial fashion such that tokens clustered within their respective modality beginning ∼150-250 ms after stimulus onset. Source-resolved ERPs further revealed modality-specific and overlapping brain regions supporting phonetic categorization. Left inferior frontal gyrus and auditory cortex showed stronger responses for sound category members compared to phonetically ambiguous tokens, whereas early visual cortices paralleled this categorical organization for graphemes. Auditory and visual categorization also recruited common visual association areas in extrastriate cortex but in opposite hemispheres (auditory = left; visual=right). Our findings reveal both auditory and visual sensory cortex supports categorical organization for phonetic labels within their respective modalities. However, a partial overlap in phoneme and grapheme processing among occipital brain areas implies the presence of an isomorphic, domain-general mapping for phonetic categories in dorsal visual system.
Collapse
|
2
|
Perron M, Liu Q, Tremblay P, Alain C. Enhancing speech perception in noise through articulation. Ann N Y Acad Sci 2024; 1537:140-154. [PMID: 38924165 DOI: 10.1111/nyas.15179] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/28/2024]
Abstract
Considerable debate exists about the interplay between auditory and motor speech systems. Some argue for common neural mechanisms, whereas others assert that there are few shared resources. In four experiments, we tested the hypothesis that priming the speech motor system by repeating syllable pairs aloud improves subsequent syllable discrimination in noise compared with a priming discrimination task involving same-different judgments via button presses. Our results consistently showed that participants who engaged in syllable repetition performed better in syllable discrimination in noise than those who engaged in the priming discrimination task. This gain in accuracy was observed for primed and new syllable pairs, highlighting increased sensitivity to phonological details. The benefits were comparable whether the priming tasks involved auditory or visual presentation. Inserting a 1-h delay between the priming tasks and the syllable-in-noise task, the benefits persisted but were confined to primed syllable pairs. Finally, we demonstrated the effectiveness of this approach in older adults. Our findings substantiate the existence of a speech production-perception relationship. They also have clinical relevance as they raise the possibility of production-based interventions to improve speech perception ability. This would be particularly relevant for older adults who often encounter difficulties in perceiving speech in noise.
Collapse
Affiliation(s)
- Maxime Perron
- Department of Psychology, University of Toronto, Toronto, Ontario, Canada
- Baycrest Academy for Research and Education, Rotman Research Institute, North York, Ontario, Canada
| | - Qiying Liu
- Department of Psychology, University of Toronto, Toronto, Ontario, Canada
- Baycrest Academy for Research and Education, Rotman Research Institute, North York, Ontario, Canada
| | - Pascale Tremblay
- CERVO Brain Research Center, Quebec City, Quebec, Canada
- École de Réadaptation, Faculté de Médecine, Université Laval, Quebec City, Quebec, Canada
| | - Claude Alain
- Department of Psychology, University of Toronto, Toronto, Ontario, Canada
- Baycrest Academy for Research and Education, Rotman Research Institute, North York, Ontario, Canada
- Institute of Medical Sciences, University of Toronto, Toronto, Ontario, Canada
- Music and Health Science Research Collaboratory, University of Toronto, Toronto, Ontario, Canada
| |
Collapse
|
3
|
Bosseler AN, Meltzoff AN, Bierer S, Huber E, Mizrahi JC, Larson E, Endevelt-Shapira Y, Taulu S, Kuhl PK. Infants' brain responses to social interaction predict future language growth. Curr Biol 2024; 34:1731-1738.e3. [PMID: 38593800 PMCID: PMC11090161 DOI: 10.1016/j.cub.2024.03.020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2023] [Revised: 02/26/2024] [Accepted: 03/13/2024] [Indexed: 04/11/2024]
Abstract
In face-to-face interactions with infants, human adults exhibit a species-specific communicative signal. Adults present a distinctive "social ensemble": they use infant-directed speech (parentese), respond contingently to infants' actions and vocalizations, and react positively through mutual eye-gaze and smiling. Studies suggest that this social ensemble is essential for initial language learning. Our hypothesis is that the social ensemble attracts attentional systems to speech and that sensorimotor systems prepare infants to respond vocally, both of which advance language learning. Using infant magnetoencephalography (MEG), we measure 5-month-old infants' neural responses during live verbal face-to-face (F2F) interaction with an adult (social condition) and during a control (nonsocial condition) in which the adult turns away from the infant to speak to another person. Using a longitudinal design, we tested whether infants' brain responses to these conditions at 5 months of age predicted their language growth at five future time points. Brain areas involved in attention (right hemisphere inferior frontal, right hemisphere superior temporal, and right hemisphere inferior parietal) show significantly higher theta activity in the social versus nonsocial condition. Critical to theory, we found that infants' neural activity in response to F2F interaction in attentional and sensorimotor regions significantly predicted future language development into the third year of life, more than 2 years after the initial measurements. We develop a view of early language acquisition that underscores the centrality of the social ensemble, and we offer new insight into the neurobiological components that link infants' language learning to their early brain functioning during social interaction.
Collapse
Affiliation(s)
- Alexis N Bosseler
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA 98195, USA
| | - Andrew N Meltzoff
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA 98195, USA; Department of Psychology, University of Washington, Seattle, WA 98195, USA
| | - Steven Bierer
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA 98195, USA
| | - Elizabeth Huber
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA 98195, USA; Department of Speech and Hearing Sciences, University of Washington, Seattle, WA 98195, USA
| | - Julia C Mizrahi
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA 98195, USA
| | - Eric Larson
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA 98195, USA
| | - Yaara Endevelt-Shapira
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA 98195, USA
| | - Samu Taulu
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA 98195, USA; Department of Physics, University of Washington, Seattle, WA 98195, USA
| | - Patricia K Kuhl
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA 98195, USA; Department of Speech and Hearing Sciences, University of Washington, Seattle, WA 98195, USA.
| |
Collapse
|
4
|
Varuzza C, D’Aiello B, Lazzaro G, Quarin F, De Rose P, Bergonzini P, Menghini D, Marini A, Vicari S. Gross, Fine and Visual-Motor Skills in Children with Language Disorder, Speech Sound Disorder and Their Combination. Brain Sci 2022; 13:brainsci13010059. [PMID: 36672041 PMCID: PMC9856286 DOI: 10.3390/brainsci13010059] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2022] [Revised: 12/16/2022] [Accepted: 12/27/2022] [Indexed: 12/30/2022] Open
Abstract
Increasing evidence shows that children with Communication Disorders (CDs) may show gross, fine, and visual-motor difficulties compared to children with typical development. Accordingly, the present study aims to characterize gross, fine and visual-motor skills in children with CDs, distinguishing children with CDs into three subgroups, i.e., with Language Disorders (LD), Speech Sound Disorders (SSD), and LD + SSD. In Experiment 1, around 60% of children with CDs (4 to 7 years; 21 with LD, 36 with SSD, and 90 with LD + SSD) showed clinical/borderline scores in balance skills, regardless of the type of communication deficit. However, children with LD, SSD, and LD + SSD did not differ in gross and fine motor skills. In Experiment 2, a higher percentage of children with CDs (4 to 7 years; 34 with LD, 62 with SSD, 148 with LD + SSD) obtained clinical/borderline scores in Visual Perception skills. Moreover, children with LD + SSD performed significantly worsen in Visual Perception and Fine Motor Coordination skills compared to children with SSD only. Our results underlined that CDs are generally associated with gross motor difficulties and that visual-motor difficulties are related to the type of communication deficit. Paying earlier attention to the motor skills of children with CDs could help clinicians design effective interventions.
Collapse
Affiliation(s)
- Cristiana Varuzza
- Child and Adolescent Neuropsychiatry Unit, Department of Neuroscience, Bambino Gesù Children’s Hospital, IRCCS, 00146 Rome, Italy
| | - Barbara D’Aiello
- Child and Adolescent Neuropsychiatry Unit, Department of Neuroscience, Bambino Gesù Children’s Hospital, IRCCS, 00146 Rome, Italy
- Department of Human Science, LUMSA University, 00193 Rome, Italy
| | - Giulia Lazzaro
- Child and Adolescent Neuropsychiatry Unit, Department of Neuroscience, Bambino Gesù Children’s Hospital, IRCCS, 00146 Rome, Italy
| | - Fabio Quarin
- Child and Adolescent Neuropsychiatry Unit, Department of Neuroscience, Bambino Gesù Children’s Hospital, IRCCS, 00146 Rome, Italy
| | - Paola De Rose
- Child and Adolescent Neuropsychiatry Unit, Department of Neuroscience, Bambino Gesù Children’s Hospital, IRCCS, 00146 Rome, Italy
| | - Paola Bergonzini
- Child and Adolescent Neuropsychiatry Unit, Department of Neuroscience, Bambino Gesù Children’s Hospital, IRCCS, 00146 Rome, Italy
| | - Deny Menghini
- Child and Adolescent Neuropsychiatry Unit, Department of Neuroscience, Bambino Gesù Children’s Hospital, IRCCS, 00146 Rome, Italy
- Correspondence:
| | - Andrea Marini
- Department of Language and Literatures, Communication, Education and Society, University of Udine, 33100 Udine, Italy
| | - Stefano Vicari
- Child and Adolescent Neuropsychiatry Unit, Department of Neuroscience, Bambino Gesù Children’s Hospital, IRCCS, 00146 Rome, Italy
- Department of Life Science and Public Health, Catholic University, 00168 Rome, Italy
| |
Collapse
|
5
|
Ross LA, Molholm S, Butler JS, Bene VAD, Foxe JJ. Neural correlates of multisensory enhancement in audiovisual narrative speech perception: a fMRI investigation. Neuroimage 2022; 263:119598. [PMID: 36049699 DOI: 10.1016/j.neuroimage.2022.119598] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/18/2022] [Revised: 08/26/2022] [Accepted: 08/28/2022] [Indexed: 11/25/2022] Open
Abstract
This fMRI study investigated the effect of seeing articulatory movements of a speaker while listening to a naturalistic narrative stimulus. It had the goal to identify regions of the language network showing multisensory enhancement under synchronous audiovisual conditions. We expected this enhancement to emerge in regions known to underlie the integration of auditory and visual information such as the posterior superior temporal gyrus as well as parts of the broader language network, including the semantic system. To this end we presented 53 participants with a continuous narration of a story in auditory alone, visual alone, and both synchronous and asynchronous audiovisual speech conditions while recording brain activity using BOLD fMRI. We found multisensory enhancement in an extensive network of regions underlying multisensory integration and parts of the semantic network as well as extralinguistic regions not usually associated with multisensory integration, namely the primary visual cortex and the bilateral amygdala. Analysis also revealed involvement of thalamic brain regions along the visual and auditory pathways more commonly associated with early sensory processing. We conclude that under natural listening conditions, multisensory enhancement not only involves sites of multisensory integration but many regions of the wider semantic network and includes regions associated with extralinguistic sensory, perceptual and cognitive processing.
Collapse
Affiliation(s)
- Lars A Ross
- The Frederick J. and Marion A. Schindler Cognitive Neurophysiology Laboratory, The Ernest J. Del Monte Institute for Neuroscience, Department of Neuroscience, University of Rochester School of Medicine and Dentistry, Rochester, New York, 14642, USA; Department of Imaging Sciences, University of Rochester Medical Center, University of Rochester School of Medicine and Dentistry, Rochester, New York, 14642, USA; The Cognitive Neurophysiology Laboratory, Departments of Pediatrics and Neuroscience, Albert Einstein College of Medicine & Montefiore Medical Center, Bronx, New York, 10461, USA.
| | - Sophie Molholm
- The Frederick J. and Marion A. Schindler Cognitive Neurophysiology Laboratory, The Ernest J. Del Monte Institute for Neuroscience, Department of Neuroscience, University of Rochester School of Medicine and Dentistry, Rochester, New York, 14642, USA; The Cognitive Neurophysiology Laboratory, Departments of Pediatrics and Neuroscience, Albert Einstein College of Medicine & Montefiore Medical Center, Bronx, New York, 10461, USA
| | - John S Butler
- The Cognitive Neurophysiology Laboratory, Departments of Pediatrics and Neuroscience, Albert Einstein College of Medicine & Montefiore Medical Center, Bronx, New York, 10461, USA; School of Mathematical Sciences, Technological University Dublin, Kevin Street Campus, Dublin, Ireland
| | - Victor A Del Bene
- The Cognitive Neurophysiology Laboratory, Departments of Pediatrics and Neuroscience, Albert Einstein College of Medicine & Montefiore Medical Center, Bronx, New York, 10461, USA; University of Alabama at Birmingham, Heersink School of Medicine, Department of Neurology, Birmingham, Alabama, 35233, USA
| | - John J Foxe
- The Frederick J. and Marion A. Schindler Cognitive Neurophysiology Laboratory, The Ernest J. Del Monte Institute for Neuroscience, Department of Neuroscience, University of Rochester School of Medicine and Dentistry, Rochester, New York, 14642, USA; The Cognitive Neurophysiology Laboratory, Departments of Pediatrics and Neuroscience, Albert Einstein College of Medicine & Montefiore Medical Center, Bronx, New York, 10461, USA.
| |
Collapse
|
6
|
Carter JA, Buder EH, Bidelman GM. Nonlinear dynamics in auditory cortical activity reveal the neural basis of perceptual warping in speech categorization. JASA EXPRESS LETTERS 2022; 2:045201. [PMID: 35434716 PMCID: PMC8984957 DOI: 10.1121/10.0009896] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/30/2021] [Accepted: 03/03/2022] [Indexed: 06/14/2023]
Abstract
Surrounding context influences speech listening, resulting in dynamic shifts to category percepts. To examine its neural basis, event-related potentials (ERPs) were recorded during vowel identification with continua presented in random, forward, and backward orders to induce perceptual warping. Behaviorally, sequential order shifted individual listeners' categorical boundary, versus random delivery, revealing perceptual warping (biasing) of the heard phonetic category dependent on recent stimulus history. ERPs revealed later (∼300 ms) activity localized to superior temporal and middle/inferior frontal gyri that predicted listeners' hysteresis/enhanced contrast magnitudes. Findings demonstrate that interactions between frontotemporal brain regions govern top-down, stimulus history effects on speech categorization.
Collapse
Affiliation(s)
- Jared A Carter
- Institute for Intelligent Systems, University of Memphis, Memphis, Tennessee 38152, USA
| | - Eugene H Buder
- School of Communication Sciences and Disorders, University of Memphis, Memphis, Tennessee 38152, USA
| | - Gavin M Bidelman
- Department of Speech, Language and Hearing Sciences, Indiana University, , Bloomington, Indiana 47408, USA , ,
| |
Collapse
|
7
|
Embodied cognition in neurodegenerative disorders: What do we know so far? A narrative review focusing on the mirror neuron system and clinical applications. J Clin Neurosci 2022; 98:66-72. [DOI: 10.1016/j.jocn.2022.01.028] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2020] [Revised: 06/24/2021] [Accepted: 01/22/2022] [Indexed: 02/04/2023]
|
8
|
Tamura S, Hirose N, Mitsudo T, Hoaki N, Nakamura I, Onitsuka T, Hirano Y. Multi-modal imaging of the auditory-larynx motor network for voicing perception. Neuroimage 2022; 251:118981. [PMID: 35150835 DOI: 10.1016/j.neuroimage.2022.118981] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2021] [Revised: 12/20/2021] [Accepted: 02/07/2022] [Indexed: 10/19/2022] Open
Abstract
Voicing is one of the most important characteristics of phonetic speech sounds. Despite its importance, voicing perception mechanisms remain largely unknown. To explore auditory-motor networks associated with voicing perception, we firstly examined the brain regions that showed common activities for voicing production and perception using functional magnetic resonance imaging. Results indicated that the auditory and speech motor areas were activated with the operculum parietale 4 (OP4) during both voicing production and perception. Secondly, we used a magnetoencephalography and examined the dynamical functional connectivity of the auditory-motor networks during a perceptual categorization task of /da/-/ta/ continuum stimuli varying in voice onset time (VOT) from 0 to 40 ms in 10 ms steps. Significant functional connectivities from the auditory cortical regions to the larynx motor area via OP4 were observed only when perceiving the stimulus with VOT 30 ms. In addition, regional activity analysis showed that the neural representation of VOT in the auditory cortical regions was mostly correlated with categorical perception of voicing but did not reflect the perception of stimulus with VOT 30 ms. We suggest that the larynx motor area, which is considered to play a crucial role in voicing production, contributes to categorical perception of voicing by complementing the temporal processing in the auditory cortical regions.
Collapse
Affiliation(s)
- Shunsuke Tamura
- Department of Neuropsychiatry, Graduate School of Medical Sciences, Kyushu University, 3-1-1 Maidashi, Higashiku, Fukuoka 812-8582, Japan.
| | - Nobuyuki Hirose
- Faculty of Information Science and Electrical Engineering, Kyushu University, Fukuoka, Japan
| | - Takako Mitsudo
- Department of Neuropsychiatry, Graduate School of Medical Sciences, Kyushu University, 3-1-1 Maidashi, Higashiku, Fukuoka 812-8582, Japan
| | | | - Itta Nakamura
- Department of Neuropsychiatry, Graduate School of Medical Sciences, Kyushu University, 3-1-1 Maidashi, Higashiku, Fukuoka 812-8582, Japan
| | - Toshiaki Onitsuka
- Department of Neuropsychiatry, Graduate School of Medical Sciences, Kyushu University, 3-1-1 Maidashi, Higashiku, Fukuoka 812-8582, Japan
| | - Yoji Hirano
- Department of Neuropsychiatry, Graduate School of Medical Sciences, Kyushu University, 3-1-1 Maidashi, Higashiku, Fukuoka 812-8582, Japan; Neural Dynamics Laboratory, Research Service, VA Boston Healthcare System, and Department of Psychiatry, Harvard Medical School, Boston, United States
| |
Collapse
|
9
|
Abstract
Ten years ago, Perspectives in Psychological Science published the Mirror Neuron Forum, in which authors debated the role of mirror neurons in action understanding, speech, imitation, and autism and asked whether mirror neurons are acquired through visual-motor learning. Subsequent research on these themes has made significant advances, which should encourage further, more systematic research. For action understanding, multivoxel pattern analysis, patient studies, and brain stimulation suggest that mirror-neuron brain areas contribute to low-level processing of observed actions (e.g., distinguishing types of grip) but not to high-level action interpretation (e.g., inferring actors' intentions). In the area of speech perception, although it remains unclear whether mirror neurons play a specific, causal role in speech perception, there is compelling evidence for the involvement of the motor system in the discrimination of speech in perceptually noisy conditions. For imitation, there is strong evidence from patient, brain-stimulation, and brain-imaging studies that mirror-neuron brain areas play a causal role in copying of body movement topography. In the area of autism, studies using behavioral and neurological measures have tried and failed to find evidence supporting the "broken-mirror theory" of autism. Furthermore, research on the origin of mirror neurons has confirmed the importance of domain-general visual-motor associative learning rather than canalized visual-motor learning, or motor learning alone.
Collapse
Affiliation(s)
- Cecilia Heyes
- All Souls College, University of Oxford
- Department of Experimental Psychology, University of Oxford
| | - Caroline Catmur
- Department of Psychology, Institute of Psychiatry, Psychology, and Neuroscience, King’s College London
| |
Collapse
|
10
|
Jenson D, Saltuklaroglu T. Sensorimotor contributions to working memory differ between the discrimination of Same and Different syllable pairs. Neuropsychologia 2021; 159:107947. [PMID: 34216594 DOI: 10.1016/j.neuropsychologia.2021.107947] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/10/2020] [Revised: 02/01/2021] [Accepted: 06/27/2021] [Indexed: 10/21/2022]
Abstract
Sensorimotor activity during speech perception is both pervasive and highly variable, changing as a function of the cognitive demands imposed by the task. The purpose of the current study was to evaluate whether the discrimination of Same (matched) and Different (unmatched) syllable pairs elicit different patterns of sensorimotor activity as stimuli are processed in working memory. Raw EEG data recorded from 42 participants were decomposed with independent component analysis to identify bilateral sensorimotor mu rhythms from 36 subjects. Time frequency decomposition of mu rhythms revealed concurrent event related desynchronization (ERD) in alpha and beta frequency bands across the peri- and post-stimulus time periods, which were interpreted as evidence of sensorimotor contributions to working memory encoding and maintenance. Left hemisphere alpha/beta ERD was stronger in Different trials than Same trials during the post-stimulus period, while right hemisphere alpha/beta ERD was stronger in Same trials than Different trials. A between-hemispheres contrast revealed no differences during Same trials, while post-stimulus alpha/beta ERD was stronger in the left hemisphere than the right during Different trials. Results were interpreted to suggest that predictive coding mechanisms lead to repetition suppression effects in Same trials. Mismatches arising from predictive coding mechanisms in Different trials shift subsequent working memory processing to the speech-dominant left hemisphere. Findings clarify how sensorimotor activity differentially supports working memory encoding and maintenance stages during speech discrimination tasks and have potential to inform sensorimotor models of speech perception and working memory.
Collapse
Affiliation(s)
- David Jenson
- Washington State University, Elson S. Floyd College of Medicine, Department of Speech and Hearing Sciences, Spokane, WA, USA.
| | - Tim Saltuklaroglu
- University of Tennessee Health Science Center, College of Health Professions, Department of Audiology and Speech-Pathology, Knoxville, TN, USA
| |
Collapse
|
11
|
Mahmud MS, Yeasin M, Bidelman GM. Data-driven machine learning models for decoding speech categorization from evoked brain responses. J Neural Eng 2021; 18. [PMID: 33690177 DOI: 10.1101/2020.08.03.234997] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2020] [Accepted: 03/09/2021] [Indexed: 05/24/2023]
Abstract
Objective.Categorical perception (CP) of audio is critical to understand how the human brain perceives speech sounds despite widespread variability in acoustic properties. Here, we investigated the spatiotemporal characteristics of auditory neural activity that reflects CP for speech (i.e. differentiates phonetic prototypes from ambiguous speech sounds).Approach.We recorded 64-channel electroencephalograms as listeners rapidly classified vowel sounds along an acoustic-phonetic continuum. We used support vector machine classifiers and stability selection to determine when and where in the brain CP was best decoded across space and time via source-level analysis of the event-related potentials.Main results. We found that early (120 ms) whole-brain data decoded speech categories (i.e. prototypical vs. ambiguous tokens) with 95.16% accuracy (area under the curve 95.14%;F1-score 95.00%). Separate analyses on left hemisphere (LH) and right hemisphere (RH) responses showed that LH decoding was more accurate and earlier than RH (89.03% vs. 86.45% accuracy; 140 ms vs. 200 ms). Stability (feature) selection identified 13 regions of interest (ROIs) out of 68 brain regions [including auditory cortex, supramarginal gyrus, and inferior frontal gyrus (IFG)] that showed categorical representation during stimulus encoding (0-260 ms). In contrast, 15 ROIs (including fronto-parietal regions, IFG, motor cortex) were necessary to describe later decision stages (later 300-800 ms) of categorization but these areas were highly associated with the strength of listeners' categorical hearing (i.e. slope of behavioral identification functions).Significance.Our data-driven multivariate models demonstrate that abstract categories emerge surprisingly early (∼120 ms) in the time course of speech processing and are dominated by engagement of a relatively compact fronto-temporal-parietal brain network.
Collapse
Affiliation(s)
- Md Sultan Mahmud
- Department of Electrical and Computer Engineering, University of Memphis, 3815 Central Avenue, Memphis, TN 38152, United States of America
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, United States of America
| | - Mohammed Yeasin
- Department of Electrical and Computer Engineering, University of Memphis, 3815 Central Avenue, Memphis, TN 38152, United States of America
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, United States of America
| | - Gavin M Bidelman
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, United States of America
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, United States of America
- University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, United States of America
| |
Collapse
|
12
|
Mahmud MS, Yeasin M, Bidelman GM. Data-driven machine learning models for decoding speech categorization from evoked brain responses. J Neural Eng 2021; 18:10.1088/1741-2552/abecf0. [PMID: 33690177 PMCID: PMC8738965 DOI: 10.1088/1741-2552/abecf0] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2020] [Accepted: 03/09/2021] [Indexed: 11/12/2022]
Abstract
Objective.Categorical perception (CP) of audio is critical to understand how the human brain perceives speech sounds despite widespread variability in acoustic properties. Here, we investigated the spatiotemporal characteristics of auditory neural activity that reflects CP for speech (i.e. differentiates phonetic prototypes from ambiguous speech sounds).Approach.We recorded 64-channel electroencephalograms as listeners rapidly classified vowel sounds along an acoustic-phonetic continuum. We used support vector machine classifiers and stability selection to determine when and where in the brain CP was best decoded across space and time via source-level analysis of the event-related potentials.Main results. We found that early (120 ms) whole-brain data decoded speech categories (i.e. prototypical vs. ambiguous tokens) with 95.16% accuracy (area under the curve 95.14%;F1-score 95.00%). Separate analyses on left hemisphere (LH) and right hemisphere (RH) responses showed that LH decoding was more accurate and earlier than RH (89.03% vs. 86.45% accuracy; 140 ms vs. 200 ms). Stability (feature) selection identified 13 regions of interest (ROIs) out of 68 brain regions [including auditory cortex, supramarginal gyrus, and inferior frontal gyrus (IFG)] that showed categorical representation during stimulus encoding (0-260 ms). In contrast, 15 ROIs (including fronto-parietal regions, IFG, motor cortex) were necessary to describe later decision stages (later 300-800 ms) of categorization but these areas were highly associated with the strength of listeners' categorical hearing (i.e. slope of behavioral identification functions).Significance.Our data-driven multivariate models demonstrate that abstract categories emerge surprisingly early (∼120 ms) in the time course of speech processing and are dominated by engagement of a relatively compact fronto-temporal-parietal brain network.
Collapse
Affiliation(s)
- Md Sultan Mahmud
- Department of Electrical and Computer Engineering, University of Memphis, 3815 Central Avenue, Memphis, TN 38152, United States of America
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, United States of America
| | - Mohammed Yeasin
- Department of Electrical and Computer Engineering, University of Memphis, 3815 Central Avenue, Memphis, TN 38152, United States of America
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, United States of America
| | - Gavin M Bidelman
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, United States of America
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, United States of America
- University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, United States of America
| |
Collapse
|
13
|
Naro A, Maggio MG, Latella D, La Rosa G, Sciarrone F, Manuli A, Calabrò RS. Does embodied cognition allow a better management of neurological diseases? A review on the link between cognitive language processing and motor function. APPLIED NEUROPSYCHOLOGY-ADULT 2021; 29:1646-1657. [PMID: 33683162 DOI: 10.1080/23279095.2021.1890595] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
Abstract
Embodied cognition (EC) refers to the interplay occurring in thinking among individual's sensorimotor capacities (i.e., the ability of the body to respond to its senses with movement), the body itself, and the environment. The aim of the present narrative review is to provide an overall understanding of whether and how motor training could lead to language recovery, consistently with EC theories (action-perception cycle, mirror neuron systems -MNS-, and embodied semantics). We therefore reviewed the works dealing with EC in terms of the link between language processing, mirror neuron system (MNS), and motor function, evaluating the potential clinical implications for better managing neurological deficits. Connections between body and mind were found, as body states influence cognitive functions, such as perception and reasoning, as well as language processing, especially in neurological disorders. In fact, abnormalities in "embodied language" were found in movement disorders and neurodegenerative diseases, negatively affecting patients' rehabilitation outcomes. Understanding the link between language processing and motor outcomes is fundamental in the rehabilitation field, given that EC can be targeted to improve patients' functional recovery and quality of life.
Collapse
Affiliation(s)
- Antonino Naro
- IRCCS Centro Neurolesi Bonino Pulejo, Messina, Italy
| | - Maria Grazia Maggio
- Studio di Psicoterapia Relazionale e Riabilitazione Cognitiva, Messina, Italy
| | | | | | | | | | | |
Collapse
|
14
|
Xu Z, Wang ZR, Li J, Hu M, Xiang MQ. Effect of Acute Moderate-Intensity Exercise on the Mirror Neuron System: Role of Cardiovascular Fitness Level. Front Psychol 2020; 11:312. [PMID: 32153482 PMCID: PMC7047835 DOI: 10.3389/fpsyg.2020.00312] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2019] [Accepted: 02/10/2020] [Indexed: 12/19/2022] Open
Abstract
Objectives The aims of this study were to use functional near-infrared spectroscopy (fNIRS) to determine whether cardiovascular fitness levels modulate the activation of the mirror neuron system (MNS) under table-setting tasks in non-exercise situation, to replicate the study that positive effect of acute moderate-intensity exercise on the MNS and investigate whether cardiovascular fitness levels modulates the effect of exercise on the activation of the MNS. Methods Thirty-six healthy college-aged participants completed a maximal graded exercise test (GXT) and were categorized as high, moderate, or low cardiovascular fitness. Participants then performed table-setting tasks including an action execution task (EXEC) and action observation task (OBS) prior to (PRE) and after (POST) either a rest condition (CTRL) or a cycling exercise condition (EXP). The EXP condition consisted of a 5-min warm-up, 15-min moderate-intensity exercise (65% VO2max), and 5-min cool-down. Results No significant differences were observed for Oxy-Hb and Deoxy-Hb between different cardiovascular fitness levels in the EXEC or OBS tasks in the non-exercise session. But there were significant improvements of oxygenated hemoglobin (Oxy-Hb) in the inferior frontal gyrus (IFG) and pre-motor area (PMC) regions under the OBS task following the acute moderate exercise. Particularly, the improvements (Post-Pre) of Δ Oxy-Hb were mainly observed in high and low fitness individuals. There was also a significant improvement of deoxygenated hemoglobin (Deoxy-Hb) in the IPL region under the OBS task. The following analysis indicated that exercise improved Δ Deoxy-Hb in high fitness individuals. Conclusion This study indicated that the activation of MNS was not modulated by the cardiovascular fitness levels in the non-exercise situation. We replicated the previous study that moderate exercise improved activation of MNS; we also provided the first empirical evidence that moderate-intensity exercise positively affects the MNS activation in college students of high and low cardiovascular fitness levels.
Collapse
Affiliation(s)
- Zebo Xu
- Department of Sports and Health, Guangzhou Sport University, Guangzhou, China.,Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong, China
| | - Zi-Rong Wang
- Department of Graduation, Guangzhou Sport University, Guangzhou, China
| | - Jin Li
- Department of Graduation, Guangzhou Sport University, Guangzhou, China
| | - Min Hu
- Department of Sports and Health, Guangzhou Sport University, Guangzhou, China
| | - Ming-Qiang Xiang
- Department of Sports and Health, Guangzhou Sport University, Guangzhou, China
| |
Collapse
|
15
|
Zhao B, Zhang G, Dang J. Temporal-Spatial-Spectral Investigation of Brain Network Dynamics in Human Speech Perception. Brain Inform 2020. [DOI: 10.1007/978-3-030-59277-6_6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022] Open
|
16
|
Jenson D, Thornton D, Harkrider AW, Saltuklaroglu T. Influences of cognitive load on sensorimotor contributions to working memory: An EEG investigation of mu rhythm activity during speech discrimination. Neurobiol Learn Mem 2019; 166:107098. [DOI: 10.1016/j.nlm.2019.107098] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2019] [Revised: 09/11/2019] [Accepted: 10/09/2019] [Indexed: 11/16/2022]
|
17
|
Tamura S, Ito K, Hirose N, Mori S. Precision of voicing perceptual identification is altered in association with voice-onset time production changes. Exp Brain Res 2019; 237:2197-2204. [DOI: 10.1007/s00221-019-05584-1] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/26/2018] [Accepted: 06/13/2019] [Indexed: 11/30/2022]
|
18
|
Abstract
Recent evidence suggests that the motor system may have a facilitatory role in speech perception during noisy listening conditions. Studies clearly show an association between activity in auditory and motor speech systems, but also hint at a causal role for the motor system in noisy speech perception. However, in the most compelling "causal" studies performance was only measured at a single signal-to-noise ratio (SNR). If listening conditions must be noisy to invoke causal motor involvement, then effects will be contingent on the SNR at which they are tested. We used articulatory suppression to disrupt motor-speech areas while measuring phonemic identification across a range of SNRs. As controls, we also measured phoneme identification during passive listening, mandible gesturing, and foot-tapping conditions. Two-parameter (threshold, slope) psychometric functions were fit to the data in each condition. Our findings indicate: (1) no effect of experimental task on psychometric function slopes; (2) a small effect of articulatory suppression, in particular, on psychometric function thresholds. The size of the latter effect was 1 dB (~5% correct) on average, suggesting, at best, a minor modulatory role of the speech motor system in perception.
Collapse
Affiliation(s)
- Ryan C Stokes
- Department of Cognitive Sciences Social and Behavioral Sciences Gateway, University of California - Irvine, Irvine, CA, 92697-5100, USA.
| | - Jonathan H Venezia
- Department of Cognitive Sciences Social and Behavioral Sciences Gateway, University of California - Irvine, Irvine, CA, 92697-5100, USA
| | - Gregory Hickok
- Department of Cognitive Sciences Social and Behavioral Sciences Gateway, University of California - Irvine, Irvine, CA, 92697-5100, USA
| |
Collapse
|
19
|
Rauschecker JP. Where did language come from? Precursor mechanisms in nonhuman primates. Curr Opin Behav Sci 2018; 21:195-204. [PMID: 30778394 PMCID: PMC6377164 DOI: 10.1016/j.cobeha.2018.06.003] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
At first glance, the monkey brain looks like a smaller version of the human brain. Indeed, the anatomical and functional architecture of the cortical auditory system in monkeys is very similar to that of humans, with dual pathways segregated into a ventral and a dorsal processing stream. Yet, monkeys do not speak. Repeated attempts to pin this inability on one particular cause have failed. A closer look at the necessary components of language, according to Darwin, reveals that all of them got a significant boost during evolution from nonhuman to human primates. The vocal-articulatory system, in particular, has developed into the most sophisticated of all human sensorimotor systems with about a dozen effectors that, in combination with each other, result in an auditory communication system like no other. This sensorimotor network possesses all the ingredients of an internal model system that permits the emergence of sequence processing, as required for phonology and syntax in modern languages.
Collapse
Affiliation(s)
- Josef P Rauschecker
- Department of Neuroscience, Georgetown University, Washington, DC 20057, USA
| |
Collapse
|
20
|
Lin X, Lei VLC, Li D, Yuan Z. Which is more costly in Chinese to English simultaneous interpreting, "pairing" or "transphrasing"? Evidence from an fNIRS neuroimaging study. NEUROPHOTONICS 2018; 5:025010. [PMID: 29876369 PMCID: PMC5987679 DOI: 10.1117/1.nph.5.2.025010] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/29/2017] [Accepted: 05/11/2018] [Indexed: 06/08/2023]
Abstract
This study examined the neural mechanism underlying two translation strategies associated with Chinese to English simultaneous interpreting (SI) targeting the left prefrontal cortex (PFC), which is generally involved in the control of interference and conflict resolution and has been identified as the brain area that plays a pivotal role in SI. Brain activation associated with the two strategies including "pairing" and "transphrasing" were compared with that from "nontranslation," which keeps the source language item unchanged in the target language production and is considered as a tactic that does not require complex cognitive operation associated with bilingual processing effort. Our findings revealed that "pairing" elicited the strongest and almost immediate brain activation in the Broca's area, and "transphrasing" resulted in the most extensive and strongest activation overall in the left PFC. By contrast, "nontranslation" induced very little brain activation in these regions. This work, which represents one of the first efforts in investigating brain activation related to translation strategies involving different levels of cognitive control, will not only pave a new avenue for better understanding of the cognitive mechanism underlying SI but also provide further insight into the role that the Broca's region plays in domain-general cognitive control.
Collapse
Affiliation(s)
- Xiaohong Lin
- University of Macau, Bioimaging Core, Faculty of Health Sciences, Macao, China
- Hangzhou Normal University, Institutes of Psychological Sciences, Hangzhou, China
| | - Victoria Lai Cheng Lei
- University of Macau, Centre for Studies of Translation, Interpreting and Cognition, Faculty of Arts and Humanities, Macao, China
| | - Defeng Li
- University of Macau, Centre for Studies of Translation, Interpreting and Cognition, Faculty of Arts and Humanities, Macao, China
| | - Zhen Yuan
- University of Macau, Bioimaging Core, Faculty of Health Sciences, Macao, China
| |
Collapse
|
21
|
Zhang M, Pratt SR, Doyle PJ, McNeil MR, Durrant JD, Roxberg J, Ortmann A. Audiological Assessment of Word Recognition Skills in Persons With Aphasia. Am J Audiol 2018; 27:1-18. [PMID: 29222555 DOI: 10.1044/2017_aja-17-0041] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2017] [Accepted: 08/01/2017] [Indexed: 11/09/2022] Open
Abstract
PURPOSE The purpose of this study was to evaluate the ability of persons with aphasia, with and without hearing loss, to complete a commonly used open-set word recognition test that requires a verbal response. Furthermore, phonotactic probabilities and neighborhood densities of word recognition errors were assessed to explore potential underlying linguistic complexities that might differentially influence performance among groups. METHOD Four groups of adult participants were tested: participants with no brain injury with normal hearing, participants with no brain injury with hearing loss, participants with brain injury with aphasia and normal hearing, and participants with brain injury with aphasia and hearing loss. The Northwestern University Auditory Test No. 6 (NU-6; Tillman & Carhart, 1966) was administered. Those participants who were unable to respond orally (repeating words as heard) were assessed with the Picture Identification Task (Wilson & Antablin, 1980), permitting a picture-pointing response instead. Error patterns from the NU-6 were assessed to determine whether phonotactic probability influenced performance. RESULTS All participants with no brain injury and 72.7% of the participants with aphasia (24 out of 33) completed the NU-6. Furthermore, all participants who were unable to complete the NU-6 were able to complete the Picture Identification Task. There were significant group differences on NU-6 performance. The 2 groups with normal hearing had significantly higher scores than the 2 groups with hearing loss, but the 2 groups with normal hearing and the 2 groups with hearing loss did not differ from one another, implying that their performance was largely determined by hearing loss rather than by brain injury or aphasia. The neighborhood density, but not phonotactic probabilities, of the participants' errors differed across groups with and without aphasia. CONCLUSIONS Because the vast majority of the participants with aphasia examined could be tested readily using an instrument such as the NU-6, clinicians should not be reticent to use this test if patients are able to repeat single words, but routine use of alternative tests is encouraged for populations of people with brain injuries.
Collapse
Affiliation(s)
- Min Zhang
- Geriatric Research, Education, and Clinical Center, VA Pittsburgh Healthcare System, PA
- Department of Communication Science and Disorders, University of Pittsburgh, PA
| | - Sheila R. Pratt
- Geriatric Research, Education, and Clinical Center, VA Pittsburgh Healthcare System, PA
- Department of Communication Science and Disorders, University of Pittsburgh, PA
| | - Patrick J. Doyle
- Geriatric Research, Education, and Clinical Center, VA Pittsburgh Healthcare System, PA
- Department of Communication Science and Disorders, University of Pittsburgh, PA
| | - Malcolm R. McNeil
- Geriatric Research, Education, and Clinical Center, VA Pittsburgh Healthcare System, PA
- Department of Communication Science and Disorders, University of Pittsburgh, PA
| | - John D. Durrant
- Geriatric Research, Education, and Clinical Center, VA Pittsburgh Healthcare System, PA
- Department of Communication Science and Disorders, University of Pittsburgh, PA
| | - Jillyn Roxberg
- Geriatric Research, Education, and Clinical Center, VA Pittsburgh Healthcare System, PA
| | - Amanda Ortmann
- Geriatric Research, Education, and Clinical Center, VA Pittsburgh Healthcare System, PA
- Department of Communication Science and Disorders, University of Pittsburgh, PA
| |
Collapse
|
22
|
Rampinini AC, Handjaras G, Leo A, Cecchetti L, Ricciardi E, Marotta G, Pietrini P. Functional and spatial segregation within the inferior frontal and superior temporal cortices during listening, articulation imagery, and production of vowels. Sci Rep 2017; 7:17029. [PMID: 29208951 PMCID: PMC5717247 DOI: 10.1038/s41598-017-17314-0] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2017] [Accepted: 11/24/2017] [Indexed: 11/09/2022] Open
Abstract
Classical models of language localize speech perception in the left superior temporal and production in the inferior frontal cortex. Nonetheless, neuropsychological, structural and functional studies have questioned such subdivision, suggesting an interwoven organization of the speech function within these cortices. We tested whether sub-regions within frontal and temporal speech-related areas retain specific phonological representations during both perception and production. Using functional magnetic resonance imaging and multivoxel pattern analysis, we showed functional and spatial segregation across the left fronto-temporal cortex during listening, imagery and production of vowels. In accordance with classical models of language and evidence from functional studies, the inferior frontal and superior temporal cortices discriminated among perceived and produced vowels respectively, also engaging in the non-classical, alternative function - i.e. perception in the inferior frontal and production in the superior temporal cortex. Crucially, though, contiguous and non-overlapping sub-regions within these hubs performed either the classical or non-classical function, the latter also representing non-linguistic sounds (i.e., pure tones). Extending previous results and in line with integration theories, our findings not only demonstrate that sensitivity to speech listening exists in production-related regions and vice versa, but they also suggest that the nature of such interwoven organisation is built upon low-level perception.
Collapse
Affiliation(s)
| | | | - Andrea Leo
- IMT School for Advanced Studies, Lucca, 55100, Italy
| | | | | | - Giovanna Marotta
- Department of Philology, Literature and Linguistics, University of Pisa, Pisa, 56100, Italy
| | | |
Collapse
|
23
|
Echoes on the motor network: how internal motor control structures afford sensory experience. Brain Struct Funct 2017; 222:3865-3888. [DOI: 10.1007/s00429-017-1484-1] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2017] [Accepted: 07/25/2017] [Indexed: 01/10/2023]
|
24
|
Rogers JC, Davis MH. Inferior Frontal Cortex Contributions to the Recognition of Spoken Words and Their Constituent Speech Sounds. J Cogn Neurosci 2017; 29:919-936. [PMID: 28129061 DOI: 10.1162/jocn_a_01096] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/06/2023]
Abstract
Speech perception and comprehension are often challenged by the need to recognize speech sounds that are degraded or ambiguous. Here, we explore the cognitive and neural mechanisms involved in resolving ambiguity in the identity of speech sounds using syllables that contain ambiguous phonetic segments (e.g., intermediate sounds between /b/ and /g/ as in "blade" and "glade"). We used an audio-morphing procedure to create a large set of natural sounding minimal pairs that contain phonetically ambiguous onset or offset consonants (differing in place, manner, or voicing). These ambiguous segments occurred in different lexical contexts (i.e., in words or pseudowords, such as blade-glade or blem-glem) and in different phonological environments (i.e., with neighboring syllables that differed in lexical status, such as blouse-glouse). These stimuli allowed us to explore the impact of phonetic ambiguity on the speed and accuracy of lexical decision responses (Experiment 1), semantic categorization responses (Experiment 2), and the magnitude of BOLD fMRI responses during attentive comprehension (Experiment 3). For both behavioral and neural measures, observed effects of phonetic ambiguity were influenced by lexical context leading to slower responses and increased activity in the left inferior frontal gyrus for high-ambiguity syllables that distinguish pairs of words, but not for equivalent pseudowords. These findings suggest lexical involvement in the resolution of phonetic ambiguity. Implications for speech perception and the role of inferior frontal regions are discussed.
Collapse
Affiliation(s)
- Jack C Rogers
- MRC Cognition & Brain Sciences Unit, Cambridge, UK.,University of Birmingham
| | | |
Collapse
|
25
|
Skipper JI, Devlin JT, Lametti DR. The hearing ear is always found close to the speaking tongue: Review of the role of the motor system in speech perception. BRAIN AND LANGUAGE 2017; 164:77-105. [PMID: 27821280 DOI: 10.1016/j.bandl.2016.10.004] [Citation(s) in RCA: 123] [Impact Index Per Article: 17.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/17/2016] [Accepted: 10/24/2016] [Indexed: 06/06/2023]
Abstract
Does "the motor system" play "a role" in speech perception? If so, where, how, and when? We conducted a systematic review that addresses these questions using both qualitative and quantitative methods. The qualitative review of behavioural, computational modelling, non-human animal, brain damage/disorder, electrical stimulation/recording, and neuroimaging research suggests that distributed brain regions involved in producing speech play specific, dynamic, and contextually determined roles in speech perception. The quantitative review employed region and network based neuroimaging meta-analyses and a novel text mining method to describe relative contributions of nodes in distributed brain networks. Supporting the qualitative review, results show a specific functional correspondence between regions involved in non-linguistic movement of the articulators, covertly and overtly producing speech, and the perception of both nonword and word sounds. This distributed set of cortical and subcortical speech production regions are ubiquitously active and form multiple networks whose topologies dynamically change with listening context. Results are inconsistent with motor and acoustic only models of speech perception and classical and contemporary dual-stream models of the organization of language and the brain. Instead, results are more consistent with complex network models in which multiple speech production related networks and subnetworks dynamically self-organize to constrain interpretation of indeterminant acoustic patterns as listening context requires.
Collapse
Affiliation(s)
- Jeremy I Skipper
- Experimental Psychology, University College London, United Kingdom.
| | - Joseph T Devlin
- Experimental Psychology, University College London, United Kingdom
| | - Daniel R Lametti
- Experimental Psychology, University College London, United Kingdom; Department of Experimental Psychology, University of Oxford, United Kingdom
| |
Collapse
|
26
|
Rosenblum LD, Dorsi J, Dias JW. The Impact and Status of Carol Fowler's Supramodal Theory of Multisensory Speech Perception. ECOLOGICAL PSYCHOLOGY 2016. [DOI: 10.1080/10407413.2016.1230373] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/31/2022]
|
27
|
Zhu L, Niu Z, Nie Y, Yang Y, Li K, Jin Z, Wei J. The Brain Effective Connectivity of Chinese during Rhyming Task. PLoS One 2016; 11:e0162158. [PMID: 27583349 PMCID: PMC5008726 DOI: 10.1371/journal.pone.0162158] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2015] [Accepted: 08/18/2016] [Indexed: 11/21/2022] Open
Abstract
With regard to brain language processing, the activation patterns have been well studied, and recently there are great interest in the connectivity models. The crucial brain areas for phonological processing involves left inferior frontal gyrus (LIFG), left inferior parietal lobule (LIPL) and left posterior middle temporal gyrus (LpMTG). Specially in Chinese processing, the left middle frontal gyrus (LMFG) is considered as an essential region. However, the connectivity pattern among these brain areas is not well understood. In this study, a rhyming experiment of Chinese was conducted, and the Dynamic causal modeling (DCM) and the Bayesian model selection (BMS) were used to examine the interaction between brain regions and choose the best model for rhyming task of Chinese. By examining the interactions, it was found that LMFG exerted inhibitory modulation on LIPL and LIFG; the phonological processing enhanced the connection from LIPL to LIFG and LMFG, which suggested the important roles of these connections for the increased phonological load; And LpMTG modulated LIFG and LMFG negatively, and LIPL positively under rhyming judgment task.
Collapse
Affiliation(s)
- Linlin Zhu
- School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China
| | - Zhendong Niu
- School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China
- * E-mail:
| | - Yaoxin Nie
- School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China
| | - Yang Yang
- Department of Linguistics, University of Hong Kong, Hong Kong
| | - Ke Li
- The 306th Hospital of PLA, Beijing, China
| | - Zhen Jin
- The 306th Hospital of PLA, Beijing, China
| | - Jieyao Wei
- School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China
| |
Collapse
|
28
|
Abstract
How are the meanings of words, events, and objects represented and organized in the brain? This question, perhaps more than any other in the field, probes some of the deepest and most foundational puzzles regarding the structure of the mind and brain. Accordingly, it has spawned a field of inquiry that is diverse and multidisciplinary, has led to the discovery of numerous empirical phenomena, and has spurred the development of a wide range of theoretical positions. This special issue brings together the most recent theoretical developments from the leaders in the field, representing a range of viewpoints on issues of fundamental significance to a theory of meaning representation. Here we introduce the special issue by way of pulling out some key themes that cut across the contributions that form this issue and situating those themes in the broader literature. The core issues around which research on conceptual representation can be organized are representational format, representational content, the organization of concepts in the brain, and the processing dynamics that govern interactions between the conceptual system and sensorimotor representations. We highlight areas in which consensus has formed; for those areas in which opinion is divided, we seek to clarify the relation of theory and evidence and to set in relief the bridging assumptions that undergird current discussions.
Collapse
Affiliation(s)
- Bradford Z Mahon
- Department of Brain and Cognitive Sciences, University of Rochester, Meliora Hall, Rochester, NY, 14627-0268, USA.
- Department of Neurosurgery, University of Rochester, Rochester, NY, USA.
- Center for Visual Science, University of Rochester, Rochester, NY, USA.
- Center for Language Sciences, University of Rochester, Rochester, NY, USA.
| | - Gregory Hickok
- Department of Cognitive Sciences, University of California, Irvine, CA, USA
| |
Collapse
|
29
|
Okada K, Rogalsky C, O'Grady L, Hanaumi L, Bellugi U, Corina D, Hickok G. An fMRI study of perception and action in deaf signers. Neuropsychologia 2016; 82:179-188. [PMID: 26796716 DOI: 10.1016/j.neuropsychologia.2016.01.015] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/19/2015] [Revised: 01/11/2016] [Accepted: 01/13/2016] [Indexed: 11/30/2022]
Abstract
Since the discovery of mirror neurons, there has been a great deal of interest in understanding the relationship between perception and action, and the role of the human mirror system in language comprehension and production. Two questions have dominated research. One concerns the role of Broca's area in speech perception. The other concerns the role of the motor system more broadly in understanding action-related language. The current study investigates both of these questions in a way that bridges research on language with research on manual actions. We studied the neural basis of observing and executing American Sign Language (ASL) object and action signs. In an fMRI experiment, deaf signers produced signs depicting actions and objects as well as observed/comprehended signs of actions and objects. Different patterns of activation were found for observation and execution although with overlap in Broca's area, providing prima facie support for the claim that the motor system participates in language perception. In contrast, we found no evidence that action related signs differentially involved the motor system compared to object related signs. These findings are discussed in the context of lesion studies of sign language execution and observation. In this broader context, we conclude that the activation in Broca's area during ASL observation is not causally related to sign language understanding.
Collapse
Affiliation(s)
- Kayoko Okada
- Department of Psychological Sciences, Whittier College, Whittier, CA, United states; Department of Cognitive Sciences, University of California, Irvine, CA, United States
| | - Corianne Rogalsky
- Department of Speech and Hearing Science, Arizona State University, Tempe, AZ, United States
| | - Lucinda O'Grady
- Laboratory for Cognitive Neuroscience, The Salk Institute for Biological Studies, San Diego, CA, United States
| | - Leila Hanaumi
- Laboratory for Cognitive Neuroscience, The Salk Institute for Biological Studies, San Diego, CA, United States
| | - Ursula Bellugi
- Laboratory for Cognitive Neuroscience, The Salk Institute for Biological Studies, San Diego, CA, United States
| | - David Corina
- Department of Linguistics, University of California, Davis, CA, United States
| | - Gregory Hickok
- Department of Cognitive Sciences, University of California, Irvine, CA, United States.
| |
Collapse
|
30
|
Smalle EHM, Rogers J, Möttönen R. Dissociating Contributions of the Motor Cortex to Speech Perception and Response Bias by Using Transcranial Magnetic Stimulation. Cereb Cortex 2015; 25:3690-8. [PMID: 25274987 PMCID: PMC4585509 DOI: 10.1093/cercor/bhu218] [Citation(s) in RCA: 35] [Impact Index Per Article: 3.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022] Open
Abstract
Recent studies using repetitive transcranial magnetic stimulation (TMS) have demonstrated that disruptions of the articulatory motor cortex impair performance in demanding speech perception tasks. These findings have been interpreted as support for the idea that the motor cortex is critically involved in speech perception. However, the validity of this interpretation has been called into question, because it is unknown whether the TMS-induced disruptions in the motor cortex affect speech perception or rather response bias. In the present TMS study, we addressed this question by using signal detection theory to calculate sensitivity (i.e., d') and response bias (i.e., criterion c). We used repetitive TMS to temporarily disrupt the lip or hand representation in the left motor cortex. Participants discriminated pairs of sounds from a "ba"-"da" continuum before TMS, immediately after TMS (i.e., during the period of motor disruption), and after a 30-min break. We found that the sensitivity for between-category pairs was reduced during the disruption of the lip representation. In contrast, disruption of the hand representation temporarily reduced response bias. This double dissociation indicates that the hand motor cortex contributes to response bias during demanding discrimination tasks, whereas the articulatory motor cortex contributes to perception of speech sounds.
Collapse
Affiliation(s)
- Eleonore H. M. Smalle
- Department of Experimental Psychology, University of Oxford, Oxford OX1 3UD, UK
- Psychological Sciences Research Institute, Institute of Neuroscience, Université Catholique de Louvain, B-1348 Louvain-la-Neuve, Belgium
| | - Jack Rogers
- Department of Experimental Psychology, University of Oxford, Oxford OX1 3UD, UK
| | - Riikka Möttönen
- Department of Experimental Psychology, University of Oxford, Oxford OX1 3UD, UK
| |
Collapse
|
31
|
Stasenko A, Bonn C, Teghipco A, Garcea FE, Sweet C, Dombovy M, McDonough J, Mahon BZ. A causal test of the motor theory of speech perception: a case of impaired speech production and spared speech perception. Cogn Neuropsychol 2015; 32:38-57. [PMID: 25951749 DOI: 10.1080/02643294.2015.1035702] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Abstract
The debate about the causal role of the motor system in speech perception has been reignited by demonstrations that motor processes are engaged during the processing of speech sounds. Here, we evaluate which aspects of auditory speech processing are affected, and which are not, in a stroke patient with dysfunction of the speech motor system. We found that the patient showed a normal phonemic categorical boundary when discriminating two non-words that differ by a minimal pair (e.g., ADA-AGA). However, using the same stimuli, the patient was unable to identify or label the non-word stimuli (using a button-press response). A control task showed that he could identify speech sounds by speaker gender, ruling out a general labelling impairment. These data suggest that while the motor system is not causally involved in perception of the speech signal, it may be used when other cues (e.g., meaning, context) are not available.
Collapse
Affiliation(s)
- Alena Stasenko
- a Department of Brain & Cognitive Sciences , University of Rochester , Rochester , NY , USA
| | | | | | | | | | | | | | | |
Collapse
|
32
|
Woollams AM. Lexical is as lexical does: computational approaches to lexical representation. LANGUAGE, COGNITION AND NEUROSCIENCE 2015; 30:395-408. [PMID: 25893204 PMCID: PMC4396497 DOI: 10.1080/23273798.2015.1005637] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
In much of neuroimaging and neuropsychology, regions of the brain have been associated with 'lexical representation', with little consideration as to what this cognitive construct actually denotes. Within current computational models of word recognition, there are a number of different approaches to the representation of lexical knowledge. Structural lexical representations, found in original theories of word recognition, have been instantiated in modern localist models. However, such a representational scheme lacks neural plausibility in terms of economy and flexibility. Connectionist models have therefore adopted distributed representations of form and meaning. Semantic representations in connectionist models necessarily encode lexical knowledge. Yet when equipped with recurrent connections, connectionist models can also develop attractors for familiar forms that function as lexical representations. Current behavioural, neuropsychological and neuroimaging evidence shows a clear role for semantic information, but also suggests some modality- and task-specific lexical representations. A variety of connectionist architectures could implement these distributed functional representations, and further experimental and simulation work is required to discriminate between these alternatives. Future conceptualisations of lexical representations will therefore emerge from a synergy between modelling and neuroscience.
Collapse
Affiliation(s)
- Anna M. Woollams
- Neuroscience and Aphasia Research Unit, School of Psychological Sciences, University of Manchester, Zochonis Building, Brunswick Street, ManchesterM13 9PL, UK
| |
Collapse
|
33
|
Myers EB. Emergence of category-level sensitivities in non-native speech sound learning. Front Neurosci 2014; 8:238. [PMID: 25152708 PMCID: PMC4125857 DOI: 10.3389/fnins.2014.00238] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2014] [Accepted: 07/20/2014] [Indexed: 11/23/2022] Open
Abstract
Over the course of development, speech sounds that are contrastive in one's native language tend to become perceived categorically: that is, listeners are unaware of variation within phonetic categories while showing excellent sensitivity to speech sounds that span linguistically meaningful phonetic category boundaries. The end stage of this developmental process is that the perceptual systems that handle acoustic-phonetic information show special tuning to native language contrasts, and as such, category-level information appears to be present at even fairly low levels of the neural processing stream. Research on adults acquiring non-native speech categories offers an avenue for investigating the interplay of category-level information and perceptual sensitivities to these sounds as speech categories emerge. In particular, one can observe the neural changes that unfold as listeners learn not only to perceive acoustic distinctions that mark non-native speech sound contrasts, but also to map these distinctions onto category-level representations. An emergent literature on the neural basis of novel and non-native speech sound learning offers new insight into this question. In this review, I will examine this literature in order to answer two key questions. First, where in the neural pathway does sensitivity to category-level phonetic information first emerge over the trajectory of speech sound learning? Second, how do frontal and temporal brain areas work in concert over the course of non-native speech sound learning? Finally, in the context of this literature I will describe a model of speech sound learning in which rapidly-adapting access to categorical information in the frontal lobes modulates the sensitivity of stable, slowly-adapting responses in the temporal lobes.
Collapse
Affiliation(s)
- Emily B Myers
- Department of Speech, Language, and Hearing Sciences, University of Connecticut Storrs, CT, USA ; Department of Psychology, University of Connecticut Storrs, CT, USA ; Haskins Laboratories New Haven, CT, USA
| |
Collapse
|
34
|
Hickok G. The architecture of speech production and the role of the phoneme in speech processing. LANGUAGE AND COGNITIVE PROCESSES 2014; 29:2-20. [PMID: 24489420 PMCID: PMC3904400 DOI: 10.1080/01690965.2013.834370] [Citation(s) in RCA: 54] [Impact Index Per Article: 5.4] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
Speech production has been studied within a number of traditions including linguistics, psycholinguistics, motor control, neuropsychology, and neuroscience. These traditions have had limited interaction, ostensibly because they target different levels of speech production or different dimensions such as representation, processing, or implementation. However, closer examination of reveals a substantial convergence of ideas across the traditions and recent proposals have suggested that an integrated approach may help move the field forward. The present article reviews one such attempt at integration, the state feedback control model and its descendent, the hierarchical state feedback control model. Also considered is how phoneme-level representations might fit in the context of the model.
Collapse
Affiliation(s)
- Gregory Hickok
- Department of Cognitive Sciences, University of California, Irvine, California, 92697, USA
| |
Collapse
|
35
|
Stasenko A, Garcea FE, Mahon BZ. What happens to the motor theory of perception when the motor system is damaged? LANGUAGE AND COGNITION 2013; 5:225-238. [PMID: 26823687 PMCID: PMC4727246 DOI: 10.1515/langcog-2013-0016] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Motor theories of perception posit that motor information is necessary for successful recognition of actions. Perhaps the most well known of this class of proposals is the motor theory of speech perception, which argues that speech recognition is fundamentally a process of identifying the articulatory gestures (i.e. motor representations) that were used to produce the speech signal. Here we review neuropsychological evidence from patients with damage to the motor system, in the context of motor theories of perception applied to both manual actions and speech. Motor theories of perception predict that patients with motor impairments will have impairments for action recognition. Contrary to that prediction, the available neuropsychological evidence indicates that recognition can be spared despite profound impairments to production. These data falsify strong forms of the motor theory of perception, and frame new questions about the dynamical interactions that govern how information is exchanged between input and output systems.
Collapse
Affiliation(s)
- Alena Stasenko
- Department of Brain & Cognitive Sciences, University of Rochester, USA
| | - Frank E. Garcea
- Department of Brain & Cognitive Sciences, University of Rochester, USA
| | - Bradford Z. Mahon
- Meliora Hall, University of Rochester, Rochester, NY 14627-0268, USA; Department of Brain & Cognitive Sciences, University of Rochester, USA; Department of Neurosurgery, University of Rochester Medical Center, USA; Center for Language Sciences, University of Rochester, USA
| |
Collapse
|
36
|
Emmorey K. The neurobiology of sign language and the mirror system hypothesis. LANGUAGE AND COGNITION 2013; 5:205-210. [PMID: 24707322 PMCID: PMC3972212 DOI: 10.1515/langcog-2013-0014] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
I suggest two puzzles for the Mirror System Hypothesis. First, there is little evidence that mirror neuron populations for words or for signs exist in Broca's area, and a mirror system is not critical for either speech or sign perception. Damage to Broca's area (or to the mirror system for human action) does not result in deficits in sign or speech perception. Second, the gesticulations of speakers are highly integrated with speech, but pantomimes and modern protosigns (conventional gestures) are not co-expressive with speech, and they do not co-occur with speech. Further, signers also produce global, imagistic gesticulations with their mouths and bodies simultaneously while signing with their hands. The expanding spiral of protosign and protospeech does not predict the integrated and co-expressive nature of modern gestures produced by signers and speakers.
Collapse
Affiliation(s)
- Karen Emmorey
- Lab for Language and Cognitive Neuroscience, 6495 Alvarado Road, Suite 200, San Diego, CA 92120, USA.
| |
Collapse
|
37
|
Grabski K, Tremblay P, Gracco VL, Girin L, Sato M. A mediating role of the auditory dorsal pathway in selective adaptation to speech: a state-dependent transcranial magnetic stimulation study. Brain Res 2013; 1515:55-65. [PMID: 23542585 DOI: 10.1016/j.brainres.2013.03.024] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2012] [Revised: 03/17/2013] [Accepted: 03/21/2013] [Indexed: 11/30/2022]
Abstract
In addition to sensory processing, recent neurobiological models of speech perception postulate the existence of a left auditory dorsal processing stream, linking auditory speech representations in the auditory cortex with articulatory representations in the motor system, through sensorimotor interaction interfaced in the supramarginal gyrus and/or the posterior part of the superior temporal gyrus. The present state-dependent transcranial magnetic stimulation study is aimed at determining whether speech recognition is indeed mediated by the auditory dorsal pathway, by examining the causal contribution of the left ventral premotor cortex, supramarginal gyrus and posterior part of the superior temporal gyrus during an auditory syllable identification/categorization task. To this aim, participants listened to a sequence of /ba/ syllables before undergoing a two forced-choice auditory syllable decision task on ambiguous syllables (ranging in the categorical boundary between /ba/ and /da/). Consistent with previous studies on selective adaptation to speech, following adaptation to /ba/, participants responses were biased towards /da/. In contrast, in a control condition without prior auditory adaptation no such bias was observed. Crucially, compared to the results observed without stimulation, single-pulse transcranial magnetic stimulation delivered at the onset of each target stimulus interacted with the initial state of each of the stimulated brain area by enhancing the adaptation effect. These results demonstrate that the auditory dorsal pathway contribute to auditory speech adaptation.
Collapse
Affiliation(s)
- Krystyna Grabski
- GIPSA-lab, Département Parole & Cognition, CNRS & Grenoble Université, France.
| | | | | | | | | |
Collapse
|
38
|
Nuzzo R. Why tongue twisters are hard to say. Nature 2013. [DOI: 10.1038/nature.2013.12471] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
39
|
Action-verb processing in Parkinson’s disease: new pathways for motor–language coupling. Brain Struct Funct 2013; 218:1355-73. [DOI: 10.1007/s00429-013-0510-1] [Citation(s) in RCA: 84] [Impact Index Per Article: 7.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2012] [Accepted: 01/23/2013] [Indexed: 10/27/2022]
|
40
|
Rogalsky C, Raphel K, Tomkovicz V, O'Grady L, Damasio H, Bellugi U, Hickok G. Neural Basis of Action Understanding: Evidence from Sign Language Aphasia. APHASIOLOGY 2013; 27:1147-1158. [PMID: 24031116 PMCID: PMC3767459 DOI: 10.1080/02687038.2013.812779] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
BACKGROUND The neural basis of action understanding is a hotly debated issue. The mirror neuron account holds that motor simulation in fronto-parietal circuits is critical to action understanding including speech comprehension, while others emphasize the ventral stream in the temporal lobe. Evidence from speech strongly supports the ventral stream account, but on the other hand, evidence from manual gesture comprehension (e.g., in limb apraxia) has led to contradictory findings. AIMS Here we present a lesion analysis of sign language comprehension. Sign language is an excellent model for studying mirror system function in that it bridges the gap between the visual-manual system in which mirror neurons are best characterized and language systems which have represented a theoretical target of mirror neuron research. METHODS & PROCEDURES Twenty-one life long deaf signers with focal cortical lesions performed two tasks: one involving the comprehension of individual signs and the other involving comprehension of signed sentences (commands). Participants' lesions, as indicated on MRI or CT scans, were mapped onto a template brain to explore the relationship between lesion location and sign comprehension measures. OUTCOMES & RESULTS Single sign comprehension was not significantly affected by left hemisphere damage. Sentence sign comprehension impairments were associated with left temporal-parietal damage. We found that damage to mirror system related regions in the left frontal lobe were not associated with deficits on either of these comprehension tasks. CONCLUSIONS We conclude that the mirror system is not critically involved in action understanding.
Collapse
Affiliation(s)
- Corianne Rogalsky
- Department of Cognitive Sciences, University of California, Irvine, Irvine, CA 92697
| | | | | | | | | | | | | |
Collapse
|
41
|
Hickok G. The cortical organization of speech processing: feedback control and predictive coding the context of a dual-stream model. JOURNAL OF COMMUNICATION DISORDERS 2012; 45:393-402. [PMID: 22766458 PMCID: PMC3468690 DOI: 10.1016/j.jcomdis.2012.06.004] [Citation(s) in RCA: 148] [Impact Index Per Article: 12.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2023]
Abstract
Speech recognition is an active process that involves some form of predictive coding. This statement is relatively uncontroversial. What is less clear is the source of the prediction. The dual-stream model of speech processing suggests that there are two possible sources of predictive coding in speech perception: the motor speech system and the lexical-conceptual system. Here I provide an overview of the dual-stream model of speech processing and then discuss evidence concerning the source of predictive coding during speech recognition. I conclude that, in contrast to recent theoretical trends, the dorsal sensory-motor stream is not a source of forward prediction that can facilitate speech recognition. Rather, it is forward prediction coming out of the ventral stream that serves this function.
Collapse
Affiliation(s)
- Gregory Hickok
- Department of Cognitive Sciences, University of California, Irvine, Irvine, CA 92697, United States.
| |
Collapse
|
42
|
Venezia JH, Saberi K, Chubb C, Hickok G. Response Bias Modulates the Speech Motor System during Syllable Discrimination. Front Psychol 2012; 3:157. [PMID: 22723787 PMCID: PMC3361017 DOI: 10.3389/fpsyg.2012.00157] [Citation(s) in RCA: 46] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2011] [Accepted: 04/30/2012] [Indexed: 11/16/2022] Open
Abstract
Recent evidence suggests that the speech motor system may play a significant role in speech perception. Repetitive transcranial magnetic stimulation (TMS) applied to a speech region of premotor cortex impaired syllable identification, while stimulation of motor areas for different articulators selectively facilitated identification of phonemes relying on those articulators. However, in these experiments performance was not corrected for response bias. It is not currently known how response bias modulates activity in these networks. The present functional magnetic resonance imaging experiment was designed to produce specific, measureable changes in response bias in a speech perception task. Minimal consonant-vowel stimulus pairs were presented between volume acquisitions for same-different discrimination. Speech stimuli were embedded in Gaussian noise at the psychophysically determined threshold level. We manipulated bias by changing the ratio of same-to-different trials: 1:3, 1:2, 1:1, 2:1, 3:1. Ratios were blocked by run and subjects were cued to the upcoming ratio at the beginning of each run. The stimuli were physically identical across runs. Response bias (criterion, C) was measured in individual subjects for each ratio condition. Group mean bias varied in the expected direction. We predicted that activation in frontal but not temporal brain regions would co-vary with bias. Group-level regression of bias scores on percent signal change revealed a fronto-parietal network of motor and sensory-motor brain regions that were sensitive to changes in response bias. We identified several pre- and post-central clusters in the left hemisphere that overlap well with TMS targets from the aforementioned studies. Importantly, activity in these regions covaried with response bias even while the perceptual targets remained constant. Thus, previous results suggesting that speech motor cortex participates directly in the perceptual analysis of speech should be called into question.
Collapse
Affiliation(s)
- Jonathan Henry Venezia
- Auditory and Language Neuroscience Laboratory, Center for Language Sciences, Center for Cognitive Neuroscience, Department of Cognitive Sciences, University of California Irvine Irvine, CA, USA
| | | | | | | |
Collapse
|
43
|
Baum SH, Martin RC, Hamilton AC, Beauchamp MS. Multisensory speech perception without the left superior temporal sulcus. Neuroimage 2012; 62:1825-32. [PMID: 22634292 DOI: 10.1016/j.neuroimage.2012.05.034] [Citation(s) in RCA: 28] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2012] [Revised: 05/10/2012] [Accepted: 05/14/2012] [Indexed: 10/28/2022] Open
Abstract
Converging evidence suggests that the left superior temporal sulcus (STS) is a critical site for multisensory integration of auditory and visual information during speech perception. We report a patient, SJ, who suffered a stroke that damaged the left tempo-parietal area, resulting in mild anomic aphasia. Structural MRI showed complete destruction of the left middle and posterior STS, as well as damage to adjacent areas in the temporal and parietal lobes. Surprisingly, SJ demonstrated preserved multisensory integration measured with two independent tests. First, she perceived the McGurk effect, an illusion that requires integration of auditory and visual speech. Second, her perception of morphed audiovisual speech with ambiguous auditory or visual information was significantly influenced by the opposing modality. To understand the neural basis for this preserved multisensory integration, blood-oxygen level dependent functional magnetic resonance imaging (BOLD fMRI) was used to examine brain responses to audiovisual speech in SJ and 23 healthy age-matched controls. In controls, bilateral STS activity was observed. In SJ, no activity was observed in the damaged left STS but in the right STS, more cortex was active in SJ than in any of the normal controls. Further, the amplitude of the BOLD response in right STS response to McGurk stimuli was significantly greater in SJ than in controls. The simplest explanation of these results is a reorganization of SJ's cortical language networks such that the right STS now subserves multisensory integration of speech.
Collapse
Affiliation(s)
- Sarah H Baum
- Department of Neurobiology and Anatomy, University of Texas Medical School at Houston, TX, USA
| | | | | | | |
Collapse
|
44
|
Abstract
Monkeys can easily form lasting central representations of visual and tactile stimuli, yet they seem unable to do the same with sounds. Humans, by contrast, are highly proficient in auditory long-term memory (LTM). These mnemonic differences within and between species raise the question of whether the human ability is supported in some way by speech and language, e.g., through subvocal reproduction of speech sounds and by covert verbal labeling of environmental stimuli. If so, the explanation could be that storing rapidly fluctuating acoustic signals requires assistance from the motor system, which is uniquely organized to chain-link rapid sequences. To test this hypothesis, we compared the ability of normal participants to recognize lists of stimuli that can be easily reproduced, labeled, or both (pseudowords, nonverbal sounds, and words, respectively) versus their ability to recognize a list of stimuli that can be reproduced or labeled only with great difficulty (reversed words, i.e., words played backward). Recognition scores after 5-min delays filled with articulatory-suppression tasks were relatively high (75-80% correct) for all sound types except reversed words; the latter yielded scores that were not far above chance (58% correct), even though these stimuli were discriminated nearly perfectly when presented as reversed-word pairs at short intrapair intervals. The combined results provide preliminary support for the hypothesis that participation of the oromotor system may be essential for laying down the memory of speech sounds and, indeed, that speech and auditory memory may be so critically dependent on each other that they had to coevolve.
Collapse
|
45
|
Aboitiz F. Gestures, vocalizations, and memory in language origins. FRONTIERS IN EVOLUTIONARY NEUROSCIENCE 2012; 4:2. [PMID: 22347184 PMCID: PMC3269654 DOI: 10.3389/fnevo.2012.00002] [Citation(s) in RCA: 48] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/28/2011] [Accepted: 01/11/2012] [Indexed: 01/09/2023]
Abstract
THIS ARTICLE DISCUSSES THE POSSIBLE HOMOLOGIES BETWEEN THE HUMAN LANGUAGE NETWORKS AND COMPARABLE AUDITORY PROJECTION SYSTEMS IN THE MACAQUE BRAIN, IN AN ATTEMPT TO RECONCILE TWO EXISTING VIEWS ON LANGUAGE EVOLUTION: one that emphasizes hand control and gestures, and the other that emphasizes auditory-vocal mechanisms. The capacity for language is based on relatively well defined neural substrates whose rudiments have been traced in the non-human primate brain. At its core, this circuit constitutes an auditory-vocal sensorimotor circuit with two main components, a "ventral pathway" connecting anterior auditory regions with anterior ventrolateral prefrontal areas, and a "dorsal pathway" connecting auditory areas with parietal areas and with posterior ventrolateral prefrontal areas via the arcuate fasciculus and the superior longitudinal fasciculus. In humans, the dorsal circuit is especially important for phonological processing and phonological working memory, capacities that are critical for language acquisition and for complex syntax processing. In the macaque, the homolog of the dorsal circuit overlaps with an inferior parietal-premotor network for hand and gesture selection that is under voluntary control, while vocalizations are largely fixed and involuntary. The recruitment of the dorsal component for vocalization behavior in the human lineage, together with a direct cortical control of the subcortical vocalizing system, are proposed to represent a fundamental innovation in human evolution, generating an inflection point that permitted the explosion of vocal language and human communication. In this context, vocal communication and gesturing have a common history in primate communication.
Collapse
Affiliation(s)
- Francisco Aboitiz
- Departamento de Psiquiatría, Facultad de Medicina y Centro Interdisciplinario de Neurociencia, Pontificia Universidad Católica de ChileSantiago, Chile
| |
Collapse
|