1
|
Nematova S, Zinszer B, Jasinska KK. Exploring audiovisual speech perception in monolingual and bilingual children in Uzbekistan. J Exp Child Psychol 2024; 239:105808. [PMID: 37972516 DOI: 10.1016/j.jecp.2023.105808] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2023] [Revised: 10/05/2023] [Accepted: 10/23/2023] [Indexed: 11/19/2023]
Abstract
This study aimed to investigate the development of audiovisual speech perception in monolingual Uzbek-speaking and bilingual Uzbek-Russian-speaking children, focusing on the impact of language experience on audiovisual speech perception and the role of visual phonetic (i.e., mouth movements corresponding to phonetic/lexical information) and temporal (i.e., timing of speech signals) cues. A total of 321 children aged 4 to 10 years in Tashkent, Uzbekistan, discriminated /ba/ and /da/ syllables across three conditions: auditory-only, audiovisual phonetic (i.e., sound accompanied by mouth movements), and audiovisual temporal (i.e., sound onset/offset accompanied by mouth opening/closing). Effects of modality (audiovisual phonetic, audiovisual temporal, or audio-only cues), age, group (monolingual or bilingual), and their interactions were tested using a Bayesian regression model. Overall, older participants performed better than younger participants. Participants performed better in the audiovisual phonetic modality compared with the auditory modality. However, no significant difference between monolingual and bilingual children was observed across all modalities. This finding stands in contrast to earlier studies. We attribute the contrasting findings of our study and the existing literature to the cross-linguistic similarity of the language pairs involved. When the languages spoken by bilinguals exhibit substantial linguistic similarity, there may be an increased necessity to disambiguate speech signals, leading to a greater reliance on audiovisual cues. The limited phonological similarity between Uzbek and Russian might have minimized bilinguals' need to rely on visual speech cues, contributing to the lack of group differences in our study.
Collapse
Affiliation(s)
- Shakhlo Nematova
- Department of Linguistics and Cognitive Science, University of Delaware, Newark, DE 19716, USA.
| | - Benjamin Zinszer
- Department of Psychology, Swarthmore College, Swarthmore, PA 19081, USA
| | - Kaja K Jasinska
- Department of Applied Psychology and Human Development, University of Toronto, Toronto, ON M5S 1A1, Canada
| |
Collapse
|
2
|
Sandhya, Vinay, V M. Perception of Incongruent Audiovisual Speech: Distribution of Modality-Specific Responses. Am J Audiol 2021; 30:968-979. [PMID: 34499528 DOI: 10.1044/2021_aja-20-00213] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022] Open
Abstract
PURPOSE Multimodal sensory integration in audiovisual (AV) speech perception is a naturally occurring phenomenon. Modality-specific responses such as auditory left, auditory right, and visual responses to dichotic incongruent AV speech stimuli help in understanding AV speech processing through each input modality. It is observed that distribution of activity in the frontal motor areas involved in speech production has been shown to correlate with how subjects perceive the same syllable differently or perceive different syllables. This study investigated the distribution of modality-specific responses to dichotic incongruent AV speech stimuli by simultaneously presenting consonant-vowel (CV) syllables with different places of articulation to the participant's left and right ears and visually. DESIGN A dichotic experimental design was adopted. Six stop CV syllables /pa/, /ta/, /ka/, /ba/, /da/, and /ga/ were assembled to create dichotic incongruent AV speech material. Participants included 40 native speakers of Norwegian (20 women, M age = 22.6 years, SD = 2.43 years; 20 men, M age = 23.7 years, SD = 2.08 years). RESULTS Findings of this study showed that, under dichotic listening conditions, velar CV syllables resulted in the highest scores in the respective ears, and this might be explained by stimulus dominance of velar consonants, as shown in previous studies. However, this study, with dichotic auditory stimuli accompanied by an incongruent video segment, demonstrated that the presentation of a visually distinct video segment possibly draws attention to the video segment in some participants, thereby reducing the overall recognition of the dominant syllable. Furthermore, the findings here suggest the possibility of lesser response times to incongruent AV stimuli in females compared with males. CONCLUSION The identification of the left audio, right audio, and visual segments in dichotic incongruent AV stimuli depends on place of articulation, stimulus dominance, and voice onset time of the CV syllables.
Collapse
Affiliation(s)
- Sandhya
- Department of Neuromedicine and Movement Science, Faculty of Medicine and Health Sciences, Norwegian University of Science and Technology, Trondheim, Norway
| | - Vinay
- Department of Neuromedicine and Movement Science, Faculty of Medicine and Health Sciences, Norwegian University of Science and Technology, Trondheim, Norway
| | - Manchaiah, V
- Department of Speech and Hearing Sciences, Lamar University, Beaumont, TX
| |
Collapse
|
3
|
Ujiie Y, Takahashi K. Own-race faces promote integrated audiovisual speech information. Q J Exp Psychol (Hove) 2021; 75:924-935. [PMID: 34427494 DOI: 10.1177/17470218211044480] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
The other-race effect indicates a perceptual advantage when processing own-race faces. This effect has been demonstrated in individuals' recognition of facial identity and emotional expressions. However, it remains unclear whether the other-race effect also exists in multisensory domains. We conducted two experiments to provide evidence for the other-race effect in facial speech recognition, using the McGurk effect. Experiment 1 tested this issue among East Asian adults, examining the magnitude of the McGurk effect during stimuli using speakers from two different races (own-race vs. other-race). We found that own-race faces induced a stronger McGurk effect than other-race faces. Experiment 2 indicated that the other-race effect was not simply due to different levels of attention being paid to the mouths of own- and other-race speakers. Our findings demonstrated that own-race faces enhance the weight of visual input during audiovisual speech perception, and they provide evidence of the own-race effect in the audiovisual interaction for speech perception in adults.
Collapse
Affiliation(s)
- Yuta Ujiie
- Research Organization of Open Innovation and Collaboration, Ritsumeikan University, Ibaraki, Japan.,Graduate School of Psychology, Chukyo University, Nagoya-shi, Japan.,Japan Society for the Promotion of Science, Tokyo, Japan.,Research and Development Initiative, Chuo University, Tokyo, Japan
| | - Kohske Takahashi
- School of Psychology, Chukyo University, Nagoya-shi, Japan.,College of Comprehensive Psychology, Ritsumeikan University, Ibaraki, Japan
| |
Collapse
|
4
|
The McGurk effect in the time of pandemic: Age-dependent adaptation to an environmental loss of visual speech cues. Psychon Bull Rev 2021; 28:992-1002. [PMID: 33443708 DOI: 10.3758/s13423-020-01852-2] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/19/2020] [Indexed: 11/08/2022]
Abstract
Seeing a person's mouth move for [ga] while hearing [ba] often results in the perception of "da." Such audiovisual integration of speech cues, known as the McGurk effect, is stable within but variable across individuals. When the visual or auditory cues are degraded, due to signal distortion or the perceiver's sensory impairment, reliance on cues via the impoverished modality decreases. This study tested whether cue-reliance adjustments due to exposure to reduced cue availability are persistent and transfer to subsequent perception of speech with all cues fully available. A McGurk experiment was administered at the beginning and after a month of mandatory face-mask wearing (enforced in Czechia during the 2020 pandemic). Responses to audio-visually incongruent stimuli were analyzed from 292 persons (ages 16-55), representing a cross-sectional sample, and 41 students (ages 19-27), representing a longitudinal sample. The extent to which the participants relied exclusively on visual cues was affected by testing time in interaction with age. After a month of reduced access to lipreading, reliance on visual cues (present at test) somewhat lowered for younger and increased for older persons. This implies that adults adapt their speech perception faculty to an altered environmental availability of multimodal cues, and that younger adults do so more efficiently. This finding demonstrates that besides sensory impairment or signal noise, which reduce cue availability and thus affect audio-visual cue reliance, having experienced a change in environmental conditions can modulate the perceiver's (otherwise relatively stable) general bias towards different modalities during speech communication.
Collapse
|
5
|
Brown VA, Hedayati M, Zanger A, Mayn S, Ray L, Dillman-Hasso N, Strand JF. What accounts for individual differences in susceptibility to the McGurk effect? PLoS One 2018; 13:e0207160. [PMID: 30418995 PMCID: PMC6231656 DOI: 10.1371/journal.pone.0207160] [Citation(s) in RCA: 25] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2018] [Accepted: 10/25/2018] [Indexed: 11/29/2022] Open
Abstract
The McGurk effect is a classic audiovisual speech illusion in which discrepant auditory and visual syllables can lead to a fused percept (e.g., an auditory /bɑ/ paired with a visual /gɑ/ often leads to the perception of /dɑ/). The McGurk effect is robust and easily replicated in pooled group data, but there is tremendous variability in the extent to which individual participants are susceptible to it. In some studies, the rate at which individuals report fusion responses ranges from 0% to 100%. Despite its widespread use in the audiovisual speech perception literature, the roots of the wide variability in McGurk susceptibility are largely unknown. This study evaluated whether several perceptual and cognitive traits are related to McGurk susceptibility through correlational analyses and mixed effects modeling. We found that an individual's susceptibility to the McGurk effect was related to their ability to extract place of articulation information from the visual signal (i.e., a more fine-grained analysis of lipreading ability), but not to scores on tasks measuring attentional control, processing speed, working memory capacity, or auditory perceptual gradiency. These results provide support for the claim that a small amount of the variability in susceptibility to the McGurk effect is attributable to lipreading skill. In contrast, cognitive and perceptual abilities that are commonly used predictors in individual differences studies do not appear to underlie susceptibility to the McGurk effect.
Collapse
Affiliation(s)
- Violet A. Brown
- Department of Psychology, Carleton College, Northfield, Minnesota, United States of America
| | - Maryam Hedayati
- Department of Psychology, Carleton College, Northfield, Minnesota, United States of America
| | - Annie Zanger
- Department of Psychology, Carleton College, Northfield, Minnesota, United States of America
| | - Sasha Mayn
- Department of Psychology, Carleton College, Northfield, Minnesota, United States of America
| | - Lucia Ray
- Department of Psychology, Carleton College, Northfield, Minnesota, United States of America
| | - Naseem Dillman-Hasso
- Department of Psychology, Carleton College, Northfield, Minnesota, United States of America
| | - Julia F. Strand
- Department of Psychology, Carleton College, Northfield, Minnesota, United States of America
| |
Collapse
|
6
|
Kubicek C, Gervain J, Lœvenbruck H, Pascalis O, Schwarzer G. Goldilocks versus Goldlöckchen: Visual speech preference for same-rhythm-class languages in 6-month-old infants. INFANT AND CHILD DEVELOPMENT 2018. [DOI: 10.1002/icd.2084] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Affiliation(s)
- Claudia Kubicek
- Department of Developmental Psychology; Justus Liebig University Giessen; Giessen Germany
| | - Judit Gervain
- CNRS, Université Paris Descartes; Sorbonne Paris Cité; Paris France
| | - Hélène Lœvenbruck
- Laboratoire de Psychologie et NeuroCognition, UMR CNRS 5105; Université Grenoble Alpes; Grenoble France
| | - Olivier Pascalis
- Laboratoire de Psychologie et NeuroCognition, UMR CNRS 5105; Université Grenoble Alpes; Grenoble France
| | - Gudrun Schwarzer
- Department of Developmental Psychology; Justus Liebig University Giessen; Giessen Germany
| |
Collapse
|
7
|
Alsius A, Paré M, Munhall KG. Forty Years After Hearing Lips and Seeing Voices: the McGurk Effect Revisited. Multisens Res 2018; 31:111-144. [PMID: 31264597 DOI: 10.1163/22134808-00002565] [Citation(s) in RCA: 52] [Impact Index Per Article: 8.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2016] [Accepted: 03/09/2017] [Indexed: 11/19/2022]
Abstract
Since its discovery 40 years ago, the McGurk illusion has been usually cited as a prototypical paradigmatic case of multisensory binding in humans, and has been extensively used in speech perception studies as a proxy measure for audiovisual integration mechanisms. Despite the well-established practice of using the McGurk illusion as a tool for studying the mechanisms underlying audiovisual speech integration, the magnitude of the illusion varies enormously across studies. Furthermore, the processing of McGurk stimuli differs from congruent audiovisual processing at both phenomenological and neural levels. This questions the suitability of this illusion as a tool to quantify the necessary and sufficient conditions under which audiovisual integration occurs in natural conditions. In this paper, we review some of the practical and theoretical issues related to the use of the McGurk illusion as an experimental paradigm. We believe that, without a richer understanding of the mechanisms involved in the processing of the McGurk effect, experimenters should be really cautious when generalizing data generated by McGurk stimuli to matching audiovisual speech events.
Collapse
Affiliation(s)
- Agnès Alsius
- Psychology Department, Queen's University, Humphrey Hall, 62 Arch St., Kingston, Ontario, K7L 3N6 Canada
| | - Martin Paré
- Psychology Department, Queen's University, Humphrey Hall, 62 Arch St., Kingston, Ontario, K7L 3N6 Canada
| | - Kevin G Munhall
- Psychology Department, Queen's University, Humphrey Hall, 62 Arch St., Kingston, Ontario, K7L 3N6 Canada
| |
Collapse
|
8
|
Irwin J, DiBlasi L. Audiovisual speech perception: A new approach and implications for clinical populations. LANGUAGE AND LINGUISTICS COMPASS 2017; 11:77-91. [PMID: 29520300 PMCID: PMC5839512 DOI: 10.1111/lnc3.12237] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/18/2015] [Accepted: 01/25/2017] [Indexed: 06/01/2023]
Abstract
This selected overview of audiovisual (AV) speech perception examines the influence of visible articulatory information on what is heard. Thought to be a cross-cultural phenomenon that emerges early in typical language development, variables that influence AV speech perception include properties of the visual and the auditory signal, attentional demands, and individual differences. A brief review of the existing neurobiological evidence on how visual information influences heard speech indicates potential loci, timing, and facilitatory effects of AV over auditory only speech. The current literature on AV speech in certain clinical populations (individuals with an autism spectrum disorder, developmental language disorder, or hearing loss) reveals differences in processing that may inform interventions. Finally, a new method of assessing AV speech that does not require obvious cross-category mismatch or auditory noise was presented as a novel approach for investigators.
Collapse
Affiliation(s)
- Julia Irwin
- LEARN Center, Haskins Laboratories Inc., USA
| | | |
Collapse
|
9
|
Abstract
Talkers automatically imitate aspects of perceived speech, a phenomenon known as phonetic convergence. Talkers have previously been found to converge to auditory and visual speech information. Furthermore, talkers converge more to the speech of a conversational partner who is seen and heard, relative to one who is just heard (Dias & Rosenblum Perception, 40, 1457-1466, 2011). A question raised by this finding is what visual information facilitates the enhancement effect. In the following experiments, we investigated the possible contributions of visible speech articulation to visual enhancement of phonetic convergence within the noninteractive context of a shadowing task. In Experiment 1, we examined the influence of the visibility of a talker on phonetic convergence when shadowing auditory speech either in the clear or in low-level auditory noise. The results suggest that visual speech can compensate for convergence that is reduced by auditory noise masking. Experiment 2 further established the visibility of articulatory mouth movements as being important to the visual enhancement of phonetic convergence. Furthermore, the word frequency and phonological neighborhood density characteristics of the words shadowed were found to significantly predict phonetic convergence in both experiments. Consistent with previous findings (e.g., Goldinger Psychological Review, 105, 251-279, 1998), phonetic convergence was greater when shadowing low-frequency words. Convergence was also found to be greater for low-density words, contrasting with previous predictions of the effect of phonological neighborhood density on auditory phonetic convergence (e.g., Pardo, Jordan, Mallari, Scanlon, & Lewandowski Journal of Memory and Language, 69, 183-195, 2013). Implications of the results for a gestural account of phonetic convergence are discussed.
Collapse
|
10
|
Variability and stability in the McGurk effect: contributions of participants, stimuli, time, and response type. Psychon Bull Rev 2016; 22:1299-307. [PMID: 25802068 DOI: 10.3758/s13423-015-0817-4] [Citation(s) in RCA: 89] [Impact Index Per Article: 11.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
In the McGurk effect, pairing incongruent auditory and visual syllables produces a percept different from the component syllables. Although it is a popular assay of audiovisual speech integration, little is known about the distribution of responses to the McGurk effect in the population. In our first experiment, we measured McGurk perception using 12 different McGurk stimuli in a sample of 165 English-speaking adults, 40 of whom were retested following a one-year interval. We observed dramatic differences both in how frequently different individuals perceived the illusion (from 0 % to 100 %) and in how frequently the illusion was perceived across different stimuli (17 % to 58 %). For individual stimuli, the distributions of response frequencies deviated strongly from normality, with 77 % of participants almost never or almost always perceiving the effect (≤10 % or ≥90 %). This deviation suggests that the mean response frequency, the most commonly reported measure of the McGurk effect, is a poor measure of individual participants' responses, and that the assumptions made by parametric statistical tests are invalid. Despite the substantial variability across individuals and stimuli, there was little change in the frequency of the effect between initial testing and a one-year retest (mean change in frequency = 2 %; test-retest correlation, r = 0.91). In a second experiment, we replicated our findings of high variability using eight new McGurk stimuli and tested the effects of open-choice versus forced-choice responding. Forced-choice responding resulted in an estimated 18 % greater frequency of the McGurk effect but similar levels of interindividual variability. Our results highlight the importance of examining individual differences in McGurk perception instead of relying on summary statistics averaged across a population. However, individual variability in the McGurk effect does not preclude its use as a stable measure of audiovisual integration.
Collapse
|
11
|
Alm M, Behne D. Do gender differences in audio-visual benefit and visual influence in audio-visual speech perception emerge with age? Front Psychol 2015; 6:1014. [PMID: 26236274 PMCID: PMC4503887 DOI: 10.3389/fpsyg.2015.01014] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/25/2014] [Accepted: 07/06/2015] [Indexed: 11/13/2022] Open
Abstract
Gender and age have been found to affect adults' audio-visual (AV) speech perception. However, research on adult aging focuses on adults over 60 years, who have an increasing likelihood for cognitive and sensory decline, which may confound positive effects of age-related AV-experience and its interaction with gender. Observed age and gender differences in AV speech perception may also depend on measurement sensitivity and AV task difficulty. Consequently both AV benefit and visual influence were used to measure visual contribution for gender-balanced groups of young (20-30 years) and middle-aged adults (50-60 years) with task difficulty varied using AV syllables from different talkers in alternative auditory backgrounds. Females had better speech-reading performance than males. Whereas no gender differences in AV benefit or visual influence were observed for young adults, visually influenced responses were significantly greater for middle-aged females than middle-aged males. That speech-reading performance did not influence AV benefit may be explained by visual speech extraction and AV integration constituting independent abilities. Contrastingly, the gender difference in visually influenced responses in middle adulthood may reflect an experience-related shift in females' general AV perceptual strategy. Although young females' speech-reading proficiency may not readily contribute to greater visual influence, between young and middle-adulthood recurrent confirmation of the contribution of visual cues induced by speech-reading proficiency may gradually shift females AV perceptual strategy toward more visually dominated responses.
Collapse
Affiliation(s)
- Magnus Alm
- Department of Psychology, Norwegian University of Science and Technology Trondheim, Norway
| | - Dawn Behne
- Department of Psychology, Norwegian University of Science and Technology Trondheim, Norway
| |
Collapse
|
12
|
Magnotti JF, Basu Mallick D, Feng G, Zhou B, Zhou W, Beauchamp MS. Similar frequency of the McGurk effect in large samples of native Mandarin Chinese and American English speakers. Exp Brain Res 2015; 233:2581-6. [PMID: 26041554 DOI: 10.1007/s00221-015-4324-7] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2014] [Accepted: 05/13/2015] [Indexed: 11/28/2022]
Abstract
Humans combine visual information from mouth movements with auditory information from the voice to recognize speech. A common method for assessing multisensory speech perception is the McGurk effect: When presented with particular pairings of incongruent auditory and visual speech syllables (e.g., the auditory speech sounds for "ba" dubbed onto the visual mouth movements for "ga"), individuals perceive a third syllable, distinct from the auditory and visual components. Chinese and American cultures differ in the prevalence of direct facial gaze and in the auditory structure of their languages, raising the possibility of cultural- and language-related group differences in the McGurk effect. There is no consensus in the literature about the existence of these group differences, with some studies reporting less McGurk effect in native Mandarin Chinese speakers than in English speakers and others reporting no difference. However, these studies sampled small numbers of participants tested with a small number of stimuli. Therefore, we collected data on the McGurk effect from large samples of Mandarin-speaking individuals from China and English-speaking individuals from the USA (total n = 307) viewing nine different stimuli. Averaged across participants and stimuli, we found similar frequencies of the McGurk effect between Chinese and American participants (48 vs. 44 %). In both groups, we observed a large range of frequencies both across participants (range from 0 to 100 %) and stimuli (15 to 83 %) with the main effect of culture and language accounting for only 0.3 % of the variance in the data. High individual variability in perception of the McGurk effect necessitates the use of large sample sizes to accurately estimate group differences.
Collapse
Affiliation(s)
- John F Magnotti
- Department of Neurosurgery, Baylor College of Medicine, 1 Baylor Plaza, Suite 104, Houston, TX, USA,
| | | | | | | | | | | |
Collapse
|
13
|
Ross LA, Del Bene VA, Molholm S, Frey HP, Foxe JJ. Sex differences in multisensory speech processing in both typically developing children and those on the autism spectrum. Front Neurosci 2015; 9:185. [PMID: 26074757 PMCID: PMC4445312 DOI: 10.3389/fnins.2015.00185] [Citation(s) in RCA: 38] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2015] [Accepted: 05/11/2015] [Indexed: 11/16/2022] Open
Abstract
BACKGROUND Previous work has revealed sizeable deficits in the abilities of children with an autism spectrum disorder (ASD) to integrate auditory and visual speech signals, with clear implications for social communication in this population. There is a strong male preponderance in ASD, with approximately four affected males for every female. The presence of sex differences in ASD symptoms suggests a sexual dimorphism in the ASD phenotype, and raises the question of whether this dimorphism extends to ASD traits in the neurotypical population. Here, we investigated possible sexual dimorphism in multisensory speech integration in both ASD and neurotypical individuals. METHODS We assessed whether males and females differed in their ability to benefit from visual speech when target words were presented under varying levels of signal-to-noise, in samples of neurotypical children and adults, and in children diagnosed with an ASD. RESULTS In typically developing (TD) children and children with ASD, females (n = 47 and n = 15, respectively) were significantly superior in their ability to recognize words under audiovisual listening conditions compared to males (n = 55 and n = 58, respectively). This sex difference was absent in our sample of neurotypical adults (n = 28 females; n = 28 males). CONCLUSIONS We propose that the development of audiovisual integration is delayed in male relative to female children, a delay that is also observed in ASD. In neurotypicals, these sex differences disappear in early adulthood when females approach their performance maximum and males "catch up." Our findings underline the importance of considering sex differences in the search for autism endophenotypes and strongly encourage increased efforts to study the underrepresented population of females within ASD.
Collapse
Affiliation(s)
- Lars A. Ross
- The Sheryl and Daniel R. Tishman Cognitive Neurophysiology Laboratory, Department of Pediatrics, Children's Evaluation and Rehabilitation Center, Albert Einstein College of Medicine & Montefiore Medical CenterBronx, NY, USA
- The Gordon F. Derner Institute of Advanced Psychological Studies, Adelphi UniversityGarden City, NY, USA
| | - Victor A. Del Bene
- The Sheryl and Daniel R. Tishman Cognitive Neurophysiology Laboratory, Department of Pediatrics, Children's Evaluation and Rehabilitation Center, Albert Einstein College of Medicine & Montefiore Medical CenterBronx, NY, USA
- Ferkauf Graduate School of Psychology, Albert Einstein College of MedicineBronx, NY, USA
| | - Sophie Molholm
- The Sheryl and Daniel R. Tishman Cognitive Neurophysiology Laboratory, Department of Pediatrics, Children's Evaluation and Rehabilitation Center, Albert Einstein College of Medicine & Montefiore Medical CenterBronx, NY, USA
- The Dominick P. Purpura Department of Neuroscience, Rose F. Kennedy Intellectual and Developmental Disabilities Research Center, Albert Einstein College of MedicineBronx, NY, USA
| | - Hans-Peter Frey
- The Sheryl and Daniel R. Tishman Cognitive Neurophysiology Laboratory, Department of Pediatrics, Children's Evaluation and Rehabilitation Center, Albert Einstein College of Medicine & Montefiore Medical CenterBronx, NY, USA
- Division of Neurocritical Care, Department of Neurology, Columbia University Medical CenterNew York, NY, USA
| | - John J. Foxe
- The Sheryl and Daniel R. Tishman Cognitive Neurophysiology Laboratory, Department of Pediatrics, Children's Evaluation and Rehabilitation Center, Albert Einstein College of Medicine & Montefiore Medical CenterBronx, NY, USA
- The Dominick P. Purpura Department of Neuroscience, Rose F. Kennedy Intellectual and Developmental Disabilities Research Center, Albert Einstein College of MedicineBronx, NY, USA
| |
Collapse
|
14
|
|
15
|
Oliveira LND, Soares AD, Chiari BM. Speechreading as a communication mediator. Codas 2014; 25:548-56. [DOI: 10.1590/s2317-17822014000100008] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2013] [Accepted: 10/28/2013] [Indexed: 11/22/2022] Open
Abstract
Purposes: To compare the speechreading between individuals with hearing impairment and with normal hearing levels to verify the factors that influence the speechreading among hearing impaired patients. Methods: Forty individuals with severe-to-profound hearing loss aged between 13 and 70 years old (study group) and 21 hearing individuals aged between 17 and 63 years old (control group) were evaluated. As a research instrument, anamnesis was used to characterize the groups; three speechreading instruments, presenting stimuli via a mute video, with a female speaker; and a vocabulary test, to verify their influence on speechreading. A descriptive and analytical statistics (ANOVA test and Pearson's correlation), adopting a significance level of 0.05 (5%). Results: A better performance was observed in the group with hearing impairment in speechreading tests than in the group with hearing individuals. By analyzing the group with hearing loss, there was a mean difference between tests (p<0.001), which also showed correlation between them. Individuals with pre-lingual hearing loss and those who underwent therapy for speechreading had a better performance for most speechreading instruments. The variables gender and schooling showed no influence on speechreading. Conclusion: Individuals with hearing impairment had better performance on speechreading tasks in comparison to people with normal hearing. Furthermore, it was found that the ability to perform speechread might be influenced by the vocabulary, period of installation of the hearing loss, and speechreading therapy.
Collapse
|
16
|
Hallam RS, Corney R. Conversation tactics in persons with normal hearing and hearing-impairment. Int J Audiol 2013; 53:174-81. [DOI: 10.3109/14992027.2013.852256] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
|
17
|
Irwin JR, Frost SJ, Mencl WE, Chen H, Fowler CA. Functional activation for imitation of seen and heard speech. JOURNAL OF NEUROLINGUISTICS 2011; 24:611-618. [PMID: 21966094 PMCID: PMC3182484 DOI: 10.1016/j.jneuroling.2011.05.001] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
This study examined fMRI activation when perceivers either passively observed or observed and imitated matched or mismatched audiovisual ("McGurk") speech stimuli. Greater activation was observed in the inferior frontal gyrus (IFG) overall for imitation than for perception of audiovisual speech and for imitation of the McGurk-type mismatched stimuli than matched audiovisual stimuli. This unique activation in the IFG during imitation of incongruent audiovisual speech may reflect activation associated with direct matching of incongruent auditory and visual stimuli or conflict between category responses. This study provides novel data about the underlying neurobiology of imitation and integration of AV speech.
Collapse
|
18
|
|
19
|
Liederman J, Gilbert K, Fisher JM, Mathews G, Frye RE, Joshi P. Are women more influenced than men by top-down semantic information when listening to disrupted speech? LANGUAGE AND SPEECH 2011; 54:33-48. [PMID: 21524011 DOI: 10.1177/0023830910388000] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
Perception is a product of the interaction between bottom-up sensory processing and top-down higher order cognitive activity. For example, when the initial phoneme of a word is obliterated and replaced with noise, listeners hear it as intact provided there is semantic context. We modified this phonemic restoration paradigm by masking (not obliterating) the initial phoneme of a target word and presenting it within a carrier phrase which was informative (I), uninformative (U), or misinformative (M). Bias in favor of top-down context was measured as the extent to which M trials mislead listeners into reporting a target word other than that which was presented (relative to U trials that have irrelevant top-down semantic context). Forty-one participants (20 men) completed 600 test trials (300 delayed report of the phrase, 300 forced choice). Relative to the U condition, women were more affected by both the I and M cues than men, at certain levels of audibility during the forced choice condition. Moreover, the semantic strength of the I carrier phrases was correlated with the rate of correct reports of the target words in women but not in men.This suggests that women can be more affected by top-down semantic context than men.
Collapse
|
20
|
Schwartz JL. A reanalysis of McGurk data suggests that audiovisual fusion in speech perception is subject-dependent. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2010; 127:1584-1594. [PMID: 20329858 DOI: 10.1121/1.3293001] [Citation(s) in RCA: 29] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2023]
Abstract
Audiovisual perception of conflicting stimuli displays a large level of intersubject variability, generally larger than pure auditory or visual data. However, it is not clear whether this actually reflects differences in integration per se or just the consequence of slight differences in unisensory perception. It is argued that the debate has been blurred by methodological problems in the analysis of experimental data, particularly when using the fuzzy-logical model of perception (FLMP) [Massaro, D. W. (1987). Speech Perception by Ear and Eye: A Paradigm for Psychological Inquiry (Laurence Erlbaum Associates, London)] shown to display overfitting abilities with McGurk stimuli [Schwartz, J. L. (2006). J. Acoust. Soc. Am. 120, 1795-1798]. A large corpus of McGurk data is reanalyzed, using a methodology based on (1) comparison of FLMP and a variant with subject-dependent weights of the auditory and visual inputs in the fusion process, weighted FLMP (WFLMP); (2) use of a Bayesian selection model criterion instead of a root mean square error fit in model assessment; and (3) systematic exploration of the number of useful parameters in the models to compare, attempting to discard poorly explicative parameters. It is shown that WFLMP performs significantly better than FLMP, suggesting that audiovisual fusion is indeed subject-dependent, some subjects being more "auditory," and others more "visual." Intersubject variability has important consequences for theoretical understanding of the fusion process, and re-education of hearing impaired people.
Collapse
Affiliation(s)
- Jean-Luc Schwartz
- Department of Speech and Cognition/Institut de la Communication Parlee, GIPSA-Lab, UMR 5216, CNRS, Grenoble University, 38402 Saint Martin d'Heres Cedex, France.
| |
Collapse
|
21
|
Feld JE, Sommers MS. Lipreading, processing speed, and working memory in younger and older adults. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2009; 52:1555-1565. [PMID: 19717657 PMCID: PMC3119632 DOI: 10.1044/1092-4388(2009/08-0137)] [Citation(s) in RCA: 33] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/28/2023]
Abstract
PURPOSE To examine several cognitive and perceptual abilities--including working memory (WM), information processing speed (PS), perceptual closure, and perceptual disembedding skill--as factors contributing to individual differences in lipreading performance and to examine how patterns in predictor variables change across age groups. METHOD Forty-three younger adults (mean age = 20.8 years, SD = 2.4) and 38 older adults (mean age = 76.8 years, SD = 5.6) completed tasks measuring lipreading ability, verbal WM, spatial WM (SWM), PS, and perceptual abilities. RESULTS Younger adults demonstrated superior lipreading ability and perceptual skills compared with older adults. In addition, younger participants exhibited longer WM spans and faster PS than did the older participants. SWM and PS accounted for a significant proportion of the variance in lipreading ability in both younger and older adults, and the pattern of predictor variables remained consistent over age groups. CONCLUSIONS These findings suggest that the large individual variability in lipreading ability can be explained, in part, by individual differences in SWM and PS. Furthermore, as both of these abilities are known to decline with age, the findings suggest that age-related impairments in either or both of these abilities may account for the poorer lipreading ability of older compared with younger adults.
Collapse
Affiliation(s)
- Julia E Feld
- Washington University in St. Louis, Psychology Department, Campus Box 1125, St. Louis, MO 63130, USA.
| | | |
Collapse
|
22
|
Kaiser A, Haller S, Schmitz S, Nitsch C. On sex/gender related similarities and differences in fMRI language research. ACTA ACUST UNITED AC 2009; 61:49-59. [DOI: 10.1016/j.brainresrev.2009.03.005] [Citation(s) in RCA: 102] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2008] [Revised: 03/29/2009] [Accepted: 03/30/2009] [Indexed: 11/27/2022]
|
23
|
Improvement in speech-reading ability by auditory training: Evidence from gender differences in normally hearing, deaf and cochlear implanted subjects. Neuropsychologia 2009; 47:972-9. [DOI: 10.1016/j.neuropsychologia.2008.10.017] [Citation(s) in RCA: 42] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2008] [Revised: 10/20/2008] [Accepted: 10/23/2008] [Indexed: 12/31/2022]
|
24
|
Mongillo EA, Irwin JR, Whalen DH, Klaiman C, Carter AS, Schultz RT. Audiovisual processing in children with and without autism spectrum disorders. J Autism Dev Disord 2008; 38:1349-58. [PMID: 18307027 DOI: 10.1007/s10803-007-0521-y] [Citation(s) in RCA: 104] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2005] [Accepted: 12/11/2007] [Indexed: 10/22/2022]
Abstract
Fifteen children with autism spectrum disorders (ASD) and twenty-one children without ASD completed six perceptual tasks designed to characterize the nature of the audiovisual processing difficulties experienced by children with ASD. Children with ASD scored significantly lower than children without ASD on audiovisual tasks involving human faces and voices, but scored similarly to children without ASD on audiovisual tasks involving nonhuman stimuli (bouncing balls). Results suggest that children with ASD may use visual information for speech differently from children without ASD. Exploratory results support an inverse association between audiovisual speech processing capacities and social impairment in children with ASD.
Collapse
Affiliation(s)
- Elizabeth A Mongillo
- Department of Psychology, University of Massachusetts Boston, 100 Morrissey Blvd., Boston, MA 02125-3393, USA.
| | | | | | | | | | | |
Collapse
|
25
|
Auer ET, Bernstein LE. Enhanced visual speech perception in individuals with early-onset hearing impairment. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2007; 50:1157-65. [PMID: 17905902 DOI: 10.1044/1092-4388(2007/080)] [Citation(s) in RCA: 75] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/17/2023]
Abstract
PURPOSE L. E. Bernstein, M. E. Demorest, and P. E. Tucker (2000) demonstrated enhanced speechreading accuracy in participants with early-onset hearing loss compared with hearing participants. Here, the authors test the generalization of Bernstein et al.'s (2000) result by testing 2 new large samples of participants. The authors also investigated correlates of speechreading ability within the early-onset hearing loss group and gender differences in speechreading ability within both participant groups. METHOD One hundred twelve individuals with early-onset hearing loss and 220 individuals with normal hearing identified 30 prerecorded sentences presented 1 at a time from visible speech information alone. RESULTS The speechreading accuracy of the participants with early-onset hearing loss (M=43.55% words correct; SD=17.48) significantly exceeded that of the participants with normal hearing (M=18.57% words correct; SD=13.18), t(330)=14.576, p<.01. Within the early-onset hearing loss participants, speechreading ability was correlated with several subjective measures of spoken communication. Effects of gender were not reliably observed. CONCLUSION The present results are consistent with the results of Bernstein et al. (2000). The need to rely on visual speech throughout life, and particularly for the acquisition of spoken language by individuals with early-onset hearing loss, can lead to enhanced speechreading ability.
Collapse
Affiliation(s)
- Edward T Auer
- Department of Speech-Language-Hearing, University of Kansas, Dole Human Development Center, 1000 Sunnyside Avenue, Room 3001, Lawrence, KS 66045-7555, USA.
| | | |
Collapse
|
26
|
Ruytjens L, Albers F, van Dijk P, Wit H, Willemsen A. Activation in Primary Auditory Cortex during Silent Lipreading Is Determined by Sex. Audiol Neurootol 2007; 12:371-7. [PMID: 17664868 DOI: 10.1159/000106480] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2006] [Accepted: 04/13/2007] [Indexed: 11/19/2022] Open
Abstract
Recent studies investigating whether the primary auditory cortex (PAC) is involved in silent lipreading gave inconsistent results. We used positron emission tomography to identify which areas in the temporal lobe process visible speech, with a focus on the PAC. Subjects were tested on lipreading numbers and only the best lipreaders were included in the study (n = 18; 9 female, 9 male). Each subject was scanned while either watching a movie with a speaker silently articulating numbers (lipreading condition) or watching a static image of the same speaker (baseline condition). Subjects were instructed to repeat internally the number seen or the number '1'. Compared to the baseline condition, silent lipreading activated temporal areas in both hemispheres with the largest activation clusters in the left hemisphere. When the whole group was examined, no activation in the PAC was found. But when investigating the two sexes separately, the female group demonstrated activation of the left PAC. There was no significant activation in the right female PAC or in the left and right male PAC. Since both groups had similar performances in lipreading, differential activity in the PAC has no effect on lipreading scores. These results may explain previous inconsistent results where no differentiation for sex was made.
Collapse
Affiliation(s)
- Liesbet Ruytjens
- Department of Otorhinolaryngology, University Medical Center Groningen, Groningen, The Netherlands.
| | | | | | | | | |
Collapse
|