1
|
Masapollo M, Nittrouer S. Immediate auditory feedback regulates inter-articulator speech coordination in service to phonetic structure. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 156:1850-1861. [PMID: 39287467 DOI: 10.1121/10.0028725] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/18/2024] [Accepted: 08/30/2024] [Indexed: 09/19/2024]
Abstract
Research has shown that talkers reliably coordinate the timing of articulator movements across variation in production rate and syllable stress, and that this precision of inter-articulator timing instantiates phonetic structure in the resulting acoustic signal. We here tested the hypothesis that immediate auditory feedback helps regulate that consistent articulatory timing control. Talkers with normal hearing recorded 480 /tV#Cat/ utterances using electromagnetic articulography, with alternative V (/ɑ/-/ɛ/) and C (/t/-/d/), across variation in production rate (fast-normal) and stress (first syllable stressed-unstressed). Utterances were split between two listening conditions: unmasked and masked. To quantify the effect of immediate auditory feedback on the coordination between the jaw and tongue-tip, the timing of tongue-tip raising onset for C, relative to the jaw opening-closing cycle for V, was obtained in each listening condition. Across both listening conditions, any manipulation that shortened the jaw opening-closing cycle reduced the latency of tongue-tip movement onset, relative to the onset of jaw opening. Moreover, tongue-tip latencies were strongly affiliated with utterance type. During auditory masking, however, tongue-tip latencies were less strongly affiliated with utterance type, demonstrating that talkers use afferent auditory signals in real-time to regulate the precision of inter-articulator timing in service to phonetic structure.
Collapse
Affiliation(s)
- Matthew Masapollo
- Department of Psychology, McGill University, 2001 McGill College, Montreal, Quebec H3A 1G1, Canada
| | - Susan Nittrouer
- Department of Speech, Language, and Hearing Sciences, University of Florida, 1225 Center Drive, Room 2150, Gainesville, Florida 32610, USA
| |
Collapse
|
2
|
Cohn M, Barreda S, Graf Estes K, Yu Z, Zellou G. Children and adults produce distinct technology- and human-directed speech. Sci Rep 2024; 14:15611. [PMID: 38971806 PMCID: PMC11227501 DOI: 10.1038/s41598-024-66313-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2024] [Accepted: 07/01/2024] [Indexed: 07/08/2024] Open
Abstract
This study compares how English-speaking adults and children from the United States adapt their speech when talking to a real person and a smart speaker (Amazon Alexa) in a psycholinguistic experiment. Overall, participants produced more effortful speech when talking to a device (longer duration and higher pitch). These differences also varied by age: children produced even higher pitch in device-directed speech, suggesting a stronger expectation to be misunderstood by the system. In support of this, we see that after a staged recognition error by the device, children increased pitch even more. Furthermore, both adults and children displayed the same degree of variation in their responses for whether "Alexa seems like a real person or not", further indicating that children's conceptualization of the system's competence shaped their register adjustments, rather than an increased anthropomorphism response. This work speaks to models on the mechanisms underlying speech production, and human-computer interaction frameworks, providing support for routinized theories of spoken interaction with technology.
Collapse
Affiliation(s)
- Michelle Cohn
- Phonetics Laboratory, Department of Linguistics, University of California, Davis, Davis, USA.
| | - Santiago Barreda
- Phonetics Laboratory, Department of Linguistics, University of California, Davis, Davis, USA
| | - Katharine Graf Estes
- Language Learning Lab, Department of Psychology, University of California, Davis, Davis, USA
| | - Zhou Yu
- Natural Language Processing (NLP) Lab, Department of Computer Science, Columbia University, New York, USA
| | - Georgia Zellou
- Phonetics Laboratory, Department of Linguistics, University of California, Davis, Davis, USA
| |
Collapse
|
3
|
Mealings K, Maggs L, Buchholz JM. The Effects of Classroom Acoustic Conditions on Teachers' Health and Well-Being: A Scoping Review. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:346-367. [PMID: 38085689 DOI: 10.1044/2023_jslhr-23-00256] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/10/2024]
Abstract
PURPOSE Teachers spend a significant amount of time in classrooms, which can be noisy environments. The aim of this scoping review was to determine what is known from the literature about the effect of classroom acoustic conditions on teachers' health and well-being. METHOD This scoping review followed the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) protocol. Four bibliographic databases were searched: ERIC, PubMed, Scopus, and Web of Science. RESULTS Thirty-three articles were deemed relevant for the review. The most studied health aspect was teachers' vocal health. The majority of results showed higher noise levels or a higher number of students have a negative effect on teachers' health and well-being (61% and 60%, respectively), while 39% showed a negative effect of longer reverberation times. Most other results showed no effect. CONCLUSIONS These results show that poor classroom acoustic conditions can have a negative effect on teachers' health and well-being. Therefore, creating classrooms with good acoustic conditions and controlling noise is vital. Limitations are discussed, and future research to better understand the relationship between classroom acoustic conditions and teachers' health and well-being is proposed. This future research will help in understanding the acoustic conditions that are needed to optimize teachers' health and well-being in the classroom.
Collapse
Affiliation(s)
- Kiri Mealings
- ECHO Lab, Macquarie University Hearing, Department of Linguistics, Macquarie University, New South Wales, Australia
| | - Lisa Maggs
- ECHO Lab, Macquarie University Hearing, Department of Linguistics, Macquarie University, New South Wales, Australia
| | - Joerg M Buchholz
- ECHO Lab, Macquarie University Hearing, Department of Linguistics, Macquarie University, New South Wales, Australia
| |
Collapse
|
4
|
Elie B, Šimko J, Turk A. Optimization-based modeling of Lombard speech articulation: Supraglottal characteristics. JASA EXPRESS LETTERS 2024; 4:015204. [PMID: 38206126 DOI: 10.1121/10.0024364] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Accepted: 12/30/2023] [Indexed: 01/12/2024]
Abstract
This paper shows that a highly simplified model of speech production based on the optimization of articulatory effort versus intelligibility can account for some observed articulatory consequences of signal-to-noise ratio. Simulations of static vowels in the presence of various background noise levels show that the model predicts articulatory and acoustic modifications of the type observed in Lombard speech. These features were obtained only when the constraint applied to articulatory effort decreases as the level of background noise increases. These results support the hypothesis that Lombard speech is listener oriented and speakers adapt their articulation in noisy environments.
Collapse
Affiliation(s)
- Benjamin Elie
- Linguistics and English Language, School of Philosophy, Psychology and Language Sciences, The University of Edinburgh, Edinburgh, Scotland, United Kingdom
| | - Juraj Šimko
- Department of Digital Humanities, Faculty of Arts, University of Helsinki, Helsinki, , ,
| | - Alice Turk
- Linguistics and English Language, School of Philosophy, Psychology and Language Sciences, The University of Edinburgh, Edinburgh, Scotland, United Kingdom
| |
Collapse
|
5
|
Cohn M, Zellou G. Selective tuning of nasal coarticulation and hyperarticulation across slow-clear, casual, and fast-clear speech styles. JASA EXPRESS LETTERS 2023; 3:125203. [PMID: 38117232 DOI: 10.1121/10.0023841] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/14/2023] [Accepted: 11/21/2023] [Indexed: 12/21/2023]
Abstract
This study investigates how California English speakers adjust nasal coarticulation and hyperarticulation on vowels across three speech styles: speaking slowly and clearly (imagining a hard-of-hearing addressee), casually (imagining a friend/family member addressee), and speaking quickly and clearly (imagining being an auctioneer). Results show covariation in speaking rate and vowel hyperarticulation across the styles. Additionally, results reveal that speakers produce more extensive anticipatory nasal coarticulation in the slow-clear speech style, in addition to a slower speech rate. These findings are interpreted in terms of accounts of coarticulation in which speakers selectively tune their production of nasal coarticulation based on the speaking style.
Collapse
Affiliation(s)
- Michelle Cohn
- Phonetics Lab, Linguistics Department, UC Davis, Davis, California 95616, ,
| | - Georgia Zellou
- Phonetics Lab, Linguistics Department, UC Davis, Davis, California 95616, ,
| |
Collapse
|
6
|
Garnier M, Smith J, Wolfe J. Lip hyper-articulation in loud voice: Effect on resonance-harmonic proximity. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 152:3695. [PMID: 36586885 DOI: 10.1121/10.0016595] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/11/2022] [Accepted: 12/01/2022] [Indexed: 06/17/2023]
Abstract
Men and women speakers were recorded while producing sustained vowels at comfortable and loud levels. Following comfortable speech, loud levels were produced in three different conditions: first without specific instruction (UL); then maintaining the same pitch as the comfortable level (PL); and finally, keeping both pitch and lip articulation constant (PAL). The sound pressure level, the fundamental frequency ( fo), the first two vocal tract resonances (R1 and R2), the lip geometry, and the larynx height were measured. For women, a closer proximity of R1 to its nearest harmonic, nfo, was observed in UL. However, no such increased proximity was found in PL, when speakers could, and did, hyper-articulate. Also, no increased proximity was observed in PAL, when lip articulation was constrained. No significant increase in R1: nfo proximity was observed in men in any of the three loud conditions. Finally, R2 was not observed significantly closer to a voice harmonic in loud speech, for neither men nor women.
Collapse
Affiliation(s)
- Maëva Garnier
- Univ. Grenoble Alpes, CNRS, Grenoble Institute of Engineering Univ. Grenoble Alpes, GIPSA-Lab, 38000 Grenoble, France
| | - John Smith
- School of Physics, UNSW Sydney, Sydney, New South Wales 2052, Australia
| | - Joe Wolfe
- School of Physics, UNSW Sydney, Sydney, New South Wales 2052, Australia
| |
Collapse
|
7
|
Piazza G, Martin CD, Kalashnikova M. The Acoustic Features and Didactic Function of Foreigner-Directed Speech: A Scoping Review. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:2896-2918. [PMID: 35914012 DOI: 10.1044/2022_jslhr-21-00609] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
PURPOSE This scoping review considers the acoustic features of a clear speech register directed to nonnative listeners known as foreigner-directed speech (FDS). We identify vowel hyperarticulation and low speech rate as the most representative acoustic features of FDS; other features, including wide pitch range and high intensity, are still under debate. We also discuss factors that may influence the outcomes and characteristics of FDS. We start by examining accommodation theories, outlining the reasons why FDS is likely to serve a didactic function by helping listeners acquire a second language (L2). We examine how this speech register adapts to listeners' identities and linguistic needs, suggesting that FDS also takes listeners' L2 proficiency into account. To confirm the didactic function of FDS, we compare it to other clear speech registers, specifically infant-directed speech and Lombard speech. CONCLUSIONS Our review reveals that research has not yet established whether FDS succeeds as a didactic tool that supports L2 acquisition. Moreover, a complex set of factors determines specific realizations of FDS, which need further exploration. We conclude by summarizing open questions and indicating directions and recommendations for future research.
Collapse
Affiliation(s)
- Giorgio Piazza
- Basque Center on Cognition, Brain and Language (BCBL), Donostia-San Sebastián, Spain
- Department of Social Sciences and Law, Universidad del País Vasco/Euskal Herriko Unibertsitatea, Donostia-San Sebastián, Spain
| | - Clara D Martin
- Basque Center on Cognition, Brain and Language (BCBL), Donostia-San Sebastián, Spain
- Ikerbasque, Basque Foundation for Science, Bilbao, Spain
| | - Marina Kalashnikova
- Basque Center on Cognition, Brain and Language (BCBL), Donostia-San Sebastián, Spain
- Ikerbasque, Basque Foundation for Science, Bilbao, Spain
| |
Collapse
|
8
|
Gutz SE, Rowe HP, Tilton-Bolowsky VE, Green JR. Speaking with a KN95 face mask: a within-subjects study on speaker adaptation and strategies to improve intelligibility. Cogn Res Princ Implic 2022; 7:73. [PMID: 35907167 PMCID: PMC9339031 DOI: 10.1186/s41235-022-00423-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2021] [Accepted: 07/18/2022] [Indexed: 11/15/2022] Open
Abstract
Mask-wearing during the COVID-19 pandemic has prompted a growing interest in the functional impact of masks on speech and communication. Prior work has shown that masks dampen sound, impede visual communication cues, and reduce intelligibility. However, more work is needed to understand how speakers change their speech while wearing a mask and to identify strategies to overcome the impact of wearing a mask. Data were collected from 19 healthy adults during a single in-person session. We investigated the effects of wearing a KN95 mask on speech intelligibility, as judged by two speech-language pathologists, examined speech kinematics and acoustics associated with mask-wearing, and explored KN95 acoustic filtering. We then considered the efficacy of three speaking strategies to improve speech intelligibility: Loud, Clear, and Slow speech. To inform speaker strategy recommendations, we related findings to self-reported speaker effort. Results indicated that healthy speakers could compensate for the presence of a mask and achieve normal speech intelligibility. Additionally, we showed that speaking loudly or clearly-and, to a lesser extent, slowly-improved speech intelligibility. However, using these strategies may require increased physical and cognitive effort and should be used only when necessary. These results can inform recommendations for speakers wearing masks, particularly those with communication disorders (e.g., dysarthria) who may struggle to adapt to a mask but can respond to explicit instructions. Such recommendations may further help non-native speakers and those communicating in a noisy environment or with listeners with hearing loss.
Collapse
Affiliation(s)
- Sarah E. Gutz
- Program in Speech and Hearing Bioscience and Technology, Harvard Medical School, Boston, MA USA
| | - Hannah P. Rowe
- Department of Communication Sciences and Disorders, MGH Institute of Health Professions, Building 79/96, 2nd floor, 13th Street, Boston, MA 02129 USA
| | - Victoria E. Tilton-Bolowsky
- Department of Communication Sciences and Disorders, MGH Institute of Health Professions, Building 79/96, 2nd floor, 13th Street, Boston, MA 02129 USA
| | - Jordan R. Green
- Program in Speech and Hearing Bioscience and Technology, Harvard Medical School, Boston, MA USA
- Department of Communication Sciences and Disorders, MGH Institute of Health Professions, Building 79/96, 2nd floor, 13th Street, Boston, MA 02129 USA
| |
Collapse
|
9
|
Trujillo JP, Levinson SC, Holler J. A multi-scale investigation of the human communication system's response to visual disruption. ROYAL SOCIETY OPEN SCIENCE 2022; 9:211489. [PMID: 35425638 PMCID: PMC9006025 DOI: 10.1098/rsos.211489] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/20/2021] [Accepted: 03/25/2022] [Indexed: 05/03/2023]
Abstract
In human communication, when the speech is disrupted, the visual channel (e.g. manual gestures) can compensate to ensure successful communication. Whether speech also compensates when the visual channel is disrupted is an open question, and one that significantly bears on the status of the gestural modality. We test whether gesture and speech are dynamically co-adapted to meet communicative needs. To this end, we parametrically reduce visibility during casual conversational interaction and measure the effects on speakers' communicative behaviour using motion tracking and manual annotation for kinematic and acoustic analyses. We found that visual signalling effort was flexibly adapted in response to a decrease in visual quality (especially motion energy, gesture rate, size, velocity and hold-time). Interestingly, speech was also affected: speech intensity increased in response to reduced visual quality (particularly in speech-gesture utterances, but independently of kinematics). Our findings highlight that multi-modal communicative behaviours are flexibly adapted at multiple scales of measurement and question the notion that gesture plays an inferior role to speech.
Collapse
Affiliation(s)
- James P. Trujillo
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD Nijmegen, The Netherlands
| | - Stephen C. Levinson
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD Nijmegen, The Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD Nijmegen, The Netherlands
| |
Collapse
|
10
|
Schiller IS, Remacle A, Durieux N, Morsomme D. Effects of Noise and a Speaker's Impaired Voice Quality on Spoken Language Processing in School-Aged Children: A Systematic Review and Meta-Analysis. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:169-199. [PMID: 34902257 DOI: 10.1044/2021_jslhr-21-00183] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
PURPOSE Background noise and voice problems among teachers can degrade listening conditions in classrooms. The aim of this literature review is to understand how these acoustic degradations affect spoken language processing in 6- to 18-year-old children. METHOD In a narrative report and meta-analysis, we systematically review studies that examined the effects of noise and/or impaired voice on children's response accuracy and response time (RT) in listening tasks. We propose the Speech Processing under Acoustic DEgradations (SPADE) framework to classify relevant findings according to three processing dimensions-speech perception, listening comprehension, and auditory working memory-and highlight potential moderators. RESULTS Thirty-one studies are included in this systematic review. Our meta-analysis shows that noise can impede children's accuracy in listening tasks across all processing dimensions (Cohen's d between -0.67 and -2.65, depending on signal-to-noise ratio) and that impaired voice lowers children's accuracy in listening comprehension tasks (d = -0.35). A handful of studies assessed RT, but results are inconclusive. The impact of noise and impaired voice can be moderated by listener, task, environmental, and exposure factors. The interaction between noise and impaired voice remains underinvestigated. CONCLUSIONS Overall, this review suggests that children have more trouble perceiving speech, processing verbal messages, and recalling verbal information when listening to speech in noise or to a speaker with dysphonia. Impoverished speech input could impede pupils' motivation and academic performance at school. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.17139377.
Collapse
Affiliation(s)
- Isabel S Schiller
- Research Unit for a Life-Course Perspective on Health & Education, Faculty of Psychology, Speech and Language Therapy, and Educational Sciences, University of Liège, Belgium
- Teaching and Research Area Work and Engineering Psychology, Institute of Psychology, RWTH Aachen University, Germany
| | - Angélique Remacle
- Research Unit for a Life-Course Perspective on Health & Education, Faculty of Psychology, Speech and Language Therapy, and Educational Sciences, University of Liège, Belgium
- Center For Research in Cognition and Neurosciences, Faculty of Psychological Science and Education, Université Libre de Bruxelles, Belgium
| | - Nancy Durieux
- Research Unit for a Life-Course Perspective on Health & Education, Faculty of Psychology, Speech and Language Therapy, and Educational Sciences, University of Liège, Belgium
| | - Dominique Morsomme
- Research Unit for a Life-Course Perspective on Health & Education, Faculty of Psychology, Speech and Language Therapy, and Educational Sciences, University of Liège, Belgium
| |
Collapse
|
11
|
Trujillo J, Özyürek A, Holler J, Drijvers L. Speakers exhibit a multimodal Lombard effect in noise. Sci Rep 2021; 11:16721. [PMID: 34408178 PMCID: PMC8373897 DOI: 10.1038/s41598-021-95791-0] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2021] [Accepted: 07/29/2021] [Indexed: 12/03/2022] Open
Abstract
In everyday conversation, we are often challenged with communicating in non-ideal settings, such as in noise. Increased speech intensity and larger mouth movements are used to overcome noise in constrained settings (the Lombard effect). How we adapt to noise in face-to-face interaction, the natural environment of human language use, where manual gestures are ubiquitous, is currently unknown. We asked Dutch adults to wear headphones with varying levels of multi-talker babble while attempting to communicate action verbs to one another. Using quantitative motion capture and acoustic analyses, we found that (1) noise is associated with increased speech intensity and enhanced gesture kinematics and mouth movements, and (2) acoustic modulation only occurs when gestures are not present, while kinematic modulation occurs regardless of co-occurring speech. Thus, in face-to-face encounters the Lombard effect is not constrained to speech but is a multimodal phenomenon where the visual channel carries most of the communicative burden.
Collapse
Affiliation(s)
- James Trujillo
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands.
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD, Nijmegen, The Netherlands.
| | - Asli Özyürek
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD, Nijmegen, The Netherlands
- Centre for Language Studies, Radboud University Nijmegen, Nijmegen, The Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD, Nijmegen, The Netherlands
| | - Linda Drijvers
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD, Nijmegen, The Netherlands
| |
Collapse
|
12
|
Cohn M, Pycha A, Zellou G. Intelligibility of face-masked speech depends on speaking style: Comparing casual, clear, and emotional speech. Cognition 2021; 210:104570. [DOI: 10.1016/j.cognition.2020.104570] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2020] [Revised: 12/15/2020] [Accepted: 12/22/2020] [Indexed: 10/22/2022]
|
13
|
Haldin C, Loevenbruck H, Hueber T, Marcon V, Piscicelli C, Perrier P, Chrispin A, Pérennou D, Baciu M. Speech rehabilitation in post-stroke aphasia using visual illustration of speech articulators: A case report study. CLINICAL LINGUISTICS & PHONETICS 2021; 35:253-276. [PMID: 32567986 DOI: 10.1080/02699206.2020.1780473] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/17/2019] [Revised: 04/24/2020] [Accepted: 06/06/2020] [Indexed: 06/11/2023]
Abstract
Recent studies on the remediation of speech disorders suggest that providing visual information of speech articulators may contribute to improve speech production. In this study, we evaluate the effectiveness of an illustration-based rehabilitation method on speech recovery of a patient with non-fluent chronic aphasia. The Ultraspeech-player software allowed visualization by the patient of reference tongue and lip movements recorded using ultrasound and video imaging. This method can improve the patient's awareness of their own lingual and labial movements, which can increase the ability to coordinate and combine articulatory gestures. The effects of this method were assessed by analyzing performance during speech tasks, the phonological processes identified in the errors made during the phoneme repetition task and the acoustic parameters derived from the speech signal. We also evaluated cognitive performance before and after rehabilitation. The integrity of visuospatial ability, short-term and working memory and some executive functions supports the effectiveness of the rehabilitation method. Our results showed that illustration-based rehabilitation technique had a beneficial effect on the patient's speech production, especially for stop and fricative consonants which are targeted (high visibility of speech articulator configurations) by the software, but also on reading abilities. Acoustic parameters indicated an improvement in the distinction between consonant categories: voiced and voiceless stops or alveolar, post-alveolar and labiodental fricatives. However, the patient showed little improvement for vowels. These results confirmed the advantage of using illustration-based rehabilitation technique and the necessity of detailed subjective and objective intra-speaker evaluation in speech production to fully evaluate speech abilities.
Collapse
Affiliation(s)
- Célise Haldin
- Laboratoire De Psychologie Et Neurocognition, UMR CNRS 5105, Université Grenoble Alpes , Grenoble, France
| | - Hélène Loevenbruck
- Laboratoire De Psychologie Et Neurocognition, UMR CNRS 5105, Université Grenoble Alpes , Grenoble, France
| | - Thomas Hueber
- GIPSA-lab, UMR CNRS 5216, Université Grenoble-Alpes , Grenoble, France
| | - Valérie Marcon
- CHU Grenoble-Alpes, Médecine Physique Et De Réadaptation , Grenoble, France
| | - Céline Piscicelli
- Laboratoire De Psychologie Et Neurocognition, UMR CNRS 5105, Université Grenoble Alpes , Grenoble, France
- CHU Grenoble-Alpes, Médecine Physique Et De Réadaptation , Grenoble, France
| | - Pascal Perrier
- GIPSA-lab, UMR CNRS 5216, Université Grenoble-Alpes , Grenoble, France
| | - Anne Chrispin
- CHU Grenoble-Alpes, Médecine Physique Et De Réadaptation , Grenoble, France
| | - Dominic Pérennou
- Laboratoire De Psychologie Et Neurocognition, UMR CNRS 5105, Université Grenoble Alpes , Grenoble, France
- CHU Grenoble-Alpes, Médecine Physique Et De Réadaptation , Grenoble, France
| | - Monica Baciu
- Laboratoire De Psychologie Et Neurocognition, UMR CNRS 5105, Université Grenoble Alpes , Grenoble, France
| |
Collapse
|
14
|
Villegas J, Perkins J, Wilson I. Effects of task and language nativeness on the Lombard effect and on its onset and offset timing. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:1855. [PMID: 33765802 DOI: 10.1121/10.0003772] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/17/2020] [Accepted: 02/24/2021] [Indexed: 06/12/2023]
Abstract
This study focuses on the differences in speech sound pressure levels (here, called speech loudness) of Lombard speech (i.e., speech produced in the presence of an energetic masker) associated with different tasks and language nativeness. Vocalizations were produced by native speakers of Japanese with normal hearing and limited English proficiency while performing four tasks: dialog, a competitive game (both communicative), soliloquy, and text passage reading (noncommunicative). Relative to the native language (L1), larger loudness increments were observed in the game and text reading when performed in the second language (L2). Communicative tasks yielded louder vocalizations and larger increments of speech loudness than did noncommunicative tasks regardless of the spoken language. The period in which speakers increased their loudness after the onset of the masker was about fourfold longer than the time in which they decreased their loudness after the offset of the masker. Results suggest that when relying on acoustic signals, speakers use similar vocalization strategies in L1 and L2, and these depend on the complexity of the task, the need for accurate pronunciation, and the presence of a listener. Results also suggest that speakers use different strategies depending on the onset or offset of an energetic masker.
Collapse
Affiliation(s)
- Julián Villegas
- Computer Arts Laboratory, University of Aizu, Aizu-Wakamatsu, Fukushima, 965-8580, Japan
| | - Jeremy Perkins
- CLR Phonetics Laboratory, University of Aizu, Aizu-Wakamatsu, Fukushima, 965-8580, Japan
| | - Ian Wilson
- CLR Phonetics Laboratory, University of Aizu, Aizu-Wakamatsu, Fukushima, 965-8580, Japan
| |
Collapse
|
15
|
Xue Y, Marxen M, Akagi M, Birkholz P. Acoustic and articulatory analysis and synthesis of shouted vowels. COMPUT SPEECH LANG 2021. [DOI: 10.1016/j.csl.2020.101156] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|