1
|
Kim J, Hazan V, Tuomainen O, Davis C. Partner-directed gaze and co-speech hand gestures: effects of age, hearing loss and noise. Front Psychol 2024; 15:1324667. [PMID: 38882511 PMCID: PMC11178134 DOI: 10.3389/fpsyg.2024.1324667] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/29/2023] [Accepted: 05/10/2024] [Indexed: 06/18/2024] Open
Abstract
Research on the adaptations talkers make to different communication conditions during interactive conversations has primarily focused on speech signals. We extended this type of investigation to two other important communicative signals, i.e., partner-directed gaze and iconic co-speech hand gestures with the aim of determining if the adaptations made by older adults differ from younger adults across communication conditions. We recruited 57 pairs of participants, comprising 57 primary talkers and 57 secondary ones. Primary talkers consisted of three groups: 19 older adults with mild Hearing Loss (older adult-HL); 17 older adults with Normal Hearing (older adult-NH); and 21 younger adults. The DiapixUK "spot the difference" conversation-based task was used to elicit conversions in participant pairs. One easy (No Barrier: NB) and three difficult communication conditions were tested. The three conditions consisted of two in which the primary talker could hear clearly, but the secondary talkers could not, due to multi-talker babble noise (BAB1) or a less familiar hearing loss simulation (HLS), and a condition in which both the primary and secondary talkers heard each other in babble noise (BAB2). For primary talkers, we measured mean number of partner-directed gazes; mean total gaze duration; and the mean number of co-speech hand gestures. We found a robust effects of communication condition that interacted with participant group. Effects of age were found for both gaze and gesture in BAB1, i.e., older adult-NH looked and gestured less than younger adults did when the secondary talker experienced babble noise. For hearing status, a difference in gaze between older adult-NH and older adult-HL was found for the BAB1 condition; for gesture this difference was significant in all three difficult communication conditions (older adult-HL gazed and gestured more). We propose the age effect may be due to a decline in older adult's attention to cues signaling how well a conversation is progressing. To explain the hearing status effect, we suggest that older adult's attentional decline is offset by hearing loss because these participants have learned to pay greater attention to visual cues for understanding speech.
Collapse
Affiliation(s)
- Jeesun Kim
- The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Sydney, NSW, Australia
| | - Valerie Hazan
- Speech Hearing and Phonetic Sciences, University College London, London, United Kingdom
| | - Outi Tuomainen
- Department of Linguistics, University of Potsdam, Potsdam, Germany
| | - Chris Davis
- The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Sydney, NSW, Australia
| |
Collapse
|
2
|
Okalidou A, Peng ZE, Banioti A, Fourakis M, Kyriafinis G. The Lombard effect in children with cochlear implants: suprasegmental aspects. CLINICAL LINGUISTICS & PHONETICS 2024:1-21. [PMID: 38679889 DOI: 10.1080/02699206.2024.2340096] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/02/2023] [Accepted: 04/02/2024] [Indexed: 05/01/2024]
Abstract
Children with cochlear implants (CI) communicate in noisy environments, such as in classrooms, where multiple talkers and reverberation are present. Speakers compensate for noise via the 'Lombard effect'. The present study examined the Lombard effect on the intensity and duration of stressed vowels in the speech of children with Cochlear Implants (CIs) as compared to children with Normal Hearing (NH), focusing on the effects of speech-shaped noise (SSN) and speech-shaped noise with reverberation (SSN+Reverberation). The sample consisted of 7 children with CIs and 7 children with NH, aged 7-12 years. Regarding intensity, a) children with CIs produced stressed vowels with an overall greater intensity across acoustic conditions as compared to NH peers, b) both groups increased their stressed vowel intensity for all vowels from Quiet to both noise conditions, and c) children with NH further increased their intensity when reverberation was added to SSN, esp. for the vowel/u/. Regarding duration, longer stressed vowels were produced by children with CIs as compared to NH in Quiet and SSN conditions but the effect was retained only for the vowels/i/,/o/and/u/when reverberation was added to noise. The SSN+Reverberation condition induced systematic lengthening in stressed vowels for children with NH. Furthermore, although greater intensity and duration ratios of stressed/unstressed syllables were observed for children with NH as compared to CIs in Quiet condition, they diminished with noise. The differences observed across groups have implications for speaking in classroom noise.
Collapse
Affiliation(s)
- Areti Okalidou
- Department of Educational & Social Policy, Graduate Program of Communication Disorders & Sciences, University of Macedonia, Thessaloniki, Greece
| | - Z Ellen Peng
- Functional Hearing Laboratory, Boys Town National Research Hospital, Omaha, Nebraska, USA
| | - Aggeliki Banioti
- Department of Educational & Social Policy, Graduate Program of Communication Disorders & Sciences, University of Macedonia, Thessaloniki, Greece
| | - Marios Fourakis
- Department of Hearing & Speech Sciences, University of Maryland, College Park, Maryland, USA
| | - Georgios Kyriafinis
- 1st University Otolaryngology Clinic of AHEPA Hospital, Medical School, Aristotle University of Thessaloniki, Thessaloniki, Greece
| |
Collapse
|
3
|
Smith NA, Hammans CA, Vallier TJ, McMurray B. Child-Directed Speech in Noise: Testing Signal- and Code-Based Phonetic Enhancement. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:72-91. [PMID: 38039984 DOI: 10.1044/2023_jslhr-23-00033] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/03/2023]
Abstract
PURPOSE Talkers adapt their speech according to the demands of their listeners and the communicative context, enhancing the properties of the signal (pitch, intensity) and/or properties of the code (enhancement of phonemic contrasts). This study asked how mothers adapt their child-directed speech (CDS) in ways that might serve the immediate goals of increasing intelligibility, as well as long-term goals of supporting speech and language development in their children. METHOD Mothers (N = 28) participated in a real-time interactive speech production/perception paradigm, in which mothers instructed their young (3- to 5-year-old) children, or an adult listener, to select the picture corresponding to a target word. The task was performed at low and high levels (56 vs. 75 dB SPL) of background noise to examine the Lombard effects of decreased audibility on speech production. RESULTS Acoustic-phonetic analyses of CDS and adult-directed speech (ADS) productions of target words and carrier phrase (e.g., "Find pig") revealed that mothers significantly enhanced the mean pitch, pitch variability, and intensity of target words in CDS, particularly at higher background noise levels and for younger children. Mothers produce CDS with a higher signal-to-noise ratio than ADS. However, limited evidence was found for phonetic enhancement of the segmental properties of speech. Although increased category separation was found in the voice onset time of stop consonants, decreased vowel category separation (an anti-enhancement effect) was observed in CDS. CONCLUSIONS Mothers readily enhance the suprasegmental signal properties of their speech in CDS, but not the acoustic-phonetic properties of phonemes. This study fails to provide evidence of phonetic enhancement in preschool children in a dyadic communication task under noisy listening conditions. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.24645423.
Collapse
Affiliation(s)
- Nicholas A Smith
- Department of Speech, Language and Hearing Sciences, University of Missouri-Columbia
- Department of Psychological Sciences, University of Missouri-Columbia
| | | | | | - Bob McMurray
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City
- Department of Linguistics, University of Iowa, Iowa City
| |
Collapse
|
4
|
Miles K, Weisser A, Kallen RW, Varlet M, Richardson MJ, Buchholz JM. Behavioral dynamics of conversation, (mis)communication and coordination in noisy environments. Sci Rep 2023; 13:20271. [PMID: 37985887 PMCID: PMC10662155 DOI: 10.1038/s41598-023-47396-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2023] [Accepted: 11/13/2023] [Indexed: 11/22/2023] Open
Abstract
During conversations people coordinate simultaneous channels of verbal and nonverbal information to hear and be heard. But the presence of background noise levels such as those found in cafes and restaurants can be a barrier to conversational success. Here, we used speech and motion-tracking to reveal the reciprocal processes people use to communicate in noisy environments. Conversations between twenty-two pairs of typical-hearing adults were elicited under different conditions of background noise, while standing or sitting around a table. With the onset of background noise, pairs rapidly adjusted their interpersonal distance and speech level, with the degree of initial change dependent on noise level and talker configuration. Following this transient phase, pairs settled into a sustaining phase in which reciprocal speech and movement-based coordination processes synergistically maintained effective communication, again with the magnitude of stability of these coordination processes covarying with noise level and talker configuration. Finally, as communication breakdowns increased at high noise levels, pairs exhibited resetting behaviors to help restore communication-decreasing interpersonal distance and/or increasing speech levels in response to communication breakdowns. Approximately 78 dB SPL defined a threshold where behavioral processes were no longer sufficient for maintaining effective conversation and communication breakdowns rapidly increased.
Collapse
Affiliation(s)
- Kelly Miles
- ECHO Laboratory, MU Hearing, and Performance and Expertise Research Centre, Macquarie University, Sydney, Australia.
| | - Adam Weisser
- ECHO Laboratory, Macquarie University, Sydney, Australia
| | - Rachel W Kallen
- Performance and Expertise Research Centre, School of Psychological Sciences, Macquarie University, Sydney, Australia
| | - Manuel Varlet
- The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Sydney, Australia
| | - Michael J Richardson
- Performance and Expertise Research Centre, School of Psychological Sciences, Macquarie University, Sydney, Australia
| | - Joerg M Buchholz
- ECHO Laboratory, MU Hearing, and Performance and Expertise Research Centre, Macquarie University, Sydney, Australia
| |
Collapse
|
5
|
Ishikawa K, Li H, Coster E. The Effect of Noise on Initiation and Maintenance of Clear Speech and Associated Mental Demand. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2023; 66:4180-4190. [PMID: 37793611 DOI: 10.1044/2023_jslhr-23-00157] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/06/2023]
Abstract
BACKGROUND/OBJECTIVES The objectives of this study were to explore (a) the influence of different types of background noise and their informational content on the ability of speakers to initiate and maintain clear speech (CS), a widely utilized technique for enhancing speech intelligibility, and (b) the impact of background noise and CS usage on speakers' mental demand. METHOD Five adult females were asked to read sentences using both habitual and CS under four distinct noise conditions: quiet, multitalker (MT) noise, reversed multitalker (RevMT) noise, and speech-shaped (SS) noise. Following this, speakers rated their perceived level of mental demand for each speaking condition using the modified NASA Task Load Index scale. A two-part listening experiment with 48 listeners was conducted to evaluate the speakers' effectiveness in initiating and maintaining CS. RESULTS Speakers initiated CS more successfully in noise than in quiet, with better performance observed in the presence of RevMT noise as compared to SS noise. Regarding the maintenance of CS, none of the speakers were successful in a quiet environment. Furthermore, the ability to maintain CS was most adversely affected in MT noise, followed by RevMT noise and SS noise. CONCLUSIONS Our findings suggest that the effect of background noise on speech production is complex and multifaceted. The noise type affected speakers' ability to initiate and maintain CS as well as the mental demand associated with the speech task. The results underscore the importance of considering the characteristics of background noise and cognitive aspects of speech production when training and evaluating speakers' performance.
Collapse
Affiliation(s)
- Keiko Ishikawa
- Department of Communication Science and Disorders, University of Kentucky, Lexington
| | - Hannah Li
- Department of Speech and Hearing Science, University of Illinois at Urbana-Champaign
| | - Elisabeth Coster
- Department of Speech and Hearing Science, University of Illinois at Urbana-Champaign
| |
Collapse
|
6
|
Nelson BS, Harmon TG, Dromey C, Clawson KD. Telling Stories in Noise: The Impact of Background Noises on Spoken Language for People With Aphasia. AMERICAN JOURNAL OF SPEECH-LANGUAGE PATHOLOGY 2023; 32:2444-2460. [PMID: 37486853 DOI: 10.1044/2023_ajslp-22-00299] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/26/2023]
Abstract
PURPOSE The purpose of this study was to determine how different background noise conditions affect the spoken language of participants with aphasia during a story retell task. METHOD Participants included 11 adults with mild to moderate aphasia and 11 age- and gender-matched controls. Participants retold stories in a silent baseline and five background noise conditions (conversation, monologue, phone call, cocktail, and pink noise). Dependent measures of speech acoustics (fundamental frequency and mean intensity), speech fluency (speech rate and disfluent words), and language production (correct information units [CIUs], lexical errors, lexical diversity, and cohesive utterances) were compared between groups and across conditions. RESULTS Background noise resulted in higher fundamental frequency (fo) and increased mean intensity for control participants across all noise conditions but only across some conditions for participants with aphasia. In relation to language production, background noise interfered significantly more with communication efficiency (i.e., percent CIUs) for participants with aphasia than the control group. For participants with aphasia, the phone call condition led to decreased lexical diversity. Across groups, condition effects generally suggested more interference on speech acoustics in conditions where continuous noise was present and more interference on language in conditions that presented continuous informational noise. CONCLUSIONS Although additional research is needed, preliminary findings suggest that background noise interferes with narrative discourse more for people with aphasia (PWA) than neurologically healthy adults. PWA may benefit from therapy that directly addresses communicating in noise. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.23681703.
Collapse
Affiliation(s)
| | - Tyson G Harmon
- Department of Communication Disorders, Brigham Young University, Provo, UT
| | - Christopher Dromey
- Department of Communication Disorders, Brigham Young University, Provo, UT
| | | |
Collapse
|
7
|
Shen C, Cooke M, Janse E. Speaking in the presence of noise: Consistency of acoustic properties in clear-Lombard speech over time. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 153:2165. [PMID: 37092911 DOI: 10.1121/10.0017769] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/22/2022] [Accepted: 03/17/2023] [Indexed: 05/03/2023]
Abstract
Individual speakers are often able to modify their speech to facilitate communication in challenging conditions, such as speaking in a noisy environment. Such vocal "enrichments" might include reductions in speech rate or increases in acoustic contrasts. However, it is unclear how consistently speakers enrich their speech over time. This study examined inter-speaker variability in the speech enrichment modifications applied by speakers. The study compared a baseline habitual speaking style to a clear-Lombard style and measured changes in acoustic differences between the two styles over sentence trials. Seventy-eight young adult participants read out sentences in the habitual and clear-Lombard speaking styles. Acoustic differences between speaking styles generally increased nonlinearly over trials, suggesting that speakers require practice before realizing their full speech enrichment potential when speaking clearly in noise with reduced auditory feedback. Using a recent objective intelligibility metric based on glimpses, the study also found that predicted intelligibility increased over trials, highlighting that communicative benefits of the clear-Lombard style are not static. These findings underline the dynamic nature of speaking styles.
Collapse
Affiliation(s)
- Chen Shen
- Centre for Language Studies, Radboud University, Nijmegen, 6500 HD, The Netherlands
| | - Martin Cooke
- Ikerbasque (Basque Science Foundation), Bilbao, 01006, Spain
| | - Esther Janse
- Centre for Language Studies, Radboud University, Nijmegen, 6500 HD, The Netherlands
| |
Collapse
|
8
|
Sachdeva S, Ruan H, Hamarneh G, Behne DM, Jongman A, Sereno JA, Wang Y. Plain-to-clear speech video conversion for enhanced intelligibility. INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY 2023; 26:163-184. [PMID: 37008883 PMCID: PMC10042924 DOI: 10.1007/s10772-023-10018-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/28/2022] [Accepted: 01/08/2023] [Indexed: 06/19/2023]
Abstract
Clearly articulated speech, relative to plain-style speech, has been shown to improve intelligibility. We examine if visible speech cues in video only can be systematically modified to enhance clear-speech visual features and improve intelligibility. We extract clear-speech visual features of English words varying in vowels produced by multiple male and female talkers. Via a frame-by-frame image-warping based video generation method with a controllable parameter (displacement factor), we apply the extracted clear-speech visual features to videos of plain speech to synthesize clear speech videos. We evaluate the generated videos using a robust, state of the art AI Lip Reader as well as human intelligibility testing. The contributions of this study are: (1) we successfully extract relevant visual cues for video modifications across speech styles, and have achieved enhanced intelligibility for AI; (2) this work suggests that universal talker-independent clear-speech features may be utilized to modify any talker's visual speech style; (3) we introduce "displacement factor" as a way of systematically scaling the magnitude of displacement modifications between speech styles; and (4) the high definition generated videos make them ideal candidates for human-centric intelligibility and perceptual training studies.
Collapse
Affiliation(s)
- Shubam Sachdeva
- Language and Brain Lab, Department of Linguistics, Simon Fraser University, Burnaby, BC Canada
| | - Haoyao Ruan
- Language and Brain Lab, Department of Linguistics, Simon Fraser University, Burnaby, BC Canada
| | - Ghassan Hamarneh
- Medical Image, Analysis Research Group, School of Computing Science, Simon Fraser University, Burnaby, BC Canada
| | - Dawn M. Behne
- NTNU Speech Lab, Department of Psychology, Norwegian University of Science and Technology, Trondheim, Norway
| | - Allard Jongman
- KU Phonetics and Psycholinguistics Lab, Department of Linguistics, University of Kansas, Lawrence, KS USA
| | - Joan A. Sereno
- KU Phonetics and Psycholinguistics Lab, Department of Linguistics, University of Kansas, Lawrence, KS USA
| | - Yue Wang
- Language and Brain Lab, Department of Linguistics, Simon Fraser University, Burnaby, BC Canada
| |
Collapse
|
9
|
Kondaurova MV, Zheng Q, Donaldson CW, Smith AF. Effect of telepractice on pediatric cochlear implant users and provider vowel space: A preliminary report. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 153:467. [PMID: 36732236 DOI: 10.1121/10.0016866] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/09/2022] [Accepted: 12/27/2022] [Indexed: 06/18/2023]
Abstract
Clear speaking styles are goal-oriented modifications in which talkers adapt acoustic-phonetic characteristics of speech to compensate for communication challenges. Do children with hearing loss and a clinical provider modify speech characteristics during telepractice to adjust for remote communication? The study examined the effect of telepractice (tele-) on vowel production in seven (mean age 4:11 years, SD 1:2 years) children with cochlear implants (CIs) and a provider. The first (F1) and second (F2) formant frequencies of /i/, /ɑ/, and /u/ vowels were measured in child and provider speech during one in-person and one tele-speech-language intervention, order counterbalanced. Child and provider vowel space areas (VSA) were calculated. The results demonstrated an increase in F2 formant frequency for /i/ vowel in child and provider speech and an increase in F1 formant frequency for /ɑ/ vowel in the provider speech during tele- compared to in-person intervention. An expansion of VSA was found in child and provider speech in tele- compared to in-person intervention. In children, the earlier age of CI activation was associated with larger VSA in both tele- and in-person intervention. The results suggest that the children and the provider adjust vowel articulation in response to remote communication during telepractice.
Collapse
Affiliation(s)
- Maria V Kondaurova
- Department of Psychological and Brain Sciences, University of Louisville, 301 Life Sciences Building, Louisville, Kentucky 40292, USA
| | - Qi Zheng
- Department of Bioinformatics and Biostatistics, University of Louisville, Louisville, Kentucky 40202, USA
| | - Cheryl W Donaldson
- The Heuser Hearing Institute and Language Academy, Louisville, Kentucky 40203, USA
| | - Alan F Smith
- Department of Otolaryngology-Head/Neck Surgery and Communicative Disorders, Speech-Language Pathology Program, University of Louisville, Louisville, Kentucky 40202, USA
| |
Collapse
|
10
|
Stenbäck V, Marsja E, Hällgren M, Lyxell B, Larsby B. Informational Masking and Listening Effort in Speech Recognition in Noise: The Role of Working Memory Capacity and Inhibitory Control in Older Adults With and Without Hearing Impairment. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:4417-4428. [PMID: 36283680 DOI: 10.1044/2022_jslhr-21-00674] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
PURPOSE The study aimed to assess the relationship between (a) speech recognition in noise, mask type, working memory capacity (WMC), and inhibitory control and (b) self-rated listening effort, speech material, and mask type, in older adults with and without hearing impairment. It was of special interest to assess the relationship between WMC, inhibitory control, and speech recognition in noise when informational maskers masked target speech. METHOD A mixed design was used. A group (N = 24) of older (Mage = 69.7 years) individuals with hearing impairment and a group of age normal-hearing adults (Mage = 59.3 years, SD = 6.5) participated in the study. The participants were presented with auditory tests in a sound-attenuated room and with cognitive tests in a quiet office. The participants were asked to rate listening effort after being presented with energetic and informational background maskers in two different speech materials used in this study (i.e., Hearing In Noise Test and Hagerman test). Linear mixed-effects models were set up to assess the effect of the two different speech materials, energetic and informational maskers, hearing ability, WMC, inhibitory control, and self-rated listening effort. RESULTS Results showed that WMC and inhibitory control were of importance for speech recognition in noise, even when controlling for pure-tone average 4 hearing thresholds and age, when the maskers were informational. Concerning listening effort, on the other hand, the results suggest that hearing ability, but not cognitive abilities, is important for self-rated listening effort in speech recognition in noise. CONCLUSIONS Speech-in-noise recognition is more dependent on WMC for older adults in informational maskers than in energetic maskers. Hearing ability is a stronger predictor than cognition for self-rated listening effort. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.21357648.
Collapse
Affiliation(s)
- Victoria Stenbäck
- Disability Research Division, Department of Behavioural Sciences and Learning, Linköping University, Sweden
- Division of Education, Teaching and Learning, Department of Behavioural Sciences and Learning, Linköping University, Sweden
| | - Erik Marsja
- Disability Research Division, Department of Behavioural Sciences and Learning, Linköping University, Sweden
| | - Mathias Hällgren
- Department of Otorhinolaryngology in Östergötland and Department of Biomedical and Clinical Sciences, Linköping University, Sweden
| | - Björn Lyxell
- Disability Research Division, Department of Behavioural Sciences and Learning, Linköping University, Sweden
- Department of Special Needs Education, University of Oslo, Norway
| | - Birgitta Larsby
- Department of Otorhinolaryngology in Östergötland and Department of Biomedical and Clinical Sciences, Linköping University, Sweden
| |
Collapse
|
11
|
Nusseck M, Immerz A, Richter B, Traser L. Vocal Behavior of Teachers Reading with Raised Voice in a Noisy Environment. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2022; 19:ijerph19158929. [PMID: 35897294 PMCID: PMC9331438 DOI: 10.3390/ijerph19158929] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/23/2022] [Revised: 07/20/2022] [Accepted: 07/21/2022] [Indexed: 01/27/2023]
Abstract
(1) Objective: Teaching is a particularly voice-demanding occupation. Voice training provided during teachers’ education is often insufficient and thus teachers are at risk of developing voice disorders. Vocal demands during teaching are not only characterized by speaking for long durations but also by speaking in noisy environments. This provokes the so-called Lombard effect, which intuitively leads to an increase in voice intensity, pitch and phonation time in laboratory studies. However, this effect has not been thoroughly investigated in realistic teaching scenarios. (2) Methods: This study thus examined how 13 experienced, but vocally untrained, teachers behaved when reading in a noisy compared to quiet background environment. The quiet and noisy conditions were provided by a live audience either listening quietly or making noise by talking to each other. By using a portable voice accumulator, the fundamental frequency, sound pressure level of the voice and the noise as well as the phonation time were recorded in both conditions. (3) Results: The results showed that the teachers mainly responded according to the Lombard effect. In addition, analysis of phonation time revealed that they failed to increase inhalation time and appeared to lose articulation through the shortening of voiceless consonants in the noisy condition. (4) Conclusions: The teachers demonstrated vocally demanding behavior when speaking in the noisy condition, which can lead to vocal fatigue and cause dysphonia. The findings underline the necessity for specific voice training in teachers’ education, and the content of such training is discussed in light of the results.
Collapse
|
12
|
Kachlicka M, Laffere A, Dick F, Tierney A. Slow phase-locked modulations support selective attention to sound. Neuroimage 2022; 252:119024. [PMID: 35231629 PMCID: PMC9133470 DOI: 10.1016/j.neuroimage.2022.119024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2022] [Revised: 02/16/2022] [Accepted: 02/19/2022] [Indexed: 11/16/2022] Open
Abstract
To make sense of complex soundscapes, listeners must select and attend to task-relevant streams while ignoring uninformative sounds. One possible neural mechanism underlying this process is alignment of endogenous oscillations with the temporal structure of the target sound stream. Such a mechanism has been suggested to mediate attentional modulation of neural phase-locking to the rhythms of attended sounds. However, such modulations are compatible with an alternate framework, where attention acts as a filter that enhances exogenously-driven neural auditory responses. Here we attempted to test several predictions arising from the oscillatory account by playing two tone streams varying across conditions in tone duration and presentation rate; participants attended to one stream or listened passively. Attentional modulation of the evoked waveform was roughly sinusoidal and scaled with rate, while the passive response did not. However, there was only limited evidence for continuation of modulations through the silence between sequences. These results suggest that attentionally-driven changes in phase alignment reflect synchronization of slow endogenous activity with the temporal structure of attended stimuli.
Collapse
Affiliation(s)
- Magdalena Kachlicka
- Department of Psychological Sciences, Birkbeck, University of London, Malet Street, London WC1E 7HX, England
| | - Aeron Laffere
- Department of Psychological Sciences, Birkbeck, University of London, Malet Street, London WC1E 7HX, England
| | - Fred Dick
- Department of Psychological Sciences, Birkbeck, University of London, Malet Street, London WC1E 7HX, England; Division of Psychology & Language Sciences, UCL, Gower Street, London WC1E 6BT, England
| | - Adam Tierney
- Department of Psychological Sciences, Birkbeck, University of London, Malet Street, London WC1E 7HX, England.
| |
Collapse
|
13
|
Mora JC, Ortega M, Mora-Plaza I, Aliaga-García C. Training the pronunciation of L2 vowels under different conditions: the use of non-lexical materials and masking noise. PHONETICA 2022; 79:1-43. [PMID: 35427446 DOI: 10.1515/phon-2022-2018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
The current study extends traditional perceptual high-variability phonetic training (HVPT) in a foreign language learning context by implementing a comprehensive training paradigm that combines perception (discrimination and identification) and production (immediate repetition) training tasks and by exploring two potentially enhancing training conditions: the use of non-lexical training stimuli and the presence of masking noise during production training. We assessed training effects on L1-Spanish/Catalan bilingual EFL learners' production of a difficult English vowel contrast (/æ/-/ʌ/). The participants (N = 62) were randomly assigned to either non-lexical (N = 24) or lexical (N = 24) training and were further subdivided into two groups, one trained in noise (N = 12) and one in silence (N = 12). An untrained control group (N = 14) was also tested. Training gains, measured through spectral distance scores (Euclidean distances) with respect to native speakers' productions of /æ/ and /ʌ/, were assessed through delayed word and sentence repetition tasks. The results showed an advantage of non-lexical training over lexical training, detrimental effects of noise for participants trained with nonwords, but not for those trained with words, and less accurate production of vowels elicited in isolated words than in words embedded in sentences, where training gains were only observable for participants trained with nonwords.
Collapse
Affiliation(s)
- Joan C Mora
- Department of Modern Languages and Literatures and English Studies, Faculty of Philology and Communication, Universitat de Barcelona, Barcelona, Spain
| | - Mireia Ortega
- Department of Modern Languages and Literatures and English Studies, Faculty of Philology and Communication, Universitat de Barcelona, Barcelona, Spain
| | - Ingrid Mora-Plaza
- Department of Modern Languages and Literatures and English Studies, Faculty of Philology and Communication, Universitat de Barcelona, Barcelona, Spain
| | - Cristina Aliaga-García
- Department of Modern Languages and Literatures and English Studies, Faculty of Philology and Communication, Universitat de Barcelona, Barcelona, Spain
| |
Collapse
|
14
|
Meemann K, Smiljanić R. Intelligibility of Noise-Adapted and Clear Speech in Energetic and Informational Maskers for Native and Nonnative Listeners. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:1263-1281. [PMID: 35235410 DOI: 10.1044/2021_jslhr-21-00175] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
PURPOSE This study explored clear speech (CS) and noise-adapted speech (NAS) intelligibility benefits for native and nonnative English listeners. It also examined how the two speaking style adaptations interact with maskers that vary from purely energetic to largely informational at different signal-to-noise ratios (SNRs). METHOD Materials consisted of 40 sentences produced by 10 young adult talkers in a conversational and a clear speaking style under two conditions: (a) in quiet and (b) in response to speech-shaped noise (SSN) played over headphones (NAS). Young adult native (Experiment 1) and nonnative (Experiment 2) English listeners heard target sentences presented in two-talker (2T) babble, six-talker (6T) babble, or SSN and at an "easier" and a "harder" SNR. RESULTS When talkers produced CS and NAS, word recognition accuracy was significantly improved for both listener groups. The largest intelligibility benefit was obtained for the CS produced in response to noise (CS+NAS). Overall accuracy was highest in 2T babble. Accuracy was higher in SSN than in 6T babble for nonnative listeners at both levels of listening difficulty but only at a more difficult SNR for native listeners. Listeners benefited from CS and NAS most in the presence of SSN and least in 2T babble. When SNRs were the same for the two listener groups, native listeners outperformed nonnative listeners in almost all listening conditions, but nonnative listeners benefited more from CS and NAS in 6T babble than native listeners did. CONCLUSIONS Combined speaking style enhancements, CS+NAS, provided the largest intelligibility increases for native and nonnative listeners in all listening conditions. The results add to the body of evidence supporting speech-oriented, behavioral therapy techniques for maximizing speech intelligibility in everyday listening situations.
Collapse
Affiliation(s)
- Kirsten Meemann
- Department of Linguistics, The University of Texas at Austin
| | - Rajka Smiljanić
- Department of Linguistics, The University of Texas at Austin
| |
Collapse
|
15
|
Scott SK. The neural control of volitional vocal production-from speech to identity, from social meaning to song. Philos Trans R Soc Lond B Biol Sci 2022; 377:20200395. [PMID: 34775825 PMCID: PMC8591378 DOI: 10.1098/rstb.2020.0395] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/05/2023] Open
Abstract
The networks of cortical and subcortical fields that contribute to speech production have benefitted from many years of detailed study, and have been used as a framework for human volitional vocal production more generally. In this article, I will argue that we need to consider speech production as an expression of the human voice in a more general sense. I will also argue that the neural control of the voice can and should be considered to be a flexible system, into which more right hemispheric networks are differentially recruited, based on the factors that are modulating vocal production. I will explore how this flexible network is recruited to express aspects of non-verbal information in the voice, such as identity and social traits. Finally, I will argue that we need to widen out the kinds of vocal behaviours that we explore, if we want to understand the neural underpinnings of the true range of sound-making capabilities of the human voice. This article is part of the theme issue 'Voice modulation: from origin and mechanism to social impact (Part II)'.
Collapse
Affiliation(s)
- Sophie K. Scott
- Institute of Cognitive Neuroscience, University College London, 17 Queen Square, London, UK
| |
Collapse
|
16
|
Tuomainen O, Taschenberger L, Rosen S, Hazan V. Speech modifications in interactive speech: effects of age, sex and noise type. Philos Trans R Soc Lond B Biol Sci 2022; 377:20200398. [PMID: 34775827 PMCID: PMC8591383 DOI: 10.1098/rstb.2020.0398] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/05/2023] Open
Abstract
When attempting to maintain conversations in noisy communicative settings, talkers typically modify their speech to make themselves understood by the listener. In this study, we investigated the impact of background interference type and talker age on speech adaptations, vocal effort and communicative success. We measured speech acoustics (articulation rate, mid-frequency energy, fundamental frequency), vocal effort (correlation between mid-frequency energy and fundamental frequency) and task completion time in 114 participants aged 8-80 years carrying out an interactive problem-solving task in good and noisy listening conditions (quiet, non-speech noise, background speech). We found greater changes in fundamental frequency and mid-frequency energy in non-speech noise than in background speech and similar reductions in articulation rate in both. However, older participants (50+ years) increased vocal effort in both background interference types, whereas younger children (less than 13 years) increased vocal effort only in background speech. The presence of background interference did not lead to longer task completion times. These results suggest that when the background interference involves a higher cognitive load, as in the case of other speech of other talkers, children and older talkers need to exert more vocal effort to ensure successful communication. We discuss these findings within the communication effort framework. This article is part of the theme issue 'Voice modulation: from origin and mechanism to social impact (Part II)'.
Collapse
Affiliation(s)
- Outi Tuomainen
- Speech Hearing and Phonetic Sciences, University College London, 2 Wakefield Street, London WC1N 1PF, UK,Department of Linguistics, University of Potsdam, Haus 14, Karl-Liebknecht-Straße 24-25, 14476 Potsdam, Germany
| | - Linda Taschenberger
- Speech Hearing and Phonetic Sciences, University College London, 2 Wakefield Street, London WC1N 1PF, UK
| | - Stuart Rosen
- Speech Hearing and Phonetic Sciences, University College London, 2 Wakefield Street, London WC1N 1PF, UK
| | - Valerie Hazan
- Speech Hearing and Phonetic Sciences, University College London, 2 Wakefield Street, London WC1N 1PF, UK
| |
Collapse
|
17
|
Keerstock S, Smiljanic R. Reading aloud in clear speech reduces sentence recognition memory and recall for native and non-native talkers. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 150:3387. [PMID: 34852619 DOI: 10.1121/10.0006732] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/29/2020] [Accepted: 09/23/2021] [Indexed: 06/13/2023]
Abstract
Speaking style variation plays a role in how listeners remember speech. Compared to conversational sentences, clearly spoken sentences were better recalled and identified as previously heard by native and non-native listeners. The present study investigated whether speaking style variation also plays a role in how talkers remember speech that they produce. Although distinctive forms of production (e.g., singing, speaking loudly) can enhance memory, the cognitive and articulatory efforts required to plan and produce listener-oriented hyper-articulated clear speech could detrimentally affect encoding and subsequent retrieval. Native and non-native English talkers' memories for sentences that they read aloud in clear and conversational speaking styles were assessed through a sentence recognition memory task (experiment 1; N = 90) and a recall task (experiment 2; N = 75). The results showed enhanced recognition memory and recall for sentences read aloud conversationally rather than clearly for both talker groups. In line with the "effortfulness" hypothesis, producing clear speech may increase the processing load diverting resources from memory encoding. Implications for the relationship between speech perception and production are discussed.
Collapse
Affiliation(s)
- Sandie Keerstock
- Department of Psychological Sciences, University of Missouri, 124 Psychology Building, 200 South 7th Street, Columbia, Missouri 65211, USA
| | - Rajka Smiljanic
- Department of Linguistics, University of Texas at Austin, 305 East 23rd Street STOP B5100, Austin, Texas 78712, USA
| |
Collapse
|
18
|
Rapp M, Cabrera D, Yadav M. Effect of voice support level and spectrum on conversational speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 150:2635. [PMID: 34717507 DOI: 10.1121/10.0006570] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/12/2021] [Accepted: 09/16/2021] [Indexed: 06/13/2023]
Abstract
One's own voice (autophony) is transmitted to the ears as direct airborne sound, bone conduction, and indirect airborne sound from reflections characterized by overall gain and spectro-temporal features. This study investigates how the spectral profile and gain of simulated indirect airborne sound, quantified as voice support (STV), affect the speaking voice of talkers. Pairs of participants performed a conversation elicitation task in anechoic conditions. The indirect airborne sound was provided in real-time via open headphones that maintain the direct airborne transmission path. Experimental conditions included high-pass, low-pass, and all-pass versions of STV, each presented at three overall gains, and a Baseline condition with no electroacoustic contribution to STV. The results show an overall speech level reduction of 0.22 dB for every additional dB of speech-weighted STV, i.e., a -0.22 dB/dB slope. There was some effect of STV spectrum on speech: slope for the high-pass condition was steeper (statistically significant) and significantly different from the all-pass slope; spectral balance (2-4 kHz vs 0-2 kHz) of speech showed an interaction effect between gender and experimental conditions. This paper's findings may inform acoustic treatments in environments where overall sound reduction is of interest for favorable ergonomics and occupational health for voice professionals.
Collapse
Affiliation(s)
- Mary Rapp
- Sydney School of Architecture, Design and Planning, University of Sydney, Sydney, New South Wales 2006, Australia
| | - Densil Cabrera
- Sydney School of Architecture, Design and Planning, University of Sydney, Sydney, New South Wales 2006, Australia
| | - Manuj Yadav
- Sydney School of Architecture, Design and Planning, University of Sydney, Sydney, New South Wales 2006, Australia
| |
Collapse
|
19
|
Keshavarzi M, Varano E, Reichenbach T. Cortical Tracking of a Background Speaker Modulates the Comprehension of a Foreground Speech Signal. J Neurosci 2021; 41:5093-5101. [PMID: 33926996 PMCID: PMC8197648 DOI: 10.1523/jneurosci.3200-20.2021] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2020] [Revised: 02/23/2021] [Accepted: 04/12/2021] [Indexed: 11/21/2022] Open
Abstract
Understanding speech in background noise is a difficult task. The tracking of speech rhythms such as the rate of syllables and words by cortical activity has emerged as a key neural mechanism for speech-in-noise comprehension. In particular, recent investigations have used transcranial alternating current stimulation (tACS) with the envelope of a speech signal to influence the cortical speech tracking, demonstrating that this type of stimulation modulates comprehension and therefore providing evidence of a functional role of the cortical tracking in speech processing. Cortical activity has been found to track the rhythms of a background speaker as well, but the functional significance of this neural response remains unclear. Here we use a speech-comprehension task with a target speaker in the presence of a distractor voice to show that tACS with the speech envelope of the target voice as well as tACS with the envelope of the distractor speaker both modulate the comprehension of the target speech. Because the envelope of the distractor speech does not carry information about the target speech stream, the modulation of speech comprehension through tACS with this envelope provides evidence that the cortical tracking of the background speaker affects the comprehension of the foreground speech signal. The phase dependency of the resulting modulation of speech comprehension is, however, opposite to that obtained from tACS with the envelope of the target speech signal. This suggests that the cortical tracking of the ignored speech stream and that of the attended speech stream may compete for neural resources.SIGNIFICANCE STATEMENT Loud environments such as busy pubs or restaurants can make conversation difficult. However, they also allow us to eavesdrop into other conversations that occur in the background. In particular, we often notice when somebody else mentions our name, even if we have not been listening to that person. However, the neural mechanisms by which background speech is processed remain poorly understood. Here we use transcranial alternating current stimulation, a technique through which neural activity in the cerebral cortex can be influenced, to show that cortical responses to rhythms in the distractor speech modulate the comprehension of the target speaker. Our results provide evidence that the cortical tracking of background speech rhythms plays a functional role in speech processing.
Collapse
Affiliation(s)
- Mahmoud Keshavarzi
- Department of Bioengineering and Centre for Neurotechnology, Imperial College London, South Kensington Campus, London, SW7 2AZ, England
| | - Enrico Varano
- Department of Bioengineering and Centre for Neurotechnology, Imperial College London, South Kensington Campus, London, SW7 2AZ, England
| | - Tobias Reichenbach
- Department of Bioengineering and Centre for Neurotechnology, Imperial College London, South Kensington Campus, London, SW7 2AZ, England
| |
Collapse
|
20
|
Harmon TG, Dromey C, Nelson B, Chapman K. Effects of Background Noise on Speech and Language in Young Adults. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2021; 64:1104-1116. [PMID: 33719537 DOI: 10.1044/2020_jslhr-20-00376] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Purpose The aim of this study was to investigate how different types of background noise that differ in their level of linguistic content affect speech acoustics, speech fluency, and language production for young adult speakers when performing a monologue discourse task. Method Forty young adults monologued by responding to open-ended questions in a silent baseline and five background noise conditions (debate, movie dialogue, contemporary music, classical music, and pink noise). Measures related to speech acoustics (intensity and frequency), speech fluency (speech rate, pausing, and disfluencies), and language production (lexical, morphosyntactic, and macrolinguistic structure) were analyzed and compared across conditions. Participants also reported on which conditions they perceived as more distracting. Results All noise conditions resulted in some change to spoken language compared with the silent baseline. Effects on speech acoustics were consistent with expected changes due to the Lombard effect (e.g., increased intensity and fundamental frequency). Effects on speech fluency showed decreased pausing and increased disfluencies. Several background noise conditions also seemed to interfere with language production. Conclusions Findings suggest that young adults present with both compensatory and interference effects when speaking in noise. Several adjustments may facilitate intelligibility when noise is present and help both speaker and listener maintain attention on the production. Other adjustments provide evidence that background noise eliciting linguistic interference has the potential to degrade spoken language even for healthy young adults, because of increased cognitive demands.
Collapse
Affiliation(s)
- Tyson G Harmon
- Department of Communication Disorders, Brigham Young University, Provo, UT
| | - Christopher Dromey
- Department of Communication Disorders, Brigham Young University, Provo, UT
| | - Brenna Nelson
- Department of Communication Disorders, Brigham Young University, Provo, UT
| | - Kacy Chapman
- Department of Communication Disorders, Brigham Young University, Provo, UT
| |
Collapse
|
21
|
Meekings S, Scott SK. Error in the Superior Temporal Gyrus? A Systematic Review and Activation Likelihood Estimation Meta-Analysis of Speech Production Studies. J Cogn Neurosci 2020; 33:422-444. [PMID: 33326327 DOI: 10.1162/jocn_a_01661] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Evidence for perceptual processing in models of speech production is often drawn from investigations in which the sound of a talker's voice is altered in real time to induce "errors." Methods of acoustic manipulation vary but are assumed to engage the same neural network and psychological processes. This paper aims to review fMRI and PET studies of altered auditory feedback and assess the strength of the evidence these studies provide for a speech error correction mechanism. Studies included were functional neuroimaging studies of speech production in neurotypical adult humans, using natural speech errors or one of three predefined speech manipulation techniques (frequency altered feedback, delayed auditory feedback, and masked auditory feedback). Seventeen studies met the inclusion criteria. In a systematic review, we evaluated whether each study (1) used an ecologically valid speech production task, (2) controlled for auditory activation caused by hearing the perturbation, (3) statistically controlled for multiple comparisons, and (4) measured behavioral compensation correlating with perturbation. None of the studies met all four criteria. We then conducted an activation likelihood estimation meta-analysis of brain coordinates from 16 studies that reported brain responses to manipulated over unmanipulated speech feedback, using the GingerALE toolbox. These foci clustered in bilateral superior temporal gyri, anterior to cortical fields typically linked to error correction. Within the limits of our analysis, we conclude that existing neuroimaging evidence is insufficient to determine whether error monitoring occurs in the posterior superior temporal gyrus regions proposed by models of speech production.
Collapse
|
22
|
Calandruccio L, Porter HL, Leibold LJ, Buss E. The Clear-Speech Benefit for School-Age Children: Speech-in-Noise and Speech-in-Speech Recognition. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:4265-4276. [PMID: 33151767 PMCID: PMC8608216 DOI: 10.1044/2020_jslhr-20-00353] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/22/2020] [Revised: 08/22/2020] [Accepted: 08/24/2020] [Indexed: 06/11/2023]
Abstract
Purpose Talkers often modify their speech when communicating with individuals who struggle to understand speech, such as listeners with hearing loss. This study evaluated the benefit of clear speech in school-age children and adults with normal hearing for speech-in-noise and speech-in-speech recognition. Method Masked sentence recognition thresholds were estimated for school-age children and adults using an adaptive procedure. In Experiment 1, the target and masker were summed and presented over a loudspeaker located directly in front of the listener. The masker was either speech-shaped noise or two-talker speech, and target sentences were produced using a clear or conversational speaking style. In Experiment 2, stimuli were presented over headphones. The two-talker speech masker was diotic (M0). Clear and conversational target sentences were presented either in-phase (T0) or out-of-phase (Tπ) between the two ears. The M0Tπ condition introduces a segregation cue that was expected to improve performance. Results For speech presented over a single loudspeaker (Experiment 1), the clear-speech benefit was independent of age for the noise masker, but it increased with age for the two-talker masker. Similar age effects for the two-talker speech masker were seen under headphones with diotic presentation (M0T0), but comparable clear-speech benefit as a function of age was observed with a binaural cue to facilitate segregation (M0Tπ). Conclusions Consistent with prior research, children showed a robust clear-speech benefit for speech-in-noise recognition. Immaturity in the ability to segregate target from masker speech may limit young children's ability to benefit from clear-speech modifications for speech-in-speech recognition under some conditions. When provided with a cue that facilitates segregation, children as young as 4-7 years of age derived a clear-speech benefit in a two-talker masker that was similar to the benefit experienced by adults.
Collapse
Affiliation(s)
- Lauren Calandruccio
- Department of Psychological Sciences, Case Western Reserve University, Cleveland, OH
| | - Heather L. Porter
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, NE
| | - Lori J. Leibold
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, NE
| | - Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, The University of North Carolina at Chapel Hill
| |
Collapse
|
23
|
Laffere A, Dick F, Holt LL, Tierney A. Attentional modulation of neural entrainment to sound streams in children with and without ADHD. Neuroimage 2020; 224:117396. [PMID: 32979522 DOI: 10.1016/j.neuroimage.2020.117396] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2020] [Revised: 08/25/2020] [Accepted: 09/14/2020] [Indexed: 01/06/2023] Open
Abstract
To extract meaningful information from complex auditory scenes like a noisy playground, rock concert, or classroom, children can direct attention to different sound streams. One means of accomplishing this might be to align neural activity with the temporal structure of a target stream, such as a specific talker or melody. However, this may be more difficult for children with ADHD, who can struggle with accurately perceiving and producing temporal intervals. In this EEG study, we found that school-aged children's attention to one of two temporally-interleaved isochronous tone 'melodies' was linked to an increase in phase-locking at the melody's rate, and a shift in neural phase that aligned the neural responses with the attended tone stream. Children's attention task performance and neural phase alignment with the attended melody were linked to performance on temporal production tasks, suggesting that children with more robust control over motor timing were better able to direct attention to the time points associated with the target melody. Finally, we found that although children with ADHD performed less accurately on the tonal attention task than typically developing children, they showed the same degree of attentional modulation of phase locking and neural phase shifts, suggesting that children with ADHD may have difficulty with attentional engagement rather than attentional selection.
Collapse
Affiliation(s)
- Aeron Laffere
- Department of Psychological Sciences, Birkbeck, University of London, Malet Street, London, WC1E 7HX, United Kingdom
| | - Fred Dick
- Department of Psychological Sciences, Birkbeck, University of London, Malet Street, London, WC1E 7HX, United Kingdom; Division of Psychology & Language Sciences, UCL, Gower Street, London, WC1E 6BT, United Kingdom
| | - Lori L Holt
- Department of Psychology, Carnegie Mellon University, 5000 Forbes Avenue, Pittsburgh, PA 15213, United States
| | - Adam Tierney
- Department of Psychological Sciences, Birkbeck, University of London, Malet Street, London, WC1E 7HX, United Kingdom.
| |
Collapse
|
24
|
Understanding Lombard speech: a review of compensation techniques towards improving speech based recognition systems. Artif Intell Rev 2020. [DOI: 10.1007/s10462-020-09907-5] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
25
|
Borrie SA, Wynn CJ, Berisha V, Lubold N, Willi MM, Coelho CA, Barrett TS. Conversational Coordination of Articulation Responds to Context: A Clinical Test Case With Traumatic Brain Injury. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:2567-2577. [PMID: 32755503 PMCID: PMC7872735 DOI: 10.1044/2020_jslhr-20-00104] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/04/2020] [Revised: 04/15/2020] [Accepted: 05/14/2020] [Indexed: 05/19/2023]
Abstract
Purpose Coordination of communicative behavior supports shared understanding in conversation. The current study brings together analysis of two speech coordination strategies, entrainment and compensation of articulation, in a preliminary investigation into whether strategy organization is shaped by a challenging communicative context-conversing with a person who has a communication disorder. Method As an initial clinical test case, an automated measure of articulatory precision was analyzed in a corpus of spoken dialogue, where a confederate conversed with participants with traumatic brain injury (n = 28) and participants with no brain injury (n = 48). Results Overall, the confederate engaged in significant entrainment and high compensation (hyperarticulation) in conversations with participants with traumatic brain injury relative to significant entrainment and low compensation (hypoarticulation) in conversations with participants with no brain injury. Furthermore, the confederate's articulatory precision changed over the course of the conversations. Conclusions Findings suggest that the organization of conversational coordination is sensitive to context, supporting synergistic models of spoken dialogue. While corpus limitations are acknowledged, these initial results point to differences in the way in which speech strategies are realized in challenging communicative contexts, highlighting a viable and important target for investigation with clinical populations. A framework for investigating speech coordination strategies in tandem and ideas for advancing this line of inquiry serve as key contributions of this work.
Collapse
Affiliation(s)
- Stephanie A. Borrie
- Department of Communicative Disorders and Deaf Education, Utah State University, Logan
| | - Camille J. Wynn
- Department of Communicative Disorders and Deaf Education, Utah State University, Logan
| | - Visar Berisha
- School of Electrical, Computer, and Energy Engineering, Arizona State University, Tempe
| | - Nichola Lubold
- School of Electrical, Computer, and Energy Engineering, Arizona State University, Tempe
| | - Megan M. Willi
- Communication Sciences and Disorders Program, California State University, Chico
| | - Carl A. Coelho
- Department of Speech, Language, and Hearing Sciences, University of Connecticut, Storrs
| | | |
Collapse
|
26
|
Pirilä S, Jokitulppo J, Niemitalo-Haapola E, Yliherva A, Rantala L. Teachers' and Children's Experiences after an Acoustic Intervention and a Noise-Controlling Workshop in Two Elementary Classrooms. Folia Phoniatr Logop 2019; 72:454-463. [PMID: 31639814 DOI: 10.1159/000503231] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2019] [Accepted: 09/05/2019] [Indexed: 11/19/2022] Open
Abstract
INTRODUCTION Treating sources of noise is a novel aspect of voice ergonomics intended to enhance the preconditions for good voice production and easy listening. OBJECTIVE To improve experiences of listening and voice ergonomics in classrooms. METHODS Participants were two female elementary school teachers with voice symptoms and their pupils (n = 50). Two interventions were performed: the acoustic intervention and then the workshop intervention where the teachers and pupils were active. Teachers' voice symptoms and pupils' and teachers' experiences of the interventions were elicited by questionnaire. RESULTS The teacher with many voice symptoms experienced more annoyance from sounds and benefitted more from the interventions. After the interventions both teachers suffered fewer voice breaks and voice symptoms such as lump and mucus in the throat. The pupils reported improvement in the teachers' voice clarity and audibility (p = 0.001). Pupils aged 12-13 years were more annoyed by sounds than those aged 8-9 years (p = 0.003). The older pupils experienced less sound annoyance after both interventions and the younger ones after the workshop intervention. CONCLUSIONS The importance of good acoustics and individuals' ability to improve voice ergonomics and listening conditions was demonstrated.
Collapse
Affiliation(s)
- Sirpa Pirilä
- Faculty of Humanities/Logopedics, Child Language Research, University of Oulu, Oulu, Finland,
| | | | - Elina Niemitalo-Haapola
- Faculty of Humanities/Logopedics, Child Language Research, University of Oulu, Oulu, Finland
| | - Anneli Yliherva
- Faculty of Humanities/Logopedics, Child Language Research, University of Oulu, Oulu, Finland
| | - Leena Rantala
- Faculty of Social Sciences, University of Tampere, Tampere, Finland
| |
Collapse
|
27
|
Abstract
Humans exhibit a high level of vocal plasticity in speech production, which allows us to acquire both native and foreign languages and dialects, and adapt to local accents in social communication. In comparison, non-human primates exhibit limited vocal plasticity, especially in adulthood, which would limit their ability to adapt to different social and environmental contexts in vocal communication. Here, we quantitatively examined the ability of adult common marmosets (Callithrix jacchus), a highly vocal New World primate species, to modulate their vocal production in social contexts. While recent studies have demonstrated vocal learning in developing marmosets, we know much less about the extent of vocal learning and plasticity in adult marmosets. We found, in the present study, that marmosets were able to adaptively modify the spectrotemporal structure of their vocalizations when they encountered interfering sounds. Our experiments showed that marmosets shifted the spectrum of their vocalizations away from the spectrum of the interfering sounds in order to avoid the overlap. More interestingly, we found that marmosets made predictive and long-lasting spectral shifts in their vocalizations after they had experienced a particular type of interfering sound. These observations provided evidence for directional control of the vocalization spectrum and long-term vocal plasticity by adult marmosets. The findings reported here have important implications for the ability of this New World primate species in voluntarily and adaptively controlling their vocal production in social communication.
Collapse
Affiliation(s)
- Lingyun Zhao
- 1 Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, The Johns Hopkins University School of Medicine , Baltimore, MD 21205 , USA
| | - Bahar Boroumand Rad
- 2 Department of Biological Sciences, Towson University , Towson, MD 21252 , USA
| | - Xiaoqin Wang
- 1 Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, The Johns Hopkins University School of Medicine , Baltimore, MD 21205 , USA
| |
Collapse
|
28
|
Zellou G, Scarborough R. Neighborhood-conditioned phonetic enhancement of an allophonic vowel split. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 145:3675. [PMID: 31255131 DOI: 10.1121/1.5113582] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/16/2019] [Accepted: 06/04/2019] [Indexed: 06/09/2023]
Abstract
Vowels are enhanced via vowel-space expansion in perceptually difficult contexts, including in words subject to greater lexical competition. Yet, vowel hyperarticulation often covaries with other acoustic adjustments, such as increased nasal coarticulation, suggesting that the goals of phonetic enhancement are not strictly to produce canonical phoneme realizations. This study explores phonetic enhancement by examining how speakers realize an allophonic vowel split in lexically challenging conditions. Specifically, in US English, /æ/ is raising before nasal codas, such that pre-nasal and pre-oral /æ/ are moving apart. Speakers produced monosyllabic words varying in phonological neighborhood density (ND), a measure of lexical difficulty, with CæN or CæC structure to a real listener interlocutor in an interactive task. Acoustic analyses reveal that speakers enhance pre-oral /æ/ by lowering it in Hi ND words; meanwhile, pre-nasal /æ/ Hi ND words are produced with greater degrees of nasalization and increased diphthongization. These patterns indicate that ND-conditioned phonetic enhancement is realized in targeted ways for distinct allophones of /æ/. Results support views of hyperarticulation in which the goal is to make words, that is, segments in their contexts, as distinct as possible.
Collapse
Affiliation(s)
- Georgia Zellou
- University of California at Davis, Linguistics Department, 469 Kerr Hall, One Shields Avenue, Davis, California 95616, USA
| | - Rebecca Scarborough
- University of Colorado, Boulder, Linguistics Department, 295 University of Colorado Boulder, Boulder, Colorado 80309, USA
| |
Collapse
|
29
|
Beechey T, Buchholz JM, Keidser G. Eliciting Naturalistic Conversations: A Method for Assessing Communication Ability, Subjective Experience, and the Impacts of Noise and Hearing Impairment. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:470-484. [PMID: 30950689 DOI: 10.1044/2018_jslhr-h-18-0107] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
Purpose The purpose of this study was to introduce a method of eliciting conversational behavior with many aspects of realism, which may be used to study the impacts of hearing impairment and noise on verbal communication; to describe the characteristics of speech and language participants produced during the task; and to assess participants' engagement and motivation while completing the task. Method Twenty young adults with normal hearing and 20 older adults with hearing impairment took part in face-to-face conversations while completing a referential communication puzzle task designed to elicit natural conversational speech production and language with a number of realistic characteristics. Participants rated the difficulty and relevance of acoustic scenes for communication and their engagement in conversations. Results The communication task elicited speech production in a natural conversational register and language with many realistic characteristics, including complex linguistic constructions and typical disfluencies found in everyday speech, and approximately balanced contributions within dyads. Subjective ratings suggest that the task is robust to learning and fatigue effects and that participants remained highly engaged throughout the experiment. All participants were able to maintain successful communication regardless of background noise level and degree of hearing impairment. Conclusions The communication task described here may be used as part of a functional assessment of the ability to communicate in the presence of noise and hearing impairment. Although existing speech assessments have many strengths, they do not take into account the inherently interactive nature of spoken communication or the effects of motivation and engagement.
Collapse
Affiliation(s)
- Timothy Beechey
- The HEARing CRC, Carlton, Victoria, Australia
- National Acoustic Laboratories, Macquarie Park, New South Wales, Australia
- Department of Linguistics, Macquarie University, Macquarie Park, New South Wales, Australia
| | - Jörg M Buchholz
- The HEARing CRC, Carlton, Victoria, Australia
- National Acoustic Laboratories, Macquarie Park, New South Wales, Australia
- Department of Linguistics, Macquarie University, Macquarie Park, New South Wales, Australia
| | - Gitte Keidser
- The HEARing CRC, Carlton, Victoria, Australia
- National Acoustic Laboratories, Macquarie Park, New South Wales, Australia
- Department of Audiology, The University of Queensland, St. Lucia, Australia
| |
Collapse
|
30
|
Hazan V, Tuomainen O, Kim J, Davis C, Sheffield B, Brungart D. Clear speech adaptations in spontaneous speech produced by young and older adults. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 144:1331. [PMID: 30424655 DOI: 10.1121/1.5053218] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/20/2018] [Accepted: 08/17/2018] [Indexed: 06/09/2023]
Abstract
The study investigated the speech adaptations by older adults (OA) with and without age-related hearing loss made to communicate effectively in challenging communicative conditions. Acoustic analyses were carried out on spontaneous speech produced during a problem-solving task (diapix) carried out by talker pairs in different listening conditions. There were 83 talkers of Southern British English. Fifty-seven talkers were OAs aged 65-84, 30 older adults with normal hearing (OANH), and 27 older adults with hearing loss (OAHL) [mean pure tone average (PTA) 0.250-4 kHz: 27.7 dB HL]. Twenty-six talkers were younger adults (YA) aged 18-26 with normal hearing. Participants were recorded while completing the diapix task with a conversational partner (YA of the same sex) when (a) both talkers heard normally (NORM), (b) the partner had a simulated hearing loss, and (c) both talkers heard babble noise. Irrespective of hearing status, there were age-related differences in some acoustic characteristics of YA and OA speech produced in NORM, most likely linked to physiological factors. In challenging conditions, while OANH talkers typically patterned with YA talkers, OAHL talkers made adaptations more consistent with an increase in vocal effort. The study suggests that even mild presbycusis in healthy OAs can affect the speech adaptations made to maintain effective communication.
Collapse
Affiliation(s)
- Valerie Hazan
- Department of Speech Hearing and Phonetic Sciences, University College London, 2 Wakefield Street, London WC1N 1PF, United Kingdom
| | - Outi Tuomainen
- Department of Speech Hearing and Phonetic Sciences, University College London, 2 Wakefield Street, London WC1N 1PF, United Kingdom
| | - Jeesun Kim
- The MARCS Institute, Western Sydney University, Locked Bag 1797, Penrith, New South Wales 2751, Australia
| | - Christopher Davis
- The MARCS Institute, Western Sydney University, Locked Bag 1797, Penrith, New South Wales 2751, Australia
| | - Benjamin Sheffield
- Audiology and Speech-Pathology Center, Walter Reed National Military Medical Center, Bethesda, 4494 North Palmer Road, Bethesda, Maryland 20889, USA
| | - Douglas Brungart
- Audiology and Speech-Pathology Center, Walter Reed National Military Medical Center, Bethesda, 4494 North Palmer Road, Bethesda, Maryland 20889, USA
| |
Collapse
|
31
|
Garnier M, Ménard L, Alexandre B. Hyper-articulation in Lombard speech: An active communicative strategy to enhance visible speech cues? THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 144:1059. [PMID: 30180713 DOI: 10.1121/1.5051321] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/22/2017] [Accepted: 08/02/2018] [Indexed: 06/08/2023]
Abstract
This study investigates the hypothesis that speakers make active use of the visual modality in production to improve their speech intelligibility in noisy conditions. Six native speakers of Canadian French produced speech in quiet conditions and in 85 dB of babble noise, in three situations: interacting face-to-face with the experimenter (AV), using the auditory modality only (AO), or reading aloud (NI, no interaction). The audio signal was recorded with the three-dimensional movements of their lips and tongue, using electromagnetic articulography. All the speakers reacted similarly to the presence vs absence of communicative interaction, showing significant speech modifications with noise exposure in both interactive and non-interactive conditions, not only for parameters directly related to voice intensity or for lip movements (very visible) but also for tongue movements (less visible); greater adaptation was observed in interactive conditions, though. However, speakers reacted differently to the availability or unavailability of visual information: only four speakers enhanced their visible articulatory movements more in the AV condition. These results support the idea that the Lombard effect is at least partly a listener-oriented adaptation. However, to clarify their speech in noisy conditions, only some speakers appear to make active use of the visual modality.
Collapse
Affiliation(s)
- Maëva Garnier
- Centre National de la Recherche Scientifique, Laboratoire Grenoble Images Parole Signal Automatique, 11 rue des Mathématiques, Grenoble Campus, Boîte Postale 46, F-38402 Saint Martin d'Hères Cedex, France
| | - Lucie Ménard
- Département de Linguistique, Laboratoire de Phonétique, Center for Research on Brain, Language, and Music, Université du Québec à Montréal, 320, Ste-Catherine Est, Montréal, Quebec H2X 1L7, Canada
| | - Boris Alexandre
- Centre National de la Recherche Scientifique, Laboratoire Grenoble Images Parole Signal Automatique, 11 rue des Mathématiques, Grenoble Campus, Boîte Postale 46, F-38402 Saint Martin d'Hères Cedex, France
| |
Collapse
|
32
|
Hazan V, Tuomainen O, Tu L, Kim J, Davis C, Brungart D, Sheffield B. How do aging and age-related hearing loss affect the ability to communicate effectively in challenging communicative conditions? Hear Res 2018; 369:33-41. [PMID: 29941310 DOI: 10.1016/j.heares.2018.06.009] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/31/2017] [Revised: 05/09/2018] [Accepted: 06/14/2018] [Indexed: 11/15/2022]
Abstract
This study investigated the relation between the intelligibility of conversational and clear speech produced by older and younger adults and (a) the acoustic profile of their speech (b) communication effectiveness. Speech samples from 30 talkers from the elderLUCID corpus were used: 10 young adults (YA), 10 older adults with normal hearing (OANH) and 10 older adults with presbycusis (OAHL). Samples were extracted from recordings made while participants completed a problem-solving cooperative task (diapix) with a conversational partner who could either hear them easily (NORM) or via a simulated hearing loss (HLS), which led talkers to naturally adopt a clear speaking style. In speech-in-noise listening experiments involving 21 young adult listeners, speech samples by OANH and OAHL were rated and perceived as less intelligible than those of YA talkers. HLS samples were more intelligible than NORM samples, with greater improvements in intelligibility across conditions seen for OA speech. The presence of presbycusis affected (a) the clear speech strategies adopted by OAHL talkers and (b) task effectiveness: OAHL talkers showed some adaptations consistent with an increase in vocal effort, and it took them significantly longer than the YA group to complete the diapix task. The relative energy in the 1-3 kHz frequency region of the long-term average spectrum was the feature that best predicted: (a) the intelligibility of speech samples, and (b) task transaction time in the HLS condition. Overall, our study suggests that spontaneous speech produced by older adults is less intelligible in babble noise, probably due to less energy present in the 1-3 kHz frequency range rich in acoustic cues. Even mild presbycusis in 'healthy aged' adults can affect the dynamic adaptations in speech that are beneficial for effective communication.
Collapse
Affiliation(s)
- Valerie Hazan
- Department of Speech Hearing and Phonetic Sciences, Chandler House, UCL, 2 Wakefield Street, London WC1N 1PF, UK.
| | - Outi Tuomainen
- Department of Speech Hearing and Phonetic Sciences, Chandler House, UCL, 2 Wakefield Street, London WC1N 1PF, UK.
| | - Lilian Tu
- Department of Speech Hearing and Phonetic Sciences, Chandler House, UCL, 2 Wakefield Street, London WC1N 1PF, UK.
| | - Jeesun Kim
- The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Locked Bag 1797, Penrith NSW 2751, Australia.
| | - Chris Davis
- The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Locked Bag 1797, Penrith NSW 2751, Australia.
| | - Douglas Brungart
- Audiology and Speech-Pathology Center, Walter Reed National Military Medical Center, Bethesda, 4494 North Palmer Road, Bethesda, MD 20889, USA.
| | - Benjamin Sheffield
- Audiology and Speech-Pathology Center, Walter Reed National Military Medical Center, Bethesda, 4494 North Palmer Road, Bethesda, MD 20889, USA.
| |
Collapse
|
33
|
Granlund S, Hazan V, Mahon M. Children's Acoustic and Linguistic Adaptations to Peers With Hearing Impairment. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2018; 61:1055-1069. [PMID: 29710271 DOI: 10.1044/2017_jslhr-s-16-0456] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/19/2016] [Accepted: 10/13/2017] [Indexed: 06/08/2023]
Abstract
PURPOSE This study aims to examine the clear speaking strategies used by older children when interacting with a peer with hearing loss, focusing on both acoustic and linguistic adaptations in speech. METHOD The Grid task, a problem-solving task developed to elicit spontaneous interactive speech, was used to obtain a range of global acoustic and linguistic measures. Eighteen 9- to 14-year-old children with normal hearing (NH) performed the task in pairs, once with a friend with NH and once with a friend with a hearing impairment (HI). RESULTS In HI-directed speech, children increased their fundamental frequency range and midfrequency intensity, decreased the number of words per phrase, and expanded their vowel space area by increasing F1 and F2 range, relative to NH-directed speech. However, participants did not appear to make changes to their articulation rate, the lexical frequency of content words, or lexical diversity when talking to their friend with HI compared with their friend with NH. CONCLUSIONS Older children show evidence of listener-oriented adaptations to their speech production; although their speech production systems are still developing, they are able to make speech adaptations to benefit the needs of a peer with HI, even without being given a specific instruction to do so. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.6118817.
Collapse
Affiliation(s)
- Sonia Granlund
- Speech, Hearing & Phonetic Sciences, University College London, United Kingdom
| | - Valerie Hazan
- Speech, Hearing & Phonetic Sciences, University College London, United Kingdom
| | - Merle Mahon
- Language & Cognition, University College London, United Kingdom
| |
Collapse
|
34
|
|
35
|
Agnew ZK, McGettigan C, Banks B, Scott SK. Group and individual variability in speech production networks during delayed auditory feedback. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 143:3009. [PMID: 29857719 PMCID: PMC5963950 DOI: 10.1121/1.5026500] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/21/2016] [Revised: 02/05/2018] [Accepted: 02/12/2018] [Indexed: 06/08/2023]
Abstract
Altering reafferent sensory information can have a profound effect on motor output. Introducing a short delay [delayed auditory feedback (DAF)] during speech production results in modulations of voice and loudness, and produces a range of speech dysfluencies. The ability of speakers to resist the effects of delayed feedback is variable yet it is unclear what neural processes underlie differences in susceptibility to DAF. Here, susceptibility to DAF is investigated by looking at the neural basis of within and between subject changes in speech fluency under 50 and 200 ms delay conditions. Using functional magnetic resonance imaging, networks involved in producing speech under two levels of DAF were identified, lying largely within networks active during normal speech production. Independent of condition, fluency ratings were associated with midbrain activity corresponding to periaqueductal grey matter. Across subject variability in ability to produce normal sounding speech under a 200 ms delay was associated with activity in ventral sensorimotor cortices, whereas ability to produce normal sounding speech under a 50 ms delay was associated with left inferior frontal gyrus activity. These data indicate whilst overlapping cortical mechanisms are engaged for speaking under different delay conditions, susceptibility to different temporal delays in speech feedback may involve different processes.
Collapse
Affiliation(s)
- Z K Agnew
- Institute for Cognitive Neuroscience, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom
| | - C McGettigan
- Institute for Cognitive Neuroscience, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom
| | - B Banks
- Institute for Cognitive Neuroscience, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom
| | - S K Scott
- Institute for Cognitive Neuroscience, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom
| |
Collapse
|
36
|
Hay J. Sociophonetics: The Role of Words, the Role of Context, and the Role of Words in Context. Top Cogn Sci 2018; 10:696-706. [PMID: 29498479 DOI: 10.1111/tops.12326] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2016] [Revised: 10/16/2017] [Accepted: 12/27/2017] [Indexed: 11/29/2022]
Abstract
This paper synthesizes a wide range of literature from sociolinguistics and cognitive psychology, to argue for a central role for the "word" as a vehicle of language variation and change. Three crucially interlinked strands of research are reviewed-the role of context in associative learning, the word-level storage of phonetic and contextual detail, and the phonetic consequences of skewed distributions of words across different contexts. I argue that the human capacity for associative learning, combined with attention to fine-phonetic detail at the level of the word, plays a significant role in predicting a range of subtle but systematically robust observed socioindexical patterns in speech production and perception.
Collapse
Affiliation(s)
- Jennifer Hay
- New Zealand Institute of Language, Brain and Behaviour, University of Canterbury
| |
Collapse
|
37
|
Smiljanic R, Gilbert RC. Acoustics of Clear and Noise-Adapted Speech in Children, Young, and Older Adults. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2017; 60:3081-3096. [PMID: 29075775 DOI: 10.1044/2017_jslhr-s-16-0130] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/04/2016] [Accepted: 05/08/2017] [Indexed: 06/07/2023]
Abstract
PURPOSE This study investigated acoustic-phonetic modifications produced in noise-adapted speech (NAS) and clear speech (CS) by children, young adults, and older adults. METHOD Ten children (11-13 years of age), 10 young adults (18-29 years of age), and 10 older adults (60-84 years of age) read sentences in conversational and clear speaking style in quiet and in noise. A number of acoustic measurements were obtained. RESULTS NAS and CS were characterized by a decrease in speaking rate and an increase in 1-3 kHz energy, sound pressure level (SPL), vowel space area (VSA), and harmonics-to-noise ratio. NAS increased fundamental frequency (F0) mean and decreased jitter and shimmer. CS increased frequency and duration of pauses. Older adults produced the slowest speaking rate, longest pauses, and smallest increase in F0 mean, 1-3 kHz energy, and SPL when speaking clearly. They produced the smallest increases in VSA in NAS and CS. Children slowed down less, increased the VSA least, increased harmonics-to-noise ratio, and decreased jitter and shimmer most in CS. Children increased mean F0 and F1 most in noise. CONCLUSIONS Findings have implications for a model of speech production in healthy speakers as well as the potential to aid in clinical decision making for individuals with speech disorders, particularly dysarthria.
Collapse
|
38
|
Smiljanic R, Gilbert RC. Intelligibility of Noise-Adapted and Clear Speech in Child, Young Adult, and Older Adult Talkers. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2017; 60:3069-3080. [PMID: 29075748 DOI: 10.1044/2017_jslhr-s-16-0165] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/25/2016] [Accepted: 04/21/2017] [Indexed: 06/07/2023]
Abstract
PURPOSE This study examined intelligibility of conversational and clear speech sentences produced in quiet and in noise by children, young adults, and older adults. Relative talker intelligibility was assessed across speaking styles. METHOD Sixty-one young adult participants listened to sentences mixed with speech-shaped noise at -5 dB signal-to-noise ratio. The analyses examined percent correct scores across conversational, clear, and noise-adapted conditions and the three talker groups. Correlation analyses examined whether talker intelligibility is consistent across speaking style adaptations. RESULTS Noise-adapted and clear speech significantly enhanced intelligibility for young adult listeners. The intelligibility improvement varied across the three talker groups. Notably, intelligibility benefit was smallest for children's speaking style modifications. Listeners also perceived speech produced in noise by older adults to be less intelligible compared to the younger talkers. Talker intelligibility was correlated strongly between conversational and clear speech in quiet, but not for conversational speech produced in quiet and in noise. CONCLUSIONS Results provide evidence that intelligibility variation related to age and communicative barrier has the potential to aid clinical decision making for individuals with speech disorders, particularly dysarthria.
Collapse
|
39
|
Tang P, Xu Rattanasone N, Yuen I, Demuth K. Phonetic enhancement of Mandarin vowels and tones: Infant-directed speech and Lombard speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:493. [PMID: 28863611 DOI: 10.1121/1.4995998] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Speech units are reported to be hyperarticulated in both infant-directed speech (IDS) and Lombard speech. Since these two registers have typically been studied separately, it is unclear if the same speech units are hyperarticulated in the same manner between these registers. The aim of the present study is to compare the effect of register on vowel and tone modification in the tonal language Mandarin Chinese. Vowel and tone productions were produced by 15 Mandarin-speaking mothers during interactions with their 12-month-old infants during a play session (IDS), in conversation with a Mandarin-speaking adult in a 70 dBA eight-talker babble noise environment (Lombard speech), and in a quiet environment (adult-directed speech). Vowel space expansion was observed in IDS and Lombard speech, however, the patterns of vowel-shift were different between the two registers. IDS displayed tone space expansion only in the utterance-final position, whereas there was no tone space expansion in Lombard speech. The overall pitch increased for all tones in both registers. The tone-bearing vowel duration also increased in both registers, but only in utterance-final position. The difference in speech modifications between these two registers is discussed in light of speakers' different communicative needs.
Collapse
Affiliation(s)
- Ping Tang
- Department of Linguistics, ARC Centre of Excellence in Cognition and its Disorders, Macquarie University, Sydney, 16 University Avenue, Australian Hearing Hub, Balaclava Road, North Ryde, New South Wales 2109 Australia
| | - Nan Xu Rattanasone
- Department of Linguistics, ARC Centre of Excellence in Cognition and its Disorders, Macquarie University, Sydney, 16 University Avenue, Australian Hearing Hub, Balaclava Road, North Ryde, New South Wales 2109 Australia
| | - Ivan Yuen
- Department of Linguistics, ARC Centre of Excellence in Cognition and its Disorders, Macquarie University, Sydney, 16 University Avenue, Australian Hearing Hub, Balaclava Road, North Ryde, New South Wales 2109 Australia
| | - Katherine Demuth
- Department of Linguistics, ARC Centre of Excellence in Cognition and its Disorders, Macquarie University, Sydney, 16 University Avenue, Australian Hearing Hub, Balaclava Road, North Ryde, New South Wales 2109 Australia
| |
Collapse
|
40
|
Graetzer S, Bottalico P, Hunter EJ. Speech produced in noise: Relationship between listening difficulty and acoustic and durational parameters. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:974. [PMID: 28863615 PMCID: PMC5648561 DOI: 10.1121/1.4997906] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/10/2016] [Revised: 05/17/2017] [Accepted: 07/20/2017] [Indexed: 06/02/2023]
Abstract
Conversational speech produced in noise can be characterised by increases in intelligibility relative to such speech produced in quiet. Listening difficulty (LD) is a metric that can be used to evaluate speech transmission performance more sensitively than intelligibility scores in situations in which performance is likely to be high. The objectives of the present study were to evaluate the LD of speech produced in different noise and style conditions, to evaluate the spectral and durational speech modifications associated with these conditions, and to determine whether any of the spectral and durational parameters predicted LD. Nineteen subjects were instructed to speak at normal and loud volumes in the presence of background noise at 40.5 dB(A) and babble noise at 61 dB(A). The speech signals were amplitude-normalised, combined with pink noise to obtain a signal-to-noise ratio of -6 dB, and presented to twenty raters who judged their LD. Vowel duration, fundamental frequency and the proportion of the spectral energy in high vs low frequencies increased with the noise level within both styles. LD was lowest when the speech was produced in the presence of high level noise and at a loud volume, indicating improved intelligibility. Spectrum balance was observed to predict LD.
Collapse
Affiliation(s)
- Simone Graetzer
- Acoustics Research Unit, School of Architecture, University of Liverpool, Liverpool, England
| | - Pasquale Bottalico
- Voice Biomechanics and Acoustics Laboratory, Department of Communicative Sciences and Disorders, Michigan State University, East Lansing, Michigan 48824, USA
| | - Eric J Hunter
- Voice Biomechanics and Acoustics Laboratory, Department of Communicative Sciences and Disorders, Michigan State University, East Lansing, Michigan 48824, USA
| |
Collapse
|
41
|
Benuš Š, Šimko J. Stability and Variability in Slovak Prosodic Boundaries. PHONETICA 2017; 73:163-193. [PMID: 28208129 DOI: 10.1159/000446350] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/03/2015] [Accepted: 04/15/2016] [Indexed: 06/06/2023]
Abstract
BACKGROUND/AIM Encoding intended meanings in the type and strength of prosodic boundaries and strategies for communicating these meanings in ambient noise use similar prosodic cues. We analyze how increasing the level of ambient noise affects the realization of Slovak prosodic boundaries. METHODS Five native speakers of Slovak read sentences, manipulating the boundary type (weak, rise, fall) and the location of pre-boundary pitch accent. Ambient noise of several levels was administered via headphones. Acoustic and articulatory data (electromagnetometry) were collected. RESULTS Under normal condition, boundary strength is signaled with longer pre-boundary rhymes, more frequent pauses, greater crossboundary f0 resets and jaw displacement. The strength of falls is realized in crossboundary features (pauses, f0 reset), and rises in pre-boundary features (rhyme duration, f0 range). Pitch-accented rhymes are strengthened in all features, but f0 range. In noise, the increase in boundary strength is weak, and falls strengthen more than rises. F0 targets for falls and rises are adjusted in addition to noiseinduced global f0 scaling and lengthening. CONCLUSION Hyper-articulation of prosodic boundaries in ambient noise is not robust and uniform; rather, durational, f0 and jaw displacement features co-create complex prosodic patterns in a complementary and synergetic manner based on affordances in normal speech.
Collapse
Affiliation(s)
- Štefan Benuš
- Constantine the Philosopher University, Nitra, Slovakia
| | | |
Collapse
|
42
|
Culling JF. Speech intelligibility in virtual restaurants. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 140:2418. [PMID: 27794329 DOI: 10.1121/1.4964401] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/22/2023]
Abstract
Speech reception thresholds (SRTs) for a target voice on the same virtual table were measured in various restaurant simulations under conditions of masking by between one and eight interferers at other tables. Results for different levels of reverberation and different simulation techniques were qualitatively similar. SRTs increased steeply with the number of interferers, reflecting progressive failure to perceptually unmask the target speech as the acoustic scene became more complex. For a single interferer, continuous noise was the most effective masker, and a single interfering voice of either gender was least effective. With two interferers, evidence of informational masking emerged as a difference in SRT between forward and reversed speech, but SRTs for all interferer types progressively converged at four and eight interferers. In simulation based on a real room, this occurred at a signal-to-noise ratio of around -5 dB.
Collapse
Affiliation(s)
- John F Culling
- School of Psychology, Cardiff University, Tower Building, Park Place, Cardiff, CF10 3AT, United Kingdom
| |
Collapse
|
43
|
Meekings S, Evans S, Lavan N, Boebinger D, Krieger-Redwood K, Cooke M, Scott SK. Distinct neural systems recruited when speech production is modulated by different masking sounds. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 140:8. [PMID: 27475128 DOI: 10.1121/1.4948587] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
When talkers speak in masking sounds, their speech undergoes a variety of acoustic and phonetic changes. These changes are known collectively as the Lombard effect. Most behavioural research and neuroimaging research in this area has concentrated on the effect of energetic maskers such as white noise on Lombard speech. Previous fMRI studies have argued that neural responses to speaking in noise are driven by the quality of auditory feedback-that is, the audibility of the speaker's voice over the masker. However, we also frequently produce speech in the presence of informational maskers such as another talker. Here, speakers read sentences over a range of maskers varying in their informational and energetic content: speech, rotated speech, speech modulated noise, and white noise. Subjects also spoke in quiet and listened to the maskers without speaking. When subjects spoke in masking sounds, their vocal intensity increased in line with the energetic content of the masker. However, the opposite pattern was found neurally. In the superior temporal gyrus, activation was most strongly associated with increases in informational, rather than energetic, masking. This suggests that the neural activations associated with speaking in noise are more complex than a simple feedback response.
Collapse
Affiliation(s)
- Sophie Meekings
- Institute of Cognitive Neuroscience, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom
| | - Samuel Evans
- Institute of Cognitive Neuroscience, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom
| | - Nadine Lavan
- Institute of Cognitive Neuroscience, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom
| | - Dana Boebinger
- Institute of Cognitive Neuroscience, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom
| | - Katya Krieger-Redwood
- Institute of Cognitive Neuroscience, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom
| | - Martin Cooke
- University of the Basque Country, Facultad de Letras, Universidad del País Vasco/EHU, Paseo de la Universidad 5, Vitoria, Alava 01006, Spain
| | - Sophie K Scott
- Psychology and Language Sciences, University College London, Gower Street, London WC1E 6BT, United Kingdom
| |
Collapse
|
44
|
Leung KKW, Jongman A, Wang Y, Sereno JA. Acoustic characteristics of clearly spoken English tense and lax vowels. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 140:45. [PMID: 27475131 DOI: 10.1121/1.4954737] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Clearly produced vowels exhibit longer duration and more extreme spectral properties than plain, conversational vowels. These features also characterize tense relative to lax vowels. This study explored the interaction of clear-speech and tensity effects by comparing clear and plain productions of three English tense-lax vowel pairs (/i-ɪ/, /ɑ-ʌ/, /u-ʊ/ in /kVd/ words). Both temporal and spectral acoustic features were examined, including vowel duration, vowel-to-word duration ratio, formant frequency, and dynamic spectral characteristics. Results revealed that the tense-lax vowel difference was generally enhanced in clear relative to plain speech, but clear-speech modifications for tense and lax vowels showed a trade-off in the use of temporal and spectral cues. While plain-to-clear vowel lengthening was greater for tense than lax vowels, clear-speech modifications in spectral change were larger for lax than tense vowels. Moreover, peripheral tense vowels showed more consistent clear-speech modifications in the temporal than spectral domain. Presumably, articulatory constraints limit the spectral variation of these extreme vowels, so clear-speech modifications resort to temporal features and reserve the primary spectral features for tensity contrasts. These findings suggest that clear-speech and tensity interactions involve compensatory modifications in different acoustic domains.
Collapse
Affiliation(s)
- Keith K W Leung
- Department of Linguistics, Simon Fraser University, 8888 University Drive, Burnaby, British Columbia V5A 1S6, Canada
| | - Allard Jongman
- Department of Linguistics, The University of Kansas, Lawrence, Kansas 66044, USA
| | - Yue Wang
- Department of Linguistics, Simon Fraser University, 8888 University Drive, Burnaby, British Columbia V5A 1S6, Canada
| | - Joan A Sereno
- Department of Linguistics, The University of Kansas, Lawrence, Kansas 66044, USA
| |
Collapse
|
45
|
Stenbäck V, Hällgren M, Larsby B. Executive functions and working memory capacity in speech communication under adverse conditions. SPEECH LANGUAGE AND HEARING 2016. [DOI: 10.1080/2050571x.2016.1196034] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
46
|
Dromey C, Scott S. The effects of noise on speech movements in young, middle-aged, and older adults. SPEECH, LANGUAGE AND HEARING 2016. [DOI: 10.1080/2050571x.2015.1133757] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
47
|
Koper N, Leston L, Baker TM, Curry C, Rosa P. Effects of ambient noise on detectability and localization of avian songs and tones by observers in grasslands. Ecol Evol 2015; 6:245-55. [PMID: 26811789 PMCID: PMC4716498 DOI: 10.1002/ece3.1847] [Citation(s) in RCA: 30] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/13/2015] [Revised: 10/20/2015] [Accepted: 10/21/2015] [Indexed: 12/04/2022] Open
Abstract
Probability of detection and accuracy of distance estimates in aural avian surveys may be affected by the presence of anthropogenic noise, and this may lead to inaccurate evaluations of the effects of noisy infrastructure on wildlife. We used arrays of speakers broadcasting recordings of grassland bird songs and pure tones to assess the probability of detection, and localization accuracy, by observers at sites with and without noisy oil and gas infrastructure in south‐central Alberta from 2012 to 2014. Probability of detection varied with species and with speaker distance from transect line, but there were few effects of noisy infrastructure. Accuracy of distance estimates for songs and tones decreased as distance to observer increased, and distance estimation error was higher for tones at sites with infrastructure noise. Our results suggest that quiet to moderately loud anthropogenic noise may not mask detection of bird songs; however, errors in distance estimates during aural surveys may lead to inaccurate estimates of avian densities calculated using distance sampling. We recommend caution when applying distance sampling if most birds are unseen, and where ambient noise varies among treatments.
Collapse
Affiliation(s)
- Nicola Koper
- Natural Resources Institute University of Manitoba 70 Dysart Road Winnipeg Manitoba R3T 2M7 Canada
| | - Lionel Leston
- Natural Resources Institute University of Manitoba 70 Dysart Road Winnipeg Manitoba R3T 2M7 Canada; Department of Biological Sciences University of Alberta CW 405 Biological Sciences Building Edmonton Alberta T6G 2E9 Canada
| | - Tyne M Baker
- Natural Resources Institute University of Manitoba 70 Dysart Road Winnipeg Manitoba R3T 2M7 Canada; TERA Environmental Consultants 815 8 Ave SW Calgary Alberta T2M 2M8 Canada
| | - Claire Curry
- Natural Resources Institute University of Manitoba 70 Dysart Road Winnipeg Manitoba R3T 2M7 Canada
| | - Patricia Rosa
- Natural Resources Institute University of Manitoba 70 Dysart Road Winnipeg Manitoba R3T 2M7 Canada
| |
Collapse
|
48
|
Rantala LM, Hakala S, Holmqvist S, Sala E. Classroom Noise and Teachers' Voice Production. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2015; 58:1397-1406. [PMID: 26089145 DOI: 10.1044/2015_jslhr-s-14-0248] [Citation(s) in RCA: 27] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/05/2014] [Accepted: 06/05/2015] [Indexed: 06/04/2023]
Abstract
PURPOSE The aim of this study was to research the associations between noise (ambient and activity noise) and objective metrics of teachers' voices in real working environments (i.e., classrooms). METHOD Thirty-two female and 8 male teachers from 14 elementary schools were randomly selected for the study. Ambient noise was measured during breaks in unoccupied classrooms and, likewise, the noise caused by pupils' activity during lessons. Voice samples were recorded before and after a working day. Voice variables measured were sound pressure level (voice SPL), fundamental frequency, jitter, shimmer, and the tilt of the sound spectrum slope (alpha ratio). RESULTS The ambient noise correlated most often with the fundamental frequency of men and voice SPL, whereas activity noise correlated with the alpha ratio and perturbation values. Teachers working in louder ambient noise spoke more loudly before work than those working in lower noise levels. Voice variables generally changed less during work among teachers working in loud activity noise than among those working in lower noise levels. CONCLUSIONS Ambient and activity noises affect teachers' voice use. Under loud ambient noise teachers seem to speak habitually loudly, and under loud activity noise teachers' ability to react to loading deteriorates.
Collapse
|
49
|
Stenbäck V, Hällgren M, Lyxell B, Larsby B. The Swedish Hayling task, and its relation to working memory, verbal ability, and speech-recognition-in-noise. Scand J Psychol 2015; 56:264-72. [PMID: 25819210 DOI: 10.1111/sjop.12206] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2014] [Accepted: 01/13/2015] [Indexed: 11/28/2022]
Abstract
Cognitive functions and speech-recognition-in-noise were evaluated with a cognitive test battery, assessing response inhibition using the Hayling task, working memory capacity (WMC) and verbal information processing, and an auditory test of speech recognition. The cognitive tests were performed in silence whereas the speech recognition task was presented in noise. Thirty young normally-hearing individuals participated in the study. The aim of the study was to investigate one executive function, response inhibition, and whether it is related to individual working memory capacity (WMC), and how speech-recognition-in-noise relates to WMC and inhibitory control. The results showed a significant difference between initiation and response inhibition, suggesting that the Hayling task taps cognitive activity responsible for executive control. Our findings also suggest that high verbal ability was associated with better performance in the Hayling task. We also present findings suggesting that individuals who perform well on tasks involving response inhibition, and WMC, also perform well on a speech-in-noise task. Our findings indicate that capacity to resist semantic interference can be used to predict performance on speech-in-noise tasks.
Collapse
Affiliation(s)
- Victoria Stenbäck
- Technical Audiology, Department of Clinical and Experimental Medicine, Linköping University, Sweden; Linnaeus Centre HEAD, Linköping University, Linköping, Sweden
| | | | | | | |
Collapse
|
50
|
Valentini-Botinhao C, Yamagishi J, King S, Maia R. Intelligibility enhancement of HMM-generated speech in additive noise by modifying Mel cepstral coefficients to increase the glimpse proportion. COMPUT SPEECH LANG 2014. [DOI: 10.1016/j.csl.2013.06.001] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
|