1
|
Bauer JF, Schindler-Gmelch L, Gerczuk M, Schuller B, Berking M. Prosody-focused feedback enhances the efficacy of anti-depressive self-statements in depressed individuals - A randomized controlled trial. Behav Res Ther 2025; 184:104667. [PMID: 39700643 DOI: 10.1016/j.brat.2024.104667] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2023] [Revised: 08/05/2024] [Accepted: 12/06/2024] [Indexed: 12/21/2024]
Abstract
This study was aimed to evaluate whether the efficacy of invoking anti-depressive self-statements to cope with depressed mood can be enhanced for depressed individuals by systematically guiding them to amplify the expression of conviction in their voice. Accordingly, we recruited N = 144 participants (48 clinically depressed individuals, 48 sub-clinically depressed individuals, and 48 non-depressed individuals). Participants were randomly assigned to an experimental or control condition. Across study conditions, participants completed a mood induction procedure, then read aloud scripted anti-depressive self-statements designed to reduce depressed mood. Participants in the experimental condition received instructions to heighten the prosodic expression of conviction in their voice; participants in the control condition received no prosodic expression instructions. Results showed that depressed participants achieved a more pronounced decrease of depressed mood in the experimental condition than in the control condition. Further, the results indicated no effects in sub-clinically depressed and non-depressed individuals. Finally, heightened conviction expressed by participants in the experimental condition was associated with lower depressed mood and diminished depressive symptom severity. Overall, our findings suggest that fostering the prosodic expression of conviction in depressed persons' voices, while they vocalize anti-depressive self-statements, represents a promising method for augmenting the efficacy of cognitive interventions for depression.
Collapse
Affiliation(s)
- Jonathan F Bauer
- Department of Clinical Psychology and Psychotherapy, Friedrich-Alexander-Universität Erlangen Nürnberg, Nägelsbachstraße 25a, 91052, Erlangen, Germany.
| | - Lena Schindler-Gmelch
- Department of Clinical Psychology and Psychotherapy, Friedrich-Alexander-Universität Erlangen Nürnberg, Nägelsbachstraße 25a, 91052, Erlangen, Germany.
| | - Maurice Gerczuk
- Chair of Health Informatics, MRI, Technical University Munich, Ismaninger Straße 22, 81675 München, Germany; Munich Center for Machine Learning, Arcisstraße 21, 80333 München, Germany.
| | - Björn Schuller
- Chair of Health Informatics, MRI, Technical University Munich, Ismaninger Straße 22, 81675 München, Germany; Munich Center for Machine Learning, Arcisstraße 21, 80333 München, Germany; GLAM - Group on Language, Audio, & Music, Imperial College London, London, SW7 2AZ, UK.
| | - Matthias Berking
- Department of Clinical Psychology and Psychotherapy, Friedrich-Alexander-Universität Erlangen Nürnberg, Nägelsbachstraße 25a, 91052, Erlangen, Germany.
| |
Collapse
|
2
|
Nakai T, Rachman L, Arias Sarah P, Okanoya K, Aucouturier JJ. Algorithmic voice transformations reveal the phonological basis of language-familiarity effects in cross-cultural emotion judgments. PLoS One 2023; 18:e0285028. [PMID: 37134091 PMCID: PMC10156011 DOI: 10.1371/journal.pone.0285028] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2022] [Accepted: 04/13/2023] [Indexed: 05/04/2023] Open
Abstract
People have a well-described advantage in identifying individuals and emotions in their own culture, a phenomenon also known as the other-race and language-familiarity effect. However, it is unclear whether native-language advantages arise from genuinely enhanced capacities to extract relevant cues in familiar speech or, more simply, from cultural differences in emotional expressions. Here, to rule out production differences, we use algorithmic voice transformations to create French and Japanese stimulus pairs that differed by exactly the same acoustical characteristics. In two cross-cultural experiments, participants performed better in their native language when categorizing vocal emotional cues and detecting non-emotional pitch changes. This advantage persisted over three types of stimulus degradation (jabberwocky, shuffled and reversed sentences), which disturbed semantics, syntax, and supra-segmental patterns, respectively. These results provide evidence that production differences are not the sole drivers of the language-familiarity effect in cross-cultural emotion perception. Listeners' unfamiliarity with the phonology of another language, rather than with its syntax or semantics, impairs the detection of pitch prosodic cues and, in turn, the recognition of expressive prosody.
Collapse
Affiliation(s)
- Tomoya Nakai
- Lyon Neuroscience Research Center (CRNL), (INSERM/CNRS/University of Lyon), Bron, France
- Center for Information and Neural Networks, National Institute of Information and Communications Technology, Suita, Japan
| | - Laura Rachman
- Department of Otorhinolaryngology, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
| | - Pablo Arias Sarah
- Lund University Cognitive Science, Lund University, Lund, Sweden
- Sciences et Technologies de la Musique et du Son (IRCAM/CNRS/Sorbonne Université), Paris, France
- School of Psychology & Neuroscience, University of Glasgow, Glasgow, United Kingdom
| | - Kazuo Okanoya
- The University of Tokyo, Graduate School of Arts and Sciences, Tokyo, Japan
- Advanced Comprehensive Research Organization, Teikyo University, Tokyo, Japan
| | - Jean-Julien Aucouturier
- Sciences et Technologies de la Musique et du Son (IRCAM/CNRS/Sorbonne Université), Paris, France
- FEMTO-ST Institute (CNRS/Université de Bourgogne Franche Comté), Besançon, France
| |
Collapse
|
3
|
Skipper JI. A voice without a mouth no more: The neurobiology of language and consciousness. Neurosci Biobehav Rev 2022; 140:104772. [PMID: 35835286 DOI: 10.1016/j.neubiorev.2022.104772] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2021] [Revised: 05/18/2022] [Accepted: 07/05/2022] [Indexed: 11/26/2022]
Abstract
Most research on the neurobiology of language ignores consciousness and vice versa. Here, language, with an emphasis on inner speech, is hypothesised to generate and sustain self-awareness, i.e., higher-order consciousness. Converging evidence supporting this hypothesis is reviewed. To account for these findings, a 'HOLISTIC' model of neurobiology of language, inner speech, and consciousness is proposed. It involves a 'core' set of inner speech production regions that initiate the experience of feeling and hearing words. These take on affective qualities, deriving from activation of associated sensory, motor, and emotional representations, involving a largely unconscious dynamic 'periphery', distributed throughout the whole brain. Responding to those words forms the basis for sustained network activity, involving 'default mode' activation and prefrontal and thalamic/brainstem selection of contextually relevant responses. Evidence for the model is reviewed, supporting neuroimaging meta-analyses conducted, and comparisons with other theories of consciousness made. The HOLISTIC model constitutes a more parsimonious and complete account of the 'neural correlates of consciousness' that has implications for a mechanistic account of mental health and wellbeing.
Collapse
|
4
|
Guerouaou N, Vaiva G, Aucouturier JJ. The shallow of your smile: the ethics of expressive vocal deep-fakes. Philos Trans R Soc Lond B Biol Sci 2022; 377:20210083. [PMID: 34775820 PMCID: PMC8591385 DOI: 10.1098/rstb.2021.0083] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2021] [Accepted: 07/28/2021] [Indexed: 11/12/2022] Open
Abstract
Rapid technological advances in artificial intelligence are creating opportunities for real-time algorithmic modulations of a person's facial and vocal expressions, or 'deep-fakes'. These developments raise unprecedented societal and ethical questions which, despite much recent public awareness, are still poorly understood from the point of view of moral psychology. We report here on an experimental ethics study conducted on a sample of N = 303 participants (predominantly young, western and educated), who evaluated the acceptability of vignettes describing potential applications of expressive voice transformation technology. We found that vocal deep-fakes were generally well accepted in the population, notably in a therapeutic context and for emotions judged otherwise difficult to control, and surprisingly, even if the user lies to their interlocutors about using them. Unlike other emerging technologies like autonomous vehicles, there was no evidence of social dilemma in which one would, for example, accept for others what they resent for themselves. The only real obstacle to the massive deployment of vocal deep-fakes appears to be situations where they are applied to a speaker without their knowing, but even the acceptability of such situations was modulated by individual differences in moral values and attitude towards science fiction. This article is part of the theme issue 'Voice modulation: from origin and mechanism to social impact (Part II)'.
Collapse
Affiliation(s)
- Nadia Guerouaou
- Science and Technology of Music and Sound, IRCAM/CNRS/Sorbonne Université, Paris, France
- Lille Neuroscience and Cognition Center (LiNC), Team PSY, INSERM U-1172/CHRU Lille, France
| | - Guillaume Vaiva
- Lille Neuroscience and Cognition Center (LiNC), Team PSY, INSERM U-1172/CHRU Lille, France
| | | |
Collapse
|
5
|
Bedoya D, Arias P, Rachman L, Liuni M, Canonne C, Goupil L, Aucouturier JJ. Even violins can cry: specifically vocal emotional behaviours also drive the perception of emotions in non-vocal music. Philos Trans R Soc Lond B Biol Sci 2021; 376:20200396. [PMID: 34719254 PMCID: PMC8558776 DOI: 10.1098/rstb.2020.0396] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
A wealth of theoretical and empirical arguments have suggested that music triggers emotional responses by resembling the inflections of expressive vocalizations, but have done so using low-level acoustic parameters (pitch, loudness, speed) that, in fact, may not be processed by the listener in reference to human voice. Here, we take the opportunity of the recent availability of computational models that allow the simulation of three specifically vocal emotional behaviours: smiling, vocal tremor and vocal roughness. When applied to musical material, we find that these three acoustic manipulations trigger emotional perceptions that are remarkably similar to those observed on speech and scream sounds, and identical across musician and non-musician listeners. Strikingly, this not only applied to singing voice with and without musical background, but also to purely instrumental material. This article is part of the theme issue ‘Voice modulation: from origin and mechanism to social impact (Part I)’.
Collapse
Affiliation(s)
- D Bedoya
- Science and Technology of Music and Sound, IRCAM/CNRS/Sorbonne Université, Paris, France
| | - P Arias
- Science and Technology of Music and Sound, IRCAM/CNRS/Sorbonne Université, Paris, France.,Department of Cognitive Science, Lund University, Lund, Sweden
| | - L Rachman
- Faculty of Medical Sciences, University of Groningen, Groningen, The Netherlands
| | - M Liuni
- Alta Voce SAS, Houilles, France
| | - C Canonne
- Science and Technology of Music and Sound, IRCAM/CNRS/Sorbonne Université, Paris, France
| | - L Goupil
- BabyDevLab, University of East London, London, UK
| | - J-J Aucouturier
- FEMTO-ST Institute, Université de Bourgogne Franche-Comté/CNRS, Besançon, France
| |
Collapse
|
6
|
Goupil L, Johansson P, Hall L, Aucouturier JJ. Vocal signals only impact speakers' own emotions when they are self-attributed. Conscious Cogn 2021; 88:103072. [PMID: 33406449 DOI: 10.1016/j.concog.2020.103072] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2019] [Revised: 11/02/2020] [Accepted: 12/12/2020] [Indexed: 11/30/2022]
Abstract
Emotions are often accompanied by vocalizations whose acoustic features provide information about the physiological state of the speaker. Here, we ask if perceiving these affective signals in one's own voice has an impact on one's own emotional state, and if it is necessary to identify these signals as self-originated for the emotional effect to occur. Participants had to deliberate out loud about how they would feel in various familiar emotional scenarios, while we covertly manipulated their voices in order to make them sound happy or sad. Perceiving the artificial affective signals in their own voice altered participants' judgements about how they would feel in these situations. Crucially, this effect disappeared when participants detected the vocal manipulation, either explicitly or implicitly. The original valence of the scenarios also modulated the vocal feedback effect. These results highlight the role of the exteroception of self-attributed affective signals in the emergence of emotional feelings.
Collapse
Affiliation(s)
- Louise Goupil
- STMS UMR 9912 (CNRS/IRCAM/SU), Paris, France; University of East London, London, UK.
| | - Petter Johansson
- Lund University Cognitive Science, Lund University, Lund, Sweden
| | - Lars Hall
- Lund University Cognitive Science, Lund University, Lund, Sweden
| | | |
Collapse
|
7
|
Arias P, Rachman L, Liuni M, Aucouturier JJ. Beyond Correlation: Acoustic Transformation Methods for the Experimental Study of Emotional Voice and Speech. EMOTION REVIEW 2020. [DOI: 10.1177/1754073920934544] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
While acoustic analysis methods have become a commodity in voice emotion research, experiments that attempt not only to describe but to computationally manipulate expressive cues in emotional voice and speech have remained relatively rare. We give here a nontechnical overview of voice-transformation techniques from the audio signal-processing community that we believe are ripe for adoption in this context. We provide sound examples of what they can achieve, examples of experimental questions for which they can be used, and links to open-source implementations. We point at a number of methodological properties of these algorithms, such as being specific, parametric, exhaustive, and real-time, and describe the new possibilities that these open for the experimental study of the emotional voice.
Collapse
Affiliation(s)
- Pablo Arias
- STMS UMR9912, IRCAM/CNRS/Sorbonne Université, France
| | - Laura Rachman
- STMS UMR9912, IRCAM/CNRS/Sorbonne Université, France
| | - Marco Liuni
- STMS UMR9912, IRCAM/CNRS/Sorbonne Université, France
| | | |
Collapse
|
8
|
Rachman L, Dubal S, Aucouturier JJ. Happy you, happy me: expressive changes on a stranger's voice recruit faster implicit processes than self-produced expressions. Soc Cogn Affect Neurosci 2020; 14:559-568. [PMID: 31044241 PMCID: PMC6545538 DOI: 10.1093/scan/nsz030] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/05/2018] [Revised: 04/09/2019] [Accepted: 04/21/2019] [Indexed: 01/09/2023] Open
Abstract
In social interactions, people have to pay attention both to the ‘what’ and ‘who’. In particular, expressive changes heard on speech signals have to be integrated with speaker identity, differentiating e.g. self- and other-produced signals. While previous research has shown that self-related visual information processing is facilitated compared to non-self stimuli, evidence in the auditory modality remains mixed. Here, we compared electroencephalography (EEG) responses to expressive changes in sequence of self- or other-produced speech sounds using a mismatch negativity (MMN) passive oddball paradigm. Critically, to control for speaker differences, we used programmable acoustic transformations to create voice deviants that differed from standards in exactly the same manner, making EEG responses to such deviations comparable between sequences. Our results indicate that expressive changes on a stranger’s voice are highly prioritized in auditory processing compared to identical changes on the self-voice. Other-voice deviants generate earlier MMN onset responses and involve stronger cortical activations in a left motor and somatosensory network suggestive of an increased recruitment of resources for less internally predictable, and therefore perhaps more socially relevant, signals.
Collapse
Affiliation(s)
- Laura Rachman
- Inserm U, CNRS UMR, Sorbonne Université UMR S, Institut du Cerveau et de la Moelle épinière, Social and Affective Neuroscience Lab, Paris, France.,Science & Technology of Music and Sound, UMR (CNRS/IRCAM/Sorbonne Université), Paris, France
| | - Stéphanie Dubal
- Inserm U, CNRS UMR, Sorbonne Université UMR S, Institut du Cerveau et de la Moelle épinière, Social and Affective Neuroscience Lab, Paris, France
| | - Jean-Julien Aucouturier
- Science & Technology of Music and Sound, UMR (CNRS/IRCAM/Sorbonne Université), Paris, France
| |
Collapse
|
9
|
Abstract
Voice synthesis is a useful method for investigating the communicative role of different acoustic features. Although many text-to-speech systems are available, researchers of human nonverbal vocalizations and bioacousticians may profit from a dedicated simple tool for synthesizing and manipulating natural-sounding vocalizations. Soundgen ( https://CRAN.R-project.org/package=soundgen ) is an open-source R package that synthesizes nonverbal vocalizations based on meaningful acoustic parameters, which can be specified from the command line or in an interactive app. This tool was validated by comparing the perceived emotion, valence, arousal, and authenticity of 60 recorded human nonverbal vocalizations (screams, moans, laughs, and so on) and their approximate synthetic reproductions. Each synthetic sound was created by manually specifying only a small number of high-level control parameters, such as syllable length and a few anchors for the intonation contour. Nevertheless, the valence and arousal ratings of synthetic sounds were similar to those of the original recordings, and the authenticity ratings were comparable, maintaining parity with the originals for less complex vocalizations. Manipulating the precise acoustic characteristics of synthetic sounds may shed light on the salient predictors of emotion in the human voice. More generally, soundgen may prove useful for any studies that require precise control over the acoustic features of nonspeech sounds, including research on animal vocalizations and auditory perception.
Collapse
Affiliation(s)
- Andrey Anikin
- Division of Cognitive Science, Department of Philosophy, Lund University, Box 192, SE-221 00, Lund, Sweden.
| |
Collapse
|
10
|
Social Cognition through the Lens of Cognitive and Clinical Neuroscience. BIOMED RESEARCH INTERNATIONAL 2018; 2018:4283427. [PMID: 30302338 PMCID: PMC6158937 DOI: 10.1155/2018/4283427] [Citation(s) in RCA: 73] [Impact Index Per Article: 10.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/15/2018] [Accepted: 08/13/2018] [Indexed: 12/13/2022]
Abstract
Social cognition refers to a set of processes, ranging from perception to decision-making, underlying the ability to decode others' intentions and behaviors to plan actions fitting with social and moral, besides individual and economic considerations. Its centrality in everyday life reflects the neural complexity of social processing and the ubiquity of social cognitive deficits in different pathological conditions. Social cognitive processes can be clustered in three domains associated with (a) perceptual processing of social information such as faces and emotional expressions (social perception), (b) grasping others' cognitive or affective states (social understanding), and (c) planning behaviors taking into consideration others', in addition to one's own, goals (social decision-making). We review these domains from the lens of cognitive neuroscience, i.e., in terms of the brain areas mediating the role of such processes in the ability to make sense of others' behavior and plan socially appropriate actions. The increasing evidence on the “social brain” obtained from healthy young individuals nowadays constitutes the baseline for detecting changes in social cognitive skills associated with physiological aging or pathological conditions. In the latter case, impairments in one or more of the abovementioned domains represent a prominent concern, or even a core facet, of neurological (e.g., acquired brain injury or neurodegenerative diseases), psychiatric (e.g., schizophrenia), and developmental (e.g., autism) disorders. To pave the way for the other papers of this issue, addressing the social cognitive deficits associated with severe acquired brain injury, we will briefly discuss the available evidence on the status of social cognition in normal aging and its breakdown in neurodegenerative disorders. Although the assessment and treatment of such impairments is a relatively novel sector in neurorehabilitation, the evidence summarized here strongly suggests that the development of remediation procedures for social cognitive skills will represent a future field of translational research in clinical neuroscience.
Collapse
|
11
|
DAVID: An open-source platform for real-time transformation of infra-segmental emotional cues in running speech. Behav Res Methods 2018; 50:323-343. [PMID: 28374144 PMCID: PMC5809549 DOI: 10.3758/s13428-017-0873-y] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/07/2023]
Abstract
We present an open-source software platform that transforms emotional cues expressed by speech signals using audio effects like pitch shifting, inflection, vibrato, and filtering. The emotional transformations can be applied to any audio file, but can also run in real time, using live input from a microphone, with less than 20-ms latency. We anticipate that this tool will be useful for the study of emotions in psychology and neuroscience, because it enables a high level of control over the acoustical and emotional content of experimental stimuli in a variety of laboratory situations, including real-time social situations. We present here results of a series of validation experiments aiming to position the tool against several methodological requirements: that transformed emotions be recognized at above-chance levels, valid in several languages (French, English, Swedish, and Japanese) and with a naturalness comparable to natural speech.
Collapse
|
12
|
Abstract
People perceive their recorded voice differently from their actively spoken voice. The uncanny valley theory proposes that as an object approaches humanlike characteristics, there is an increase in the sense of familiarity; however, eventually a point is reached where the object becomes strangely similar and makes us feel uneasy. The feeling of discomfort experienced when people hear their recorded voice may correspond to the floor of the proposed uncanny valley. To overcome the feeling of eeriness of own-voice recordings, previous studies have suggested equalization of the recorded voice with various types of filters, such as step, bandpass, and low-pass, yet the effectiveness of these filters has not been evaluated. To address this, the aim of experiment 1 was to identify what type of voice recording was the most representative of one’s own voice. The voice recordings were presented in five different conditions: unadjusted recorded voice, step filtered voice, bandpass filtered voice, low-pass filtered voice, and a voice for which the participants freely adjusted the parameters. We found large individual differences in the most representative own-voice filter. In order to consider roles of sense of agency, experiment 2 investigated if lip-synching would influence the rating of own voice. The result suggested lip-synching did not affect own voice ratings. In experiment 3, based on the assumption that the voices used in previous experiments corresponded to continuous representations of non-own voice to own voice, the existence of an uncanny valley was examined. Familiarity, eeriness, and the sense of own voice were rated. The result did not support the existence of an uncanny valley. Taken together, the experiments led us to the following conclusions: there is no general filter that can represent own voice for everyone, sense of agency has no effect on own voice rating, and the uncanny valley does not exist for own voice, specifically.
Collapse
|
13
|
Schepman A, Rodway P, Cornmell L, Smith B, de Sa SL, Borwick C, Belfon-Thompson E. Right-ear precedence and vocal emotion contagion: The role of the left hemisphere. Laterality 2018; 23:290-317. [DOI: 10.1080/1357650x.2017.1360902] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Affiliation(s)
- Astrid Schepman
- Department of Psychology, University of Chester, Chester, UK
| | - Paul Rodway
- Department of Psychology, University of Chester, Chester, UK
| | - Louise Cornmell
- Department of Psychology, University of Chester, Chester, UK
| | - Bethany Smith
- Department of Psychology, University of Chester, Chester, UK
| | | | - Ciara Borwick
- Department of Psychology, University of Chester, Chester, UK
| | | |
Collapse
|
14
|
Villain M, Cosin C, Glize B, Berthoz S, Swendsen J, Sibon I, Mayo W. Affective Prosody and Depression After Stroke: A Pilot Study. Stroke 2016; 47:2397-400. [PMID: 27507865 DOI: 10.1161/strokeaha.116.013852] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/20/2016] [Accepted: 07/12/2016] [Indexed: 11/16/2022]
Abstract
BACKGROUND AND PURPOSE Poststroke depression (PSD) is a frequent complication of stroke with detrimental consequences in terms of quality of life and functional outcomes. In individuals with major depression, several studies have demonstrated an alteration of affective prosody. The aim of this study is to identify prosodic markers that may be predictive of PSD. METHODS Patient voices were recorded at baseline and 3 months after stroke. We extracted prosodic parameters, including fundamental frequency, percentage of voice breaks, and shimmer. Depression and anxiety symptoms were assessed 3 months later. RESULTS Among the 49 patients included in the study, 22.5% developed PSD 3 months after stroke. A significant decrease was observed concerning the fundamental frequency among patients who developed PSD. Discriminant analysis demonstrated that initial voice breaks coupled with shimmer are strongly predictive of subsequent PSD. CONCLUSIONS Early alterations of affective prosody are associated with a higher risk of PSD 3 months after a stroke. This new physiological approach overcomes traditional barriers associated with clinical instruments and contributes to the prediction of this disorder.
Collapse
Affiliation(s)
- Marie Villain
- From the Université de Bordeaux, INCIA, CNRS UMR5287, Talence, France (M.V., C.C., J.S., I.S., W.M.); Ecole Pratique des Hautes Etudes, Paris, France (M.V., C.C., J.S.); CHU Bordeaux, Pôle de Neurosciences Cliniques, Hôpital Pellegrin, Bordeaux, France (B.G., I.S.); and CESP, Université Paris Sud, Institut Mutualiste Montsouris, Paris, France (S.B.)
| | - Charlotte Cosin
- From the Université de Bordeaux, INCIA, CNRS UMR5287, Talence, France (M.V., C.C., J.S., I.S., W.M.); Ecole Pratique des Hautes Etudes, Paris, France (M.V., C.C., J.S.); CHU Bordeaux, Pôle de Neurosciences Cliniques, Hôpital Pellegrin, Bordeaux, France (B.G., I.S.); and CESP, Université Paris Sud, Institut Mutualiste Montsouris, Paris, France (S.B.).
| | - Bertrand Glize
- From the Université de Bordeaux, INCIA, CNRS UMR5287, Talence, France (M.V., C.C., J.S., I.S., W.M.); Ecole Pratique des Hautes Etudes, Paris, France (M.V., C.C., J.S.); CHU Bordeaux, Pôle de Neurosciences Cliniques, Hôpital Pellegrin, Bordeaux, France (B.G., I.S.); and CESP, Université Paris Sud, Institut Mutualiste Montsouris, Paris, France (S.B.)
| | - Sylvie Berthoz
- From the Université de Bordeaux, INCIA, CNRS UMR5287, Talence, France (M.V., C.C., J.S., I.S., W.M.); Ecole Pratique des Hautes Etudes, Paris, France (M.V., C.C., J.S.); CHU Bordeaux, Pôle de Neurosciences Cliniques, Hôpital Pellegrin, Bordeaux, France (B.G., I.S.); and CESP, Université Paris Sud, Institut Mutualiste Montsouris, Paris, France (S.B.)
| | - Joel Swendsen
- From the Université de Bordeaux, INCIA, CNRS UMR5287, Talence, France (M.V., C.C., J.S., I.S., W.M.); Ecole Pratique des Hautes Etudes, Paris, France (M.V., C.C., J.S.); CHU Bordeaux, Pôle de Neurosciences Cliniques, Hôpital Pellegrin, Bordeaux, France (B.G., I.S.); and CESP, Université Paris Sud, Institut Mutualiste Montsouris, Paris, France (S.B.)
| | - Igor Sibon
- From the Université de Bordeaux, INCIA, CNRS UMR5287, Talence, France (M.V., C.C., J.S., I.S., W.M.); Ecole Pratique des Hautes Etudes, Paris, France (M.V., C.C., J.S.); CHU Bordeaux, Pôle de Neurosciences Cliniques, Hôpital Pellegrin, Bordeaux, France (B.G., I.S.); and CESP, Université Paris Sud, Institut Mutualiste Montsouris, Paris, France (S.B.)
| | - Willy Mayo
- From the Université de Bordeaux, INCIA, CNRS UMR5287, Talence, France (M.V., C.C., J.S., I.S., W.M.); Ecole Pratique des Hautes Etudes, Paris, France (M.V., C.C., J.S.); CHU Bordeaux, Pôle de Neurosciences Cliniques, Hôpital Pellegrin, Bordeaux, France (B.G., I.S.); and CESP, Université Paris Sud, Institut Mutualiste Montsouris, Paris, France (S.B.)
| |
Collapse
|
15
|
Emergency medical triage decisions are swayed by computer-manipulated cues of physical dominance in caller's voice. Sci Rep 2016; 6:30219. [PMID: 27456205 PMCID: PMC4960535 DOI: 10.1038/srep30219] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2016] [Accepted: 06/30/2016] [Indexed: 01/11/2023] Open
Abstract
In humans as well as other animals, displays of body strength such as power postures or deep masculine voices are associated with prevalence in conflicts of interest and facilitated access to resources. We conduct here an ecological and highly critical test of this hypothesis in a domain that, on first thought, would appear to be shielded from such influences: access to emergency medical care. Using acoustic manipulations of vocal masculinity, we systematically varied the perceived level of physical dominance of mock patients calling a medical call center simulator. Callers whose voice were perceived as indicative of physical dominance (i.e. those with low fundamental and formant frequency voices) obtained a higher grade of response, a higher evaluation of medical emergency and longer attention from physicians than callers with strictly identical medical needs whose voice signaled lower physical dominance. Strikingly, while the effect was important for physician participants, it was virtually non-existent when calls were processed by non-medically-trained phone operators. This finding demonstrates an unprecedented degree of vulnerability of telephone-based medical decisions to extra-medical factors carried by vocal cues, and shows that it may not simply be assumed that more medical training will shield decisions from such influences.
Collapse
|