1
|
Ozaki Y, Tierney A, Pfordresher PQ, McBride JM, Benetos E, Proutskova P, Chiba G, Liu F, Jacoby N, Purdy SC, Opondo P, Fitch WT, Hegde S, Rocamora M, Thorne R, Nweke F, Sadaphal DP, Sadaphal PM, Hadavi S, Fujii S, Choo S, Naruse M, Ehara U, Sy L, Parselelo ML, Anglada-Tort M, Hansen NC, Haiduk F, Færøvik U, Magalhães V, Krzyżanowski W, Shcherbakova O, Hereld D, Barbosa BS, Varella MAC, van Tongeren M, Dessiatnitchenko P, Zar SZ, El Kahla I, Muslu O, Troy J, Lomsadze T, Kurdova D, Tsope C, Fredriksson D, Arabadjiev A, Sarbah JP, Arhine A, Meachair TÓ, Silva-Zurita J, Soto-Silva I, Millalonco NEM, Ambrazevičius R, Loui P, Ravignani A, Jadoul Y, Larrouy-Maestri P, Bruder C, Teyxokawa TP, Kuikuro U, Natsitsabui R, Sagarzazu NB, Raviv L, Zeng M, Varnosfaderani SD, Gómez-Cañón JS, Kolff K, der Nederlanden CVB, Chhatwal M, David RM, Setiawan IPG, Lekakul G, Borsan VN, Nguqu N, Savage PE. Globally, songs and instrumental melodies are slower and higher and use more stable pitches than speech: A Registered Report. SCIENCE ADVANCES 2024; 10:eadm9797. [PMID: 38748798 PMCID: PMC11095461 DOI: 10.1126/sciadv.adm9797] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/15/2023] [Accepted: 04/19/2024] [Indexed: 05/19/2024]
Abstract
Both music and language are found in all known human societies, yet no studies have compared similarities and differences between song, speech, and instrumental music on a global scale. In this Registered Report, we analyzed two global datasets: (i) 300 annotated audio recordings representing matched sets of traditional songs, recited lyrics, conversational speech, and instrumental melodies from our 75 coauthors speaking 55 languages; and (ii) 418 previously published adult-directed song and speech recordings from 209 individuals speaking 16 languages. Of our six preregistered predictions, five were strongly supported: Relative to speech, songs use (i) higher pitch, (ii) slower temporal rate, and (iii) more stable pitches, while both songs and speech used similar (iv) pitch interval size and (v) timbral brightness. Exploratory analyses suggest that features vary along a "musi-linguistic" continuum when including instrumental melodies and recited lyrics. Our study provides strong empirical evidence of cross-cultural regularities in music and speech.
Collapse
Affiliation(s)
- Yuto Ozaki
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
| | - Adam Tierney
- Department of Psychological Sciences, Birkbeck, University of London, London, UK
| | - Peter Q. Pfordresher
- Department of Psychology, University at Buffalo, State University of New York, Buffalo, NY, USA
| | - John M. McBride
- Center for Algorithmic and Robotized Synthesis, Institute for Basic Science, Ulsan, South Korea
| | - Emmanouil Benetos
- School of Electronic Engineering and Computer Science, Queen Mary University of London, London, UK
| | - Polina Proutskova
- School of Electronic Engineering and Computer Science, Queen Mary University of London, London, UK
| | - Gakuto Chiba
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
| | - Fang Liu
- School of Psychology and Clinical Language Sciences, University of Reading, Reading, UK
| | - Nori Jacoby
- Computational Auditory Perception Group, Max-Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany
| | - Suzanne C. Purdy
- School of Psychology, University of Auckland, Auckland, New Zealand
- Centre for Brain Research and Eisdell Moore Centre for Hearing and Balance Research, University of Auckland, Auckland, New Zealand
| | - Patricia Opondo
- School of Arts, Music Discipline, University of KwaZulu Natal, Durban, South Africa
| | - W. Tecumseh Fitch
- Department of Behavioral and Cognitive Biology, University of Vienna, Vienna, Austria
| | - Shantala Hegde
- Music Cognition Lab, Department of Clinical Psychology, National Institute of Mental Health and Neuro Sciences, Bangalore, Karnataka, India
| | - Martín Rocamora
- Universidad de la República, Montevideo, Uruguay
- Music Technology Group, Universitat Pompeu Fabra, Barcelona, Spain
| | - Rob Thorne
- School of Music, Victoria University of Wellington, Wellington, New Zealand
| | - Florence Nweke
- Department of Creative Arts, University of Lagos, Lagos, Nigeria
- Department of Music, Mountain Top University, Ogun, Nigeria
| | - Dhwani P. Sadaphal
- Department of Behavioral and Cognitive Biology, University of Vienna, Vienna, Austria
| | | | - Shafagh Hadavi
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
| | - Shinya Fujii
- Faculty of Environment and Information Studies, Keio University, Fujisawa, Kanagawa, Japan
| | - Sangbuem Choo
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
| | - Marin Naruse
- Faculty of Policy Management, Keio University, Fujisawa, Kanagawa, Japan
| | | | - Latyr Sy
- Independent researcher, Tokyo, Japan
- Independent researcher, Dakar, Sénégal
| | - Mark Lenini Parselelo
- Memorial University of Newfoundland, St. John’s, NL, Canada
- Department of Music and Dance, Kenyatta University, Nairobi, Kenya
| | | | - Niels Chr. Hansen
- Aarhus Institute of Advanced Studies, Aarhus University, Aarhus, Denmark
- Centre of Excellence in Music, Mind, Body and Brain, University of Jyväskylä, Jyväskylä, Finland
- Interacting Minds Centre, School of Culture and Society, Aarhus University, Aarhus, Denmark
- Royal Academy of Music Aarhus/Aalborg, Aarhus, Denmark
| | - Felix Haiduk
- Department of Behavioral and Cognitive Biology, University of Vienna, Vienna, Austria
- Department of General Psychology, University of Padua, Padua, Italy
| | - Ulvhild Færøvik
- Institute of Biological and Medical Psychology, Department of Psychology, University of Bergen, Bergen, Norway
| | - Violeta Magalhães
- Centre of Linguistics of the University of Porto (CLUP), Porto, Portugal
- Faculty of Arts and Humanities of the University of Porto (FLUP), Porto, Portugal
- School of Education of the Polytechnic of Porto (ESE IPP), Porto, Portugal
| | - Wojciech Krzyżanowski
- Adam Mickiewicz University, Faculty of Art Studies, Musicology Institute, Poznań, Poland
| | | | - Diana Hereld
- Department of Psychiatry, UCLA Semel Institute for Neuroscience and Human Behavior, Los Angeles, CA, USA
| | | | | | | | | | - Su Zar Zar
- Headmistress, The Royal Music Academy, Yangon, Myanmar
| | - Iyadh El Kahla
- Department of Cultural Policy, University of Hildesheim, Hildesheim, Germany
| | - Olcay Muslu
- Centre for the Study of Higher Education, University of Kent, Canterbury, UK
- MIRAS, Centre for Cultural Sustainability, Istanbul, Turkey
| | - Jakelin Troy
- Director, Indigenous Research, Office of the Deputy Vice-Chancellor (Research); Department of Linguistics, Faculty of Arts and Social Sciences, The University of Sydney, Camperdown, NSW, Australia
| | - Teona Lomsadze
- International Research Center for Traditional Polyphony of the Tbilisi State Conservatoire, Tbilisi, Georgia
- Georgian Studies Fellow, University of Oxford, Oxford, UK
| | - Dilyana Kurdova
- South-West University Neofit Rilski, Blagoevgrad, Bulgaria
- Phoenix Perpeticum Foundation, Sofia, Bulgaria
| | | | | | - Aleksandar Arabadjiev
- Department of Folk Music Research and Ethnomusicology, University of Music and Performing Arts–MDW, Wien, Austria
| | | | - Adwoa Arhine
- Department of Music, University of Ghana, Accra, Ghana
| | - Tadhg Ó Meachair
- Department of Ethnomusicology and Folklore, Indiana University, Bloomington, IN, USA
| | - Javier Silva-Zurita
- Department of Humanities and Arts, University of Los Lagos, Osorno, Chile
- Millennium Nucleus on Musical and Sound Cultures (CMUS NCS 2022-16), Santiago, Chile
| | - Ignacio Soto-Silva
- Department of Humanities and Arts, University of Los Lagos, Osorno, Chile
- Millennium Nucleus on Musical and Sound Cultures (CMUS NCS 2022-16), Santiago, Chile
| | | | | | - Psyche Loui
- Music, Imaging and Neural Dynamics Lab, Northeastern University, Boston, MA, USA
| | - Andrea Ravignani
- Department of Human Neurosciences, Sapienza University of Rome, Rome, Italy
- Max Planck Institute for Psycholinguistics, Nijmegen, Netherlands
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University, Aarhus, Denmark & The Royal Academy of Music Aarhus/Aalborg, Aarhus, Denmark
| | - Yannick Jadoul
- Department of Human Neurosciences, Sapienza University of Rome, Rome, Italy
- Max Planck Institute for Psycholinguistics, Nijmegen, Netherlands
| | - Pauline Larrouy-Maestri
- Music Department, Max-Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany
- Max Planck—NYU Center for Language, Music, and Emotion (CLaME), New York, NY, USA
| | - Camila Bruder
- Music Department, Max-Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany
| | - Tutushamum Puri Teyxokawa
- Txemim Puri Project–Puri Language Research, Vitalization and Teaching/Recording and Preservation of Puri History and Culture, Rio de Janeiro, Brasil
| | | | | | | | - Limor Raviv
- Max Planck Institute for Psycholinguistics, Nijmegen, Netherlands
- cSCAN, University of Glasgow, Glasgow, UK
| | - Minyu Zeng
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
- Rhode Island School of Design, Providence, RI, USA
| | - Shahaboddin Dabaghi Varnosfaderani
- Institute for English and American Studies (IEAS), Goethe University of Frankfurt am Main, Frankfurt am Main, Germany
- Cognitive and Developmental Psychology Unit, Centre, for Cognitive Science, University of Kaiserslautern-Landau (RPTU), Kaiserslautern, Germany
| | | | - Kayla Kolff
- Institute of Cognitive Science, University of Osnabrück, Osnabrück, Germany
| | | | - Meyha Chhatwal
- Department of Psychology, University of Toronto Mississauga, Mississauga, ON, Canada
| | - Ryan Mark David
- Department of Psychology, University of Toronto Mississauga, Mississauga, ON, Canada
| | | | - Great Lekakul
- Faculty of Fine Arts, Chiang Mai University, Chiang Mai, Thailand
| | - Vanessa Nina Borsan
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
- Université de Lille, CNRS, Centrale Lille, UMR 9189 CRIStAL, F-59000 Lille, France
| | - Nozuko Nguqu
- School of Arts, Music Discipline, University of KwaZulu Natal, Durban, South Africa
| | - Patrick E. Savage
- School of Psychology, University of Auckland, Auckland, New Zealand
- Faculty of Environment and Information Studies, Keio University, Fujisawa, Kanagawa, Japan
| |
Collapse
|
2
|
Kamiloğlu RG, Sauter DA. Sounds like a fight: listeners can infer behavioural contexts from spontaneous nonverbal vocalisations. Cogn Emot 2024; 38:277-295. [PMID: 37997898 PMCID: PMC11057848 DOI: 10.1080/02699931.2023.2285854] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2022] [Accepted: 11/13/2023] [Indexed: 11/25/2023]
Abstract
When we hear another person laugh or scream, can we tell the kind of situation they are in - for example, whether they are playing or fighting? Nonverbal expressions are theorised to vary systematically across behavioural contexts. Perceivers might be sensitive to these putative systematic mappings and thereby correctly infer contexts from others' vocalisations. Here, in two pre-registered experiments, we test the prediction that listeners can accurately deduce production contexts (e.g. being tickled, discovering threat) from spontaneous nonverbal vocalisations, like sighs and grunts. In Experiment 1, listeners (total n = 3120) matched 200 nonverbal vocalisations to one of 10 contexts using yes/no response options. Using signal detection analysis, we show that listeners were accurate at matching vocalisations to nine of the contexts. In Experiment 2, listeners (n = 337) categorised the production contexts by selecting from 10 response options in a forced-choice task. By analysing unbiased hit rates, we show that participants categorised all 10 contexts at better-than-chance levels. Together, these results demonstrate that perceivers can infer contexts from nonverbal vocalisations at rates that exceed that of random selection, suggesting that listeners are sensitive to systematic mappings between acoustic structures in vocalisations and behavioural contexts.
Collapse
Affiliation(s)
- Roza G. Kamiloğlu
- Department of Psychology, University of Amsterdam, Amsterdam, the Netherlands
- Department of Experimental and Applied Psychology, Vrije Universiteit Amsterdam, Amsterdam, The Netherlands
| | - Disa A. Sauter
- Department of Psychology, University of Amsterdam, Amsterdam, the Netherlands
| |
Collapse
|
3
|
Loosening the leash: The unique emotional canvas of human screams. Behav Brain Sci 2023; 46:e10. [PMID: 36799052 DOI: 10.1017/s0140525x22000851] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/18/2023]
Abstract
We use screams to explore ideas presented in the target article. Evolving first in animals as a response to predation, screams reveal more complex social use in nonhuman primates and, in humans, uniquely, are associated with a much greater variety of emotional contexts including fear, anger, surprise, and happiness. This expansion, and the potential for manipulation, promotes listener social vigilance.
Collapse
|
4
|
Debracque C, Gruber T, Lacoste R, Meguerditchian A, Grandjean D. Cerebral Activity in Female Baboons ( Papio anubis) During the Perception of Conspecific and Heterospecific Agonistic Vocalizations: a Functional Near Infrared Spectroscopy Study. AFFECTIVE SCIENCE 2022; 3:783-791. [PMID: 36519140 PMCID: PMC9743891 DOI: 10.1007/s42761-022-00164-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/09/2022] [Accepted: 10/27/2022] [Indexed: 05/02/2023]
Abstract
UNLABELLED The "voice areas" in the superior temporal cortex have been identified in both humans and non-human primates as selective to conspecific vocalizations only (i.e., expressed by members of our own species), suggesting its old evolutionary roots across the primate lineage. With respect to non-human primate species, it remains unclear whether the listening of vocal emotions from conspecifics leads to similar or different cerebral activations when compared to heterospecific calls (i.e., expressed by another primate species) triggered by the same emotion. Using a neuroimaging technique rarely employed in monkeys so far, functional Near Infrared Spectroscopy, the present study investigated in three lightly anesthetized female baboons (Papio anubis), temporal cortex activities during exposure to agonistic vocalizations from conspecifics and from other primates (chimpanzees-Pan troglodytes), and energy matched white noises in order to control for this low-level acoustic feature. Permutation test analyses on the extracted OxyHemoglobin signal revealed great inter-individual differences on how conspecific and heterospecific vocal stimuli were processed in baboon brains with a cortical response recorded either in the right or the left temporal cortex. No difference was found between emotional vocalizations and their energy-matched white noises. Despite the phylogenetic gap between Homo sapiens and African monkeys, modern humans and baboons both showed a highly heterogeneous brain process for the perception of vocal and emotional stimuli. The results of this study do not exclude that old evolutionary mechanisms for vocal emotional processing may be shared and inherited from our common ancestor. SUPPLEMENTARY INFORMATION The online version contains supplementary material available at 10.1007/s42761-022-00164-z.
Collapse
Affiliation(s)
- Coralie Debracque
- Neuroscience of Emotion and Affective Dynamics Lab, Faculty of Psychology and Educational Sciences and Swiss Center for Affective Sciences, University of Geneva, Campus Biotech, Chemin Des Mines 9, 1202 Geneva, Switzerland
| | - Thibaud Gruber
- Neuroscience of Emotion and Affective Dynamics Lab, Faculty of Psychology and Educational Sciences and Swiss Center for Affective Sciences, University of Geneva, Campus Biotech, Chemin Des Mines 9, 1202 Geneva, Switzerland
| | - Romain Lacoste
- Station de Primatologie-Celphedia, CNRS UARS846, Rousset-Sur-Arc, France
| | - Adrien Meguerditchian
- Station de Primatologie-Celphedia, CNRS UARS846, Rousset-Sur-Arc, France
- Laboratoire de Psychologie Cognitive UMR7290, CNRS, Université Aix-Marseille, Marseille, France
| | - Didier Grandjean
- Neuroscience of Emotion and Affective Dynamics Lab, Faculty of Psychology and Educational Sciences and Swiss Center for Affective Sciences, University of Geneva, Campus Biotech, Chemin Des Mines 9, 1202 Geneva, Switzerland
| |
Collapse
|
5
|
Abstract
Like most human non-verbal vocalizations, laughter is produced by speakers of all languages, across all known societies. But despite this obvious fact (or perhaps because of it), there is little comparative research examining the structural and functional similarity of laughter across speakers from different cultures. Here, we describe existing research examining (i) the perception of laughter across disparate cultures, (ii) conversation analysis examining how laughter manifests itself during discourse across different languages, and (iii) computational methods developed for automatically detecting laughter in spoken language databases. Together, these three areas of investigation provide clues regarding universals and cultural variations in laughter production and perception, and offer methodological tools that can be useful for future large-scale cross-cultural studies. We conclude by providing suggestions for areas of research and predictions of what we should expect to discover. Overall, we highlight how important questions regarding human vocal communication across cultures can be addressed through the examination of spontaneous and volitional laughter. This article is part of the theme issue ‘Cracking the laugh code: laughter through the lens of biology, psychology and neuroscience’.
Collapse
Affiliation(s)
- Gregory A Bryant
- Department of Communication, Center for Behavior, Evolution, and Culture, University of California, Los Angeles, 2225 Rolfe Hall, Los Angeles, CA 90095, USA
| | - Constance M Bainbridge
- Department of Communication, Center for Behavior, Evolution, and Culture, University of California, Los Angeles, 2225 Rolfe Hall, Los Angeles, CA 90095, USA
| |
Collapse
|
6
|
Grollero D, Petrolini V, Viola M, Morese R, Lettieri G, Cecchetti L. The structure underlying core affect and perceived affective qualities of human vocal bursts. Cogn Emot 2022; 37:1-17. [PMID: 36300588 DOI: 10.1080/02699931.2022.2139661] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/03/2022]
Abstract
Vocal bursts are non-linguistic affectively-laden sounds with a crucial function in human communication, yet their affective structure is still debated. Studies showed that ratings of valence and arousal follow a V-shaped relationship in several kinds of stimuli: high arousal ratings are more likely to go on a par with very negative or very positive valence. Across two studies, we asked participants to listen to 1,008 vocal bursts and judge both how they felt when listening to the sound (i.e. core affect condition), and how the speaker felt when producing it (i.e. perception of affective quality condition). We show that a V-shaped fit outperforms a linear model in explaining the valence-arousal relationship across conditions and studies, even after equating the number of exemplars across emotion categories. Also, although subjective experience can be significantly predicted using affective quality ratings, core affect scores are significantly lower in arousal, less extreme in valence, more variable between individuals, and less reproducible between studies. Nonetheless, stimuli rated with opposite valence between conditions range from 11% (study 1) to 17% (study 2). Lastly, we demonstrate that ambiguity in valence (i.e. high between-participants variability) explains violations of the V-shape and relates to higher arousal.
Collapse
Affiliation(s)
- Demetrio Grollero
- Social and Affective Neuroscience (SANe) Group, MoMiLab, IMT School for Advanced Studies Lucca, Lucca, Italy
| | - Valentina Petrolini
- Lindy Lab - Language in Neurodiversity, Department of Linguistics and Basque Studies, University of the Basque Country (UPV/EHU), Vitoria-Gasteiz, Spain
| | - Marco Viola
- Department of Philosophy and Education, University of Turin, Turin, Italy
| | - Rosalba Morese
- Faculty of Communication, Culture and Society, Università della Svizzera Italiana, Lugano, Switzerland
- Faculty of Biomedical Sciences, Università della Svizzera Italiana, Lugano, Switzerland
| | - Giada Lettieri
- Social and Affective Neuroscience (SANe) Group, MoMiLab, IMT School for Advanced Studies Lucca, Lucca, Italy
- Crossmodal Perception and Plasticity Laboratory, IPSY, University of Louvain, Louvain-la-Neuve, Belgium
| | - Luca Cecchetti
- Social and Affective Neuroscience (SANe) Group, MoMiLab, IMT School for Advanced Studies Lucca, Lucca, Italy
| |
Collapse
|
7
|
Liu L, Götz A, Lorette P, Tyler MD. How Tone, Intonation and Emotion Shape the Development of Infants’ Fundamental Frequency Perception. Front Psychol 2022; 13:906848. [PMID: 35719494 PMCID: PMC9204181 DOI: 10.3389/fpsyg.2022.906848] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2022] [Accepted: 05/10/2022] [Indexed: 12/02/2022] Open
Abstract
Fundamental frequency (ƒ0), perceived as pitch, is the first and arguably most salient auditory component humans are exposed to since the beginning of life. It carries multiple linguistic (e.g., word meaning) and paralinguistic (e.g., speakers’ emotion) functions in speech and communication. The mappings between these functions and ƒ0 features vary within a language and differ cross-linguistically. For instance, a rising pitch can be perceived as a question in English but a lexical tone in Mandarin. Such variations mean that infants must learn the specific mappings based on their respective linguistic and social environments. To date, canonical theoretical frameworks and most empirical studies do not view or consider the multi-functionality of ƒ0, but typically focus on individual functions. More importantly, despite the eventual mastery of ƒ0 in communication, it is unclear how infants learn to decompose and recognize these overlapping functions carried by ƒ0. In this paper, we review the symbioses and synergies of the lexical, intonational, and emotional functions that can be carried by ƒ0 and are being acquired throughout infancy. On the basis of our review, we put forward the Learnability Hypothesis that infants decompose and acquire multiple ƒ0 functions through native/environmental experiences. Under this hypothesis, we propose representative cases such as the synergy scenario, where infants use visual cues to disambiguate and decompose the different ƒ0 functions. Further, viable ways to test the scenarios derived from this hypothesis are suggested across auditory and visual modalities. Discovering how infants learn to master the diverse functions carried by ƒ0 can increase our understanding of linguistic systems, auditory processing and communication functions.
Collapse
Affiliation(s)
- Liquan Liu
- MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, NSW, Australia
- Center for Multilingualism in Society Across the Lifespan, University of Oslo, Oslo, Norway
- Australian Research Council Centre of Excellence for the Dynamics of Language, Canberra, ACT, Australia
- *Correspondence: Liquan Liu,
| | - Antonia Götz
- MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, NSW, Australia
- Department of Linguistics, University of Potsdam, Potsdam, Germany
| | - Pernelle Lorette
- Department of English Linguistics, University of Mannheim, Mannheim, Germany
| | - Michael D. Tyler
- MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, NSW, Australia
- Australian Research Council Centre of Excellence for the Dynamics of Language, Canberra, ACT, Australia
| |
Collapse
|
8
|
Carlier C, Niemeijer K, Mestdagh M, Bauwens M, Vanbrabant P, Geurts L, van Waterschoot T, Kuppens P. In Search of State and Trait Emotion Markers in Mobile-Sensed Language: Field Study. JMIR Ment Health 2022; 9:e31724. [PMID: 35147507 PMCID: PMC8881775 DOI: 10.2196/31724] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/03/2021] [Revised: 09/21/2021] [Accepted: 10/08/2021] [Indexed: 01/20/2023] Open
Abstract
BACKGROUND Emotions and mood are important for overall well-being. Therefore, the search for continuous, effortless emotion prediction methods is an important field of study. Mobile sensing provides a promising tool and can capture one of the most telling signs of emotion: language. OBJECTIVE The aim of this study is to examine the separate and combined predictive value of mobile-sensed language data sources for detecting both momentary emotional experience as well as global individual differences in emotional traits and depression. METHODS In a 2-week experience sampling method study, we collected self-reported emotion ratings and voice recordings 10 times a day, continuous keyboard activity, and trait depression severity. We correlated state and trait emotions and depression and language, distinguishing between speech content (spoken words), speech form (voice acoustics), writing content (written words), and writing form (typing dynamics). We also investigated how well these features predicted state and trait emotions using cross-validation to select features and a hold-out set for validation. RESULTS Overall, the reported emotions and mobile-sensed language demonstrated weak correlations. The most significant correlations were found between speech content and state emotions and between speech form and state emotions, ranging up to 0.25. Speech content provided the best predictions for state emotions. None of the trait emotion-language correlations remained significant after correction. Among the emotions studied, valence and happiness displayed the most significant correlations and the highest predictive performance. CONCLUSIONS Although using mobile-sensed language as an emotion marker shows some promise, correlations and predictive R2 values are low.
Collapse
Affiliation(s)
- Chiara Carlier
- Department of Psychology and Educational Sciences, Katholieke Universiteit Leuven, Leuven, Belgium
| | - Koen Niemeijer
- Department of Psychology and Educational Sciences, Katholieke Universiteit Leuven, Leuven, Belgium
| | - Merijn Mestdagh
- Department of Psychology and Educational Sciences, Katholieke Universiteit Leuven, Leuven, Belgium
| | - Michael Bauwens
- Department of Smart Organisations, University College Leuven-Limburg, Heverlee, Belgium
| | - Peter Vanbrabant
- Department of Smart Organisations, University College Leuven-Limburg, Heverlee, Belgium
| | - Luc Geurts
- Department of Computer Science, Katholieke Universiteit Leuven, Leuven, Belgium
| | - Toon van Waterschoot
- Department of Electrical Engineering, Katholieke Universiteit Leuven, Leuven, Belgium
| | - Peter Kuppens
- Department of Psychology and Educational Sciences, Katholieke Universiteit Leuven, Leuven, Belgium
| |
Collapse
|
9
|
Pisanski K, Bryant GA, Cornec C, Anikin A, Reby D. Form follows function in human nonverbal vocalisations. ETHOL ECOL EVOL 2022. [DOI: 10.1080/03949370.2022.2026482] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Affiliation(s)
- Katarzyna Pisanski
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
- CNRS French National Centre for Scientific Research, DDL Dynamics of Language Lab, University of Lyon 2, Lyon 69007, France
| | - Gregory A. Bryant
- Department of Communication, Center for Behavior, Evolution, and Culture, University of California, Los Angeles, California, USA
| | - Clément Cornec
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
| | - Andrey Anikin
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
- Division of Cognitive Science, Lund University, Lund 22100, Sweden
| | - David Reby
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
| |
Collapse
|
10
|
Pell MD, Sethi S, Rigoulot S, Rothermich K, Liu P, Jiang X. Emotional voices modulate perception and predictions about an upcoming face. Cortex 2022; 149:148-164. [DOI: 10.1016/j.cortex.2021.12.017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2021] [Revised: 09/15/2021] [Accepted: 01/05/2022] [Indexed: 11/26/2022]
|
11
|
Lin Y, Wu C, Limb CJ, Lu H, Feng IJ, Peng S, Deroche MLD, Chatterjee M. Voice emotion recognition by Mandarin-speaking pediatric cochlear implant users in Taiwan. Laryngoscope Investig Otolaryngol 2022; 7:250-258. [PMID: 35155805 PMCID: PMC8823186 DOI: 10.1002/lio2.732] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/29/2021] [Accepted: 12/29/2021] [Indexed: 11/06/2022] Open
Abstract
OBJECTIVES To explore the effects of obligatory lexical tone learning on speech emotion recognition and the cross-culture differences between United States and Taiwan for speech emotion understanding in children with cochlear implant. METHODS This cohort study enrolled 60 cochlear-implanted (cCI) Mandarin-speaking, school-aged children who underwent cochlear implantation before 5 years of age and 53 normal-hearing children (cNH) in Taiwan. The emotion recognition and the sensitivity of fundamental frequency (F0) changes for those school-aged cNH and cCI (6-17 years old) were examined in a tertiary referred center. RESULTS The mean emotion recognition score of the cNH group was significantly better than the cCI. Female speakers' vocal emotions are more easily to be recognized than male speakers' emotion. There was a significant effect of age at test on voice recognition performance. The average score of cCI with full-spectrum speech was close to the average score of cNH with eight-channel narrowband vocoder speech. The average performance of voice emotion recognition across speakers for cCI could be predicted by their sensitivity to changes in F0. CONCLUSIONS Better pitch discrimination ability comes with better voice emotion recognition for Mandarin-speaking cCI. Besides the F0 cues, cCI are likely to adapt their voice emotion recognition by relying more on secondary cues such as intensity and duration. Although cross-culture differences exist for the acoustic features of voice emotion, Mandarin-speaking cCI and their English-speaking cCI peer expressed a positive effect for age at test on emotion recognition, suggesting the learning effect and brain plasticity. Therefore, further device/processor development to improve presentation of pitch information and more rehabilitative efforts are needed to improve the transmission and perception of voice emotion in Mandarin. LEVEL OF EVIDENCE 3.
Collapse
Affiliation(s)
- Yung‐Song Lin
- Department of OtolaryngologyChi Mei Medical CenterTainanTaiwan
- Department of OtolaryngologySchool of Medicine, College of Medicine, Taipei Medical UniversityTaipeiTaiwan
| | - Che‐Ming Wu
- Department of OtorhinolaryngologyNew Taipei Municipal TuCheng Hospital (built and operated by Chang Gung Medical Foundation)New Taipei CityTaiwan
- Department of OtorhinolaryngologyChang Gung Memorial HospitalTaoyuanTaiwan
- School of Medicine, Chang Gung UniversityTaoyuanTaiwan
| | - Charles J. Limb
- School of Medicine, University of California San FranciscoSan FranciscoCaliforniaUSA
| | - Hui‐Ping Lu
- Center of Speech and Hearing, Department of OtolaryngologyChi Mei Medical CenterTainanTaiwan
| | - I. Jung Feng
- Institute of Precision Medicine, National Sun Yat‐sen UniversityKaohsiungTaiwan
| | - Shu‐Chen Peng
- Center for Devices and Radiological HealthUnited States Food and Drug AdministrationSilver SpringMarylandUSA
| | | | | |
Collapse
|
12
|
Bryant GA. Vocal communication across cultures: theoretical and methodological issues. Philos Trans R Soc Lond B Biol Sci 2022; 377:20200387. [PMID: 34775828 PMCID: PMC8591381 DOI: 10.1098/rstb.2020.0387] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2021] [Accepted: 08/03/2021] [Indexed: 11/12/2022] Open
Abstract
The study of human vocal communication has been conducted primarily in Western, educated, industrialized, rich, democratic (WEIRD) societies. Recently, cross-cultural investigations in several domains of voice research have been expanding into more diverse populations. Theoretically, it is important to understand how universals and cultural variations interact in vocal production and perception, but cross-cultural voice research presents many methodological challenges. Experimental methods typically used in WEIRD societies are often not possible to implement in many populations such as rural, small-scale societies. Moreover, theoretical and methodological issues are often unnecessarily intertwined. Here, I focus on three areas of cross-cultural voice modulation research: (i) vocal signalling of formidability and dominance, (ii) vocal emotions, and (iii) production and perception of infant-directed speech. Research in these specific areas illustrates challenges that apply more generally across the human behavioural sciences but also reveals promise as we develop our understanding of the evolution of human communication. This article is part of the theme issue 'Voice modulation: from origin and mechanism to social impact (Part II)'.
Collapse
Affiliation(s)
- Gregory A. Bryant
- Department of Communication, Center for Behavior, Evolution, and Culture, University of California, Los Angeles, 2225 Rolfe Hall, Los Angeles, CA 90095-1563, USA
| |
Collapse
|
13
|
Hughes SM, Puts DA. Vocal modulation in human mating and competition. Philos Trans R Soc Lond B Biol Sci 2021; 376:20200388. [PMID: 34719246 DOI: 10.1098/rstb.2020.0388] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
The human voice is dynamic, and people modulate their voices across different social interactions. This article presents a review of the literature examining natural vocal modulation in social contexts relevant to human mating and intrasexual competition. Altering acoustic parameters during speech, particularly pitch, in response to mating and competitive contexts can influence social perception and indicate certain qualities of the speaker. For instance, a lowered voice pitch is often used to exert dominance, display status and compete with rivals. Changes in voice can also serve as a salient medium for signalling a person's attraction to another, and there is evidence to support the notion that attraction and/or romantic interest can be distinguished through vocal tones alone. Individuals can purposely change their vocal behaviour in attempt to sound more attractive and to facilitate courtship success. Several findings also point to the effectiveness of vocal change as a mechanism for communicating relationship status. As future studies continue to explore vocal modulation in the arena of human mating, we will gain a better understanding of how and why vocal modulation varies across social contexts and its impact on receiver psychology. This article is part of the theme issue 'Voice modulation: from origin and mechanism to social impact (Part I)'.
Collapse
Affiliation(s)
- Susan M Hughes
- Psychology Department, Albright College, Reading, PA 19612, USA
| | - David A Puts
- Department of Anthropology, Pennsylvania State University, University Park, PA 16802, USA
| |
Collapse
|
14
|
Pinheiro AP, Anikin A, Conde T, Sarzedas J, Chen S, Scott SK, Lima CF. Emotional authenticity modulates affective and social trait inferences from voices. Philos Trans R Soc Lond B Biol Sci 2021; 376:20200402. [PMID: 34719249 PMCID: PMC8558771 DOI: 10.1098/rstb.2020.0402] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 05/12/2021] [Indexed: 01/31/2023] Open
Abstract
The human voice is a primary tool for verbal and nonverbal communication. Studies on laughter emphasize a distinction between spontaneous laughter, which reflects a genuinely felt emotion, and volitional laughter, associated with more intentional communicative acts. Listeners can reliably differentiate the two. It remains unclear, however, if they can detect authenticity in other vocalizations, and whether authenticity determines the affective and social impressions that we form about others. Here, 137 participants listened to laughs and cries that could be spontaneous or volitional and rated them on authenticity, valence, arousal, trustworthiness and dominance. Bayesian mixed models indicated that listeners detect authenticity similarly well in laughter and crying. Speakers were also perceived to be more trustworthy, and in a higher arousal state, when their laughs and cries were spontaneous. Moreover, spontaneous laughs were evaluated as more positive than volitional ones, and we found that the same acoustic features predicted perceived authenticity and trustworthiness in laughter: high pitch, spectral variability and less voicing. For crying, associations between acoustic features and ratings were less reliable. These findings indicate that emotional authenticity shapes affective and social trait inferences from voices, and that the ability to detect authenticity in vocalizations is not limited to laughter. This article is part of the theme issue 'Voice modulation: from origin and mechanism to social impact (Part I)'.
Collapse
Affiliation(s)
- Ana P. Pinheiro
- CICPSI, Faculdade de Psicologia, Universidade de Lisboa, Alameda da Universidade, 1649-013 Lisboa, Portugal
| | - Andrey Anikin
- Equipe de Neuro-Ethologie Sensorielle (ENES)/Centre de Recherche em Neurosciences de Lyon (CRNL), University of Lyon/Saint-Etienne, CNRS UMR5292, INSERM UMR_S 1028, 42023 Saint-Etienne, France
- Division of Cognitive Science, Lund University, 221 00 Lund, Sweden
| | - Tatiana Conde
- CICPSI, Faculdade de Psicologia, Universidade de Lisboa, Alameda da Universidade, 1649-013 Lisboa, Portugal
| | - João Sarzedas
- CICPSI, Faculdade de Psicologia, Universidade de Lisboa, Alameda da Universidade, 1649-013 Lisboa, Portugal
| | - Sinead Chen
- National Taiwan University, Taipei City, 10617 Taiwan
| | - Sophie K. Scott
- Institute of Cognitive Neuroscience, University College London, London WC1N 3AZ, UK
| | - César F. Lima
- Institute of Cognitive Neuroscience, University College London, London WC1N 3AZ, UK
- Instituto Universitário de Lisboa (ISCTE-IUL), Avenida das Forças Armadas, 1649-026 Lisboa, Portugal
| |
Collapse
|
15
|
Affiliation(s)
- Sasha L. Winkler
- Department of Anthropology, University of California, Los Angeles, CA, USA
- UCLA Center for Behavior, Evolution, and Culture
| | - Gregory A. Bryant
- UCLA Center for Behavior, Evolution, and Culture
- Department of Communication, University of California, Los Angeles, CA, USA
| |
Collapse
|
16
|
Abstract
Neurocognitive models (e.g., Schirmer & Kotz, 2006) have helped to characterize how listeners incrementally derive meaning from vocal expressions of emotion in spoken language, what neural mechanisms are involved at different processing stages, and their relative time course. But how can these insights be applied to communicative situations in which prosody serves a predominantly interpersonal function? This comment examines recent data highlighting the dynamic interplay of prosody and language, when vocal attributes serve the sociopragmatic goals of the speaker or reveal interpersonal information that listeners use to construct a mental representation of what is being communicated. Our comment serves as a beacon to researchers interested in how the neurocognitive system “makes sense” of socioemotive aspects of prosody.
Collapse
Affiliation(s)
- Marc D. Pell
- School of Communication Sciences and Disorders, McGill University, Canada
| | - Sonja A. Kotz
- Department of Neuropsychology and Psychopharmacology, Maastricht University, The Netherlands
- Department of Neuropsychology, Max Planck Institute for Human Cognitive and Brain Sciences, Germany
| |
Collapse
|
17
|
Scherer KR. Comment: Advances in Studying the Vocal Expression of Emotion: Current Contributions and Further Options. EMOTION REVIEW 2021. [DOI: 10.1177/1754073920949671] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
I consider the five contributions in this special section as evidence that the research area dealing with the vocal expression of emotion is advancing rapidly, both in terms of the number of pertinent empirical studies and with respect to an ever increasing sophistication of methodology. I provide some suggestions on promising areas for future interdisciplinary research, including work on emotion expression in singing and the potential of vocal symptoms of emotional disorder. As to the popular discussion of the respective role of universality versus language/culture differences, I suggest to move on from exclusively studying the accuracy of recognition in judgment studies to a more differentiated approach adding production aspects, taking into account the multiple vocal and acoustic features that interact to communicate emotion.
Collapse
Affiliation(s)
- Klaus R. Scherer
- Department of Psychology, University of Geneva, Switzerland
- Department of Psychology, Ludwig-Maximilians-University of Munich, Germany
| |
Collapse
|