1
|
Lavan N, Rinke P, Scharinger M. The time course of person perception from voices in the brain. Proc Natl Acad Sci U S A 2024; 121:e2318361121. [PMID: 38889147 DOI: 10.1073/pnas.2318361121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2023] [Accepted: 04/26/2024] [Indexed: 06/20/2024] Open
Abstract
When listeners hear a voice, they rapidly form a complex first impression of who the person behind that voice might be. We characterize how these multivariate first impressions from voices emerge over time across different levels of abstraction using electroencephalography and representational similarity analysis. We find that for eight perceived physical (gender, age, and health), trait (attractiveness, dominance, and trustworthiness), and social characteristics (educatedness and professionalism), representations emerge early (~80 ms after stimulus onset), with voice acoustics contributing to those representations between ~100 ms and 400 ms. While impressions of person characteristics are highly correlated, we can find evidence for highly abstracted, independent representations of individual person characteristics. These abstracted representationse merge gradually over time. That is, representations of physical characteristics (age, gender) arise early (from ~120 ms), while representations of some trait and social characteristics emerge later (~360 ms onward). The findings align with recent theoretical models and shed light on the computations underpinning person perception from voices.
Collapse
Affiliation(s)
- Nadine Lavan
- Department of Biological and Experimental Psychology, School of Biological and Behavioural Sciences, Queen Mary University of London, London E1 4NS, United Kingdom
| | - Paula Rinke
- Research Group Phonetics, Institute of German Linguistics, Philipps-University Marburg, Marburg 35037, Germany
| | - Mathias Scharinger
- Research Group Phonetics, Institute of German Linguistics, Philipps-University Marburg, Marburg 35037, Germany
- Research Center "Deutscher Sprachatlas", Philipps-University Marburg, Marburg 35037, Germany
- Center for Mind, Brain & Behavior, Universities of Marburg & Gießen, Marburg 35032, Germany
| |
Collapse
|
2
|
Roche JM, Asaro K, Morris BJ, Morgan SD. Gender stereotypes and social perception of vocal confidence is mitigated by salience of socio-indexical cues to gender. Front Psychol 2023; 14:1125164. [PMID: 38155698 PMCID: PMC10753021 DOI: 10.3389/fpsyg.2023.1125164] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2022] [Accepted: 11/09/2023] [Indexed: 12/30/2023] Open
Abstract
Introduction Socio-indexical cues to gender and vocal affect often interact and sometimes lead listeners to make differential judgements of affective intent based on the gender of the speaker. Previous research suggests that rising intonation is a common cue that both women and men produce to communicate lack of confidence, but listeners are more sensitive to this cue when it is produced by women. Some speech perception theories assume that listeners will track conditional statistics of speech and language cues (e.g., frequency of the socio-indexical cues to gender and affect) in their listening and communication environments during speech perception. It is currently less clear if these conditional statistics will impact listener ratings when context varies (e.g., number of talkers). Methods To test this, we presented listeners with vocal utterances from one female and one male-pitched voice (single talker condition) or many female/male-pitched voices (4 female voices; 4 female voices pitch-shifted to a male range) to examine how they impacted perceptions of talker confidence. Results Results indicated that when one voice was evaluated, listeners defaulted to the gender stereotype that the female voice using rising intonation (a cue to lack of confidence) was less confident than the male-pitched voice (using the same cue). However, in the multi-talker condition, this effect went away and listeners equally rated the confidence of the female and male-pitched voices. Discussion Findings support dual process theories of information processing, such that listeners may rely on heuristics when speech perception is devoid of context, but when there are no differentiating qualities across talkers (regardless of gender), listeners may be ideal adapters who focus on only the relevant cues.
Collapse
Affiliation(s)
- Jennifer M. Roche
- Educational Psychology, Kent State University, Kent, OH, United States
- Speech Pathology & Audiology, Kent State University, Kent, OH, United States
| | - Katie Asaro
- Educational Psychology, Kent State University, Kent, OH, United States
| | - Bradley J. Morris
- Educational Psychology, Kent State University, Kent, OH, United States
| | - Shae D. Morgan
- Otolaryngology, Head and Neck Surgery and Communicative Disorders, University of Louisville, Louisville, KY, United States
| |
Collapse
|
3
|
Mauchand M, Pell MD. Complain like you mean it! How prosody conveys suffering even about innocuous events. BRAIN AND LANGUAGE 2023; 244:105305. [PMID: 37562118 DOI: 10.1016/j.bandl.2023.105305] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/23/2022] [Revised: 07/28/2023] [Accepted: 07/30/2023] [Indexed: 08/12/2023]
Abstract
When complaining, speakers can use their voice to convey a feeling of pain, even when describing innocuous events. Rapid detection of emotive and identity features of the voice may constrain how the semantic content of complaints is processed, as indexed by N400 and P600 effects evoked by the final, pain-related word. Twenty-six participants listened to statements describing painful and innocuous events expressed in a neutral or complaining voice, produced by ingroup and outgroup accented speakers. Participants evaluated how hurt the speaker felt under EEG monitoring. Principal Component Analysis of Event-Related Potentials from the final word onset demonstrated N400 and P600 increases when complainers described innocuous vs. painful events in a neutral voice, but these effects were altered when utterances were expressed in a complaining voice. Independent of prosody, N400 amplitudes increased for complaints spoken in outgroup vs. ingroup accents. Results demonstrate that prosody and accent constrain the processing of spoken complaints as proposed in a parallel-constraint-satisfaction model.
Collapse
Affiliation(s)
- Maël Mauchand
- McGill University, School of Communication Sciences and Disorders, Montréal, Québec, Canada.
| | - Marc D Pell
- McGill University, School of Communication Sciences and Disorders, Montréal, Québec, Canada
| |
Collapse
|
4
|
Lavan N. The Time Course of Person Perception From Voices: A Behavioral Study. Psychol Sci 2023:9567976231161565. [PMID: 37227791 DOI: 10.1177/09567976231161565] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/27/2023] Open
Abstract
Listeners spontaneously form impressions of a person from their voice: Is someone old or young? Trustworthy or untrustworthy? Some studies suggest that these impressions emerge rapidly (e.g., < 400 ms for traits), but it is unclear just how rapidly different impressions can emerge and whether the time courses differ across characteristics. I presented 618 adult listeners with voice recordings ranging from 25 ms to 800 ms in duration and asked them to rate physical (age, sex, health), trait (trustworthiness, dominance, attractiveness), and social (educatedness, poshness, professionalism) characteristics. I then used interrater agreement as an index for impression formation. Impressions of physical characteristics and dominance emerged fastest, showing high agreement after only 25 ms of exposure. In contrast, agreement for trait and social characteristics was initially low to moderate and gradually increased. Such a staggered time course suggests that there could be a temporo-perceptual hierarchy for person perception in which faster impressions could influence later ones.
Collapse
Affiliation(s)
- Nadine Lavan
- Department of Biological and Experimental Psychology, School of Biological and Behavioural Sciences, Queen Mary University of London
| |
Collapse
|
5
|
Fu H, Yu H, Wang X, Lu X, Zhu C. A Semi-Supervised Speech Deception Detection Algorithm Combining Acoustic Statistical Features and Time-Frequency Two-Dimensional Features. Brain Sci 2023; 13:brainsci13050725. [PMID: 37239197 DOI: 10.3390/brainsci13050725] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2023] [Revised: 04/23/2023] [Accepted: 04/24/2023] [Indexed: 05/28/2023] Open
Abstract
Human lying is influenced by cognitive neural mechanisms in the brain, and conducting research on lie detection in speech can help to reveal the cognitive mechanisms of the human brain. Inappropriate deception detection features can easily lead to dimension disaster and make the generalization ability of the widely used semi-supervised speech deception detection model worse. Because of this, this paper proposes a semi-supervised speech deception detection algorithm combining acoustic statistical features and time-frequency two-dimensional features. Firstly, a hybrid semi-supervised neural network based on a semi-supervised autoencoder network (AE) and a mean-teacher network is established. Secondly, the static artificial statistical features are input into the semi-supervised AE to extract more robust advanced features, and the three-dimensional (3D) mel-spectrum features are input into the mean-teacher network to obtain features rich in time-frequency two-dimensional information. Finally, a consistency regularization method is introduced after feature fusion, effectively reducing the occurrence of over-fitting and improving the generalization ability of the model. This paper carries out experiments on the self-built corpus for deception detection. The experimental results show that the highest recognition accuracy of the algorithm proposed in this paper is 68.62% which is 1.2% higher than the baseline system and effectively improves the detection accuracy.
Collapse
Affiliation(s)
- Hongliang Fu
- Key Laboratory of Food Information Processing and Control, Ministry of Education, Henan University of Technology, Zhengzhou 450001, China
- Henan Engineering Laboratory of Grain IOT Technology, Henan University of Technology, Zhengzhou 450001, China
| | - Hang Yu
- Key Laboratory of Food Information Processing and Control, Ministry of Education, Henan University of Technology, Zhengzhou 450001, China
| | - Xuemei Wang
- Key Laboratory of Food Information Processing and Control, Ministry of Education, Henan University of Technology, Zhengzhou 450001, China
- Henan Engineering Laboratory of Grain IOT Technology, Henan University of Technology, Zhengzhou 450001, China
| | - Xiangying Lu
- Key Laboratory of Food Information Processing and Control, Ministry of Education, Henan University of Technology, Zhengzhou 450001, China
| | - Chunhua Zhu
- Key Laboratory of Food Information Processing and Control, Ministry of Education, Henan University of Technology, Zhengzhou 450001, China
- Henan Engineering Laboratory of Grain IOT Technology, Henan University of Technology, Zhengzhou 450001, China
| |
Collapse
|
6
|
Ji Y, Hu Y, Jiang X. Segmental and suprasegmental encoding of speaker confidence in Wuxi dialect vowels. Front Psychol 2022; 13:1028106. [PMID: 36578688 PMCID: PMC9791101 DOI: 10.3389/fpsyg.2022.1028106] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2022] [Accepted: 11/15/2022] [Indexed: 12/14/2022] Open
Abstract
Introduction Wuxi dialect is a variation of Wu dialect spoken in eastern China and is characterized by a rich tonal system. Compared with standard Mandarin speakers, those of Wuxi dialect as their mother tongue can be more efficient in varying vocal cues to encode communicative meanings in speech communication. While literature has demonstrated that speakers encode high vs. low confidence in global prosodic cues at the sentence level, it is unknown how speakers' intended confidence is encoded at a more local, phonetic level. This study aimed to explore the effects of speakers' intended confidence on both prosodic and formant features of vowels in two lexical tones (the flat tone and the contour tone) of Wuxi dialect. Methods Words of a single vowel were spoken in confident, unconfident, or neutral tone of voice by native Wuxi dialect speakers using a standard elicitation procedure. Linear-mixed effects modeling and parametric bootstrapping testing were performed. Results The results showed that (1) the speakers raised both F1 and F2 in the confident level (compared with the neutral-intending expression). Additionally, F1 can distinguish between the confident and unconfident expressions; (2) Compared with the neutral-intending expression, the speakers raised mean f0, had a greater variation of f0 and prolonged pronunciation time in the unconfident level while they raised mean intensity, had a greater variation of intensity and prolonged pronunciation time in the confident level. (3) The speakers modulated mean f0 and mean intensity to a larger extent on the flat tone than the contour tone to differentiate between levels of confidence in the voice, while they modulated f0 and intensity range more only on the contour tone. Discussion These findings shed new light on the mechanisms of segmental and suprasegmental encoding of speaker confidence and lack of confidence at the vowel level, highlighting the interplay of lexical tone and vocal expression in speech communication.
Collapse
|
7
|
Mauchand M, Pell MD. Listen to my feelings! How prosody and accent drive the empathic relevance of complaining speech. Neuropsychologia 2022; 175:108356. [PMID: 36037914 DOI: 10.1016/j.neuropsychologia.2022.108356] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2022] [Revised: 08/04/2022] [Accepted: 08/22/2022] [Indexed: 10/15/2022]
Abstract
Interpersonal communication often involves sharing our feelings with others; complaining, for example, aims to elicit empathy in listeners by vocally expressing a speaker's suffering. Despite the growing neuroscientific interest in the phenomenon of empathy, few have investigated how it is elicited in real time by vocal signals (prosody), and how this might be affected by interpersonal factors, such as a speaker's cultural background (based on their accent). To investigate the neural processes at play when hearing spoken complaints, twenty-six French participants listened to complaining and neutral utterances produced by in-group French and out-group Québécois (i.e., French-Canadian) speakers. Participants rated how hurt the speaker felt while their cerebral activity was monitored with electroencephalography (EEG). Principal Component Analysis of Event-Related Potentials (ERPs) taken at utterance onset showed culture-dependent time courses of emotive prosody processing. The high motivational relevance of ingroup complaints increased the P200 response compared to all other utterance types; in contrast, outgroup complaints selectively elicited an early posterior negativity in the same time window, followed by an increased N400 (due to ongoing effort to derive affective meaning from outgroup voices). Ingroup neutral utterances evoked a late negativity which may reflect re-analysis of emotively less salient, but culturally relevant ingroup speech. Results highlight the time-course of neurocognitive responses that contribute to emotive speech processing for complaints, establishing the critical role of prosody as well as social-relational factors (i.e., cultural identity) on how listeners are likely to "empathize" with a speaker.
Collapse
Affiliation(s)
- Maël Mauchand
- McGill University, School of Communication Sciences and Disorders, Montréal, Québec, Canada.
| | - Marc D Pell
- McGill University, School of Communication Sciences and Disorders, Montréal, Québec, Canada
| |
Collapse
|
8
|
Shang A, Bieszczad KM. Epigenetic mechanisms regulate cue memory underlying discriminative behavior. Neurosci Biobehav Rev 2022; 141:104811. [PMID: 35961385 DOI: 10.1016/j.neubiorev.2022.104811] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2022] [Revised: 06/15/2022] [Accepted: 08/01/2022] [Indexed: 12/01/2022]
Abstract
The burgeoning field of neuroepigenetics has introduced chromatin modification as an important interface between experience and brain function. For example, epigenetic mechanisms like histone acetylation and DNA methylation operate throughout a lifetime to powerfully regulate gene expression in the brain that is required for experiences to be transformed into long-term memories. This review highlights emerging evidence from sensory models of memory that converge on the premise that epigenetic regulation of activity-dependent transcription in the sensory brain facilitates highly precise memory recall. Chromatin modifications may be key for neurophysiological responses to transient sensory cue features experienced in the "here and now" to be recapitulated over the long term. We conclude that the function of epigenetic control of sensory system neuroplasticity is to regulate the amount and type of sensory information retained in long-term memories by regulating neural representations of behaviorally relevant cues that guide behavior. This is of broad importance in the neuroscience field because there are few circumstances in which behavioral acts are devoid of an initiating sensory experience.
Collapse
Affiliation(s)
- Andrea Shang
- Dept. of Psychology - Behavioral and Systems Neuroscience, Rutgers University - New Brunswick, 152 Frelinghuysen Road, Piscataway, NJ 08854, USA
| | - Kasia M Bieszczad
- Dept. of Psychology - Behavioral and Systems Neuroscience, Rutgers University - New Brunswick, 152 Frelinghuysen Road, Piscataway, NJ 08854, USA; Rutgers Center for Cognitive Science (RuCCS), Rutgers University, Piscataway, NJ 08854, USA; Department of Otolaryngology - Head and Neck Surgery, Rutgers Robert Wood Johnson Medical School, New Brunswick, NJ 08854, USA.
| |
Collapse
|
9
|
Roche JM, Morgan SD, Fisk S. Gender stereotypes drive perceptual differences of vocal confidence. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 151:3031. [PMID: 35649917 DOI: 10.1121/10.0010382] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/30/2021] [Accepted: 04/15/2022] [Indexed: 06/15/2023]
Abstract
One's ability to express confidence is critical to achieve one's goals in a social context-such as commanding respect from others, establishing higher social status, and persuading others. How individuals perceive confidence may be shaped by the socio-indexical cues produced by the speaker. In the current production/perception study, we asked four speakers (two cisgender women/men) to answer trivia questions under three speaking contexts: natural, overconfident, and underconfident (i.e., lack of confidence). An evaluation of the speakers' acoustics indicated that the speakers significantly varied their acoustic cues as a function of speaking context and that the women and men had significantly different acoustic cues. The speakers' answers to the trivia questions in the three contexts (natural, overconfident, underconfident) were then presented to listeners (N = 26) in a social judgment task using a computer mouse-tracking paradigm. Listeners were sensitive to the speakers' acoustic modulations of confidence and differentially interpreted these cues based on the perceived gender of the speaker, thereby impacting listeners' cognition and social decision making. We consider, then, how listeners' social judgments about confidence were impacted by gender stereotypes about women and men from social, heuristic-based processes.
Collapse
Affiliation(s)
- Jennifer M Roche
- Schools of Health Sciences and Lifespan Development and Educational Sciences, Kent State University, 800 Summit Street, Kent, Ohio 44224, USA
| | - Shae D Morgan
- Department of Otolaryngology Head and Neck Surgery and Communicative Disorders, University of Louisville, 401 East Chestnut Street, Suite 170, Louisville, Kentucky 40202, USA
| | - Susan Fisk
- Department of Sociology, Kent State University, 800 Summit Street, Kent, Ohio 44224, USA
| |
Collapse
|
10
|
|
11
|
Schirmer A, Lai O, McGlone F, Cham C, Lau D. Gentle Stroking Elicits Somatosensory ERP that Differentiates Between Hairy and Glabrous Skin. Soc Cogn Affect Neurosci 2022; 17:864-875. [PMID: 35277720 PMCID: PMC9433843 DOI: 10.1093/scan/nsac012] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2021] [Revised: 01/14/2022] [Accepted: 02/20/2022] [Indexed: 11/14/2022] Open
Abstract
Here we asked whether, similar to visual and auditory event-related potentials (ERPs), somatosensory ERPs reflect affect. Participants were stroked on hairy or glabrous skin at five stroking velocities (0.5, 1, 3, 10 and 20 cm/s). For stroking of hairy skin, pleasantness ratings related to velocity in an inverted u-shaped manner. ERPs showed a negativity at 400 ms following touch onset over somatosensory cortex contra-lateral to the stimulation site. This negativity, referred to as sN400, was larger for intermediate than for faster and slower velocities and positively predicted pleasantness ratings. For stroking of glabrous skin, pleasantness showed again an inverted u-shaped relation with velocity and, additionally, increased linearly with faster stroking. The sN400 revealed no quadratic effect and instead was larger for faster velocities. Its amplitude failed to significantly predict pleasantness. In sum, as was reported for other senses, a touch’s affective value modulates the somatosensory ERP. Notably, however, this ERP and associated subjective pleasantness dissociate between hairy and glabrous skin underscoring functional differences between the skin with which we typically receive touch and the skin with which we typically reach out to touch.
Collapse
Affiliation(s)
- Annett Schirmer
- Department of Psychology, The Chinese University of Hong Kong, Hong Kong SAR
- The Brain and Mind Institute, The Chinese University of Hong Kong, Shatin, Hong Kong SAR
| | - Oscar Lai
- Department of Psychology, The Chinese University of Hong Kong, Hong Kong SAR
| | - Francis McGlone
- School of Natural Sciences & Psychology, Liverpool John Moores University, UK
- Institute of Psychology, Health & Society, University of Liverpool, UK
| | - Clare Cham
- Department of Psychology, The Chinese University of Hong Kong, Hong Kong SAR
| | - Darwin Lau
- Department of Mechanical and Automation Engineering, The Chinese University of Hong Kong SAR
| |
Collapse
|
12
|
Pell MD, Sethi S, Rigoulot S, Rothermich K, Liu P, Jiang X. Emotional voices modulate perception and predictions about an upcoming face. Cortex 2022; 149:148-164. [DOI: 10.1016/j.cortex.2021.12.017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2021] [Revised: 09/15/2021] [Accepted: 01/05/2022] [Indexed: 11/26/2022]
|
13
|
The neural basis of authenticity recognition in laughter and crying. Sci Rep 2021; 11:23750. [PMID: 34887461 PMCID: PMC8660868 DOI: 10.1038/s41598-021-03131-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2021] [Accepted: 11/22/2021] [Indexed: 01/28/2023] Open
Abstract
Deciding whether others' emotions are genuine is essential for successful communication and social relationships. While previous fMRI studies suggested that differentiation between authentic and acted emotional expressions involves higher-order brain areas, the time course of authenticity discrimination is still unknown. To address this gap, we tested the impact of authenticity discrimination on event-related potentials (ERPs) related to emotion, motivational salience, and higher-order cognitive processing (N100, P200 and late positive complex, the LPC), using vocalised non-verbal expressions of sadness (crying) and happiness (laughter) in a 32-participant, within-subject study. Using a repeated measures 2-factor (authenticity, emotion) ANOVA, we show that N100's amplitude was larger in response to authentic than acted vocalisations, particularly in cries, while P200's was larger in response to acted vocalisations, particularly in laughs. We suggest these results point to two different mechanisms: (1) a larger N100 in response to authentic vocalisations is consistent with its link to emotional content and arousal (putatively larger amplitude for genuine emotional expressions); (2) a larger P200 in response to acted ones is in line with evidence relating it to motivational salience (putatively larger for ambiguous emotional expressions). Complementarily, a significant main effect of emotion was found on P200 and LPC amplitudes, in that the two were larger for laughs than cries, regardless of authenticity. Overall, we provide the first electroencephalographic examination of authenticity discrimination and propose that authenticity processing of others' vocalisations is initiated early, along that of their emotional content or category, attesting for its evolutionary relevance for trust and bond formation.
Collapse
|
14
|
Side- and similarity-biases during confidence conformity. PLoS One 2021; 16:e0253577. [PMID: 34270563 PMCID: PMC8284640 DOI: 10.1371/journal.pone.0253577] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2020] [Accepted: 05/31/2021] [Indexed: 11/20/2022] Open
Abstract
Memory conformity may develop when people are confronted with distinct memories reported by others in social situations and knowingly/unknowingly adhere to these exogenous memories. Earlier research on memory conformity suggests that (1) subjects were more likely to conform to confederate with high confidence; (2) subjects with low confidence on their memory accuracy were more likely to conform, and; (3) this subjective confidence could be adjusted by social manipulations. Nonetheless, it remains unclear how the confidence levels of ours and others may interact and produce a combined effect on our degree of conformity. More importantly, is memory conformity, defined by a complete adoption of the opposite side, the result of a gradual accumulation of subtler changes at the confidence level, i.e., a buildup of confidence conformity? Here, we followed participant’s confidence transformation quantitatively over three confederate sessions in a memory test. After studying a set of human motion videos, participants had to answer simultaneously whether a target or lure video had appeared before by indicating their side (i.e., Yes/No) and their associated confidence rating. Participants were allowed to adjust their responses as they were being shown randomly-generated confederates’ answers and confidence values. Results show that participants indeed demonstrated confidence conformity. Interestingly, they tended to become committed to their side early on and gain confidence gradually over subsequent sessions. This polarizing behaviour may be explained by two kinds of preferences: (1) Participant’s confidence enhancement towards same-sided confederates was greater in magnitude compared to the decrement towards an opposite-sided confederate; and (2) Participants had the most effective confidence boost when the same-sided confederates shared similar, but not considerably different, confidence level to theirs. In other words, humans exhibit side- and similarity-biases during confidence conformity.
Collapse
|
15
|
Guyer JJ, Briñol P, Vaughan-Johnston TI, Fabrigar LR, Moreno L, Petty RE. Paralinguistic Features Communicated through Voice can Affect Appraisals of Confidence and Evaluative Judgments. JOURNAL OF NONVERBAL BEHAVIOR 2021; 45:479-504. [PMID: 34744233 PMCID: PMC8553728 DOI: 10.1007/s10919-021-00374-2] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 05/25/2021] [Indexed: 11/07/2022]
Abstract
This article unpacks the basic mechanisms by which paralinguistic features communicated through the voice can affect evaluative judgments and persuasion. Special emphasis is placed on exploring the rapidly emerging literature on vocal features linked to appraisals of confidence (e.g., vocal pitch, intonation, speech rate, loudness, etc.), and their subsequent impact on information processing and meta-cognitive processes of attitude change. The main goal of this review is to advance understanding of the different psychological processes by which paralinguistic markers of confidence can affect attitude change, specifying the conditions under which they are more likely to operate. In sum, we highlight the importance of considering basic mechanisms of attitude change to predict when and why appraisals of paralinguistic markers of confidence can lead to more or less persuasion.
Collapse
Affiliation(s)
- Joshua J. Guyer
- Department of Social Psychology and Methodology, Universidad Autónoma de Madrid, Madrid, Spain
| | - Pablo Briñol
- Department of Social Psychology and Methodology, Universidad Autónoma de Madrid, Madrid, Spain
| | | | | | - Lorena Moreno
- Department of Social Psychology and Methodology, Universidad Autónoma de Madrid, Madrid, Spain
| | - Richard E. Petty
- Department of Psychology, The Ohio State University, Columbus, USA
| |
Collapse
|
16
|
Caballero JA, Mauchand M, Jiang X, Pell MD. Cortical processing of speaker politeness: Tracking the dynamic effects of voice tone and politeness markers. Soc Neurosci 2021; 16:423-438. [PMID: 34102955 DOI: 10.1080/17470919.2021.1938667] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
Abstract
Information in the tone of voice alters social impressions and underlying brain activity as listeners evaluate the interpersonal relevance of utterances. Here, we presented requests that expressed politeness distinctions through the voice (polite/rude) and explicit linguistic markers (half of the requests began with Please). Thirty participants performed a social perception task (rating friendliness) while their electroencephalogram was recorded. Behaviorally, vocal politeness strategies had a much stronger influence on the perceived friendliness than the linguistic marker. Event-related potentials revealed rapid effects of (im)polite voices on cortical activity prior to ~300 ms; P200 amplitudes increased for polite versus rude voices, suggesting that the speaker's polite stance was registered as more salient in our task. At later stages, politeness distinctions encoded by the speaker's voice and their use of Please interacted, modulating activity in the N400 (300-500 ms) and late positivity (600-800 ms) time windows. Patterns of results suggest that initial attention deployment to politeness cues is rapidly influenced by the motivational significance of a speaker's voice. At later stages, processes for integrating vocal and lexical information resulted in increased cognitive effort to reevaluate utterances with ambiguous/contradictory cues. The potential influence of social anxiety on the P200 effect is also discussed.
Collapse
Affiliation(s)
- Jonathan A Caballero
- School of Communication Sciences and Disorders 2001 McGill College, McGill University, Montréal, Québec, Canada
| | - Maël Mauchand
- School of Communication Sciences and Disorders 2001 McGill College, McGill University, Montréal, Québec, Canada
| | - Xiaoming Jiang
- Shanghai International Studies University, Institute of Linguistics (IoL), Shanghai, China
| | - Marc D Pell
- School of Communication Sciences and Disorders 2001 McGill College, McGill University, Montréal, Québec, Canada
| |
Collapse
|
17
|
Foucart A, Hartsuiker RJ. Are foreign-accented speakers that 'incredible'? The impact of the speaker's indexical properties on sentence processing. Neuropsychologia 2021; 158:107902. [PMID: 34052231 DOI: 10.1016/j.neuropsychologia.2021.107902] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2020] [Revised: 03/16/2021] [Accepted: 05/16/2021] [Indexed: 12/31/2022]
Abstract
This study investigated the impact of the speaker's identity generated by the voice on sentence processing. We examined the relation between ERP components associated with the processing of the voice (N100 and P200) from voice onset and those associated with sentence processing (N400 and late positivity) from critical word onset. We presented Dutch native speakers with sentences containing true (and known) information, unknown (but true) information or information violating world knowledge and had them perform a truth evaluation task. Sentences were spoken either in a native or a foreign accent. Truth evaluation judgments were not different for statements spoken by the native-accented and the foreign-accented speakers. Reduced N100 and P200 were observed in response to the foreign speaker's voice compared to the native speaker's. While statements containing unknown information or world knowledge violations generated a larger N400 than true statements in the native condition, they were not significantly different in the foreign condition, suggesting shallower processing of foreign-accented speech. The N100 was a significant predictor for the N400 in that the reduced N100 observed for the foreign speaker compared to the native speaker was related to a smaller N400 effect. These finding suggest that the impression of the speaker that listeners rapidly form from the voice affects semantic processing, which confirms that speaker's identity and language comprehension cannot be dissociated.
Collapse
Affiliation(s)
- Alice Foucart
- - Centro de Ciencia Cognitiva C3, Facultad de Lenguas y Educación, Universidad Nebrija, Spain.
| | | |
Collapse
|
18
|
Abstract
Neurocognitive models (e.g., Schirmer & Kotz, 2006) have helped to characterize how listeners incrementally derive meaning from vocal expressions of emotion in spoken language, what neural mechanisms are involved at different processing stages, and their relative time course. But how can these insights be applied to communicative situations in which prosody serves a predominantly interpersonal function? This comment examines recent data highlighting the dynamic interplay of prosody and language, when vocal attributes serve the sociopragmatic goals of the speaker or reveal interpersonal information that listeners use to construct a mental representation of what is being communicated. Our comment serves as a beacon to researchers interested in how the neurocognitive system “makes sense” of socioemotive aspects of prosody.
Collapse
Affiliation(s)
- Marc D. Pell
- School of Communication Sciences and Disorders, McGill University, Canada
| | - Sonja A. Kotz
- Department of Neuropsychology and Psychopharmacology, Maastricht University, The Netherlands
- Department of Neuropsychology, Max Planck Institute for Human Cognitive and Brain Sciences, Germany
| |
Collapse
|
19
|
Immediate online use of prosody reveals the ironic intentions of a speaker: neurophysiological evidence. COGNITIVE AFFECTIVE & BEHAVIORAL NEUROSCIENCE 2021; 21:74-92. [DOI: 10.3758/s13415-020-00849-7] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 11/08/2020] [Indexed: 12/12/2022]
|
20
|
Nonverbal auditory communication - Evidence for integrated neural systems for voice signal production and perception. Prog Neurobiol 2020; 199:101948. [PMID: 33189782 DOI: 10.1016/j.pneurobio.2020.101948] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2020] [Revised: 10/12/2020] [Accepted: 11/04/2020] [Indexed: 12/24/2022]
Abstract
While humans have developed a sophisticated and unique system of verbal auditory communication, they also share a more common and evolutionarily important nonverbal channel of voice signaling with many other mammalian and vertebrate species. This nonverbal communication is mediated and modulated by the acoustic properties of a voice signal, and is a powerful - yet often neglected - means of sending and perceiving socially relevant information. From the viewpoint of dyadic (involving a sender and a signal receiver) voice signal communication, we discuss the integrated neural dynamics in primate nonverbal voice signal production and perception. Most previous neurobiological models of voice communication modelled these neural dynamics from the limited perspective of either voice production or perception, largely disregarding the neural and cognitive commonalities of both functions. Taking a dyadic perspective on nonverbal communication, however, it turns out that the neural systems for voice production and perception are surprisingly similar. Based on the interdependence of both production and perception functions in communication, we first propose a re-grouping of the neural mechanisms of communication into auditory, limbic, and paramotor systems, with special consideration for a subsidiary basal-ganglia-centered system. Second, we propose that the similarity in the neural systems involved in voice signal production and perception is the result of the co-evolution of nonverbal voice production and perception systems promoted by their strong interdependence in dyadic interactions.
Collapse
|
21
|
Vergis N, Jiang X, Pell MD. Neural responses to interpersonal requests: Effects of imposition and vocally-expressed stance. Brain Res 2020; 1740:146855. [DOI: 10.1016/j.brainres.2020.146855] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2019] [Revised: 04/02/2020] [Accepted: 04/23/2020] [Indexed: 02/07/2023]
|
22
|
Rigoulot S, Jiang X, Vergis N, Pell MD. Neurophysiological correlates of sexually evocative speech. Biol Psychol 2020; 154:107909. [DOI: 10.1016/j.biopsycho.2020.107909] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2019] [Revised: 05/14/2020] [Accepted: 05/20/2020] [Indexed: 12/11/2022]
|
23
|
Jiang X, Gossack-Keenan K, Pell MD. To believe or not to believe? How voice and accent information in speech alter listener impressions of trust. Q J Exp Psychol (Hove) 2019; 73:55-79. [DOI: 10.1177/1747021819865833] [Citation(s) in RCA: 22] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/19/2023]
Abstract
Our decision to believe what another person says can be influenced by vocally expressed confidence in speech and by whether the speaker–listener are members of the same social group. The dynamic effects of these two information sources on neurocognitive processes that promote believability impressions from vocal cues are unclear. Here, English Canadian listeners were presented personal statements ( She has access to the building) produced in a confident or doubtful voice by speakers of their own dialect (in-group) or speakers from two different “out-groups” (regional or foreign-accented English). Participants rated how believable the speaker is for each statement and event-related potentials (ERPs) were analysed from utterance onset. Believability decisions were modulated by both the speaker’s vocal confidence level and their perceived in-group status. For in-group speakers, ERP effects revealed an early differentiation of vocally expressed confidence (i.e., N100, P200), highlighting the motivational significance of doubtful voices for drawing believability inferences. These early effects on vocal confidence perception were qualitatively different or absent when speakers had an accent; evaluating out-group voices was associated with increased demands on contextual integration and re-analysis of a non-native representation of believability (i.e., increased N400, late negativity response). Accent intelligibility and experience with particular out-group accents each influenced how vocal confidence was processed for out-group speakers. The N100 amplitude was sensitive to out-group attitudes and predicted actual believability decisions for certain out-group speakers. We propose a neurocognitive model in which vocal identity information (social categorization) dynamically influences how vocal expressions are decoded and used to derive social inferences during person perception.
Collapse
Affiliation(s)
- Xiaoming Jiang
- School of Communication Sciences and Disorders, McGill University, Montréal, Québec, Canada
- Department of Psychology, Tongji University, Shanghai, China
| | - Kira Gossack-Keenan
- School of Communication Sciences and Disorders, McGill University, Montréal, Québec, Canada
| | - Marc D Pell
- School of Communication Sciences and Disorders, McGill University, Montréal, Québec, Canada
| |
Collapse
|
24
|
Roche JM, Morgan SD, Fissel Brannick S, Bryndel K. Acoustic correlates of female confidence: A production and comprehension study. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 145:3410. [PMID: 31255138 DOI: 10.1121/1.5109661] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/24/2018] [Accepted: 05/09/2019] [Indexed: 06/09/2023]
Abstract
In the current study, an interactive approach is used to explore possible contributors to the misattributions listeners make about female talker expression of confidence. To do this, the expression and identification of confidence was evaluated through the evaluation of talker- (e.g., talker knowledge and affective acoustic modulation) and listener-specific factors (e.g., interaction between talker acoustic cues and listener knowledge). Talker and listener contexts were manipulated by implementing a social constraint for talkers and withholding information from listeners. Results indicated that listeners were sensitive to acoustic information produced by the female talkers in this study. However, when world knowledge and acoustics competed, judgments of talker confidence by listeners were less accurate. In fact, acoustic cues to female talker confidence were more accurately used by listeners as a cue to perceived confidence when relevant world knowledge was missing. By targeting speech dynamics between female talkers and both female and male listeners, the current study provides a better understanding of how confidence is realized acoustically and, perhaps more importantly, how those cues may be interpreted/misinterpreted by listeners.
Collapse
Affiliation(s)
- Jennifer M Roche
- School of Health Sciences, Kent State University, 1325 Theatre Drive, CPA A126, Kent, Ohio 44242, USA
| | - Shae D Morgan
- Department of Otolaryngology Head and Neck Surgery and Communicative Disorders, University of Louisville, 627 South Preston Street, Suite 220, Louisville, Kentucky 40292, USA
| | - Schea Fissel Brannick
- Speech-Language Pathology Program, Midwestern University, Glendale, 619555 North 59th Avenue, Glendale, Arizona 85308, USA
| | - Krystal Bryndel
- School of Health Sciences, Kent State University, 1325 Theatre Drive, CPA A126, Kent, Ohio 44242, USA
| |
Collapse
|
25
|
Hellbernd N, Sammler D. Neural bases of social communicative intentions in speech. Soc Cogn Affect Neurosci 2019; 13:604-615. [PMID: 29771359 PMCID: PMC6022564 DOI: 10.1093/scan/nsy034] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2017] [Accepted: 05/13/2018] [Indexed: 11/15/2022] Open
Abstract
Our ability to understand others’ communicative intentions in speech is key to successful social interaction. Indeed, misunderstanding an ‘excuse me’ as apology, while meant as criticism, may have important consequences. Recent behavioural studies have provided evidence that prosody, that is, vocal tone, is an important indicator for speakers’ intentions. Using a novel audio-morphing paradigm, the present functional magnetic resonance imaging study examined the neurocognitive mechanisms that allow listeners to ‘read’ speakers’ intents from vocal prosodic patterns. Participants categorized prosodic expressions that gradually varied in their acoustics between criticism, doubt, and suggestion. Categorizing typical exemplars of the three intentions induced activations along the ventral auditory stream, complemented by amygdala and mentalizing system. These findings likely depict the stepwise conversion of external perceptual information into abstract prosodic categories and internal social semantic concepts, including the speaker’s mental state. Ambiguous tokens, in turn, involved cingulo-opercular areas known to assist decision-making in case of conflicting cues. Auditory and decision-making processes were flexibly coupled with the amygdala, depending on prosodic typicality, indicating enhanced categorization efficiency of overtly relevant, meaningful prosodic signals. Altogether, the results point to a model in which auditory prosodic categorization and socio-inferential conceptualization cooperate to translate perceived vocal tone into a coherent representation of the speaker’s intent.
Collapse
Affiliation(s)
- Nele Hellbernd
- Otto Hahn Group Neural Bases of Intonation in Speech and Music, Max Planck Institute for Human Cognitive and Brain Sciences, Stephanstraße 1a, D-04103 Leipzig, Germany
| | - Daniela Sammler
- Otto Hahn Group Neural Bases of Intonation in Speech and Music, Max Planck Institute for Human Cognitive and Brain Sciences, Stephanstraße 1a, D-04103 Leipzig, Germany
| |
Collapse
|
26
|
Paulmann S, Weinstein N, Zougkou K. Now listen to this! Evidence from a cross-spliced experimental design contrasting pressuring and supportive communications. Neuropsychologia 2019; 124:192-201. [DOI: 10.1016/j.neuropsychologia.2018.12.011] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2018] [Revised: 12/13/2018] [Accepted: 12/14/2018] [Indexed: 02/04/2023]
|
27
|
Jiang X, Sanford R, Pell MD. Neural architecture underlying person perception from in-group and out-group voices. Neuroimage 2018; 181:582-597. [DOI: 10.1016/j.neuroimage.2018.07.042] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2017] [Revised: 07/04/2018] [Accepted: 07/16/2018] [Indexed: 01/02/2023] Open
|
28
|
Mahrholz G, Belin P, McAleer P. Judgements of a speaker's personality are correlated across differing content and stimulus type. PLoS One 2018; 13:e0204991. [PMID: 30286148 PMCID: PMC6171871 DOI: 10.1371/journal.pone.0204991] [Citation(s) in RCA: 24] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2017] [Accepted: 09/18/2018] [Indexed: 11/19/2022] Open
Abstract
It has previously been shown that first impressions of a speaker's personality, whether accurate or not, can be judged from short utterances of vowels and greetings, as well as from prolonged sentences and readings of complex paragraphs. From these studies, it is established that listeners' judgements are highly consistent with one another, suggesting that different people judge personality traits in a similar fashion, with three key personality traits being related to measures of valence (associated with trustworthiness), dominance, and attractiveness. Yet, particularly in voice perception, limited research has established the reliability of such personality judgements across stimulus types of varying lengths. Here we investigate whether first impressions of trustworthiness, dominance, and attractiveness of novel speakers are related when a judgement is made on hearing both one word and one sentence from the same speaker. Secondly, we test whether what is said, thus adjusting content, influences the stability of personality ratings. 60 Scottish voices (30 females) were recorded reading two texts: one of ambiguous content and one with socially-relevant content. One word (~500 ms) and one sentence (~3000 ms) were extracted from each recording for each speaker. 181 participants (138 females) rated either male or female voices across both content conditions (ambiguous, socially-relevant) and both stimulus types (word, sentence) for one of the three personality traits (trustworthiness, dominance, attractiveness). Pearson correlations showed personality ratings between words and sentences were strongly correlated, with no significant influence of content. In short, when establishing an impression of a novel speaker, judgments of three key personality traits are highly related whether you hear one word or one sentence, irrespective of what they are saying. This finding is consistent with initial personality judgments serving as elucidators of approach or avoidance behaviour, without modulation by time or content. All data and sounds are available on OSF (osf.io/s3cxy).
Collapse
Affiliation(s)
- Gaby Mahrholz
- School of Psychology, University of Glasgow, Glasgow, United Kingdom
| | - Pascal Belin
- Institut des Neurosciences de la Timone, UMR 7289, CNRS and Université Aix-Marseille, Marseille, France
| | - Phil McAleer
- School of Psychology, University of Glasgow, Glasgow, United Kingdom
| |
Collapse
|
29
|
The right touch: Stroking of CT-innervated skin promotes vocal emotion processing. COGNITIVE AFFECTIVE & BEHAVIORAL NEUROSCIENCE 2018; 17:1129-1140. [PMID: 28933047 PMCID: PMC5709431 DOI: 10.3758/s13415-017-0537-5] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Research has revealed a special mechanoreceptor, called C-tactile (CT) afferent, that is situated in hairy skin and that seems relevant for the processing of social touch. We pursued a possible role of this receptor in the perception of other social signals such as a person’s voice. Participants completed three sessions in which they heard surprised and neutral vocal and nonvocal sounds and detected rare sound repetitions. In a given session, participants received no touch or soft brushstrokes to the arm (CT innervated) or palm (CT free). Event-related potentials elicited to sounds revealed that stroking to the arm facilitated the integration of vocal and emotional information. The late positive potential was greater for surprised vocal relative to neutral vocal and nonvocal sounds, and this effect was greater for arm touch relative to both palm touch and no touch. Together, these results indicate that stroking to the arm facilitates the allocation of processing resources to emotional voices, thus supporting the possibility that CT stimulation benefits social perception cross-modally.
Collapse
|
30
|
Zougkou K, Weinstein N, Paulmann S. ERP correlates of motivating voices: quality of motivation and time-course matters. Soc Cogn Affect Neurosci 2018; 12:1687-1700. [PMID: 28525641 PMCID: PMC5647802 DOI: 10.1093/scan/nsx064] [Citation(s) in RCA: 22] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2016] [Accepted: 04/23/2017] [Indexed: 11/12/2022] Open
Abstract
Here, we conducted the first study to explore how motivations expressed through speech are processed in real-time. Participants listened to sentences spoken in two types of well-studied motivational tones (autonomy-supportive and controlling), or a neutral tone of voice. To examine this, listeners were presented with sentences that either signaled motivations through prosody (tone of voice) and words simultaneously (e.g. ‘You absolutely have to do it my way’ spoken in a controlling tone of voice), or lacked motivationally biasing words (e.g. ‘Why don’t we meet again tomorrow’ spoken in a motivational tone of voice). Event-related brain potentials (ERPs) in response to motivations conveyed through words and prosody showed that listeners rapidly distinguished between motivations and neutral forms of communication as shown in enhanced P2 amplitudes in response to motivational when compared with neutral speech. This early detection mechanism is argued to help determine the importance of incoming information. Once assessed, motivational language is continuously monitored and thoroughly evaluated. When compared with neutral speech, listening to controlling (but not autonomy-supportive) speech led to enhanced late potential ERP mean amplitudes, suggesting that listeners are particularly attuned to controlling messages. The importance of controlling motivation for listeners is mirrored in effects observed for motivations expressed through prosody only. Here, an early rapid appraisal, as reflected in enhanced P2 amplitudes, is only found for sentences spoken in controlling (but not autonomy-supportive) prosody. Once identified as sounding pressuring, the message seems to be preferentially processed, as shown by enhanced late potential amplitudes in response to controlling prosody. Taken together, results suggest that motivational and neutral language are differentially processed; further, the data suggest that listening to cues signaling pressure and control cannot be ignored and lead to preferential, and more in-depth processing mechanisms.
Collapse
Affiliation(s)
- Konstantina Zougkou
- Department of Psychology and Centre for Brain Science, University of Essex, Colchester CO43SQ
| | - Netta Weinstein
- School of Psychology, Cardiff University, Cardiff CF10 3AT, UK
| | - Silke Paulmann
- Department of Psychology and Centre for Brain Science, University of Essex, Colchester CO43SQ
| |
Collapse
|
31
|
Sen A, Isaacowitz D, Schirmer A. Age differences in vocal emotion perception: on the role of speaker age and listener sex. Cogn Emot 2017; 32:1189-1204. [DOI: 10.1080/02699931.2017.1393399] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
Affiliation(s)
- Antarika Sen
- Neurobiology and Aging Programme, National University of Singapore, Singapore, Singapore
| | | | - Annett Schirmer
- Department of Psychology, The Chinese University of Hong Kong, Hong Kong, Hong Kong
- The Mind and Brain Institute, The Chinese University of Hong Kong, Hong Kong, Hong Kong
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| |
Collapse
|
32
|
Jiang X, Sanford R, Pell MD. Neural systems for evaluating speaker (Un)believability. Hum Brain Mapp 2017; 38:3732-3749. [PMID: 28462535 DOI: 10.1002/hbm.23630] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2017] [Revised: 04/13/2017] [Accepted: 04/17/2017] [Indexed: 12/11/2022] Open
Abstract
Our voice provides salient cues about how confident we sound, which promotes inferences about how believable we are. However, the neural mechanisms involved in these social inferences are largely unknown. Employing functional magnetic resonance imaging, we examined the brain networks and individual differences underlying the evaluation of speaker believability from vocal expressions. Participants (n = 26) listened to statements produced in a confident, unconfident, or "prosodically unmarked" (neutral) voice, and judged how believable the speaker was on a 4-point scale. We found frontal-temporal networks were activated for different levels of confidence, with the left superior and inferior frontal gyrus more activated for confident statements, the right superior temporal gyrus for unconfident expressions, and bilateral cerebellum for statements in a neutral voice. Based on listener's believability judgment, we observed increased activation in the right superior parietal lobule (SPL) associated with higher believability, while increased left posterior central gyrus (PoCG) was associated with less believability. A psychophysiological interaction analysis found that the anterior cingulate cortex and bilateral caudate were connected to the right SPL when higher believability judgments were made, while supplementary motor area was connected with the left PoCG when lower believability judgments were made. Personal characteristics, such as interpersonal reactivity and the individual tendency to trust others, modulated the brain activations and the functional connectivity when making believability judgments. In sum, our data pinpoint neural mechanisms that are involved when inferring one's believability from a speaker's voice and establish ways that these mechanisms are modulated by individual characteristics of a listener. Hum Brain Mapp 38:3732-3749, 2017. © 2017 Wiley Periodicals, Inc.
Collapse
Affiliation(s)
- Xiaoming Jiang
- School of Communication Sciences and Disorders, McGill University, Montréal, Canada
| | - Ryan Sanford
- McConnell Brain Imaging Center, Montréal Neurological Institute, McGill University, Montréal, Canada
| | - Marc D Pell
- School of Communication Sciences and Disorders, McGill University, Montréal, Canada.,McConnell Brain Imaging Center, Montréal Neurological Institute, McGill University, Montréal, Canada
| |
Collapse
|
33
|
Schirmer A, Adolphs R. Emotion Perception from Face, Voice, and Touch: Comparisons and Convergence. Trends Cogn Sci 2017; 21:216-228. [PMID: 28173998 DOI: 10.1016/j.tics.2017.01.001] [Citation(s) in RCA: 140] [Impact Index Per Article: 20.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2016] [Revised: 12/23/2016] [Accepted: 01/03/2017] [Indexed: 11/30/2022]
Abstract
Historically, research on emotion perception has focused on facial expressions, and findings from this modality have come to dominate our thinking about other modalities. Here we examine emotion perception through a wider lens by comparing facial with vocal and tactile processing. We review stimulus characteristics and ensuing behavioral and brain responses and show that audition and touch do not simply duplicate visual mechanisms. Each modality provides a distinct input channel and engages partly nonoverlapping neuroanatomical systems with different processing specializations (e.g., specific emotions versus affect). Moreover, processing of signals across the different modalities converges, first into multi- and later into amodal representations that enable holistic emotion judgments.
Collapse
Affiliation(s)
- Annett Schirmer
- Chinese University of Hong Kong, Hong Kong; Max Planck Institute for Human Cognitive and Brain Sciences, Germany; National University of Singapore, Singapore.
| | - Ralph Adolphs
- California Institute of Technology, Pasadena, CA, USA.
| |
Collapse
|
34
|
Rogers T, Ten Brinke L, Carney DR. Unacquainted callers can predict which citizens will vote over and above citizens' stated self-predictions. Proc Natl Acad Sci U S A 2016; 113:6449-53. [PMID: 27217566 PMCID: PMC4988572 DOI: 10.1073/pnas.1525688113] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/04/2023] Open
Abstract
People are regularly asked to report on their likelihoods of carrying out consequential future behaviors, including complying with medical advice, completing educational assignments, and voting in upcoming elections. Despite these stated self-predictions being notoriously unreliable, they are used to inform many strategic decisions. We report two studies examining stated self-prediction about whether citizens will vote. We find that most self-predicted voters do not actually vote despite saying they will, and that campaign callers can discern which self-predicted voters will not actually vote. In study 1 (n = 4,463), self-predicted voters rated by callers as "100% likely to vote" were 2 times more likely to actually vote than those rated unlikely to vote. Study 2 (n = 3,064) replicated this finding and further demonstrated that callers' prediction accuracy was mediated by citizens' nonverbal signals of uncertainty and deception. Strangers can use nonverbal signals to improve predictions of follow through on self-reported intentions-an insight of potential value for politics, medicine, and education.
Collapse
Affiliation(s)
- Todd Rogers
- Harvard Kennedy School, Harvard University, Cambridge, MA 02138;
| | - Leanne Ten Brinke
- Haas School of Business and Department of Psychology, University of California, Berkeley, CA 94720
| | - Dana R Carney
- Haas School of Business, University of California, Berkeley, CA 94720
| |
Collapse
|
35
|
Jiang X, Pell MD. Neural responses towards a speaker's feeling of (un)knowing. Neuropsychologia 2015; 81:79-93. [PMID: 26700458 DOI: 10.1016/j.neuropsychologia.2015.12.008] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2015] [Revised: 11/16/2015] [Accepted: 12/11/2015] [Indexed: 10/22/2022]
Abstract
During interpersonal communication, listeners must rapidly evaluate verbal and vocal cues to arrive at an integrated meaning about the utterance and about the speaker, including a representation of the speaker's 'feeling of knowing' (i.e., how confident they are in relation to the utterance). In this study, we investigated the time course and neural responses underlying a listener's ability to evaluate speaker confidence from combined verbal and vocal cues. We recorded real-time brain responses as listeners judged statements conveying three levels of confidence with the speaker's voice (confident, close-to-confident, unconfident), which were preceded by meaning-congruent lexical phrases (e.g. I am positive, Most likely, Perhaps). Event-related potentials to utterances with combined lexical and vocal cues about speaker confidence were compared to responses elicited by utterances without the verbal phrase in a previous study (Jiang and Pell, 2015). Utterances with combined cues about speaker confidence elicited reduced, N1, P2 and N400 responses when compared to corresponding utterances without the phrase. When compared to confident statements, close-to-confident and unconfident expressions elicited reduced N1 and P2 responses and a late positivity from 900 to 1250 ms; unconfident and close-to-confident expressions were differentiated later in the 1250-1600 ms time window. The effect of lexical phrases on confidence processing differed for male and female participants, with evidence that female listeners incorporated information from the verbal and vocal channels in a distinct manner. Individual differences in trait empathy and trait anxiety also moderated neural responses during confidence processing. Our findings showcase the cognitive processing mechanisms and individual factors governing how we infer a speaker's mental (knowledge) state from the speech signal.
Collapse
Affiliation(s)
- Xiaoming Jiang
- School of Communication Sciences and Disorders and Center for Research in Brain, Language and Music, McGill University, Montréal, Canada.
| | - Marc D Pell
- School of Communication Sciences and Disorders and Center for Research in Brain, Language and Music, McGill University, Montréal, Canada.
| |
Collapse
|
36
|
Jiang X, Zhou X. Who is respectful? Effects of social context and individual empathic ability on ambiguity resolution during utterance comprehension. Front Psychol 2015; 6:1588. [PMID: 26557102 PMCID: PMC4615935 DOI: 10.3389/fpsyg.2015.01588] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2015] [Accepted: 10/01/2015] [Indexed: 11/13/2022] Open
Abstract
Verbal communication is often ambiguous. By employing the event-related potential (ERP) technique, this study investigated how a comprehender resolves referential ambiguity by using information concerning the social status of communicators. Participants read a conversational scenario which included a minimal conversational context describing a speaker and two other persons of the same or different social status and a directly quoted utterance. A singular, second-person pronoun in the respectful form (nin/nin-de in Chinese) in the utterance could be ambiguous with respect to which of the two persons was the addressee (the “Ambiguous condition”). Alternatively, the pronoun was not ambiguous either because one of the two persons was of higher social status and hence should be the addressee according to social convention (the “Status condition”) or because a word referring to the status of a person was additionally inserted before the pronoun to help indicate the referent of the pronoun (the “Referent condition”). Results showed that the perceived ambiguity decreased over the Ambiguous, Status, and Referent conditions. Electrophysiologically, the pronoun elicited an increased N400 in the Referent than in the Status and the Ambiguous conditions, reflecting an increased integration demand due to the necessity of linking the pronoun to both its antecedent and the status word. Relative to the Referent condition, a late, sustained positivity was elicited for the Status condition starting from 600 ms, while a more delayed, anterior negativity was elicited for the Ambiguous condition. Moreover, the N400 effect was modulated by individuals' sensitivity to the social status information, while the late positivity effect was modulated by individuals' empathic ability. These findings highlight the neurocognitive flexibility of contextual bias in referential processing during utterance comprehension.
Collapse
Affiliation(s)
- Xiaoming Jiang
- Center for Brain and Cognitive Sciences and Department of Psychology, Peking University Beijing, China ; School of Communication Sciences and Disorders, McGill University Montréal, QC, Canada
| | - Xiaolin Zhou
- Center for Brain and Cognitive Sciences and Department of Psychology, Peking University Beijing, China ; Key Laboratory of Machine Perception and Key Laboratory of Computational Linguistics (Ministry of Education), Peking University Beijing, China ; Beijing Key Laboratory of Behavior and Mental Health, Peking University Beijing, China ; IDG McGovern Institute for Brain Research at PKU, Peking University Beijing, China
| |
Collapse
|
37
|
Pell MD, Rothermich K, Liu P, Paulmann S, Sethi S, Rigoulot S. Preferential decoding of emotion from human non-linguistic vocalizations versus speech prosody. Biol Psychol 2015; 111:14-25. [PMID: 26307467 DOI: 10.1016/j.biopsycho.2015.08.008] [Citation(s) in RCA: 76] [Impact Index Per Article: 8.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2015] [Revised: 08/04/2015] [Accepted: 08/19/2015] [Indexed: 11/26/2022]
Abstract
This study used event-related brain potentials (ERPs) to compare the time course of emotion processing from non-linguistic vocalizations versus speech prosody, to test whether vocalizations are treated preferentially by the neurocognitive system. Participants passively listened to vocalizations or pseudo-utterances conveying anger, sadness, or happiness as the EEG was recorded. Simultaneous effects of vocal expression type and emotion were analyzed for three ERP components (N100, P200, late positive component). Emotional vocalizations and speech were differentiated very early (N100) and vocalizations elicited stronger, earlier, and more differentiated P200 responses than speech. At later stages (450-700ms), anger vocalizations evoked a stronger late positivity (LPC) than other vocal expressions, which was similar but delayed for angry speech. Individuals with high trait anxiety exhibited early, heightened sensitivity to vocal emotions (particularly vocalizations). These data provide new neurophysiological evidence that vocalizations, as evolutionarily primitive signals, are accorded precedence over speech-embedded emotions in the human voice.
Collapse
Affiliation(s)
- M D Pell
- School of Communication Sciences and Disorders, McGill University, Montreal, Canada; International Laboratory for Brain, Music, and Sound Research, Montreal, Canada.
| | - K Rothermich
- School of Communication Sciences and Disorders, McGill University, Montreal, Canada
| | - P Liu
- School of Communication Sciences and Disorders, McGill University, Montreal, Canada
| | - S Paulmann
- Department of Psychology and Centre for Brain Science, University of Essex, Colchester, United Kingdom
| | - S Sethi
- School of Communication Sciences and Disorders, McGill University, Montreal, Canada
| | - S Rigoulot
- International Laboratory for Brain, Music, and Sound Research, Montreal, Canada
| |
Collapse
|