1
|
Adl Zarrabi A, Jeulin M, Bardet P, Commère P, Naccache L, Aucouturier JJ, Ponsot E, Villain M. A simple psychophysical procedure separates representational and noise components in impairments of speech prosody perception after right-hemisphere stroke. Sci Rep 2024; 14:15194. [PMID: 38956187 PMCID: PMC11219855 DOI: 10.1038/s41598-024-64295-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2023] [Accepted: 06/06/2024] [Indexed: 07/04/2024] Open
Abstract
After a right hemisphere stroke, more than half of the patients are impaired in their capacity to produce or comprehend speech prosody. Yet, and despite its social-cognitive consequences for patients, aprosodia following stroke has received scant attention. In this report, we introduce a novel, simple psychophysical procedure which, by combining systematic digital manipulations of speech stimuli and reverse-correlation analysis, allows estimating the internal sensory representations that subtend how individual patients perceive speech prosody, and the level of internal noise that govern behavioral variability in how patients apply these representations. Tested on a sample of N = 22 right-hemisphere stroke survivors and N = 21 age-matched controls, the representation + noise model provides a promising alternative to the clinical gold standard for evaluating aprosodia (MEC): both parameters strongly associate with receptive, and not expressive, aprosodia measured by MEC within the patient group; they have better sensitivity than MEC for separating high-functioning patients from controls; and have good specificity with respect to non-prosody-related impairments of auditory attention and processing. Taken together, individual differences in either internal representation, internal noise, or both, paint a potent portrait of the variety of sensory/cognitive mechanisms that can explain impairments of prosody processing after stroke.
Collapse
Affiliation(s)
- Aynaz Adl Zarrabi
- Université de Franche-Comté, SUPMICROTECH, CNRS, Institut FEMTO-ST, 25000, Besançon, France
| | - Mélissa Jeulin
- Department of Physical Medicine & Rehabilitation, APHP/Hôpital Pitié-Salpêtrière, 75013, Paris, France
| | - Pauline Bardet
- Department of Physical Medicine & Rehabilitation, APHP/Hôpital Pitié-Salpêtrière, 75013, Paris, France
| | - Pauline Commère
- Department of Physical Medicine & Rehabilitation, APHP/Hôpital Pitié-Salpêtrière, 75013, Paris, France
| | - Lionel Naccache
- Department of Physical Medicine & Rehabilitation, APHP/Hôpital Pitié-Salpêtrière, 75013, Paris, France
- Paris Brain Institute (ICM), Inserm, CNRS, PICNIC-Lab, 75013, Paris, France
| | | | - Emmanuel Ponsot
- Science & Technology of Music and Sound, IRCAM/CNRS/Sorbonne Université, 75004, Paris, France
| | - Marie Villain
- Department of Physical Medicine & Rehabilitation, APHP/Hôpital Pitié-Salpêtrière, 75013, Paris, France.
- Paris Brain Institute (ICM), Inserm, CNRS, PICNIC-Lab, 75013, Paris, France.
| |
Collapse
|
2
|
Osses A, Spinelli E, Meunier F, Gaudrain E, Varnet L. Prosodic cues to word boundaries in a segmentation task assessed using reverse correlation. JASA EXPRESS LETTERS 2023; 3:095205. [PMID: 37756550 DOI: 10.1121/10.0021022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/08/2023] [Accepted: 09/03/2023] [Indexed: 09/29/2023]
Abstract
When listening to speech sounds, listeners are able to exploit acoustic features that mark the boundaries between successive words, the so-called segmentation cues. These cues are typically investigated by directly manipulating features that are hypothetically related to segmentation. The current study uses a different approach based on reverse correlation, where the stimulus manipulations are based on minimal assumptions. The method was evaluated using pairs of phonemically identical sentences in French, whose prosody was changed by introducing random f0 trajectories and segment durations. Our results support a prominent perceptual role of the f0 rise and vowel duration at the beginning of content words.
Collapse
Affiliation(s)
- Alejandro Osses
- Laboratoire des Systèmes Perceptifs, Département d'Études Cognitives, École Normale Supérieure, PSL University, CNRS, Paris, France
| | - Elsa Spinelli
- Laboratoire de Psychologie et NeuroCognition, Université Grenoble Alpes, Grenoble, France
| | | | - Etienne Gaudrain
- Lyon Neuroscience Research Center, CNRS, Inserm, Université Lyon 1, Lyon, , , , ,
| | - Léo Varnet
- Laboratoire des Systèmes Perceptifs, Département d'Études Cognitives, École Normale Supérieure, PSL University, CNRS, Paris, France
| |
Collapse
|
3
|
Yan S, Soladié C, Aucouturier JJ, Seguier R. Combining GAN with reverse correlation to construct personalized facial expressions. PLoS One 2023; 18:e0290612. [PMID: 37624781 PMCID: PMC10456187 DOI: 10.1371/journal.pone.0290612] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2023] [Accepted: 08/11/2023] [Indexed: 08/27/2023] Open
Abstract
Recent deep-learning techniques have made it possible to manipulate facial expressions in digital photographs or videos, however, these techniques still lack fine and personalized ways to control their creation. Moreover, current technologies are highly dependent on large labeled databases, which limits the range and complexity of expressions that can be modeled. Thus, these technologies cannot deal with non-basic emotions. In this paper, we propose a novel interdisciplinary approach combining the Generative Adversarial Network (GAN) with a technique inspired by cognitive sciences, psychophysical reverse correlation. Reverse correlation is a data-driven method able to extract an observer's 'mental representation' of what a given facial expression should look like. Our approach can generate 1) personalized facial expression prototypes, 2) of basic emotions, and non-basic emotions that are not available in existing databases, and 3) without the need for expertise. Personalized prototypes obtained with reverse correlation can then be applied to manipulate facial expressions. In addition, our system challenges the universality of facial expression prototypes by proposing the concepts of dominant and complementary action units to describe facial expression prototypes. The evaluations we conducted on a limited number of emotions validate the effectiveness of our proposed method. The code is available at https://github.com/yansen0508/Mental-Deep-Reverse-Engineering.
Collapse
Affiliation(s)
- Sen Yan
- CentraleSupelec, IETR, Rennes, France
| | | | | | | |
Collapse
|
4
|
Wang L, Ong JH, Ponsot E, Hou Q, Jiang C, Liu F. Mental representations of speech and musical pitch contours reveal a diversity of profiles in autism spectrum disorder. AUTISM : THE INTERNATIONAL JOURNAL OF RESEARCH AND PRACTICE 2023; 27:629-646. [PMID: 35848413 PMCID: PMC10074762 DOI: 10.1177/13623613221111207] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
LAY ABSTRACT As a key auditory attribute of sounds, pitch is ubiquitous in our everyday listening experience involving language, music and environmental sounds. Given its critical role in auditory processing related to communication, numerous studies have investigated pitch processing in autism spectrum disorder. However, the findings have been mixed, reporting either enhanced, typical or impaired performance among autistic individuals. By investigating top-down comparisons of internal mental representations of pitch contours in speech and music, this study shows for the first time that, while autistic individuals exhibit diverse profiles of pitch processing compared to non-autistic individuals, their mental representations of pitch contours are typical across domains. These findings suggest that pitch-processing mechanisms are shared across domains in autism spectrum disorder and provide theoretical implications for using music to improve speech for those autistic individuals who have language problems.
Collapse
Affiliation(s)
- Li Wang
- University of Reading, UK
- The Chinese University of Hong Kong, Hong
Kong
| | | | | | - Qingqi Hou
- Nanjing Normal University of Special
Education, China
| | | | | |
Collapse
|
5
|
Wass SV, Goupil L. Studying the Developing Brain in Real-World Contexts: Moving From Castles in the Air to Castles on the Ground. Front Integr Neurosci 2022; 16:896919. [PMID: 35910339 PMCID: PMC9326302 DOI: 10.3389/fnint.2022.896919] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2022] [Accepted: 06/17/2022] [Indexed: 11/13/2022] Open
Abstract
Most current research in cognitive neuroscience uses standardized non-ecological experiments to study the developing brain. But these approaches do a poor job of mimicking the real-world, and thus can only provide a distorted picture of how cognitive operations and brain development unfold outside of the lab. Here we consider future research avenues which may lead to a better appreciation of how developing brains dynamically interact with a complex real-world environment, and how cognition develops over time. We raise several problems faced by current mainstream methods in the field, before briefly reviewing novel promising approaches that alleviate some of these issues. First, we consider research that examines perception by measuring entrainment between brain activity and temporal patterns in naturalistic stimuli. Second, we consider research that examines our ability to parse our continuous experience into discrete events, and how this ability develops over time. Third, we consider the role of children as active agents in selecting what they sample from the environment from one moment to the next. Fourth, we consider new approaches that measure how mutual influences between children and others are instantiated in suprapersonal brain networks. Finally, we discuss how we may reduce adult biases when designing developmental studies. Together, these approaches have great potential to further our understanding of how the developing brain learns to process information, and to control complex real-world behaviors.
Collapse
Affiliation(s)
- Sam V. Wass
- Department of Psychology, University of East London, London, United Kingdom
| | - Louise Goupil
- LPNC, Université Grenoble Alpes/CNRS, Grenoble, France
| |
Collapse
|
6
|
Pisanski K, Bryant GA, Cornec C, Anikin A, Reby D. Form follows function in human nonverbal vocalisations. ETHOL ECOL EVOL 2022. [DOI: 10.1080/03949370.2022.2026482] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Affiliation(s)
- Katarzyna Pisanski
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
- CNRS French National Centre for Scientific Research, DDL Dynamics of Language Lab, University of Lyon 2, Lyon 69007, France
| | - Gregory A. Bryant
- Department of Communication, Center for Behavior, Evolution, and Culture, University of California, Los Angeles, California, USA
| | - Clément Cornec
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
| | - Andrey Anikin
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
- Division of Cognitive Science, Lund University, Lund 22100, Sweden
| | - David Reby
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
| |
Collapse
|
7
|
Pruvost-Robieux E, André-Obadia N, Marchi A, Sharshar T, Liuni M, Gavaret M, Aucouturier JJ. It’s not what you say, it’s how you say it: a retrospective study of the impact of prosody on own-name P300 in comatose patients. Clin Neurophysiol 2022; 135:154-161. [DOI: 10.1016/j.clinph.2021.12.015] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2021] [Revised: 12/16/2021] [Accepted: 12/18/2021] [Indexed: 02/05/2023]
|
8
|
Goupil L, Ponsot E, Richardson D, Reyes G, Aucouturier JJ. Listeners' perceptions of the certainty and honesty of a speaker are associated with a common prosodic signature. Nat Commun 2021; 12:861. [PMID: 33558510 PMCID: PMC7870677 DOI: 10.1038/s41467-020-20649-4] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2019] [Accepted: 11/20/2020] [Indexed: 02/07/2023] Open
Abstract
The success of human cooperation crucially depends on mechanisms enabling individuals to detect unreliability in their conspecifics. Yet, how such epistemic vigilance is achieved from naturalistic sensory inputs remains unclear. Here we show that listeners' perceptions of the certainty and honesty of other speakers from their speech are based on a common prosodic signature. Using a data-driven method, we separately decode the prosodic features driving listeners' perceptions of a speaker's certainty and honesty across pitch, duration and loudness. We find that these two kinds of judgments rely on a common prosodic signature that is perceived independently from individuals' conceptual knowledge and native language. Finally, we show that listeners extract this prosodic signature automatically, and that this impacts the way they memorize spoken words. These findings shed light on a unique auditory adaptation that enables human listeners to quickly detect and react to unreliability during linguistic interactions.
Collapse
Affiliation(s)
- Louise Goupil
- STMS UMR 9912 (CNRS/IRCAM/SU), Paris, France.
- University of East London, London, UK.
| | - Emmanuel Ponsot
- Laboratoire des Systèmes Perceptifs, Département d'Études Cognitives, École Normale Supérieure, PSL University, CNRS, Paris, France
- Hearing Technology - WAVES, Department of Information Technology, Ghent University, Ghent, Belgium
| | | | | | - Jean-Julien Aucouturier
- STMS UMR 9912 (CNRS/IRCAM/SU), Paris, France
- FEMTO-ST (FEMTO-ST UMR 6174, CNRS/UBFC/ENSMM/UTBM, Besançon, France
| |
Collapse
|
9
|
Schyns PG, Zhan J, Jack RE, Ince RAA. Revealing the information contents of memory within the stimulus information representation framework. Philos Trans R Soc Lond B Biol Sci 2020; 375:20190705. [PMID: 32248774 PMCID: PMC7209912 DOI: 10.1098/rstb.2019.0705] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
The information contents of memory are the cornerstone of the most influential models in cognition. To illustrate, consider that in predictive coding, a prediction implies that specific information is propagated down from memory through the visual hierarchy. Likewise, recognizing the input implies that sequentially accrued sensory evidence is successfully matched with memorized information (categorical knowledge). Although the existing models of prediction, memory, sensory representation and categorical decision are all implicitly cast within an information processing framework, it remains a challenge to precisely specify what this information is, and therefore where, when and how the architecture of the brain dynamically processes it to produce behaviour. Here, we review a framework that addresses these challenges for the studies of perception and categorization–stimulus information representation (SIR). We illustrate how SIR can reverse engineer the information contents of memory from behavioural and brain measures in the context of specific cognitive tasks that involve memory. We discuss two specific lessons from this approach that generally apply to memory studies: the importance of task, to constrain what the brain does, and of stimulus variations, to identify the specific information contents that are memorized, predicted, recalled and replayed. This article is part of the Theo Murphy meeting issue ‘Memory reactivation: replaying events past, present and future’.
Collapse
Affiliation(s)
- Philippe G Schyns
- Institute of Neuroscience and Psychology, University of Glasgow, Scotland G12 8QB, UK.,School of Psychology, University of Glasgow, Scotland G12 8QB, UK
| | - Jiayu Zhan
- Institute of Neuroscience and Psychology, University of Glasgow, Scotland G12 8QB, UK
| | - Rachael E Jack
- School of Psychology, University of Glasgow, Scotland G12 8QB, UK
| | - Robin A A Ince
- Institute of Neuroscience and Psychology, University of Glasgow, Scotland G12 8QB, UK
| |
Collapse
|