1
|
Lazzari G, Sacheli LM, Benoit CE, Lega C, van Vugt FT. Pleasantness makes a good time: musical consonance shapes interpersonal synchronization in dyadic joint action. Front Hum Neurosci 2024; 18:1472632. [PMID: 39502786 PMCID: PMC11534602 DOI: 10.3389/fnhum.2024.1472632] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2024] [Accepted: 09/30/2024] [Indexed: 11/08/2024] Open
Abstract
Introduction Music making is a process by which humans across cultures come together to create patterns of sounds that are aesthetically pleasing. What remains unclear is how this aesthetic outcome affects the sensorimotor interaction between participants. Method Here we approach this question using an interpersonal sensorimotor synchronization paradigm to test whether the quality of a jointly created chord (consonant vs. dissonant) affects movement coordination. We recruited non-musician participants in dyads to perform a dyadic synchronization-continuation task (dSCT): on each trial, participants first synchronized their movements to a metronome (synchronization phase) and then continued tapping together at the same tempo without the metronome (continuation phase). Each tap yielded a note and participants heard both their own and that of their partner, thus creating a chord that was varied to be either consonant (Perf5 or Maj6) or dissonant (Min2 or Maj2). For each trial, participants also rated the pleasure they felt in creating the sounds together. Additionally, they completed questionnaires about social closeness to the other participant, musical reward sensitivity and musical training. Results Results showed that participants' taps were closer in time when they jointly created consonant (high pleasure) vs. dissonant (low pleasure) chords, and that pleasure experienced by the dyad in each trial predicted interpersonal synchronization. However, consonance did not affect individual synchronization with the metronome or individual tapping when the metronome was discontinued. The effect of consonance on synchronization was greater in dyads who reported feeling less close prior to the task. Discussion Together, these results highlight the role of consonance in shaping the temporal coordination of our actions with others. More broadly, this work shows that the aesthetic outcome of what we create together affects joint behaviors.
Collapse
Affiliation(s)
- Giorgio Lazzari
- Department of Brain and Behavioral Sciences, University of Pavia, Pavia, Italy
| | | | - Charles-Etienne Benoit
- Inter-University Laboratory of Human Movement Biology, Univ Lyon, University Claude Bernard Lyon 1, Villeurbanne, France
| | - Carlotta Lega
- Department of Brain and Behavioral Sciences, University of Pavia, Pavia, Italy
| | - Floris T. van Vugt
- Centre for Research on Brain, Language and Music (CRBLM), Montreal, QC, Canada
- Psychology Department, University of Montreal, Montreal, QC, Canada
- International Laboratory for Brain, Music and Sound Research (BRAMS), Montreal, QC, Canada
| |
Collapse
|
2
|
Jo HS, Hsieh TH, Chien WC, Shaw FZ, Liang SF, Kung CC. Probing the neural dynamics of musicians' and non-musicians' consonant/dissonant perception: Joint analyses of electrical encephalogram (EEG) and functional magnetic resonance imaging (fMRI). Neuroimage 2024; 298:120784. [PMID: 39147290 DOI: 10.1016/j.neuroimage.2024.120784] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/25/2023] [Revised: 08/06/2024] [Accepted: 08/12/2024] [Indexed: 08/17/2024] Open
Abstract
The perception of two (or more) simultaneous musical notes, depending on their pitch interval(s), could be broadly categorized as consonant or dissonant. Previous literature has suggested that musicians and non-musicians adopt different strategies when discerning music intervals: while musicians rely on the frequency ratios between the two fundamental frequencies, such as "perfect fifth" (3:2) as consonant and "tritone" (45:32) as dissonant intervals; non-musicians may rely on the presence of 'roughness' or 'beats', generated by the difference of fundamental frequencies, as the key elements of 'dissonance'. The separate Event-Related Potential (ERP) differences in N1 and P2 along the midline electrodes provided evidence congruent with such 'separate reliances'. To replicate and to extend, in this study we reran the previous experiment, and separately collected fMRI data of the same protocol (with sparse sampling modifications). The behavioral and EEG results largely corresponded to our previous finding. The fMRI results, with the joint analyses by univariate, psycho-physiological interaction, and representational similarity analysis (RSA) approaches, further reinforce the involvement of central midline-related brain regions, such as ventromedial prefrontal and dorsal anterior cingulate cortex, in consonant/dissonance judgments. The final spatiotemporal searchlight RSA provided convincing evidence that the medial prefrontal cortex, along with the bilateral superior temporal cortex, is the joint locus of midline N1 and dorsal anterior cingulate cortex for the P2 effect (for musicians). Together, these analyses reaffirm that musicians rely more on experience-driven knowledge for consonance/dissonance perception; but also demonstrate the advantages of multiple analyses in constraining the findings from both EEG and fMRI.
Collapse
Affiliation(s)
- Han Shin Jo
- Institute of Medical Informatics, National Cheng Kung University (NCKU), Tainan, 70101, Taiwan
| | - Tsung-Hao Hsieh
- Department of Computer Science and Information Engineering, NCKU, Tainan, 70101, Taiwan; Department of Computer Science, Tunghai University, Taichung, 407224, Taiwan
| | - Wei-Che Chien
- Department of Computer Science and Information Engineering, NCKU, Tainan, 70101, Taiwan
| | - Fu-Zen Shaw
- Department of Psychology, NCKU, Tainan, 70101, Taiwan; Mind Research and Imaging Center, NCKU, Tainan, 70101, Taiwan
| | - Sheng-Fu Liang
- Institute of Medical Informatics, National Cheng Kung University (NCKU), Tainan, 70101, Taiwan; Department of Computer Science and Information Engineering, NCKU, Tainan, 70101, Taiwan
| | - Chun-Chia Kung
- Department of Psychology, NCKU, Tainan, 70101, Taiwan; Mind Research and Imaging Center, NCKU, Tainan, 70101, Taiwan.
| |
Collapse
|
3
|
Bravo F, Glogowski J, Stamatakis EA, Herfert K. Dissonant music engages early visual processing. Proc Natl Acad Sci U S A 2024; 121:e2320378121. [PMID: 39008675 PMCID: PMC11287129 DOI: 10.1073/pnas.2320378121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2023] [Accepted: 06/04/2024] [Indexed: 07/17/2024] Open
Abstract
The neuroscientific examination of music processing in audio-visual contexts offers a valuable framework to assess how auditory information influences the emotional encoding of visual information. Using fMRI during naturalistic film viewing, we investigated the neural mechanisms underlying the effect of music on valence inferences during mental state attribution. Thirty-eight participants watched the same short-film accompanied by systematically controlled consonant or dissonant music. Subjects were instructed to think about the main character's intentions. The results revealed that increasing levels of dissonance led to more negatively valenced inferences, displaying the profound emotional impact of musical dissonance. Crucially, at the neuroscientific level and despite music being the sole manipulation, dissonance evoked the response of the primary visual cortex (V1). Functional/effective connectivity analysis showed a stronger coupling between the auditory ventral stream (AVS) and V1 in response to tonal dissonance and demonstrated the modulation of early visual processing via top-down feedback inputs from the AVS to V1. These V1 signal changes indicate the influence of high-level contextual representations associated with tonal dissonance on early visual cortices, serving to facilitate the emotional interpretation of visual information. Our results highlight the significance of employing systematically controlled music, which can isolate emotional valence from the arousal dimension, to elucidate the brain's sound-to-meaning interface and its distributive crossmodal effects on early visual encoding during naturalistic film viewing.
Collapse
Affiliation(s)
- Fernando Bravo
- Department of Preclinical Imaging and Radiopharmacy, University of Tübingen, Tübingen72076, Germany
- Cognition and Consciousness Imaging Group, Division of Anaesthesia, Department of Medicine, University of Cambridge, Addenbrooke’s Hospital, CambridgeCB2 0SP, United Kingdom
- Department of Clinical Neurosciences, University of Cambridge, Addenbrooke’s Hospital, CambridgeCB2 0SP, United Kingdom
- Institut für Kunst- und Musikwissenschaft, Division of Musicology, Technische Universität Dresden, Dresden01219, Germany
| | - Jana Glogowski
- Department of Psychology, Humboldt-Universität zu Berlin, Berlin12489, Germany
| | - Emmanuel Andreas Stamatakis
- Cognition and Consciousness Imaging Group, Division of Anaesthesia, Department of Medicine, University of Cambridge, Addenbrooke’s Hospital, CambridgeCB2 0SP, United Kingdom
- Department of Clinical Neurosciences, University of Cambridge, Addenbrooke’s Hospital, CambridgeCB2 0SP, United Kingdom
| | - Kristina Herfert
- Department of Preclinical Imaging and Radiopharmacy, University of Tübingen, Tübingen72076, Germany
| |
Collapse
|
4
|
Wöhrle SD, Reuter C, Rupp A, Andermann M. Neuromagnetic representation of musical roundness in chord progressions. Front Neurosci 2024; 18:1383554. [PMID: 38650622 PMCID: PMC11034485 DOI: 10.3389/fnins.2024.1383554] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2024] [Accepted: 03/26/2024] [Indexed: 04/25/2024] Open
Abstract
Introduction Musical roundness perception relies on consonance/dissonance within a rule-based harmonic context, but also on individual characteristics of the listener. The present work tackles these aspects in a combined psychoacoustic and neurophysiological study, taking into account participant's musical aptitude. Methods Our paradigm employed cadence-like four-chord progressions, based on Western music theory. Chord progressions comprised naturalistic and artificial sounds; moreover, their single chords varied regarding consonance/dissonance and harmonic function. Thirty participants listened to the chord progressions while their cortical activity was measured with magnetoencephalography; afterwards, they rated the individual chord progressions with respect to their perceived roundness. Results Roundness ratings differed according to the degree of dissonance in the dominant chord at the progression's third position; this effect was pronounced in listeners with high musical aptitude. Interestingly, a corresponding pattern occurred in the neuromagnetic N1m response to the fourth chord (i.e., at the progression's resolution), again with somewhat stronger differentiation among musical listeners. The N1m magnitude seemed to increase during chord progressions that were considered particularly round, with the maximum difference after the final chord; here, however, the musical aptitude effect just missed significance. Discussion The roundness of chord progressions is reflected in participant's psychoacoustic ratings and in their transient cortical activity, with stronger differentiation among listeners with high musical aptitude. The concept of roundness might help to reframe consonance/dissonance to a more holistic, gestalt-like understanding that covers chord relations in Western music.
Collapse
Affiliation(s)
- Sophie D. Wöhrle
- Section of Biomagnetism, Department of Neurology, Heidelberg University Hospital, Heidelberg, Germany
| | - Christoph Reuter
- Musicological Department (Acoustics/Music Psychology), University of Vienna, Vienna, Austria
| | - André Rupp
- Section of Biomagnetism, Department of Neurology, Heidelberg University Hospital, Heidelberg, Germany
| | - Martin Andermann
- Section of Biomagnetism, Department of Neurology, Heidelberg University Hospital, Heidelberg, Germany
| |
Collapse
|
5
|
Marjieh R, Harrison PMC, Lee H, Deligiannaki F, Jacoby N. Timbral effects on consonance disentangle psychoacoustic mechanisms and suggest perceptual origins for musical scales. Nat Commun 2024; 15:1482. [PMID: 38369535 PMCID: PMC11258268 DOI: 10.1038/s41467-024-45812-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2022] [Accepted: 12/11/2023] [Indexed: 02/20/2024] Open
Abstract
The phenomenon of musical consonance is an essential feature in diverse musical styles. The traditional belief, supported by centuries of Western music theory and psychological studies, is that consonance derives from simple (harmonic) frequency ratios between tones and is insensitive to timbre. Here we show through five large-scale behavioral studies, comprising 235,440 human judgments from US and South Korean populations, that harmonic consonance preferences can be reshaped by timbral manipulations, even as far as to induce preferences for inharmonic intervals. We show how such effects may suggest perceptual origins for diverse scale systems ranging from the gamelan's slendro scale to the tuning of Western mean-tone and equal-tempered scales. Through computational modeling we show that these timbral manipulations dissociate competing psychoacoustic mechanisms underlying consonance, and we derive an updated computational model combining liking of harmonicity, disliking of fast beats (roughness), and liking of slow beats. Altogether, this work showcases how large-scale behavioral experiments can inform classical questions in auditory perception.
Collapse
Affiliation(s)
- Raja Marjieh
- Department of Psychology, Princeton University, Princeton, NJ, USA.
- Max Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany.
| | - Peter M C Harrison
- Max Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany.
- Centre for Music and Science, University of Cambridge, Cambridge, UK.
| | - Harin Lee
- Max Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| | - Fotini Deligiannaki
- Max Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany
- German Aerospace Center (DLR), Institute for AI Safety and Security, Bonn, Germany
| | - Nori Jacoby
- Max Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany.
| |
Collapse
|
6
|
Shan T, Cappelloni MS, Maddox RK. Subcortical responses to music and speech are alike while cortical responses diverge. Sci Rep 2024; 14:789. [PMID: 38191488 PMCID: PMC10774448 DOI: 10.1038/s41598-023-50438-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2023] [Accepted: 12/20/2023] [Indexed: 01/10/2024] Open
Abstract
Music and speech are encountered daily and are unique to human beings. Both are transformed by the auditory pathway from an initial acoustical encoding to higher level cognition. Studies of cortex have revealed distinct brain responses to music and speech, but differences may emerge in the cortex or may be inherited from different subcortical encoding. In the first part of this study, we derived the human auditory brainstem response (ABR), a measure of subcortical encoding, to recorded music and speech using two analysis methods. The first method, described previously and acoustically based, yielded very different ABRs between the two sound classes. The second method, however, developed here and based on a physiological model of the auditory periphery, gave highly correlated responses to music and speech. We determined the superiority of the second method through several metrics, suggesting there is no appreciable impact of stimulus class (i.e., music vs speech) on the way stimulus acoustics are encoded subcortically. In this study's second part, we considered the cortex. Our new analysis method resulted in cortical music and speech responses becoming more similar but with remaining differences. The subcortical and cortical results taken together suggest that there is evidence for stimulus-class dependent processing of music and speech at the cortical but not subcortical level.
Collapse
Affiliation(s)
- Tong Shan
- Department of Biomedical Engineering, University of Rochester, Rochester, NY, USA
- Del Monte Institute for Neuroscience, University of Rochester, Rochester, NY, USA
- Center for Visual Science, University of Rochester, Rochester, NY, USA
| | - Madeline S Cappelloni
- Department of Biomedical Engineering, University of Rochester, Rochester, NY, USA
- Del Monte Institute for Neuroscience, University of Rochester, Rochester, NY, USA
- Center for Visual Science, University of Rochester, Rochester, NY, USA
| | - Ross K Maddox
- Department of Biomedical Engineering, University of Rochester, Rochester, NY, USA.
- Del Monte Institute for Neuroscience, University of Rochester, Rochester, NY, USA.
- Center for Visual Science, University of Rochester, Rochester, NY, USA.
- Department of Neuroscience, University of Rochester, Rochester, NY, USA.
| |
Collapse
|
7
|
Bowling DL. Biological principles for music and mental health. Transl Psychiatry 2023; 13:374. [PMID: 38049408 PMCID: PMC10695969 DOI: 10.1038/s41398-023-02671-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/20/2022] [Revised: 10/30/2023] [Accepted: 11/17/2023] [Indexed: 12/06/2023] Open
Abstract
Efforts to integrate music into healthcare systems and wellness practices are accelerating but the biological foundations supporting these initiatives remain underappreciated. As a result, music-based interventions are often sidelined in medicine. Here, I bring together advances in music research from neuroscience, psychology, and psychiatry to bridge music's specific foundations in human biology with its specific therapeutic applications. The framework I propose organizes the neurophysiological effects of music around four core elements of human musicality: tonality, rhythm, reward, and sociality. For each, I review key concepts, biological bases, and evidence of clinical benefits. Within this framework, I outline a strategy to increase music's impact on health based on standardizing treatments and their alignment with individual differences in responsivity to these musical elements. I propose that an integrated biological understanding of human musicality-describing each element's functional origins, development, phylogeny, and neural bases-is critical to advancing rational applications of music in mental health and wellness.
Collapse
Affiliation(s)
- Daniel L Bowling
- Department of Psychiatry and Behavioral Sciences, Stanford University, School of Medicine, Stanford, CA, USA.
- Center for Computer Research in Music and Acoustics (CCRMA), Stanford University, School of Humanities and Sciences, Stanford, CA, USA.
| |
Collapse
|
8
|
Brandt AK. Calls of the wild: Exploring the evolutionary roots of consonance and dissonance: Comment on "Consonance and dissonance perception: A critical review of historical sources, multidisciplinary findings, and main hypotheses" by N. Di Stefano, P. Vuust and E. Brattico. Phys Life Rev 2023; 47:68-70. [PMID: 37734283 DOI: 10.1016/j.plrev.2023.09.005] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2023] [Accepted: 09/07/2023] [Indexed: 09/23/2023]
Affiliation(s)
- Anthony K Brandt
- The Shepherd School of Music, Rice University, United States of America.
| |
Collapse
|
9
|
Bingham MA, Cummins ML, Tong A, Purcell P, Sangari A, Sood A, Schlesinger JJ. Effects of altering harmonic structure on the recognition of simulated auditory arterial pressure alarms. Br J Anaesth 2023; 131:e178-e180. [PMID: 37758624 DOI: 10.1016/j.bja.2023.08.037] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2023] [Revised: 08/20/2023] [Accepted: 08/29/2023] [Indexed: 09/29/2023] Open
Affiliation(s)
- Molly A Bingham
- Department of Biomedical Engineering, Vanderbilt University, Nashville, TN, USA
| | - Mabel L Cummins
- Department of Neuroscience, Vanderbilt University, Nashville, TN, USA
| | - Anqy Tong
- Department of Neuroscience, Vanderbilt University, Nashville, TN, USA
| | | | - Ayush Sangari
- Renaissance School of Medicine, Stony Brook University, Stony Brook, NY, USA
| | - Aditya Sood
- Emory University School of Medicine, Atlanta, GA, USA
| | - Joseph J Schlesinger
- Division of Critical Care Medicine, Department of Anesthesiology, Vanderbilt University Medical Center, Nashville, TN, USA.
| |
Collapse
|
10
|
Rizzi R, Bidelman GM. Duplex perception reveals brainstem auditory representations are modulated by listeners' ongoing percept for speech. Cereb Cortex 2023; 33:10076-10086. [PMID: 37522248 PMCID: PMC10502779 DOI: 10.1093/cercor/bhad266] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2023] [Revised: 06/27/2023] [Accepted: 07/10/2023] [Indexed: 08/01/2023] Open
Abstract
So-called duplex speech stimuli with perceptually ambiguous spectral cues to one ear and isolated low- versus high-frequency third formant "chirp" to the opposite ear yield a coherent percept supporting their phonetic categorization. Critically, such dichotic sounds are only perceived categorically upon binaural integration. Here, we used frequency-following responses (FFRs), scalp-recorded potentials reflecting phase-locked subcortical activity, to investigate brainstem responses to fused speech percepts and to determine whether FFRs reflect binaurally integrated category-level representations. We recorded FFRs to diotic and dichotic stop-consonants (/da/, /ga/) that either did or did not require binaural fusion to properly label along with perceptually ambiguous sounds without clear phonetic identity. Behaviorally, listeners showed clear categorization of dichotic speech tokens confirming they were heard with a fused, phonetic percept. Neurally, we found FFRs were stronger for categorically perceived speech relative to category-ambiguous tokens but also differentiated phonetic categories for both diotically and dichotically presented speech sounds. Correlations between neural and behavioral data further showed FFR latency predicted the degree to which listeners labeled tokens as "da" versus "ga." The presence of binaurally integrated, category-level information in FFRs suggests human brainstem processing reflects a surprisingly abstract level of the speech code typically circumscribed to much later cortical processing.
Collapse
Affiliation(s)
- Rose Rizzi
- Department of Speech, Language, and Hearing Sciences, Indiana University, Bloomington, IN, United States
- Program in Neuroscience, Indiana University, Bloomington, IN, United States
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, United States
| | - Gavin M Bidelman
- Department of Speech, Language, and Hearing Sciences, Indiana University, Bloomington, IN, United States
- Program in Neuroscience, Indiana University, Bloomington, IN, United States
- Cognitive Science Program, Indiana University, Bloomington, IN, United States
| |
Collapse
|
11
|
Benítez-Burraco A, Nikolsky A. The (Co)Evolution of Language and Music Under Human Self-Domestication. HUMAN NATURE (HAWTHORNE, N.Y.) 2023; 34:229-275. [PMID: 37097428 PMCID: PMC10354115 DOI: 10.1007/s12110-023-09447-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 03/27/2023] [Indexed: 04/26/2023]
Abstract
Together with language, music is perhaps the most distinctive behavioral trait of the human species. Different hypotheses have been proposed to explain why only humans perform music and how this ability might have evolved in our species. In this paper, we advance a new model of music evolution that builds on the self-domestication view of human evolution, according to which the human phenotype is, at least in part, the outcome of a process similar to domestication in other mammals, triggered by the reduction in reactive aggression responses to environmental changes. We specifically argue that self-domestication can account for some of the cognitive changes, and particularly for the behaviors conducive to the complexification of music through a cultural mechanism. We hypothesize four stages in the evolution of music under self-domestication forces: (1) collective protomusic; (2) private, timbre-oriented music; (3) small-group, pitch-oriented music; and (4) collective, tonally organized music. This line of development encompasses the worldwide diversity of music types and genres and parallels what has been hypothesized for languages. Overall, music diversity might have emerged in a gradual fashion under the effects of the enhanced cultural niche construction as shaped by the progressive decrease in reactive (i.e., impulsive, triggered by fear or anger) aggression and the increase in proactive (i.e., premeditated, goal-directed) aggression.
Collapse
Affiliation(s)
- Antonio Benítez-Burraco
- Department of Spanish Language, Linguistics and Literary Theory (Linguistics), Faculty of Philology, University of Seville, Seville, Spain.
- Departamento de Lengua Española, Facultad de Filología, Área de Lingüística General, Lingüística y Teoría de la Literatura, Universidad de Sevilla, C/ Palos de la Frontera s/n, Sevilla, 41007, España.
| | | |
Collapse
|
12
|
Rizzi R, Bidelman GM. Duplex perception reveals brainstem auditory representations are modulated by listeners' ongoing percept for speech. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.05.09.540018. [PMID: 37214801 PMCID: PMC10197666 DOI: 10.1101/2023.05.09.540018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/24/2023]
Abstract
So-called duplex speech stimuli with perceptually ambiguous spectral cues to one ear and isolated low- vs. high-frequency third formant "chirp" to the opposite ear yield a coherent percept supporting their phonetic categorization. Critically, such dichotic sounds are only perceived categorically upon binaural integration. Here, we used frequency-following responses (FFRs), scalp-recorded potentials reflecting phase-locked subcortical activity, to investigate brainstem responses to fused speech percepts and to determine whether FFRs reflect binaurally integrated category-level representations. We recorded FFRs to diotic and dichotic stop-consonants (/da/, /ga/) that either did or did not require binaural fusion to properly label along with perceptually ambiguous sounds without clear phonetic identity. Behaviorally, listeners showed clear categorization of dichotic speech tokens confirming they were heard with a fused, phonetic percept. Neurally, we found FFRs were stronger for categorically perceived speech relative to category-ambiguous tokens but also differentiated phonetic categories for both diotically and dichotically presented speech sounds. Correlations between neural and behavioral data further showed FFR latency predicted the degree to which listeners labeled tokens as "da" vs. "ga". The presence of binaurally integrated, category-level information in FFRs suggests human brainstem processing reflects a surprisingly abstract level of the speech code typically circumscribed to much later cortical processing.
Collapse
Affiliation(s)
- Rose Rizzi
- Department of Speech, Language, and Hearing Sciences, Indiana University, Bloomington, IN, USA
- Program in Neuroscience, Indiana University, Bloomington, IN, USA
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, USA
| | - Gavin M. Bidelman
- Department of Speech, Language, and Hearing Sciences, Indiana University, Bloomington, IN, USA
- Program in Neuroscience, Indiana University, Bloomington, IN, USA
- Cognitive Science Program, Indiana University, Bloomington, IN, USA
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, USA
| |
Collapse
|
13
|
Di Stefano N, Vuust P, Brattico E. Consonance and dissonance perception. A critical review of the historical sources, multidisciplinary findings, and main hypotheses. Phys Life Rev 2022; 43:273-304. [PMID: 36372030 DOI: 10.1016/j.plrev.2022.10.004] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2022] [Accepted: 10/17/2022] [Indexed: 11/05/2022]
Abstract
Revealed more than two millennia ago by Pythagoras, consonance and dissonance (C/D) are foundational concepts in music theory, perception, and aesthetics. The search for the biological, acoustical, and cultural factors that affect C/D perception has resulted in descriptive accounts inspired by arithmetic, musicological, psychoacoustical or neurobiological frameworks without reaching a consensus. Here, we review the key historical sources and modern multidisciplinary findings on C/D and integrate them into three main hypotheses: the vocal similarity hypothesis (VSH), the psychocultural hypothesis (PH), and the sensorimotor hypothesis (SH). By illustrating the hypotheses-related findings, we highlight their major conceptual, methodological, and terminological shortcomings. Trying to provide a unitary framework for C/D understanding, we put together multidisciplinary research on human and animal vocalizations, which converges to suggest that auditory roughness is associated with distress/danger and, therefore, elicits defensive behavioral reactions and neural responses that indicate aversion. We therefore stress the primacy of vocality and roughness as key factors in the explanation of C/D phenomenon, and we explore the (neuro)biological underpinnings of the attraction-aversion mechanisms that are triggered by C/D stimuli. Based on the reviewed evidence, while the aversive nature of dissonance appears as solidly rooted in the multidisciplinary findings, the attractive nature of consonance remains a somewhat speculative claim that needs further investigation. Finally, we outline future directions for empirical research in C/D, especially regarding cross-modal and cross-cultural approaches.
Collapse
Affiliation(s)
- Nicola Di Stefano
- Institute for Cognitive Sciences and Technologies (ISTC), National Research Council of Italy (CNR), Via San Martino della Battaglia 44, 00185 Rome, Italy.
| | - Peter Vuust
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University Royal Academy of Music Aarhus/Aalborg (RAMA), 8000 Aarhus, Denmark.
| | - Elvira Brattico
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University Royal Academy of Music Aarhus/Aalborg (RAMA), 8000 Aarhus, Denmark; Department of Education, Psychology, Communication, University of Bari Aldo Moro, 70122 Bari, Italy.
| |
Collapse
|
14
|
Richardson ML, Guérit F, Gransier R, Wouters J, Carlyon RP, Middlebrooks JC. Temporal Pitch Sensitivity in an Animal Model: Psychophysics and Scalp Recordings : Temporal Pitch Sensitivity in Cat. J Assoc Res Otolaryngol 2022; 23:491-512. [PMID: 35668206 PMCID: PMC9437162 DOI: 10.1007/s10162-022-00849-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/24/2021] [Accepted: 04/11/2022] [Indexed: 01/28/2023] Open
Abstract
Cochlear implant (CI) users show limited sensitivity to the temporal pitch conveyed by electric stimulation, contributing to impaired perception of music and of speech in noise. Neurophysiological studies in cats suggest that this limitation is due, in part, to poor transmission of the temporal fine structure (TFS) by the brainstem pathways that are activated by electrical cochlear stimulation. It remains unknown, however, how that neural limit might influence perception in the same animal model. For that reason, we developed non-invasive psychophysical and electrophysiological measures of temporal (i.e., non-spectral) pitch processing in the cat. Normal-hearing (NH) cats were presented with acoustic pulse trains consisting of band-limited harmonic complexes that simulated CI stimulation of the basal cochlea while removing cochlear place-of-excitation cues. In the psychophysical procedure, trained cats detected changes from a base pulse rate to a higher pulse rate. In the scalp-recording procedure, the cortical-evoked acoustic change complex (ACC) and brainstem-generated frequency following response (FFR) were recorded simultaneously in sedated cats for pulse trains that alternated between the base and higher rates. The range of perceptual sensitivity to temporal pitch broadly resembled that of humans but was shifted to somewhat higher rates. The ACC largely paralleled these perceptual patterns, validating its use as an objective measure of temporal pitch sensitivity. The phase-locked FFR, in contrast, showed strong brainstem encoding for all tested pulse rates. These measures demonstrate the cat's perceptual sensitivity to pitch in the absence of cochlear-place cues and may be valuable for evaluating neural mechanisms of temporal pitch perception in the feline animal model of stimulation by a CI or novel auditory prostheses.
Collapse
Affiliation(s)
- Matthew L Richardson
- Department of Otolaryngology, Center for Hearing Research, University of California at Irvine, Irvine, CA, USA.
| | - François Guérit
- Cambridge Hearing Group, MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
| | - Robin Gransier
- Department of Neurosciences, ExpORL, KU Leuven, Leuven, Belgium
| | - Jan Wouters
- Department of Neurosciences, ExpORL, KU Leuven, Leuven, Belgium
| | - Robert P Carlyon
- Cambridge Hearing Group, MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
| | - John C Middlebrooks
- Department of Otolaryngology, Center for Hearing Research, University of California at Irvine, Irvine, CA, USA
- Departments of Neurobiology & Behavior, Biomedical Engineering, Cognitive Sciences, University of California at Irvine, Irvine, CA, USA
| |
Collapse
|
15
|
Cheng FY, Xu C, Gold L, Smith S. Rapid Enhancement of Subcortical Neural Responses to Sine-Wave Speech. Front Neurosci 2022; 15:747303. [PMID: 34987356 PMCID: PMC8721138 DOI: 10.3389/fnins.2021.747303] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2021] [Accepted: 12/02/2021] [Indexed: 01/15/2023] Open
Abstract
The efferent auditory nervous system may be a potent force in shaping how the brain responds to behaviorally significant sounds. Previous human experiments using the frequency following response (FFR) have shown efferent-induced modulation of subcortical auditory function online and over short- and long-term time scales; however, a contemporary understanding of FFR generation presents new questions about whether previous effects were constrained solely to the auditory subcortex. The present experiment used sine-wave speech (SWS), an acoustically-sparse stimulus in which dynamic pure tones represent speech formant contours, to evoke FFRSWS. Due to the higher stimulus frequencies used in SWS, this approach biased neural responses toward brainstem generators and allowed for three stimuli (/bɔ/, /bu/, and /bo/) to be used to evoke FFRSWSbefore and after listeners in a training group were made aware that they were hearing a degraded speech stimulus. All SWS stimuli were rapidly perceived as speech when presented with a SWS carrier phrase, and average token identification reached ceiling performance during a perceptual training phase. Compared to a control group which remained naïve throughout the experiment, training group FFRSWS amplitudes were enhanced post-training for each stimulus. Further, linear support vector machine classification of training group FFRSWS significantly improved post-training compared to the control group, indicating that training-induced neural enhancements were sufficient to bolster machine learning classification accuracy. These results suggest that the efferent auditory system may rapidly modulate auditory brainstem representation of sounds depending on their context and perception as non-speech or speech.
Collapse
Affiliation(s)
- Fan-Yin Cheng
- Department of Speech, Language, and Hearing Sciences, University of Texas at Austin, Austin, TX, United States
| | - Can Xu
- Department of Speech, Language, and Hearing Sciences, University of Texas at Austin, Austin, TX, United States
| | - Lisa Gold
- Department of Speech, Language, and Hearing Sciences, University of Texas at Austin, Austin, TX, United States
| | - Spencer Smith
- Department of Speech, Language, and Hearing Sciences, University of Texas at Austin, Austin, TX, United States
| |
Collapse
|
16
|
Raposo FA, Martins de Matos D, Ribeiro R. Learning Low-Dimensional Semantics for Music and Language via Multi-Subject fMRI. Neuroinformatics 2022; 20:451-461. [PMID: 34993852 DOI: 10.1007/s12021-021-09560-5] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/02/2021] [Indexed: 11/24/2022]
Abstract
Embodied Cognition (EC) states that semantics is encoded in the brain as firing patterns of neural circuits, which are learned according to the statistical structure of human multimodal experience. However, each human brain is idiosyncratically biased, according to its subjective experience, making this biological semantic machinery noisy with respect to semantics inherent to media, such as music and language. We propose to represent media semantics using low-dimensional vector embeddings by jointly modeling the functional Magnetic Resonance Imaging (fMRI) activity of several brains via Generalized Canonical Correlation Analysis (GCCA). We evaluate the semantic richness of the resulting latent space in appropriate semantic classification tasks: music genres and language topics. We show that the resulting unsupervised representations outperform the original high-dimensional fMRI voxel spaces in these downstream tasks while being more computationally efficient. Furthermore, we show that joint modeling of several subjects increases the semantic richness of the learned latent vector spaces as the number of subjects increases. Quantitative results and corresponding statistical significance testing demonstrate the instantiation of music and language semantics in the brain, thereby providing further evidence for multimodal embodied cognition as well as a method for extraction of media semantics from multi-subject brain dynamics.
Collapse
Affiliation(s)
- Francisco Afonso Raposo
- INESC-ID Lisboa, R. Alves Redol 9, Lisboa, 1000-029, Portugal. .,Instituto Superior Técnico, Universidade de Lisboa, Av. Rovisco Pais, Lisboa, 1049-001, Portugal.
| | - David Martins de Matos
- INESC-ID Lisboa, R. Alves Redol 9, Lisboa, 1000-029, Portugal.,Instituto Superior Técnico, Universidade de Lisboa, Av. Rovisco Pais, Lisboa, 1049-001, Portugal
| | - Ricardo Ribeiro
- INESC-ID Lisboa, R. Alves Redol 9, Lisboa, 1000-029, Portugal.,Instituto Universitário de Lisboa (ISCTE-IUL), Av. das Forças Armadas, Lisboa, 1649-026, Portugal
| |
Collapse
|
17
|
Spence C, Di Stefano N. Crossmodal Harmony: Looking for the Meaning of Harmony Beyond Hearing. Iperception 2022; 13:20416695211073817. [PMID: 35186248 PMCID: PMC8850342 DOI: 10.1177/20416695211073817] [Citation(s) in RCA: 16] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2021] [Revised: 11/20/2021] [Accepted: 12/23/2021] [Indexed: 12/02/2022] Open
Abstract
The notion of harmony was first developed in the context of metaphysics before being applied to the domain of music. However, in recent centuries, the term has often been used to describe especially pleasing combinations of colors by those working in the visual arts too. Similarly, the harmonization of flavors is nowadays often invoked as one of the guiding principles underpinning the deliberate pairing of food and drink. However, beyond the various uses of the term to describe and construct pleasurable unisensory perceptual experiences, it has also been suggested that music and painting may be combined harmoniously (e.g., see the literature on "color music"). Furthermore, those working in the area of "sonic seasoning" sometimes describe certain sonic compositions as harmonizing crossmodally with specific flavor sensations. In this review, we take a critical look at the putative meaning(s) of the term "harmony" when used in a crossmodal, or multisensory, context. Furthermore, we address the question of whether the term's use outside of a strictly unimodal auditory context should be considered literally or merely metaphorically (i.e., as a shorthand to describe those combinations of sensory stimuli that, for whatever reason, appear to go well together, and hence which can be processed especially fluently).
Collapse
Affiliation(s)
- Charles Spence
- Crossmodal Research Laboratory, University of Oxford, Oxford, UK
| | - Nicola Di Stefano
- Institute for Cognitive Sciences and Technologies, National Research Council of Italy (CNR), Rome, Italy
| |
Collapse
|
18
|
Carcagno S, Plack CJ. Relations between speech-reception, psychophysical temporal processing, and subcortical electrophysiological measures of auditory function in humans. Hear Res 2022; 417:108456. [PMID: 35149333 PMCID: PMC8935383 DOI: 10.1016/j.heares.2022.108456] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/07/2021] [Revised: 01/05/2022] [Accepted: 01/27/2022] [Indexed: 11/04/2022]
|
19
|
Camarena A, Manchala G, Papadopoulos J, O’Connell SR, Goldsworthy RL. Pleasantness Ratings of Musical Dyads in Cochlear Implant Users. Brain Sci 2021; 12:brainsci12010033. [PMID: 35053777 PMCID: PMC8773901 DOI: 10.3390/brainsci12010033] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2021] [Revised: 12/21/2021] [Accepted: 12/22/2021] [Indexed: 11/28/2022] Open
Abstract
Cochlear implants have been used to restore hearing to more than half a million people around the world. The restored hearing allows most recipients to understand spoken speech without relying on visual cues. While speech comprehension in quiet is generally high for recipients, many complain about the sound of music. The present study examines consonance and dissonance perception in nine cochlear implant users and eight people with no known hearing loss. Participants completed web-based assessments to characterize low-level psychophysical sensitivities to modulation and pitch, as well as higher-level measures of musical pleasantness and speech comprehension in background noise. The underlying hypothesis is that sensitivity to modulation and pitch, in addition to higher levels of musical sophistication, relate to higher-level measures of music and speech perception. This hypothesis tested true with strong correlations observed between measures of modulation and pitch with measures of consonance ratings and speech recognition. Additionally, the cochlear implant users who were the most sensitive to modulations and pitch, and who had higher musical sophistication scores, had similar pleasantness ratings as those with no known hearing loss. The implication is that better coding and focused rehabilitation for modulation and pitch sensitivity will broadly improve perception of music and speech for cochlear implant users.
Collapse
Affiliation(s)
- Andres Camarena
- Auditory Research Center, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, Los Angeles, CA 90033, USA; (A.C.); (G.M.); (J.P.); (S.R.O.)
| | - Grace Manchala
- Auditory Research Center, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, Los Angeles, CA 90033, USA; (A.C.); (G.M.); (J.P.); (S.R.O.)
| | - Julianne Papadopoulos
- Auditory Research Center, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, Los Angeles, CA 90033, USA; (A.C.); (G.M.); (J.P.); (S.R.O.)
- Thornton School of Music, University of Southern California, Los Angeles, CA 90089, USA
| | - Samantha R. O’Connell
- Auditory Research Center, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, Los Angeles, CA 90033, USA; (A.C.); (G.M.); (J.P.); (S.R.O.)
| | - Raymond L. Goldsworthy
- Auditory Research Center, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, Los Angeles, CA 90033, USA; (A.C.); (G.M.); (J.P.); (S.R.O.)
- Correspondence:
| |
Collapse
|
20
|
Neural correlates of acoustic dissonance in music: The role of musicianship, schematic and veridical expectations. PLoS One 2021; 16:e0260728. [PMID: 34852008 PMCID: PMC8635369 DOI: 10.1371/journal.pone.0260728] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2021] [Accepted: 11/15/2021] [Indexed: 11/19/2022] Open
Abstract
In western music, harmonic expectations can be fulfilled or broken by unexpected chords. Musical irregularities in the absence of auditory deviance elicit well-studied neural responses (e.g. ERAN, P3, N5). These responses are sensitive to schematic expectations (induced by syntactic rules of chord succession) and veridical expectations about predictability (induced by experimental regularities). However, the cognitive and sensory contributions to these responses and their plasticity as a result of musical training remains under debate. In the present study, we explored whether the neural processing of pure acoustic violations is affected by schematic and veridical expectations. Moreover, we investigated whether these two factors interact with long-term musical training. In Experiment 1, we registered the ERPs elicited by dissonant clusters placed either at the middle or the ending position of chord cadences. In Experiment 2, we presented to the listeners with a high proportion of cadences ending in a dissonant chord. In both experiments, we compared the ERPs of musicians and non-musicians. Dissonant clusters elicited distinctive neural responses (an early negativity, the P3 and the N5). While the EN was not affected by syntactic rules, the P3a and P3b were larger for dissonant closures than for middle dissonant chords. Interestingly, these components were larger in musicians than in non-musicians, while the N5 was the opposite. Finally, the predictability of dissonant closures in our experiment did not modulate any of the ERPs. Our study suggests that, at early time windows, dissonance is processed based on acoustic deviance independently of syntactic rules. However, at longer latencies, listeners may be able to engage integration mechanisms and further processes of attentional and structural analysis dependent on musical hierarchies, which are enhanced in musicians.
Collapse
|
21
|
Shukla B, Bidelman GM. Enhanced brainstem phase-locking in low-level noise reveals stochastic resonance in the frequency-following response (FFR). Brain Res 2021; 1771:147643. [PMID: 34473999 PMCID: PMC8490316 DOI: 10.1016/j.brainres.2021.147643] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2021] [Revised: 08/23/2021] [Accepted: 08/28/2021] [Indexed: 11/29/2022]
Abstract
In nonlinear systems, the inclusion of low-level noise can paradoxically improve signal detection, a phenomenon known as stochastic resonance (SR). SR has been observed in human hearing whereby sensory thresholds (e.g., signal detection and discrimination) are enhanced in the presence of noise. Here, we asked whether subcortical auditory processing (neural phase locking) shows evidence of SR. We recorded brainstem frequency-following-responses (FFRs) in young, normal-hearing listeners to near-electrophysiological-threshold (40 dB SPL) complex tones composed of 10 iso-amplitude harmonics of 150 Hz fundamental frequency (F0) presented concurrent with low-level noise (+20 to -20 dB SNRs). Though variable and weak across ears, some listeners showed improvement in auditory detection thresholds with subthreshold noise confirming SR psychophysically. At the neural level, low-level FFRs were initially eradicated by noise (expected masking effect) but were surprisingly reinvigorated at select masker levels (local maximum near ∼ 35 dB SPL). These data suggest brainstem phase-locking to near threshold periodic stimuli is enhanced in optimal levels of noise, the hallmark of SR. Our findings provide novel evidence for stochastic resonance in the human auditory brainstem and suggest that under some circumstances, noise can actually benefit both the behavioral and neural encoding of complex sounds.
Collapse
Affiliation(s)
- Bhanu Shukla
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA
| | - Gavin M Bidelman
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA; University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA.
| |
Collapse
|
22
|
Krishnan A, Suresh CH, Gandour JT. Cortical hemisphere preference and brainstem ear asymmetry reflect experience-dependent functional modulation of pitch. BRAIN AND LANGUAGE 2021; 221:104995. [PMID: 34303110 PMCID: PMC8559596 DOI: 10.1016/j.bandl.2021.104995] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/19/2021] [Revised: 05/07/2021] [Accepted: 07/07/2021] [Indexed: 06/13/2023]
Abstract
Temporal attributes of pitch processing at cortical and subcortical levels are differentially weighted and well-coordinated. The question is whether language experience induces functional modulation of hemispheric preference complemented by brainstem ear symmetry for pitch processing. Brainstem frequency-following and cortical pitch responses were recorded concurrently from Mandarin and English participants. A Mandarin syllable with a rising pitch contour was presented to both ears with monaural stimulation. At the cortical level, left ear stimulation in the Chinese group revealed an experience-dependent response for pitch processing in the right hemisphere, consistent with a functionalaccount. The English group revealed a contralateral hemisphere preference consistent with a structuralaccount. At the brainstem level, Chinese participants showed a functional leftward ear asymmetry, whereas English were consistent with a structural account. Overall, language experience modulates both cortical hemispheric preference and brainstem ear asymmetry in a complementary manner to optimize processing of temporal attributes of pitch.
Collapse
Affiliation(s)
- Ananthanarayan Krishnan
- Department of Speech Language Hearing Sciences, Purdue University, Lyles Porter Hall, 715 Clinic Drive, West Lafayette, IN 47907, USA.
| | - Chandan H Suresh
- Department of Speech Language Hearing Sciences, Purdue University, Lyles Porter Hall, 715 Clinic Drive, West Lafayette, IN 47907, USA; Department of Communication Disorders, California State, University, 5151 State University Drive, Los Angeles, CA 90032, USA.
| | - Jackson T Gandour
- Department of Speech Language Hearing Sciences, Purdue University, Lyles Porter Hall, 715 Clinic Drive, West Lafayette, IN 47907, USA.
| |
Collapse
|
23
|
Homma NY, Bajo VM. Lemniscal Corticothalamic Feedback in Auditory Scene Analysis. Front Neurosci 2021; 15:723893. [PMID: 34489635 PMCID: PMC8417129 DOI: 10.3389/fnins.2021.723893] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2021] [Accepted: 07/30/2021] [Indexed: 12/15/2022] Open
Abstract
Sound information is transmitted from the ear to central auditory stations of the brain via several nuclei. In addition to these ascending pathways there exist descending projections that can influence the information processing at each of these nuclei. A major descending pathway in the auditory system is the feedback projection from layer VI of the primary auditory cortex (A1) to the ventral division of medial geniculate body (MGBv) in the thalamus. The corticothalamic axons have small glutamatergic terminals that can modulate thalamic processing and thalamocortical information transmission. Corticothalamic neurons also provide input to GABAergic neurons of the thalamic reticular nucleus (TRN) that receives collaterals from the ascending thalamic axons. The balance of corticothalamic and TRN inputs has been shown to refine frequency tuning, firing patterns, and gating of MGBv neurons. Therefore, the thalamus is not merely a relay stage in the chain of auditory nuclei but does participate in complex aspects of sound processing that include top-down modulations. In this review, we aim (i) to examine how lemniscal corticothalamic feedback modulates responses in MGBv neurons, and (ii) to explore how the feedback contributes to auditory scene analysis, particularly on frequency and harmonic perception. Finally, we will discuss potential implications of the role of corticothalamic feedback in music and speech perception, where precise spectral and temporal processing is essential.
Collapse
Affiliation(s)
- Natsumi Y. Homma
- Center for Integrative Neuroscience, University of California, San Francisco, San Francisco, CA, United States
- Coleman Memorial Laboratory, Department of Otolaryngology – Head and Neck Surgery, University of California, San Francisco, San Francisco, CA, United States
| | - Victoria M. Bajo
- Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
24
|
Deviant consonance and dissonance capture attention differently only when task demand is high: An ERP study with three-stimulus oddball paradigm. Int J Psychophysiol 2021; 166:1-8. [PMID: 33932475 DOI: 10.1016/j.ijpsycho.2021.04.008] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2020] [Revised: 03/22/2021] [Accepted: 04/25/2021] [Indexed: 11/22/2022]
Abstract
The current study investigated whether consonance (imperfect consonance: major third) and dissonance (minor second) would capture attention differently when they occurred as chords (combinations of two tones) that were deviant from their context. In addition, we also examined how task demand would modulate these chords' attentional capture. For this investigation, we used an auditory three-stimulus oddball paradigm in which these chords were presented as deviant stimuli (5% each) among frequent standard (80%) and infrequent target (10%) pure tones. The task difficulty was manipulated by changing pitch intervals between standard and target tones. The results showed that these chords elicited dual-peak P3a, and that consonance enhanced the late phase of P3a compared to dissonance, only when the task demand was high. These findings revealed that deviant consonance and dissonance captured attention differently; in particular, consonance captured attention more strongly than dissonance, and this effect was induced by high task demand. This attentional capture difference between the chord categories was induced through enhanced focus of attention on the pitch dimension of oddball stimuli. In addition, the deviant chords might have been processed by a mechanism similar to that which processes novel stimuli, and these chords' differences might have affected not the novelty detection process, but a process which orients attentional resources to deviant chords, which were recognized as novel stimuli.
Collapse
|
25
|
Bestelmeyer PEG, Mühl C. Individual differences in voice adaptability are specifically linked to voice perception skill. Cognition 2021; 210:104582. [PMID: 33450447 DOI: 10.1016/j.cognition.2021.104582] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2020] [Revised: 01/01/2021] [Accepted: 01/03/2021] [Indexed: 10/22/2022]
Abstract
There are remarkable individual differences in the ability to recognise individuals by the sound of their voice. Theoretically, this ability is thought to depend on the coding accuracy of voices in a low-dimensional "voice-space". Here we were interested in how adaptive coding of voice identity relates to this variability in skill. In two adaptation experiments we explored first whether the aftereffect size to two familiar vocal identities can predict voice perception ability and second, whether this effect stems from general auditory skill (e.g. discrimination ability for tuning and tempo). Experiment 1 demonstrated that contrastive aftereffect sizes for voice identity predicted voice perception ability. In Experiment 2, we replicated this finding and further established that this effect is unrelated to general auditory abilities or general adaptability of listeners. Our results highlight the important functional role of adaptive coding in voice expertise and suggest that human voice perception is a highly specialised and distinct auditory ability.
Collapse
Affiliation(s)
| | - Constanze Mühl
- School of Psychology, Bangor University, Bangor, Gwynedd, UK
| |
Collapse
|
26
|
Losorelli S, Kaneshiro B, Musacchia GA, Blevins NH, Fitzgerald MB. Factors influencing classification of frequency following responses to speech and music stimuli. Hear Res 2020; 398:108101. [PMID: 33142106 DOI: 10.1016/j.heares.2020.108101] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/10/2019] [Revised: 09/25/2020] [Accepted: 10/19/2020] [Indexed: 01/08/2023]
Abstract
Successful mapping of meaningful labels to sound input requires accurate representation of that sound's acoustic variances in time and spectrum. For some individuals, such as children or those with hearing loss, having an objective measure of the integrity of this representation could be useful. Classification is a promising machine learning approach which can be used to objectively predict a stimulus label from the brain response. This approach has been previously used with auditory evoked potentials (AEP) such as the frequency following response (FFR), but a number of key issues remain unresolved before classification can be translated into clinical practice. Specifically, past efforts at FFR classification have used data from a given subject for both training and testing the classifier. It is also unclear which components of the FFR elicit optimal classification accuracy. To address these issues, we recorded FFRs from 13 adults with normal hearing in response to speech and music stimuli. We compared labeling accuracy of two cross-validation classification approaches using FFR data: (1) a more traditional method combining subject data in both the training and testing set, and (2) a "leave-one-out" approach, in which subject data is classified based on a model built exclusively from the data of other individuals. We also examined classification accuracy on decomposed and time-segmented FFRs. Our results indicate that the accuracy of leave-one-subject-out cross validation approaches that obtained in the more conventional cross-validation classifications while allowing a subject's results to be analysed with respect to normative data pooled from a separate population. In addition, we demonstrate that classification accuracy is highest when the entire FFR is used to train the classifier. Taken together, these efforts contribute key steps toward translation of classification-based machine learning approaches into clinical practice.
Collapse
Affiliation(s)
- Steven Losorelli
- Department of Otolaryngology Head and Neck Surgery, Stanford University School of Medicine, Palo Alto, CA, USA.
| | - Blair Kaneshiro
- Department of Otolaryngology Head and Neck Surgery, Stanford University School of Medicine, Palo Alto, CA, USA.
| | - Gabriella A Musacchia
- Department of Otolaryngology Head and Neck Surgery, Stanford University School of Medicine, Palo Alto, CA, USA; Department of Audiology, University of the Pacific, San Francisco, CA, USA.
| | - Nikolas H Blevins
- Department of Otolaryngology Head and Neck Surgery, Stanford University School of Medicine, Palo Alto, CA, USA.
| | - Matthew B Fitzgerald
- Department of Otolaryngology Head and Neck Surgery, Stanford University School of Medicine, Palo Alto, CA, USA.
| |
Collapse
|
27
|
Weber A, Busbridge S, Governo R. Evaluation of the Efficacy of Musical Vibroacupuncture in Pain Relief: A Randomized Controlled Pilot Study. Neuromodulation 2020; 24:1475-1482. [PMID: 33029913 DOI: 10.1111/ner.13281] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2020] [Revised: 08/05/2020] [Accepted: 08/24/2020] [Indexed: 12/17/2022]
Abstract
OBJECTIVES To investigate if skin vibration employing consonant frequencies emitted by skin transducers attached to a combination of acupuncture points and according to musical harmony (musical chord) produces more significant pain relief compared to just a single frequency. MATERIALS AND METHODS Skin vibrostimulation produced by five electromagnet transducers was applied at five acupoints traditionally used to pain relief and anxiety in 13 pain-free healthy volunteers using the cold pressor test (CPT). The study consisted of three randomized sessions conducted on alternate days, with participants receiving either simultaneous frequencies of 32, 48, and 64 Hz that equate those used in a musical chord, hereby defined as musical vibroacupuncture (MVA), a single frequency of 32 Hz, set as vibroacupuncture (VA) and sham procedure (SP). CPT scores for pain thresholds and pain tolerance were assessed using repeated-measures ANOVAs. Pain intensity was evaluated using a numerical rating scale (NRS), while sensory and affective aspects of pain were rated using the short-form McGill Pain Questionnaire (SF-MPQ) and State-Trait Anxiety Inventory (STAI) Y-Form. RESULTS Pain thresholds did not vary significantly between trials. Pain tolerance scores were markedly higher in MVA compared to baseline (p = 0.0043) or SP (p = 0.006) but not for VA. Pain intensity for MVA also differed significantly from the baseline (p = 0.007) or SP (p = 0.027), but not for VA. No significant differences were found in SF-MPQ and STAI questionnaires. CONCLUSIONS These results suggest that MVA effectively increased pain tolerance and reduced pain intensity when compared with all groups, although not significant to the VA group.
Collapse
Affiliation(s)
- Augusto Weber
- School of Computing, Engineering and Mathematics, University of Brighton, Brighton, East Sussex, UK
| | - Simon Busbridge
- School of Computing, Engineering and Mathematics, University of Brighton, Brighton, East Sussex, UK
| | - Ricardo Governo
- Brighton and Sussex Medical School (BSMS), University of Sussex, Brighton, East Sussex, UK
| |
Collapse
|
28
|
Kim SJ, Yoo GE, Shin Y, Cho S. Gait training for adults with cerebral palsy following harmonic modification in rhythmic auditory stimulation. Ann N Y Acad Sci 2020; 1473:11-19. [DOI: 10.1111/nyas.14306] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2019] [Revised: 01/02/2020] [Accepted: 01/08/2020] [Indexed: 12/14/2022]
Affiliation(s)
- Soo Ji Kim
- Music Therapy Education, Graduate School of EducationEwha Womans University Seoul Korea
| | - Ga Eul Yoo
- Department of Music Therapy, Graduate SchoolEwha Womans University Seoul Korea
| | - Yoon‐Kyum Shin
- Department and Research Institute of Rehabilitation MedicineYonsei University College of Medicine Seoul Korea
- Brain Korea 21 PLUS Project for Medical ScienceYonsei University College of Medicine Seoul Korea
| | - Sung‐Rae Cho
- Department and Research Institute of Rehabilitation MedicineYonsei University College of Medicine Seoul Korea
- Brain Korea 21 PLUS Project for Medical ScienceYonsei University College of Medicine Seoul Korea
- Rehabilitation Institute of Neuromuscular DiseaseYonsei University College of Medicine Seoul Korea
| |
Collapse
|
29
|
Evidence of both brainstem and auditory cortex involvement in categorical perception for Chinese lexical tones. Neuroreport 2020; 31:359-364. [DOI: 10.1097/wnr.0000000000001414] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
30
|
Harrison PMC, Pearce MT. Simultaneous consonance in music perception and composition. Psychol Rev 2020; 127:216-244. [PMID: 31868392 PMCID: PMC7032667 DOI: 10.1037/rev0000169] [Citation(s) in RCA: 21] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/21/2019] [Revised: 06/13/2019] [Accepted: 09/02/2019] [Indexed: 11/08/2022]
Abstract
Simultaneous consonance is a salient perceptual phenomenon corresponding to the perceived pleasantness of simultaneously sounding musical tones. Various competing theories of consonance have been proposed over the centuries, but recently a consensus has developed that simultaneous consonance is primarily driven by harmonicity perception. Here we question this view, substantiating our argument by critically reviewing historic consonance research from a broad variety of disciplines, reanalyzing consonance perception data from 4 previous behavioral studies representing more than 500 participants, and modeling three Western musical corpora representing more than 100,000 compositions. We conclude that simultaneous consonance is a composite phenomenon that derives in large part from three phenomena: interference, periodicity/harmonicity, and cultural familiarity. We formalize this conclusion with a computational model that predicts a musical chord's simultaneous consonance from these three features, and release this model in an open-source R package, incon, alongside 15 other computational models also evaluated in this paper. We hope that this package will facilitate further psychological and musicological research into simultaneous consonance. (PsycINFO Database Record (c) 2020 APA, all rights reserved).
Collapse
Affiliation(s)
- Peter M C Harrison
- School of Electronic Engineering and Computer Science, Queen Mary University of London
| | - Marcus T Pearce
- School of Electronic Engineering and Computer Science, Queen Mary University of London
| |
Collapse
|
31
|
Yano H, Takiguchi T, Nakagawa S. Cortical Patterns for Prediction of Subjective Preference Induced by Chords .. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2020; 2019:5168-5171. [PMID: 31947022 DOI: 10.1109/embc.2019.8857941] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
To extract an effective feature in prediction of subjective impressions from single-trial neurophysiological recordings, the spatial filter that extracts brain activities related to impressions were constructed using the common spatial pattern (CSP). We focus on subjective preference induced by chords composed of 3 notes with different frequency ratio. Magnetic cortical activities while hearing chords and comparative judgment on pair of them were measured. The predictive model that predicts the scale value of preference was trained using the CSP-based feature for each participant. The result of the evaluation experiment shows that the CSP-based feature improved the mean prediction accuracy in all participants, compared with the other features without spatially filtering. Furthermore, the capability of construction of a spatial filter that extracts cortical activities varying with degree of preference using the comparative judgments was indicated.
Collapse
|
32
|
Proverbio AM, Benedetto F, Guazzone M. Shared neural mechanisms for processing emotions in music and vocalizations. Eur J Neurosci 2019; 51:1987-2007. [DOI: 10.1111/ejn.14650] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2019] [Revised: 11/21/2019] [Accepted: 12/05/2019] [Indexed: 12/21/2022]
Affiliation(s)
- Alice Mado Proverbio
- Department of Psychology University of Milano‐Bicocca Milan Italy
- Milan Center for Neuroscience Milan Italy
| | - Francesco Benedetto
- Department of Psychology University of Milano‐Bicocca Milan Italy
- Milan Center for Neuroscience Milan Italy
| | - Martina Guazzone
- Department of Psychology University of Milano‐Bicocca Milan Italy
- Milan Center for Neuroscience Milan Italy
| |
Collapse
|
33
|
|
34
|
Carcagno S, Lakhani S, Plack CJ. Consonance perception beyond the traditional existence region of pitch. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:2279. [PMID: 31671967 DOI: 10.1121/1.5127845] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/20/2019] [Accepted: 09/12/2019] [Indexed: 06/10/2023]
Abstract
Some theories posit that the perception of consonance is based on neural periodicity detection, which is dependent on accurate phase locking of auditory nerve fibers to features of the stimulus waveform. In the current study, 15 listeners were asked to rate the pleasantness of complex tone dyads (2 note chords) forming various harmonic intervals and bandpass filtered in a high-frequency region (all components >5.8 kHz), where phase locking to the rapid stimulus fine structure is thought to be severely degraded or absent. The two notes were presented to opposite ears. Consonant intervals (minor third and perfect fifth) received higher ratings than dissonant intervals (minor second and tritone). The results could not be explained in terms of phase locking to the slower waveform envelope because the preference for consonant intervals was higher when the stimuli were harmonic, compared to a condition in which they were made inharmonic by shifting their component frequencies by a constant offset, so as to preserve their envelope periodicity. Overall the results indicate that, if phase locking is indeed absent at frequencies greater than ∼5 kHz, neural periodicity detection is not necessary for the perception of consonance.
Collapse
Affiliation(s)
- Samuele Carcagno
- Department of Psychology, Lancaster University, Lancaster, LA1 4YF, United Kingdom
| | - Saday Lakhani
- Department of Psychology, Lancaster University, Lancaster, LA1 4YF, United Kingdom
| | - Christopher J Plack
- Department of Psychology, Lancaster University, Lancaster, LA1 4YF, United Kingdom
| |
Collapse
|
35
|
Pagès-Portabella C, Toro JM. Dissonant endings of chord progressions elicit a larger ERAN than ambiguous endings in musicians. Psychophysiology 2019; 57:e13476. [PMID: 31512751 DOI: 10.1111/psyp.13476] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2018] [Revised: 07/30/2019] [Accepted: 08/08/2019] [Indexed: 11/29/2022]
Abstract
In major-minor tonal music, the hierarchical relationships and patterns of tension/release are essential for its composition and experience. For most listeners, tension leads to an expectation of resolution. Thus, when musical expectations are broken, they are usually perceived as erroneous and elicit specific neural responses such as the early right anterior negativity (ERAN). In the present study, we explored if different degrees of musical violations are processed differently after long-term musical training in comparison to day-to-day exposure. We registered the ERPs elicited by listening to unexpected chords in both musicians and nonmusicians. More specifically, we compared the responses of strong violations by unexpected dissonant endings and mild violations by unexpected but consonant endings (Neapolitan chords). Our results show that, irrespective of training, irregular endings elicited the ERAN. However, the ERAN for dissonant endings was larger in musicians than in nonmusicians. More importantly, we observed a modulation of the neural responses by the degree of violation only in musicians. In this group, the amplitude of the ERAN was larger for strong than for mild violations. These results suggest an early sensitivity of musicians to dissonance, which is processed as less expected than tonal irregularities. We also found that irregular endings elicited a P3 only in musicians. Our study suggests that, even though violations of harmonic expectancies are detected by all listeners, musical training modulates how different violations of the musical context are processed.
Collapse
Affiliation(s)
- Carlota Pagès-Portabella
- Language & Comparative Cognition Group, Center for Brain & Cognition, Universitat Pompeu Fabra, Barcelona, Spain
| | - Juan M Toro
- Language & Comparative Cognition Group, Center for Brain & Cognition, Universitat Pompeu Fabra, Barcelona, Spain.,Institució Catalana de Recerca i Estudis Avançats, Barcelona, Spain
| |
Collapse
|
36
|
Spitzer ER, Landsberger DM, Friedmann DR, Galvin JJ. Pleasantness Ratings for Harmonic Intervals With Acoustic and Electric Hearing in Unilaterally Deaf Cochlear Implant Patients. Front Neurosci 2019; 13:922. [PMID: 31551686 PMCID: PMC6733976 DOI: 10.3389/fnins.2019.00922] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2019] [Accepted: 08/16/2019] [Indexed: 11/13/2022] Open
Abstract
Background Harmony is an important part of tonal music that conveys context, form and emotion. Two notes sounded simultaneously form a harmonic interval. In normal-hearing (NH) listeners, some harmonic intervals (e.g., minor 2nd, tritone, major 7th) typically sound more dissonant than others (e.g., octave, major 3rd, 4th). Because of the limited spectro-temporal resolution afforded by cochlear implants (CIs), music perception is generally poor. However, CI users may still be sensitive to relative dissonance across intervals. In this study, dissonance ratings for harmonic intervals were measured in 11 unilaterally deaf CI patients, in whom ratings from the CI could be compared to those from the normal ear. Methods Stimuli consisted of pairs of equal amplitude MIDI piano tones. Intervals spanned a range of two octaves relative to two root notes (F3 or C4). Dissonance was assessed in terms of subjective pleasantness ratings for intervals presented to the NH ear alone, the CI ear alone, and both ears together (NH + CI). Ratings were collected for both root notes for within- and across-octave intervals (1–12 and 13–24 semitones). Participants rated the pleasantness of each interval by clicking on a line anchored with “least pleasant” and “most pleasant.” A follow-up experiment repeated the task with a smaller stimulus set. Results With NH-only listening, within-octave intervals minor 2nd, major 2nd, and major 7th were rated least pleasant; major 3rd, 5th, and octave were rated most pleasant. Across-octave counterparts were similarly rated. With CI-only listening, ratings were consistently lower and showed a reduced range. Mean ratings were highly correlated between NH-only and CI-only listening (r = 0.845, p < 0.001). Ratings were similar between NH-only and NH + CI listening, with no significant binaural enhancement/interference. The follow-up tests showed that ratings were reliable for the least and most pleasant intervals. Discussion Although pleasantness ratings were less differentiated for the CI ear than the NH ear, there were similarities between the two listening modes. Given the lack of spectro-temporal detail needed for harmonicity-based distinctions, temporal envelope interactions (within and across channels) associated with a perception of roughness may contribute to dissonance perception for harmonic intervals with CI-only listening.
Collapse
Affiliation(s)
- Emily R Spitzer
- Department of Otolaryngology-Head and Neck Surgery, New York University School of Medicine, New York, NY, United States
| | - David M Landsberger
- Department of Otolaryngology-Head and Neck Surgery, New York University School of Medicine, New York, NY, United States
| | - David R Friedmann
- Department of Otolaryngology-Head and Neck Surgery, New York University School of Medicine, New York, NY, United States
| | | |
Collapse
|
37
|
Llanos F, Xie Z, Chandrasekaran B. Biometric identification of listener identity from frequency following responses to speech. J Neural Eng 2019; 16:056004. [PMID: 31039552 DOI: 10.1088/1741-2552/ab1e01] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/31/2022]
Abstract
OBJECTIVE We investigate the biometric specificity of the frequency following response (FFR), an EEG marker of early auditory processing that reflects phase-locked activity from neural ensembles in the auditory cortex and subcortex (Chandrasekaran and Kraus 2010, Bidelman, 2015a, 2018, Coffey et al 2017b). Our objective is two-fold: demonstrate that the FFR contains information beyond stimulus properties and broad group-level markers, and to assess the practical viability of the FFR as a biometric across different sounds, auditory experiences, and recording days. APPROACH We trained the hidden Markov model (HMM) to decode listener identity from FFR spectro-temporal patterns across multiple frequency bands. Our dataset included FFRs from twenty native speakers of English or Mandarin Chinese (10 per group) listening to Mandarin Chinese tones across three EEG sessions separated by days. We decoded subject identity within the same auditory context (same tone and session) and across different stimuli and recording sessions. MAIN RESULTS The HMM decoded listeners for averaging sizes as small as one single FFR. However, model performance improved for larger averaging sizes (e.g. 25 FFRs), similarity in auditory context (same tone and day), and lack of familiarity with the sounds (i.e. native English relative to native Chinese listeners). Our results also revealed important biometric contributions from frequency bands in the cortical and subcortical EEG. SIGNIFICANCE Our study provides the first deep and systematic biometric characterization of the FFR and provides the basis for biometric identification systems incorporating this neural signal.
Collapse
Affiliation(s)
- Fernando Llanos
- Department of Communication Sciences and Disorders, University of Pittsburgh, Pittsburgh, PA 15213, United States of America
| | | | | |
Collapse
|
38
|
Mencke I, Omigie D, Wald-Fuhrmann M, Brattico E. Atonal Music: Can Uncertainty Lead to Pleasure? Front Neurosci 2019; 12:979. [PMID: 30670941 PMCID: PMC6331456 DOI: 10.3389/fnins.2018.00979] [Citation(s) in RCA: 22] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2018] [Accepted: 12/07/2018] [Indexed: 01/25/2023] Open
Abstract
In recent years, the field of neuroaesthetics has gained considerable attention with music being a favored object of study. The majority of studies concerning music have, however, focused on the experience of Western tonal music (TM), which is characterized by tonal hierarchical organization, a high degree of consonance, and a tendency to provide the listener with a tonic as a reference point during the listening experience. We argue that a narrow focus on Western TM may have led to a one-sided view regarding the qualities of the aesthetic experience of music since Western art music from the 20th and 21st century like atonal music (AM) – while lacking a tonal hierarchical structure, and while being highly dissonant and hard to predict – is nevertheless enjoyed by a group of avid listeners. We propose a research focus that investigates, in particular, the experience of AM as a novel and compelling way with which to enhance our understanding of both the aesthetic appreciation of music and the role of predictive models in the context of musical pleasure. We use music theoretical analysis and music information retrieval methods to demonstrate how AM presents the listener with a highly uncertain auditory environment. Specifically, an analysis of a corpus of 100 musical segments is used to illustrate how tonal classical music and AM differ quantitatively in terms of both key and pulse clarity values. We then examine person related, extrinsic and intrinsic factors, that point to potential mechanisms underlying the appreciation and pleasure derived from AM. We argue that personality traits like “openness to experience,” the framing of AM as art, and the mere exposure effect are key components of such mechanisms. We further argue that neural correlates of uncertainty estimation could represent a central mechanism for engaging with AM and that such contexts engender a comparatively weak predictive model in the listener. Finally we argue that in such uncertain contexts, correct predictions may be more subjectively rewarding than prediction errors since they signal to the individual that their predictive model is improving.
Collapse
Affiliation(s)
- Iris Mencke
- Department of Music, Max Planck Institute for Empirical Aesthetics, Frankfurt, Germany.,Center for Music in the Brain, Department of Clinical Medicine, Aarhus University and The Royal Academy of Music, Aarhus, Denmark
| | - Diana Omigie
- Department of Music, Max Planck Institute for Empirical Aesthetics, Frankfurt, Germany.,Department of Psychology, Goldsmiths, University of London, London, United Kingdom
| | - Melanie Wald-Fuhrmann
- Department of Music, Max Planck Institute for Empirical Aesthetics, Frankfurt, Germany
| | - Elvira Brattico
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University and The Royal Academy of Music, Aarhus, Denmark
| |
Collapse
|
39
|
Early neural responses underlie advantages for consonance over dissonance. Neuropsychologia 2018; 117:188-198. [PMID: 29885961 PMCID: PMC6092559 DOI: 10.1016/j.neuropsychologia.2018.06.005] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2017] [Revised: 06/01/2018] [Accepted: 06/06/2018] [Indexed: 12/05/2022]
Abstract
Consonant musical intervals tend to be more readily processed than dissonant intervals. In the present study, we explore the neural basis for this difference by registering how the brain responds after changes in consonance and dissonance, and how formal musical training modulates these responses. Event-related brain potentials (ERPs) were registered while participants were presented with sequences of consonant intervals interrupted by a dissonant interval, or sequences of dissonant intervals interrupted by a consonant interval. Participants were musicians and non-musicians. Our results show that brain responses triggered by changes in a consonant context differ from those triggered in a dissonant context. Changes in a sequence of consonant intervals are rapidly processed independently of musical expertise, as revealed by a change-related mismatch negativity (MMN, a component of the ERPs triggered by an odd stimulus in a sequence of stimuli) elicited in both musicians and non-musicians. In contrast, changes in a sequence of dissonant intervals elicited a late MMN only in participants with prolonged musical training. These different neural responses might form the basis for the processing advantages observed for consonance over dissonance and provide information about how formal musical training modulates them. We registered ERPs after deviant intervals in consonant and dissonant sequences. Violations of consonant sequences are detected independently of musical expertise. Musical training modulates responses to violations of dissonant sequences. These neural responses might form the basis for a processing advantage of consonance.
Collapse
|
40
|
Trulla LL, Di Stefano N, Giuliani A. Computational Approach to Musical Consonance and Dissonance. Front Psychol 2018; 9:381. [PMID: 29670552 PMCID: PMC5893895 DOI: 10.3389/fpsyg.2018.00381] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2017] [Accepted: 03/08/2018] [Indexed: 11/21/2022] Open
Abstract
In sixth century BC, Pythagoras discovered the mathematical foundation of musical consonance and dissonance. When auditory frequencies in small-integer ratios are combined, the result is a harmonious perception. In contrast, most frequency combinations result in audible, off-centered by-products labeled “beating” or “roughness;” these are reported by most listeners to sound dissonant. In this paper, we consider second-order beats, a kind of beating recognized as a product of neural processing, and demonstrate that the data-driven approach of Recurrence Quantification Analysis (RQA) allows for the reconstruction of the order in which interval ratios are ranked in music theory and harmony. We take advantage of computer-generated sounds containing all intervals over the span of an octave. To visualize second-order beats, we use a glissando from the unison to the octave. This procedure produces a profile of recurrence values that correspond to subsequent epochs along the original signal. We find that the higher recurrence peaks exactly match the epochs corresponding to just intonation frequency ratios. This result indicates a link between consonance and the dynamical features of the signal. Our findings integrate a new element into the existing theoretical models of consonance, thus providing a computational account of consonance in terms of dynamical systems theory. Finally, as it considers general features of acoustic signals, the present approach demonstrates a universal aspect of consonance and dissonance perception and provides a simple mathematical tool that could serve as a common framework for further neuro-psychological and music theory research.
Collapse
Affiliation(s)
| | - Nicola Di Stefano
- Institute of Philosophy of Scientific and Technological Practice and Laboratory of Developmental Neuroscience, Università Campus Bio-Medico di Roma, Rome, Italy
| | - Alessandro Giuliani
- Environment and Health Department, National Institute of Health, Rome, Italy
| |
Collapse
|
41
|
Bidelman GM. Subcortical sources dominate the neuroelectric auditory frequency-following response to speech. Neuroimage 2018; 175:56-69. [PMID: 29604459 DOI: 10.1016/j.neuroimage.2018.03.060] [Citation(s) in RCA: 151] [Impact Index Per Article: 25.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/05/2018] [Accepted: 03/26/2018] [Indexed: 11/16/2022] Open
Abstract
Frequency-following responses (FFRs) are neurophonic potentials that provide a window into the encoding of complex sounds (e.g., speech/music), auditory disorders, and neuroplasticity. While the neural origins of the FFR remain debated, renewed controversy has reemerged after demonstration that FFRs recorded via magnetoencephalography (MEG) are dominated by cortical rather than brainstem structures as previously assumed. Here, we recorded high-density (64 ch) FFRs via EEG and applied state-of-the art source imaging techniques to multichannel data (discrete dipole modeling, distributed imaging, independent component analysis, computational simulations). Our data confirm a mixture of generators localized to bilateral auditory nerve (AN), brainstem inferior colliculus (BS), and bilateral primary auditory cortex (PAC). However, frequency-specific scrutiny of source waveforms showed the relative contribution of these nuclei to the aggregate FFR varied across stimulus frequencies. Whereas AN and BS sources produced robust FFRs up to ∼700 Hz, PAC showed weak phase-locking with little FFR energy above the speech fundamental (100 Hz). Notably, CLARA imaging further showed PAC activation was eradicated for FFRs >150 Hz, above which only subcortical sources remained active. Our results show (i) the site of FFR generation varies critically with stimulus frequency; and (ii) opposite the pattern observed in MEG, subcortical structures make the largest contribution to electrically recorded FFRs (AN ≥ BS > PAC). We infer that cortical dominance observed in previous neuromagnetic data is likely due to the bias of MEG to superficial brain tissue, underestimating subcortical structures that drive most of the speech-FFR. Cleanly separating subcortical from cortical FFRs can be achieved by ensuring stimulus frequencies are >150-200 Hz, above the phase-locking limit of cortical neurons.
Collapse
Affiliation(s)
- Gavin M Bidelman
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA; Univeristy of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA.
| |
Collapse
|
42
|
A Conceptual Framework Encompassing the Psychoneuroimmunoendocrinological Influences of Listening to Music in Patients With Heart Failure. Holist Nurs Pract 2018; 32:81-89. [DOI: 10.1097/hnp.0000000000000253] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/25/2023]
|
43
|
Abstract
Do you know that our soul is composed of harmony? Leonardo Da Vinci Despite evidence for music-specific mechanisms at the level of pitch-pattern representations, the most fascinating aspect of music is its transmodality. Recent psychological and neuroscientific evidence suggest that music is unique in the coupling of perception, cognition, action and emotion. This potentially explains why music has been since time immemorial almost inextricably linked to healing processes and should continue to be.
Collapse
Affiliation(s)
- Paulo E Andrade
- Department of Psychology, Goldsmiths, University of London, London, UK
| | | |
Collapse
|
44
|
Abstract
The foundations of human music have long puzzled philosophers, mathematicians, psychologists, and neuroscientists. Although virtually all cultures uses combinations of tones as a basis for musical expression, why humans favor some tone combinations over others has been debated for millennia. Here we show that our attraction to specific tone combinations played simultaneously (chords) is predicted by their spectral similarity to voiced speech sounds. This connection between auditory aesthetics and a primary characteristic of vocalization adds to other evidence that tonal preferences arise from the biological advantages of social communication mediated by speech and language. Musical chords are combinations of two or more tones played together. While many different chords are used in music, some are heard as more attractive (consonant) than others. We have previously suggested that, for reasons of biological advantage, human tonal preferences can be understood in terms of the spectral similarity of tone combinations to harmonic human vocalizations. Using the chromatic scale, we tested this theory further by assessing the perceived consonance of all possible dyads, triads, and tetrads within a single octave. Our results show that the consonance of chords is predicted by their relative similarity to voiced speech sounds. These observations support the hypothesis that the relative attraction of musical tone combinations is due, at least in part, to the biological advantages that accrue from recognizing and responding to conspecific vocal stimuli.
Collapse
|
45
|
Cook ND. Calculation of the acoustical properties of triadic harmonies. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:3748. [PMID: 29289060 DOI: 10.1121/1.5018342] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
The author reports that the harmonic "tension" and major/minor "valence" of pitch combinations can be calculated directly from acoustical properties without relying on concepts from traditional harmony theory. The capability to compute the well-known types of harmonic triads means that their perception is not simply a consequence of learning an arbitrary cultural "idiom" handed down from the Italian Renaissance. On the contrary, for typical listeners familiar with diatonic music, attention to certain, definable, acoustical features underlies the perception of the valence (modality) and the inherent tension (instability) of three-tone harmonies.
Collapse
Affiliation(s)
- Norman D Cook
- Department of Informatics, Kansai University, 2-1 Reizenji, Takatsuki, Osaka, 569-1095, Japan
| |
Collapse
|
46
|
Effects of free choice and outcome valence on the sense of agency: evidence from measures of intentional binding and feelings of control. Exp Brain Res 2017; 236:129-139. [DOI: 10.1007/s00221-017-5112-3] [Citation(s) in RCA: 40] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2017] [Accepted: 10/16/2017] [Indexed: 01/31/2023]
|
47
|
Wang Q(J, Spence C. “A sweet smile”: the modulatory role of emotion in how extrinsic factors influence taste evaluation. Cogn Emot 2017; 32:1052-1061. [DOI: 10.1080/02699931.2017.1386623] [Citation(s) in RCA: 29] [Impact Index Per Article: 4.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/30/2022]
Affiliation(s)
- Qian (Janice) Wang
- Crossmodal Research Laboratory, Department of Experimental Psychology, University of Oxford, Oxford, UK
| | - Charles Spence
- Crossmodal Research Laboratory, Department of Experimental Psychology, University of Oxford, Oxford, UK
| |
Collapse
|
48
|
Abstract
OBJECTIVE Cochlear implantation is associated with poor music perception and enjoyment. Reducing music complexity has been shown to enhance music enjoyment in cochlear implant (CI) recipients. In this study, we assess the impact of harmonic series reduction on music enjoyment. STUDY DESIGN Prospective analysis of music enjoyment in normal-hearing (NH) individuals and CI recipients. SETTING Single tertiary academic medical center. PATIENTS NH adults (N = 20) and CI users (N = 8) rated the Happy Birthday song on three validated enjoyment modalities-musicality, pleasantness, and naturalness. INTERVENTION Subjective rating of music excerpts. MAIN OUTCOME MEASURES Participants listened to seven different instruments play the melody, each with five levels of harmonic reduction (Full, F3+F2+F1+F0, F2+F1+F0, F1+F0, F0). NH participants listened to the segments both with and without CI simulation. Linear mixed effect models (LME) and likelihood ratio tests were used to assess the impact of harmonic reduction on enjoyment. RESULTS NH listeners without simulation rated segments with the first four harmonics (F3+F2+F1+F0) most pleasant and natural (p <0.001, p = 0.004). NH listeners with simulation rated the first harmonic alone (F0) most pleasant and natural (p <0.001, p = 0.003). Their ratings demonstrated a positive linear relationship between harmonic reduction and both pleasantness (slope estimate = 0.030, SE = 0.004, p <0.001, LME) and naturalness (slope estimate = 0.012, SE = 0.003, p = 0.003, LME). CI recipients also found the first harmonic alone (F0) to be most pleasant (p = 0.003), with a positive linear relationship between harmonic reduction and pleasantness (slope estimate = 0.029, SE = 0.008, p <0.001, LME). CONCLUSION Harmonic series reduction increases music enjoyment in CI and NH individuals with or without CI simulation. Therefore, minimization of the harmonics may be a useful strategy for enhancing musical enjoyment among both NH and CI listeners.
Collapse
|
49
|
Kim SG, Lepsien J, Fritz TH, Mildner T, Mueller K. Dissonance encoding in human inferior colliculus covaries with individual differences in dislike of dissonant music. Sci Rep 2017; 7:5726. [PMID: 28720776 PMCID: PMC5516034 DOI: 10.1038/s41598-017-06105-2] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2017] [Accepted: 06/09/2017] [Indexed: 12/20/2022] Open
Abstract
Harmony is one of the most fundamental elements of music that evokes emotional response. The inferior colliculus (IC) has been known to detect poor agreement of harmonics of sound, that is, dissonance. Electrophysiological evidence has implicated a relationship between a sustained auditory response mainly from the brainstem and unpleasant emotion induced by dissonant harmony. Interestingly, an individual’s dislike of dissonant harmony of an individual correlated with a reduced sustained auditory response. In the current paper, we report novel evidence based on functional magnetic resonance imaging (fMRI) for such a relationship between individual variability in dislike of dissonance and the IC activation. Furthermore, for the first time, we show how dissonant harmony modulates functional connectivity of the IC and its association with behaviourally reported unpleasantness. The current findings support important contributions of low level auditory processing and corticofugal interaction in musical harmony preference.
Collapse
Affiliation(s)
- Seung-Goo Kim
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.
| | - Jöran Lepsien
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| | - Thomas Hans Fritz
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.,Institute for Psychoacoustics and Electronic Music, University of Ghent, Ghent, Belgium
| | - Toralf Mildner
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| | - Karsten Mueller
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| |
Collapse
|
50
|
Omote A, Jasmin K, Tierney A. Successful non-native speech perception is linked to frequency following response phase consistency. Cortex 2017; 93:146-154. [PMID: 28654816 PMCID: PMC5542039 DOI: 10.1016/j.cortex.2017.05.005] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2017] [Revised: 04/19/2017] [Accepted: 05/11/2017] [Indexed: 11/28/2022]
Abstract
Some people who attempt to learn a second language in adulthood meet with greater success than others. The causes driving these individual differences in second language learning skill continue to be debated. In particular, it remains controversial whether robust auditory perception can provide an advantage for non-native speech perception. Here, we tested English speech perception in native Japanese speakers through the use of frequency following responses, the evoked gamma band response, and behavioral measurements. Participants whose neural responses featured less timing jitter from trial to trial performed better on perception of English consonants than participants with more variable neural timing. Moreover, this neural metric predicted consonant perception to a greater extent than did age of arrival and length of residence in the UK, and neural jitter predicted independent variance in consonant perception after these demographic variables were accounted for. Thus, difficulties with auditory perception may be one source of problems learning second languages in adulthood.
Collapse
Affiliation(s)
- Akihiro Omote
- Department of Psychological Sciences, Birkbeck, University of London, London, United Kingdom
| | - Kyle Jasmin
- Department of Psychological Sciences, Birkbeck, University of London, London, United Kingdom
| | - Adam Tierney
- Department of Psychological Sciences, Birkbeck, University of London, London, United Kingdom.
| |
Collapse
|