1
|
Fairhurst MT, Tajadura-Jiménez A, Keller PE, Deroy O. You, me, and us: Maintaining self-other distinction enhances coordination, agency, and affect. iScience 2023; 26:108253. [PMID: 38025777 PMCID: PMC10679890 DOI: 10.1016/j.isci.2023.108253] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2023] [Revised: 08/08/2023] [Accepted: 10/16/2023] [Indexed: 12/01/2023] Open
Abstract
Coordinating our actions with others changes how we behave and feel. Here, we provide evidence that interacting with others rests on a balance between self-other integration and segregation. Using a group walking paradigm, participants were instructed to synchronize with a metronome while listening to the sounds of 8 virtual partners. By manipulating the similarity and synchronicity of the partners' steps to the participant's own, our novel auditory task disentangles the effects of synchrony and self-other similarity and examines their contribution to both collective and individual awareness. We measured temporal coordination (step timing regularity and synchrony with the metronome), gait patterns, and subjective reports about sense of self and group cohesion. The main findings show that coordination is best when participants hear distinct but synchronous virtual others, leading to greater subjective feelings of agency, strength, dominance, and happiness.
Collapse
Affiliation(s)
- Merle T. Fairhurst
- Centre for Tactile Internet with Human-in-the-Loop (CeTI), Faculty of Electrical and Computer Engineering, Technische Universität Dresden, Dresden, Germany
- Munich Centre for Neuroscience, Ludwig Maximilian University, Munich, Germany
| | - Ana Tajadura-Jiménez
- i_mBODY Lab, DEI Interactive Systems Group, Department of Computer Science and Engineering, Universidad Carlos III de Madrid, Spain
- UCL Interaction Centre (UCLIC), University College London, London, United Kingdom
| | - Peter E. Keller
- The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Sydney, Australia
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg, Aarhus, Denmark
| | - Ophelia Deroy
- Munich Centre for Neuroscience, Ludwig Maximilian University, Munich, Germany
- Faculty of Philosophy, Ludwig Maximilian University, Munich, Germany
- Institute of Philosophy, School of Advanced Study, University of London, London, United Kingdom
| |
Collapse
|
2
|
Anglada-Tort M, Harrison PMC, Lee H, Jacoby N. Large-scale iterated singing experiments reveal oral transmission mechanisms underlying music evolution. Curr Biol 2023; 33:1472-1486.e12. [PMID: 36958332 DOI: 10.1016/j.cub.2023.02.070] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/03/2022] [Revised: 12/24/2022] [Accepted: 02/23/2023] [Indexed: 03/25/2023]
Abstract
Speech and song have been transmitted orally for countless human generations, changing over time under the influence of biological, cognitive, and cultural pressures. Cross-cultural regularities and diversities in human song are thought to emerge from this transmission process, but testing how underlying mechanisms contribute to musical structures remains a key challenge. Here, we introduce an automatic online pipeline that streamlines large-scale cultural transmission experiments using a sophisticated and naturalistic modality: singing. We quantify the evolution of 3,424 melodies orally transmitted across 1,797 participants in the United States and India. This approach produces a high-resolution characterization of how oral transmission shapes melody, revealing the emergence of structures that are consistent with widespread musical features observed cross-culturally (small pitch sets, small pitch intervals, and arch-shaped melodic contours). We show how the emergence of these structures is constrained by individual biases in our participants-vocal constraints, working memory, and cultural exposure-which determine the size, shape, and complexity of evolving melodies. However, their ultimate effect on population-level structures depends on social dynamics taking place during cultural transmission. When participants recursively imitate their own productions (individual transmission), musical structures evolve slowly and heterogeneously, reflecting idiosyncratic musical biases. When participants instead imitate others' productions (social transmission), melodies rapidly shift toward homogeneous structures, reflecting shared structural biases that may underpin cross-cultural variation. These results provide the first quantitative characterization of the rich collection of biases that oral transmission imposes on music evolution, giving us a new understanding of how human song structures emerge via cultural transmission.
Collapse
Affiliation(s)
- Manuel Anglada-Tort
- Computational Auditory Perception Group, Max Planck Institute for Empirical Aesthetics, Grüneburgweg 14, Frankfurt am Main 60322, Germany; Faculty of Music, University of Oxford, St Aldate's, Oxford OX1 1DB, UK.
| | - Peter M C Harrison
- Computational Auditory Perception Group, Max Planck Institute for Empirical Aesthetics, Grüneburgweg 14, Frankfurt am Main 60322, Germany; Faculty of Music, University of Cambridge, 11 West Road, Cambridge CB3 9DP, UK
| | - Harin Lee
- Computational Auditory Perception Group, Max Planck Institute for Empirical Aesthetics, Grüneburgweg 14, Frankfurt am Main 60322, Germany; Max Planck Institute for Human Cognitive and Brain Sciences, Stephanstraße 1a, Leipzig 04103, Germany
| | - Nori Jacoby
- Computational Auditory Perception Group, Max Planck Institute for Empirical Aesthetics, Grüneburgweg 14, Frankfurt am Main 60322, Germany
| |
Collapse
|
3
|
Thomassen S, Hartung K, Einhäuser W, Bendixen A. Low-high-low or high-low-high? Pattern effects on sequential auditory scene analysis. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 152:2758. [PMID: 36456271 DOI: 10.1121/10.0015054] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/20/2022] [Accepted: 10/17/2022] [Indexed: 06/17/2023]
Abstract
Sequential auditory scene analysis (ASA) is often studied using sequences of two alternating tones, such as ABAB or ABA_, with "_" denoting a silent gap, and "A" and "B" sine tones differing in frequency (nominally low and high). Many studies implicitly assume that the specific arrangement (ABAB vs ABA_, as well as low-high-low vs high-low-high within ABA_) plays a negligible role, such that decisions about the tone pattern can be governed by other considerations. To explicitly test this assumption, a systematic comparison of different tone patterns for two-tone sequences was performed in three different experiments. Participants were asked to report whether they perceived the sequences as originating from a single sound source (integrated) or from two interleaved sources (segregated). Results indicate that core findings of sequential ASA, such as an effect of frequency separation on the proportion of integrated and segregated percepts, are similar across the different patterns during prolonged listening. However, at sequence onset, the integrated percept was more likely to be reported by the participants in ABA_low-high-low than in ABA_high-low-high sequences. This asymmetry is important for models of sequential ASA, since the formation of percepts at onset is an integral part of understanding how auditory interpretations build up.
Collapse
Affiliation(s)
- Sabine Thomassen
- Cognitive Systems Lab, Faculty of Natural Sciences, Chemnitz University of Technology, 09107 Chemnitz, Germany
| | - Kevin Hartung
- Cognitive Systems Lab, Faculty of Natural Sciences, Chemnitz University of Technology, 09107 Chemnitz, Germany
| | - Wolfgang Einhäuser
- Physics of Cognition Group, Faculty of Natural Sciences, Chemnitz University of Technology, 09107 Chemnitz, Germany
| | - Alexandra Bendixen
- Cognitive Systems Lab, Faculty of Natural Sciences, Chemnitz University of Technology, 09107 Chemnitz, Germany
| |
Collapse
|
4
|
Cho AY, Kidd G. Auditory motion as a cue for source segregation and selection in a "cocktail party" listening environment. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 152:1684. [PMID: 36182296 PMCID: PMC9489258 DOI: 10.1121/10.0013990] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/15/2023]
Abstract
Source motion was examined as a cue for segregating concurrent speech or noise sources. In two different headphone-based tasks-motion detection (MD) and speech-on-speech masking (SI)-one source among three was designated as the target only by imposing sinusoidal variation in azimuth during the stimulus presentation. For MD, the lstener was asked which of the three concurrent sources was in motion during the trial. For SI, the listener was asked to report the words spoken by the moving speech source. MD performance improved as the amplitude of the sinusoidal motion (i.e., displacement in azimuth) increased over the range of values tested (±5° to ±30°) for both modulated noise and speech targets, with better performance found for speech. SI performance also improved as the amplitude of target motion increased. Furthermore, SI performance improved as word position progressed throughout the sentence. Performance on the MD task was correlated with performance on SI task across individual subjects. For the SI conditions tested here, these findings are consistent with the proposition that listeners first detect the moving target source, then focus attention on the target location as the target sentence unfolds.
Collapse
Affiliation(s)
- Adrian Y Cho
- Speech and Hearing Bioscience and Technology Program, Harvard University, Cambridge, Massachusetts 02138, USA
| | - Gerald Kidd
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| |
Collapse
|
5
|
Brace KM, Sussman ES. The role of attention and explicit knowledge in perceiving bistable auditory input. Psychophysiology 2021; 58:e13875. [PMID: 34110020 DOI: 10.1111/psyp.13875] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2020] [Revised: 04/19/2021] [Accepted: 05/14/2021] [Indexed: 11/26/2022]
Abstract
The auditory system frequently encounters ambiguous sound input that can be perceived in multiple ways. The current study investigated the role of explicit knowledge in modulating how sounds are represented in auditory memory for a bistable sound sequence that could be perceived equally as integrated or segregated. We hypothesized that the dominant percept of the bistable sequence would suppress representation of the alternative perceptual organization as a function of how much top-down knowledge the listener had about the structure of the sequence. Performance measures and event-related brain potentials were compared when participants had explicit knowledge about one perceptual organization in the first half of the experiment to when they had explicit knowledge of both in the second half. We hypothesized that knowledge would modify the brain response to the alternative percept of the bistable sequence. However, that did not occur. When participants were performing one task, with no explicit knowledge of the bistable structure of the sequence, both integrated and segregated percepts were represented in auditory working memory. This demonstrates that explicit knowledge about the sounds is not a necessary factor for deriving and maintaining representations of multiple sound organizations within a complex sound environment. Passive attention operates in parallel with active or selective attention to maintain consistent representations of the environment, representations that may or may not be useful for task performance. It suggests a highly adaptive system useful in everyday listening situations where the listener has no prior knowledge about how the sound environment is structured.
Collapse
Affiliation(s)
- Kelin M Brace
- Department of Neuroscience, Albert Einstein College of Medicine, Bronx, NY, USA
| | - Elyse S Sussman
- Department of Neuroscience, Albert Einstein College of Medicine, Bronx, NY, USA.,Department of Otorhinolaryngology-Head & Neck Surgery, Albert Einstein College of Medicine, Bronx, NY, USA
| |
Collapse
|
6
|
Abstract
OBJECTIVES Children with hearing loss (HL), in spite of early cochlear implantation, often struggle considerably with language acquisition. Previous research has shown a benefit of rhythmic training on linguistic skills in children with HL, suggesting that improving rhythmic capacities could help attenuating language difficulties. However, little is known about general rhythmic skills of children with HL and how they relate to speech perception. The aim of this study is twofold: (1) to assess the abilities of children with HL in different rhythmic sensorimotor synchronization tasks compared to a normal-hearing control group and (2) to investigate a possible relation between sensorimotor synchronization abilities and speech perception abilities in children with HL. DESIGN A battery of sensorimotor synchronization tests with stimuli of varying acoustic and temporal complexity was used: a metronome, different musical excerpts, and complex rhythmic patterns. Synchronization abilities were assessed in 32 children (aged from 5 to 10 years) with a severe to profound HL mainly fitted with one or two cochlear implants (n = 28) or with hearing aids (n = 4). Working memory and sentence repetition abilities were also assessed. Performance was compared to an age-matched control group of 24 children with normal hearing. The comparison took into account variability in working memory capacities. For children with HL only, we computed linear regressions on speech, sensorimotor synchronization, and working memory abilities, including device-related variables such as onset of device use, type of device, and duration of use. RESULTS Compared to the normal-hearing group, children with HL performed poorly in all sensorimotor synchronization tasks, but the effect size was greater for complex as compared to simple stimuli. Group differences in working memory did not explain this result. Linear regression analysis revealed that working memory, synchronization to complex rhythms performances, age, and duration of device use predicted the number of correct syllables produced in a sentence repetition task. CONCLUSION Despite early cochlear implantation or hearing aid use, hearing impairment affects the quality of temporal processing of acoustic stimuli in congenitally deaf children. This deficit seems to be more severe with stimuli of increasing rhythmic complexity highlighting a difficulty in structuring sounds according to a temporal hierarchy.
Collapse
|
7
|
Greenlaw KM, Puschmann S, Coffey EBJ. Decoding of Envelope vs. Fundamental Frequency During Complex Auditory Stream Segregation. NEUROBIOLOGY OF LANGUAGE (CAMBRIDGE, MASS.) 2020; 1:268-287. [PMID: 37215227 PMCID: PMC10158587 DOI: 10.1162/nol_a_00013] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/26/2019] [Accepted: 04/25/2020] [Indexed: 05/24/2023]
Abstract
Hearing-in-noise perception is a challenging task that is critical to human function, but how the brain accomplishes it is not well understood. A candidate mechanism proposes that the neural representation of an attended auditory stream is enhanced relative to background sound via a combination of bottom-up and top-down mechanisms. To date, few studies have compared neural representation and its task-related enhancement across frequency bands that carry different auditory information, such as a sound's amplitude envelope (i.e., syllabic rate or rhythm; 1-9 Hz), and the fundamental frequency of periodic stimuli (i.e., pitch; >40 Hz). Furthermore, hearing-in-noise in the real world is frequently both messier and richer than the majority of tasks used in its study. In the present study, we use continuous sound excerpts that simultaneously offer predictive, visual, and spatial cues to help listeners separate the target from four acoustically similar simultaneously presented sound streams. We show that while both lower and higher frequency information about the entire sound stream is represented in the brain's response, the to-be-attended sound stream is strongly enhanced only in the slower, lower frequency sound representations. These results are consistent with the hypothesis that attended sound representations are strengthened progressively at higher level, later processing stages, and that the interaction of multiple brain systems can aid in this process. Our findings contribute to our understanding of auditory stream separation in difficult, naturalistic listening conditions and demonstrate that pitch and envelope information can be decoded from single-channel EEG data.
Collapse
Affiliation(s)
- Keelin M. Greenlaw
- Department of Psychology, Concordia University, Montreal, QC, Canada
- International Laboratory for Brain, Music and Sound Research (BRAMS)
- The Centre for Research on Brain, Language and Music (CRBLM)
| | | | | |
Collapse
|
8
|
Perceptual fusion of musical notes by native Amazonians suggests universal representations of musical intervals. Nat Commun 2020; 11:2786. [PMID: 32493923 PMCID: PMC7270137 DOI: 10.1038/s41467-020-16448-6] [Citation(s) in RCA: 24] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2019] [Accepted: 04/23/2020] [Indexed: 01/31/2023] Open
Abstract
Music perception is plausibly constrained by universal perceptual mechanisms adapted to natural sounds. Such constraints could arise from our dependence on harmonic frequency spectra for segregating concurrent sounds, but evidence has been circumstantial. We measured the extent to which concurrent musical notes are misperceived as a single sound, testing Westerners as well as native Amazonians with limited exposure to Western music. Both groups were more likely to mistake note combinations related by simple integer ratios as single sounds (‘fusion’). Thus, even with little exposure to Western harmony, acoustic constraints on sound segregation appear to induce perceptual structure on note combinations. However, fusion did not predict aesthetic judgments of intervals in Westerners, or in Amazonians, who were indifferent to consonance/dissonance. The results suggest universal perceptual mechanisms that could help explain cross-cultural regularities in musical systems, but indicate that these mechanisms interact with culture-specific influences to produce musical phenomena such as consonance. Music varies across cultures, but some features are widespread, consistent with biological constraints. Here, the authors report that both Western and native Amazonian listeners perceptually fuse concurrent notes related by simple-integer ratios, suggestive of one such biological constraint.
Collapse
|
9
|
Einhäuser W, da Silva LFO, Bendixen A. Intraindividual Consistency Between Auditory and Visual Multistability. Perception 2020; 49:119-138. [PMID: 31888418 DOI: 10.1177/0301006619896282] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
We compared perceptual multistability across modalities, using a visual plaid pattern (composed of two transparently overlaid drifting gratings) and auditory streaming (elicited by a repeating “ABA_” tone sequence). Both stimuli can be perceived as integrated (one plaid pattern, one stream comprising “A” and “B” tones) or segregated (two individual gratings, two tone streams). In the segregated case, either stream or grating can be perceived in the foreground. We queried auditory and visual perception with these three response options. We found that perceptual dominance of the integrated states was correlated across modalities: Participants who were more likely to perceive the plaid were also more likely to perceive the integrated auditory stream. When presenting both stimuli simultaneously and querying the auditory percept, eye-movement data showed that perceiving auditory and visual integration is related on a moment-by-moment basis. This suggests that in part common mechanisms underlie multistability in visual and auditory perception.
Collapse
Affiliation(s)
- Wolfgang Einhäuser
- Physics of Cognition Group, Institute of Physics, Chemnitz University of Technology, Germany
| | - Lucas F O da Silva
- Physics of Cognition Group, Institute of Physics, Chemnitz University of Technology, Germany; Cognitive Systems Lab, Institute of Physics, Chemnitz University of Technology, Germany
| | - Alexandra Bendixen
- Cognitive Systems Lab, Institute of Physics, Chemnitz University of Technology, Germany
| |
Collapse
|
10
|
Zoefel B, Allard I, Anil M, Davis MH. Perception of Rhythmic Speech Is Modulated by Focal Bilateral Transcranial Alternating Current Stimulation. J Cogn Neurosci 2019; 32:226-240. [PMID: 31659922 DOI: 10.1162/jocn_a_01490] [Citation(s) in RCA: 16] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/19/2022]
Abstract
Several recent studies have used transcranial alternating current stimulation (tACS) to demonstrate a causal role of neural oscillatory activity in speech processing. In particular, it has been shown that the ability to understand speech in a multi-speaker scenario or background noise depends on the timing of speech presentation relative to simultaneously applied tACS. However, it is possible that tACS did not change actual speech perception but rather auditory stream segregation. In this study, we tested whether the phase relation between tACS and the rhythm of degraded words, presented in silence, modulates word report accuracy. We found strong evidence for a tACS-induced modulation of speech perception, but only if the stimulation was applied bilaterally using ring electrodes (not for unilateral left hemisphere stimulation with square electrodes). These results were only obtained when data were analyzed using a statistical approach that was identified as optimal in a previous simulation study. The effect was driven by a phasic disruption of word report scores. Our results suggest a causal role of neural entrainment for speech perception and emphasize the importance of optimizing stimulation protocols and statistical approaches for brain stimulation research.
Collapse
|
11
|
Coffey EBJ, Arseneau-Bruneau I, Zhang X, Zatorre RJ. The Music-In-Noise Task (MINT): A Tool for Dissecting Complex Auditory Perception. Front Neurosci 2019; 13:199. [PMID: 30930734 PMCID: PMC6427094 DOI: 10.3389/fnins.2019.00199] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2018] [Accepted: 02/20/2019] [Indexed: 11/30/2022] Open
Abstract
The ability to segregate target sounds in noisy backgrounds is relevant both to neuroscience and to clinical applications. Recent research suggests that hearing-in-noise (HIN) problems are solved using combinations of sub-skills that are applied according to task demand and information availability. While evidence is accumulating for a musician advantage in HIN, the exact nature of the reported training effect is not fully understood. Existing HIN tests focus on tasks requiring understanding of speech in the presence of competing sound. Because visual, spatial and predictive cues are not systematically considered in these tasks, few tools exist to investigate the most relevant components of cognitive processes involved in stream segregation. We present the Music-In-Noise Task (MINT) as a flexible tool to expand HIN measures beyond speech perception, and for addressing research questions pertaining to the relative contributions of HIN sub-skills, inter-individual differences in their use, and their neural correlates. The MINT uses a match-mismatch trial design: in four conditions (Baseline, Rhythm, Spatial, and Visual) subjects first hear a short instrumental musical excerpt embedded in an informational masker of "multi-music" noise, followed by either a matching or scrambled repetition of the target musical excerpt presented in silence; the four conditions differ according to the presence or absence of additional cues. In a fifth condition (Prediction), subjects hear the excerpt in silence as a target first, which helps to anticipate incoming information when the target is embedded in masking sound. Data from samples of young adults show that the MINT has good reliability and internal consistency, and demonstrate selective benefits of musicianship in the Prediction, Rhythm, and Visual subtasks. We also report a performance benefit of multilingualism that is separable from that of musicianship. Average MINT scores were correlated with scores on a sentence-in-noise perception task, but only accounted for a relatively small percentage of the variance, indicating that the MINT is sensitive to additional factors and can provide a complement and extension of speech-based tests for studying stream segregation. A customizable version of the MINT is made available for use and extension by the scientific community.
Collapse
Affiliation(s)
- Emily B. J. Coffey
- Department of Psychology, Concordia University, Montreal, QC, Canada
- Laboratory for Brain, Music and Sound Research (BRAMS), Montreal, QC, Canada
- Centre for Research on Brain, Language and Music (CRBLM), Montreal, QC, Canada
- Centre for Interdisciplinary Research in Music Media and Technology (CIRMMT), Montreal, QC, Canada
| | - Isabelle Arseneau-Bruneau
- Laboratory for Brain, Music and Sound Research (BRAMS), Montreal, QC, Canada
- Centre for Research on Brain, Language and Music (CRBLM), Montreal, QC, Canada
- Centre for Interdisciplinary Research in Music Media and Technology (CIRMMT), Montreal, QC, Canada
- Montreal Neurological Institute, McGill University, Montreal, QC, Canada
| | - Xiaochen Zhang
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China
| | - Robert J. Zatorre
- Laboratory for Brain, Music and Sound Research (BRAMS), Montreal, QC, Canada
- Centre for Research on Brain, Language and Music (CRBLM), Montreal, QC, Canada
- Centre for Interdisciplinary Research in Music Media and Technology (CIRMMT), Montreal, QC, Canada
- Montreal Neurological Institute, McGill University, Montreal, QC, Canada
| |
Collapse
|
12
|
Jeong E, Ryu H, Shin JH, Kwon GH, Jo G, Lee JY. High Oxygen Exchange to Music Indicates Auditory Distractibility in Acquired Brain Injury: An fNIRS Study with a Vector-Based Phase Analysis. Sci Rep 2018; 8:16737. [PMID: 30425287 PMCID: PMC6233191 DOI: 10.1038/s41598-018-35172-2] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2018] [Accepted: 10/31/2018] [Indexed: 01/30/2023] Open
Abstract
Attention deficits due to auditory distractibility are pervasive among patients with acquired brain injury (ABI). It remains unclear, however, whether attention deficits following ABI specific to auditory modality are associated with altered haemodynamic responses. Here, we examined cerebral haemodynamic changes using functional near-infrared spectroscopy combined with a topological vector-based analysis method. A total of thirty-seven participants (22 healthy adults, 15 patients with ABI) performed a melodic contour identification task (CIT) that simulates auditory distractibility. Findings demonstrated that the melodic CIT was able to detect auditory distractibility in patients with ABI. The rate-corrected score showed that the ABI group performed significantly worse than the non-ABI group in both CIT1 (target contour identification against environmental sounds) and CIT2 (target contour identification against target-like distraction). Phase-associated response intensity during the CITs was greater in the ABI group than in the non-ABI group. Moreover, there existed a significant interaction effect in the left dorsolateral prefrontal cortex (DLPFC) during CIT1 and CIT2. These findings indicated that stronger hemodynamic responses involving oxygen exchange in the left DLPFC can serve as a biomarker for evaluating and monitoring auditory distractibility, which could potentially lead to the discovery of the underlying mechanism that causes auditory attention deficits in patients with ABI.
Collapse
Affiliation(s)
- Eunju Jeong
- Department of Arts and Technology, Hanyang University, Seoul, 04763, Republic of Korea.
- Division of Industrial Information Studies, Hanyang University, Seoul, 04763, Republic of Korea.
| | - Hokyoung Ryu
- Department of Arts and Technology, Hanyang University, Seoul, 04763, Republic of Korea
- Graduate School of Technology and Innovation Management, Hanyang University, Seoul, 04763, Republic of Korea
| | - Joon-Ho Shin
- Department of Neurorehabilitation, National Rehabilitation Center, Ministry of Health and Welfare, Seoul, 01022, Republic of Korea
| | - Gyu Hyun Kwon
- Department of Arts and Technology, Hanyang University, Seoul, 04763, Republic of Korea
- Graduate School of Technology and Innovation Management, Hanyang University, Seoul, 04763, Republic of Korea
| | - Geonsang Jo
- Department of Arts and Technology, Hanyang University, Seoul, 04763, Republic of Korea
| | - Ji-Yeong Lee
- Department of Neurorehabilitation, National Rehabilitation Center, Ministry of Health and Welfare, Seoul, 01022, Republic of Korea
| |
Collapse
|
13
|
Gandras K, Grimm S, Bendixen A. Electrophysiological Correlates of Speaker Segregation and Foreground-Background Selection in Ambiguous Listening Situations. Neuroscience 2018; 389:19-29. [PMID: 28735101 DOI: 10.1016/j.neuroscience.2017.07.021] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2017] [Revised: 07/10/2017] [Accepted: 07/10/2017] [Indexed: 11/15/2022]
Abstract
In everyday listening environments, a main task for our auditory system is to follow one out of multiple speakers talking simultaneously. The present study was designed to find electrophysiological indicators of two central processes involved - segregating the speech mixture into distinct speech sequences corresponding to the two speakers, and then attending to one of the speech sequences. We generated multistable speech stimuli that were set up to create ambiguity as to whether only one or two speakers are talking. Thereby we were able to investigate three perceptual alternatives (no segregation, segregated - speaker A in the foreground, segregated - speaker B in the foreground) without any confounding stimulus changes. Participants listened to a continuously repeating sequence of syllables, which were uttered alternately by two human speakers, and indicated whether they perceived the sequence as an inseparable mixture or as originating from two separate speakers. In the latter case, they distinguished which speaker was in their attentional foreground. Our data show a long-lasting event-related potential (ERP) modulation starting at 130ms after stimulus onset, which can be explained by the perceptual organization of the two speech sequences into attended foreground and ignored background streams. Our paradigm extends previous work with pure-tone sequences toward speech stimuli and adds the possibility to obtain neural correlates of the difficulty to segregate a speech mixture into distinct streams.
Collapse
Affiliation(s)
- Katharina Gandras
- Department of Psychology, Cluster of Excellence "Hearing4all", European Medical School, Carl von Ossietzky University of Oldenburg, D-26111 Oldenburg, Germany.
| | - Sabine Grimm
- Department of Physics, School of Natural Sciences, Chemnitz University of Technology, D-09126 Chemnitz, Germany.
| | - Alexandra Bendixen
- Department of Psychology, Cluster of Excellence "Hearing4all", European Medical School, Carl von Ossietzky University of Oldenburg, D-26111 Oldenburg, Germany; Department of Physics, School of Natural Sciences, Chemnitz University of Technology, D-09126 Chemnitz, Germany.
| |
Collapse
|
14
|
Wollman I, Morillon B. Organizational principles of multidimensional predictions in human auditory attention. Sci Rep 2018; 8:13466. [PMID: 30194376 PMCID: PMC6128843 DOI: 10.1038/s41598-018-31878-5] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2018] [Accepted: 08/17/2018] [Indexed: 11/09/2022] Open
Abstract
Anticipating the future rests upon our ability to exploit contextual cues and to formulate valid internal models or predictions. It is currently unknown how multiple predictions combine to bias perceptual information processing, and in particular whether this is determined by physiological constraints, behavioral relevance (task demands), or past knowledge (perceptual expertise). In a series of behavioral auditory experiments involving musical experts and non-musicians, we investigated the respective and combined contribution of temporal and spectral predictions in multiple detection tasks. We show that temporal and spectral predictions alone systematically increase perceptual sensitivity, independently of task demands or expertise. When combined, however, spectral predictions benefit more to non-musicians and dominate over temporal ones, and the extent of the spectrotemporal synergistic interaction depends on task demands. This suggests that the hierarchy of dominance primarily reflects the tonotopic organization of the auditory system and that expertise or attention only have a secondary modulatory influence.
Collapse
Affiliation(s)
- Indiana Wollman
- Montreal Neurological Institute, McGill University, Montreal, Canada
- CIRMMT, Schulich School of Music, McGill University, Montreal, Canada
| | - Benjamin Morillon
- Montreal Neurological Institute, McGill University, Montreal, Canada.
- Aix Marseille Univ, Inserm, INS, Inst Neurosci Syst, Marseille, France.
| |
Collapse
|
15
|
Clark CN, Golden HL, McCallion O, Nicholas JM, Cohen MH, Slattery CF, Paterson RW, Fletcher PD, Mummery CJ, Rohrer JD, Crutch SJ, Warren JD. Music models aberrant rule decoding and reward valuation in dementia. Soc Cogn Affect Neurosci 2018; 13:192-202. [PMID: 29186630 PMCID: PMC5827340 DOI: 10.1093/scan/nsx140] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2017] [Revised: 11/06/2017] [Accepted: 11/19/2017] [Indexed: 01/03/2023] Open
Abstract
Aberrant rule- and reward-based processes underpin abnormalities of socio-emotional behaviour in major dementias. However, these processes remain poorly characterized. Here we used music to probe rule decoding and reward valuation in patients with frontotemporal dementia (FTD) syndromes and Alzheimer's disease (AD) relative to healthy age-matched individuals. We created short melodies that were either harmonically resolved ('finished') or unresolved ('unfinished'); the task was to classify each melody as finished or unfinished (rule processing) and rate its subjective pleasantness (reward valuation). Results were adjusted for elementary pitch and executive processing; neuroanatomical correlates were assessed using voxel-based morphometry. Relative to healthy older controls, patients with behavioural variant FTD showed impairments of both musical rule decoding and reward valuation, while patients with semantic dementia showed impaired reward valuation but intact rule decoding, patients with AD showed impaired rule decoding but intact reward valuation and patients with progressive non-fluent aphasia performed comparably to healthy controls. Grey matter associations with task performance were identified in anterior temporal, medial and lateral orbitofrontal cortices, previously implicated in computing diverse biological and non-biological rules and rewards. The processing of musical rules and reward distils cognitive and neuroanatomical mechanisms relevant to complex socio-emotional dysfunction in major dementias.
Collapse
Affiliation(s)
- Camilla N Clark
- Dementia Research Centre, UCL Institute of Neurology, University College London, London, UK
| | - Hannah L Golden
- Dementia Research Centre, UCL Institute of Neurology, University College London, London, UK
| | - Oliver McCallion
- Oxford University Clinical Academic Graduate School, University of Oxford, Oxford, UK
| | - Jennifer M Nicholas
- Dementia Research Centre, UCL Institute of Neurology, University College London, London, UK
- London School of Hygiene and Tropical Medicine, University of London, London, UK
| | - Miriam H Cohen
- Dementia Research Centre, UCL Institute of Neurology, University College London, London, UK
| | - Catherine F Slattery
- Dementia Research Centre, UCL Institute of Neurology, University College London, London, UK
| | - Ross W Paterson
- Dementia Research Centre, UCL Institute of Neurology, University College London, London, UK
| | - Phillip D Fletcher
- Dementia Research Centre, UCL Institute of Neurology, University College London, London, UK
| | - Catherine J Mummery
- Dementia Research Centre, UCL Institute of Neurology, University College London, London, UK
| | - Jonathan D Rohrer
- Dementia Research Centre, UCL Institute of Neurology, University College London, London, UK
| | - Sebastian J Crutch
- Dementia Research Centre, UCL Institute of Neurology, University College London, London, UK
| | - Jason D Warren
- Dementia Research Centre, UCL Institute of Neurology, University College London, London, UK
| |
Collapse
|
16
|
Coffey EBJ, Chepesiuk AMP, Herholz SC, Baillet S, Zatorre RJ. Neural Correlates of Early Sound Encoding and their Relationship to Speech-in-Noise Perception. Front Neurosci 2017; 11:479. [PMID: 28890684 PMCID: PMC5575455 DOI: 10.3389/fnins.2017.00479] [Citation(s) in RCA: 47] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2017] [Accepted: 08/11/2017] [Indexed: 01/05/2023] Open
Abstract
Speech-in-noise (SIN) perception is a complex cognitive skill that affects social, vocational, and educational activities. Poor SIN ability particularly affects young and elderly populations, yet varies considerably even among healthy young adults with normal hearing. Although SIN skills are known to be influenced by top-down processes that can selectively enhance lower-level sound representations, the complementary role of feed-forward mechanisms and their relationship to musical training is poorly understood. Using a paradigm that minimizes the main top-down factors that have been implicated in SIN performance such as working memory, we aimed to better understand how robust encoding of periodicity in the auditory system (as measured by the frequency-following response) contributes to SIN perception. Using magnetoencephalograpy, we found that the strength of encoding at the fundamental frequency in the brainstem, thalamus, and cortex is correlated with SIN accuracy. The amplitude of the slower cortical P2 wave was previously also shown to be related to SIN accuracy and FFR strength; we use MEG source localization to show that the P2 wave originates in a temporal region anterior to that of the cortical FFR. We also confirm that the observed enhancements were related to the extent and timing of musicianship. These results are consistent with the hypothesis that basic feed-forward sound encoding affects SIN perception by providing better information to later processing stages, and that modifying this process may be one mechanism through which musical training might enhance the auditory networks that subserve both musical and language functions.
Collapse
Affiliation(s)
- Emily B J Coffey
- Department of Neurology and Neurosurgery, Montreal Neurological Institute, McGill UniversityMontréal, QC, Canada.,Laboratory for Brain, Music and Sound ResearchMontréal, QC, Canada.,Centre for Research on Brain, Language and MusicMontréal, QC, Canada
| | - Alexander M P Chepesiuk
- Department of Neurology and Neurosurgery, Montreal Neurological Institute, McGill UniversityMontréal, QC, Canada
| | - Sibylle C Herholz
- Department of Neurology and Neurosurgery, Montreal Neurological Institute, McGill UniversityMontréal, QC, Canada.,Laboratory for Brain, Music and Sound ResearchMontréal, QC, Canada.,Centre for Research on Brain, Language and MusicMontréal, QC, Canada.,German Center for Neurodegenerative DiseasesBonn, Germany
| | - Sylvain Baillet
- Centre for Research on Brain, Language and MusicMontréal, QC, Canada.,McConnell Brain Imaging Centre, Montreal Neurological Institute, McGill UniversityMontréal, QC, Canada
| | - Robert J Zatorre
- Department of Neurology and Neurosurgery, Montreal Neurological Institute, McGill UniversityMontréal, QC, Canada.,Laboratory for Brain, Music and Sound ResearchMontréal, QC, Canada.,Centre for Research on Brain, Language and MusicMontréal, QC, Canada
| |
Collapse
|
17
|
Speech-in-noise perception in musicians: A review. Hear Res 2017; 352:49-69. [PMID: 28213134 DOI: 10.1016/j.heares.2017.02.006] [Citation(s) in RCA: 83] [Impact Index Per Article: 11.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/26/2016] [Revised: 02/01/2017] [Accepted: 02/05/2017] [Indexed: 11/23/2022]
Abstract
The ability to understand speech in the presence of competing sound sources is an important neuroscience question in terms of how the nervous system solves this computational problem. It is also a critical clinical problem that disproportionally affects the elderly, children with language-related learning disorders, and those with hearing loss. Recent evidence that musicians have an advantage on this multifaceted skill has led to the suggestion that musical training might be used to improve or delay the decline of speech-in-noise (SIN) function. However, enhancements have not been universally reported, nor have the relative contributions of different bottom-up versus top-down processes, and their relation to preexisting factors been disentangled. This information that would be helpful to establish whether there is a real effect of experience, what exactly is its nature, and how future training-based interventions might target the most relevant components of cognitive processes. These questions are complicated by important differences in study design and uneven coverage of neuroimaging modality. In this review, we aim to systematize recent results from studies that have specifically looked at musician-related differences in SIN by their study design properties, to summarize the findings, and to identify knowledge gaps for future work.
Collapse
|
18
|
Boroujeni FM, Heidari F, Rouzbahani M, Kamali M. Comparison of auditory stream segregation in sighted and early blind individuals. Neurosci Lett 2017; 638:218-221. [PMID: 27986498 DOI: 10.1016/j.neulet.2016.12.022] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2016] [Revised: 12/04/2016] [Accepted: 12/12/2016] [Indexed: 12/01/2022]
Abstract
An important characteristic of the auditory system is the capacity to analyze complex sounds and make decisions on the source of the constituent parts of these sounds. Blind individuals compensate for the lack of visual information by an increase input from other sensory modalities, including increased auditory information. The purpose of the current study was to compare the fission boundary (FB) threshold of sighted and early blind individuals through spectral aspects using a psychoacoustic auditory stream segregation (ASS) test. This study was conducted on 16 sighted and 16 early blind adult individuals. The applied stimuli were presented sequentially as the pure tones A and B and as a triplet ABA-ABA pattern at the intensity of 40dBSL. The A tone frequency was selected as the basis at values of 500, 1000, and 2000Hz. The B tone was presented with the difference of a 4-100% above the basis tone frequency. Blind individuals had significantly lower FB thresholds than sighted people. FB was independent of the frequency of the tone A when expressed as the difference in the number of equivalent rectangular bandwidths (ERBs). Early blindness may increase perceptual separation of the acoustic stimuli to form accurate representations of the world.
Collapse
Affiliation(s)
- Fatemeh Moghadasi Boroujeni
- Department of Audiology, School of Rehabilitation Sciences, Iran University of Medical Sciences, Tehran, Iran; Department of Audiology, School of Rehabilitation, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Fatemeh Heidari
- Department of Audiology, School of Rehabilitation Sciences, Iran University of Medical Sciences, Tehran, Iran
| | - Masoumeh Rouzbahani
- Department of Audiology, School of Rehabilitation Sciences, Iran University of Medical Sciences, Tehran, Iran.
| | - Mohammad Kamali
- Department of Basic Sciences in Rehabilitation, School of Rehabilitation Sciences, Iran University of Medical Sciences, Tehran, Iran
| |
Collapse
|
19
|
Thomassen S, Bendixen A. Subjective perceptual organization of a complex auditory scene. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 141:265. [PMID: 28147594 DOI: 10.1121/1.4973806] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Empirical research on the sequential decomposition of an auditory scene primarily relies on interleaved sound mixtures of only two tone sequences (e.g., ABAB…). This oversimplifies the sound decomposition problem by limiting the number of putative perceptual organizations. The current study used a sound mixture composed of three different tones (ABCABC…) that could be perceptually organized in many different ways. Participants listened to these sequences and reported their subjective perception by continuously choosing one out of 12 visually presented perceptual organization alternatives. Different levels of frequency and spatial separation were implemented to check whether participants' perceptual reports would be systematic and plausible. As hypothesized, while perception switched back and forth in each condition between various perceptual alternatives (multistability), spatial as well as frequency separation generally raised the proportion of segregated and reduced the proportion of integrated alternatives. During segregated percepts, in contrast to the hypothesis, many participants had a tendency to perceive two streams in the foreground, rather than reporting alternatives with a clear foreground-background differentiation. Finally, participants perceived the organization with intermediate feature values (e.g., middle tones of the pattern) segregated in the foreground slightly less often than similar alternatives with outer feature values (e.g., higher tones).
Collapse
Affiliation(s)
- Sabine Thomassen
- Auditory Psychophysiology Lab, Department of Psychology, Carl von Ossietzky University of Oldenburg, Ammerländer Heerstrasse 114-118, D-26129 Oldenburg, Germany
| | - Alexandra Bendixen
- Auditory Psychophysiology Lab, Department of Psychology, Carl von Ossietzky University of Oldenburg, Ammerländer Heerstrasse 114-118, D-26129 Oldenburg, Germany
| |
Collapse
|
20
|
Brosowsky NP, Mondor TA. Multistable perception of ambiguous melodies and the role of musical expertise. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 140:866. [PMID: 27586718 DOI: 10.1121/1.4960450] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Whereas visual demonstrations of multistability are ubiquitous, there are few auditory examples. The purpose of the current study was to determine whether simultaneously presented melodies, such as underlie the scale illusion [Deutsch (1975). J. Acoust. Soc. Am. 57(5), 1156-1160], can elicit multiple mutually exclusive percepts, and whether reported perceptions are mediated by musical expertise. Participants listened to target melodies and reported whether the target was embedded in subsequent test melodies. Target sequences were created such that they would only be heard if the listener interpreted the test melody according to various perceptual cues. Critically, and in contrast with previous examinations of the scale illusion, an objective measure of target detection was obtained by including target-absent test melodies. As a result, listeners could reliably identify target sequences from different perceptual organizations when presented with the same test melody on different trials. This result demonstrates an ability to alternate between mutually exclusive percepts of an unchanged stimulus. However, only perceptual organizations consistent with frequency and spatial cues were available and musical expertise did mediate target detection, limiting the organizations available to non-musicians. The current study provides the first known demonstration of auditory multistability using simultaneously presented melodies and provides a unique experimental method for measuring auditory perceptual competition.
Collapse
Affiliation(s)
- Nicholaus P Brosowsky
- Department of Psychology, The Graduate Center of the City University of New York, 365 5th Avenue, New York, New York 10016, USA
| | - Todd A Mondor
- University of Manitoba, Winnipeg, Manitoba, R3T 2N2, Canada
| |
Collapse
|
21
|
Nemoto I, Yuhara R. Ambiguity involving two illusory melodies induced by a simple configuration of tones. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2016; 2015:6692-5. [PMID: 26737828 DOI: 10.1109/embc.2015.7319928] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Auditory scene analysis is essential in daily life to extract necessary information from complex acoustic environment and also from intricate development of music compositions. Auditory illusions and ambiguity are important factors in auditory scene analysis and have been studied extensively. We here report a novel form of ambiguity involving two illusory melodies implied by a very simple stimulus consisting of two sustained tones of different frequencies and an intermittently repeated tone of a frequency between the sustained tones. The measured time elapsed before a first perception change showed that illusion, ambiguity and disambiguation actually took place. We anticipate that the proposed illusion and ambiguity will provide a well-controlled approach for behavioral and neurophysiological studies of music recognition because of the simplicity of stimulus.
Collapse
|
22
|
Clark CN, Downey LE, Warren JD. Brain disorders and the biological role of music. Soc Cogn Affect Neurosci 2015; 10:444-52. [PMID: 24847111 PMCID: PMC4350491 DOI: 10.1093/scan/nsu079] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2013] [Revised: 03/07/2014] [Accepted: 05/14/2014] [Indexed: 12/16/2022] Open
Abstract
Despite its evident universality and high social value, the ultimate biological role of music and its connection to brain disorders remain poorly understood. Recent findings from basic neuroscience have shed fresh light on these old problems. New insights provided by clinical neuroscience concerning the effects of brain disorders promise to be particularly valuable in uncovering the underlying cognitive and neural architecture of music and for assessing candidate accounts of the biological role of music. Here we advance a new model of the biological role of music in human evolution and the link to brain disorders, drawing on diverse lines of evidence derived from comparative ethology, cognitive neuropsychology and neuroimaging studies in the normal and the disordered brain. We propose that music evolved from the call signals of our hominid ancestors as a means mentally to rehearse and predict potentially costly, affectively laden social routines in surrogate, coded, low-cost form: essentially, a mechanism for transforming emotional mental states efficiently and adaptively into social signals. This biological role of music has its legacy today in the disordered processing of music and mental states that characterizes certain developmental and acquired clinical syndromes of brain network disintegration.
Collapse
Affiliation(s)
- Camilla N Clark
- Dementia Research Centre, UCL Institute of Neurology, University College London, London WC1N 3BG, UK
| | - Laura E Downey
- Dementia Research Centre, UCL Institute of Neurology, University College London, London WC1N 3BG, UK
| | - Jason D Warren
- Dementia Research Centre, UCL Institute of Neurology, University College London, London WC1N 3BG, UK
| |
Collapse
|
23
|
Steele SA, Tranchina D, Rinzel J. An alternating renewal process describes the buildup of perceptual segregation. Front Comput Neurosci 2015; 8:166. [PMID: 25620927 PMCID: PMC4286718 DOI: 10.3389/fncom.2014.00166] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2014] [Accepted: 12/02/2014] [Indexed: 12/05/2022] Open
Abstract
For some ambiguous scenes perceptual conflict arises between integration and segregation. Initially, all stimulus features seem integrated. Then abruptly, perhaps after a few seconds, a segregated percept emerges. For example, segregation of acoustic features into streams may require several seconds. In behavioral experiments, when a subject's reports of stream segregation are averaged over repeated trials, one obtains a buildup function, a smooth time course for segregation probability. The buildup function has been said to reflect an underlying mechanism of evidence accumulation or adaptation. During long duration stimuli perception may alternate between integration and segregation. We present a statistical model based on an alternating renewal process (ARP) that generates buildup functions without an accumulative process. In our model, perception alternates during a trial between different groupings, as in perceptual bistability, with random and independent dominance durations sampled from different percept-specific probability distributions. Using this theory, we describe the short-term dynamics of buildup observed on short trials in terms of the long-term statistics of percept durations for the two alternating perceptual organizations. Our statistical-dynamics model describes well the buildup functions and alternations in simulations of pseudo-mechanistic neuronal network models with percept-selective populations competing through mutual inhibition. Even though the competition model can show history dependence through slow adaptation, our statistical switching model, that neglects history, predicts well the buildup function. We propose that accumulation is not a necessary feature to produce buildup. Generally, if alternations between two states exhibit independent durations with stationary statistics then the associated buildup function can be described by the statistical dynamics of an ARP.
Collapse
Affiliation(s)
- Sara A Steele
- Center for Neural Science, New York University New York, NY, USA
| | - Daniel Tranchina
- Courant Institute for Mathematical Sciences, New York University New York, NY, USA ; Department of Biology, New York University New York, NY, USA
| | - John Rinzel
- Center for Neural Science, New York University New York, NY, USA ; Courant Institute for Mathematical Sciences, New York University New York, NY, USA
| |
Collapse
|
24
|
Abstract
Recent developments in the cognitive neuroscience of music suggest that a further review of the topic of amusia is timely. In this chapter, we first consider previous taxonomies of amusia and propose a fresh framework for understanding the amusias, essentially as disorders of cognitive information processing. We critically review current cognitive and neuroanatomic findings in the published literature on amusia. We assess the extent to which the clinical and neuropsychologic evidence in amusia can be reconciled; both with the information-processing framework we propose, and with the picture of the brain organization of music and language processing emerging from cognitive neuroscience and functional neuroimaging studies. The balance of evidence suggests that the amusias can be understood as disorders of musical object cognition targeting separable levels of an information-processing hierarchy and underpinned by specific brain network dysfunction. The neuroanatomic associations of the amusias show substantial overlap with brain networks that process speech; however, this convergence leaves scope for separable brain mechanisms based on altered connectivity and dynamics across culprit networks. The study of the amusias contributes to an increasingly complex picture of the musical brain that transcends any simple dichotomy between music and speech or other complex sounds.
Collapse
Affiliation(s)
- Camilla N Clark
- Dementia Research Centre, UCL Institute of Neurology, University College London, Queen Square, London, United Kingdom
| | - Hannah L Golden
- Dementia Research Centre, UCL Institute of Neurology, University College London, Queen Square, London, United Kingdom
| | - Jason D Warren
- Dementia Research Centre, UCL Institute of Neurology, University College London, Queen Square, London, United Kingdom.
| |
Collapse
|
25
|
Deike S, Denham SL, Sussman E. Probing auditory scene analysis. Front Neurosci 2014; 8:293. [PMID: 25309314 PMCID: PMC4162357 DOI: 10.3389/fnins.2014.00293] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2014] [Accepted: 08/27/2014] [Indexed: 11/13/2022] Open
Affiliation(s)
- Susann Deike
- Special Lab Non-Invasive Brain Imaging, Leibniz Institute for Neurobiology Magdeburg, Germany
| | - Susan L Denham
- Cognition Institute, University of Plymouth Plymouth, UK ; School of Psychology, University of Plymouth Plymouth, UK
| | - Elyse Sussman
- Department of Neuroscience, Albert Einstein College of Medicine of Yeshiva University Bronx, NY, USA ; Department of Otorhinolaryngology-Head and Neck Surgery, Albert Einstein College of Medicine of Yeshiva University Bronx, NY, USA
| |
Collapse
|
26
|
Bayat A, Farhadi M, Pourbakht A, Sadjedi H, Emamdjomeh H, Kamali M, Mirmomeni G. A comparison of auditory perception in hearing-impaired and normal-hearing listeners: an auditory scene analysis study. IRANIAN RED CRESCENT MEDICAL JOURNAL 2014; 15:e9477. [PMID: 24719695 PMCID: PMC3971787 DOI: 10.5812/ircmj.9477] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/05/2012] [Accepted: 03/10/2013] [Indexed: 11/16/2022]
Abstract
Background Auditory scene analysis (ASA) is the process by which the auditory system separates individual sounds in natural-world situations. ASA is a key function of auditory system, and contributes to speech discrimination in noisy backgrounds. It is known that sensorineural hearing loss (SNHL) detrimentally affects auditory function in complex environments, but relatively few studies have focused on the influence of SNHL on higher level processes which are likely involved in auditory perception in different situations. Objectives The purpose of the current study was to compare the auditory system ability of normally hearing and SNHL subjects using the ASA examination. Materials and Methods A total of 40 right-handed adults (age range: 18 - 45 years) participated in this study. The listeners were divided equally into control and mild to moderate SNHL groups. ASA ability was measured using an ABA-ABA sequence. The frequency of the "A" was kept constant at 500, 1000, 2000 or 4000 Hz, while the frequency of the "B" was set at 3 to 80 percent above the" A" tone. For ASA threshold detection, the frequency of the B stimulus was decreased until listeners reported that they could no longer hear two separate sounds. Results The ASA performance was significantly better for controls than the SNHL group; these differences were more obvious at higher frequencies. We found no significant differences between ASA ability as a function of tone durations in both groups. Conclusions The present study indicated that SNHL may cause a reduction in perceptual separation of the incoming acoustic information to form accurate representations of our acoustic world.
Collapse
Affiliation(s)
- Arash Bayat
- Department of Audiology, Iran University of Medical Sciences, Tehran, IR Iran
- Department of Audiology, Ahvaz Jundishapur University of Medical Sciences, Ahvaz, IR Iran
| | - Mohammad Farhadi
- Department and Research Center of Otolaryngology, Head and Neck Surgery, Hazrat Rasoul Akram Hospital, Iran University of Medical Sciences, Tehran, IR Iran
| | - Akram Pourbakht
- Department of Audiology, Iran University of Medical Sciences, Tehran, IR Iran
- Rehabilitation Research Center, Iran University of Medical Sciences, Tehran, IR Iran
- Corresponding Author: Akram Pourbakht, Department of Audiology, School of Rehabilitation, Tehran University of Medical Sciences, Tehran, IR Iran. Tel: +98-212250541, Fax: +98-216670051, E-mail:
| | - Hamed Sadjedi
- Department of Electronics, Engineering faculty, Shahed University, Tehran, IR Iran
| | - Hesam Emamdjomeh
- Department and Research Center of Otolaryngology, Head and Neck Surgery, Hazrat Rasoul Akram Hospital, Iran University of Medical Sciences, Tehran, IR Iran
| | - Mohammad Kamali
- Department of Rehabilitation Management, Iran University of Medical Sciences, Tehran, IR Iran
| | - Golshan Mirmomeni
- Department of Audiology, Ahvaz Jundishapur University of Medical Sciences, Ahvaz, IR Iran
| |
Collapse
|
27
|
Bendixen A. Predictability effects in auditory scene analysis: a review. Front Neurosci 2014; 8:60. [PMID: 24744695 PMCID: PMC3978260 DOI: 10.3389/fnins.2014.00060] [Citation(s) in RCA: 73] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2014] [Accepted: 03/14/2014] [Indexed: 12/02/2022] Open
Abstract
Many sound sources emit signals in a predictable manner. The idea that predictability can be exploited to support the segregation of one source's signal emissions from the overlapping signals of other sources has been expressed for a long time. Yet experimental evidence for a strong role of predictability within auditory scene analysis (ASA) has been scarce. Recently, there has been an upsurge in experimental and theoretical work on this topic resulting from fundamental changes in our perspective on how the brain extracts predictability from series of sensory events. Based on effortless predictive processing in the auditory system, it becomes more plausible that predictability would be available as a cue for sound source decomposition. In the present contribution, empirical evidence for such a role of predictability in ASA will be reviewed. It will be shown that predictability affects ASA both when it is present in the sound source of interest (perceptual foreground) and when it is present in other sound sources that the listener wishes to ignore (perceptual background). First evidence pointing toward age-related impairments in the latter capacity will be addressed. Moreover, it will be illustrated how effects of predictability can be shown by means of objective listening tests as well as by subjective report procedures, with the latter approach typically exploiting the multi-stable nature of auditory perception. Critical aspects of study design will be delineated to ensure that predictability effects can be unambiguously interpreted. Possible mechanisms for a functional role of predictability within ASA will be discussed, and an analogy with the old-plus-new heuristic for grouping simultaneous acoustic signals will be suggested.
Collapse
Affiliation(s)
- Alexandra Bendixen
- Auditory Psychophysiology Lab, Department of Psychology, Cluster of Excellence "Hearing4all," European Medical School, Carl von Ossietzky University of Oldenburg Oldenburg, Germany
| |
Collapse
|
28
|
|
29
|
Abstract
The fundamental perceptual unit in hearing is the 'auditory object'. Similar to visual objects, auditory objects are the computational result of the auditory system's capacity to detect, extract, segregate and group spectrotemporal regularities in the acoustic environment; the multitude of acoustic stimuli around us together form the auditory scene. However, unlike the visual scene, resolving the component objects within the auditory scene crucially depends on their temporal structure. Neural correlates of auditory objects are found throughout the auditory system. However, neural responses do not become correlated with a listener's perceptual reports until the level of the cortex. The roles of different neural structures and the contribution of different cognitive states to the perception of auditory objects are not yet fully understood.
Collapse
|
30
|
Teki S, Chait M, Kumar S, Shamma S, Griffiths TD. Segregation of complex acoustic scenes based on temporal coherence. eLife 2013; 2:e00699. [PMID: 23898398 PMCID: PMC3721234 DOI: 10.7554/elife.00699] [Citation(s) in RCA: 46] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2013] [Accepted: 06/16/2013] [Indexed: 11/13/2022] Open
Abstract
In contrast to the complex acoustic environments we encounter everyday, most studies of auditory segregation have used relatively simple signals. Here, we synthesized a new stimulus to examine the detection of coherent patterns (‘figures’) from overlapping ‘background’ signals. In a series of experiments, we demonstrate that human listeners are remarkably sensitive to the emergence of such figures and can tolerate a variety of spectral and temporal perturbations. This robust behavior is consistent with the existence of automatic auditory segregation mechanisms that are highly sensitive to correlations across frequency and time. The observed behavior cannot be explained purely on the basis of adaptation-based models used to explain the segregation of deterministic narrowband signals. We show that the present results are consistent with the predictions of a model of auditory perceptual organization based on temporal coherence. Our data thus support a role for temporal coherence as an organizational principle underlying auditory segregation. DOI:http://dx.doi.org/10.7554/eLife.00699.001 Even when seated in the middle of a crowded restaurant, we are still able to distinguish the speech of the person sitting opposite us from the conversations of fellow diners and a host of other background noise. While we generally perform this task almost effortlessly, it is unclear how the brain solves what is in reality a complex information processing problem. In the 1970s, researchers began to address this question using stimuli consisting of simple tones. When subjects are played a sequence of alternating high and low frequency tones, they perceive them as two independent streams of sound. Similar experiments in macaque monkeys reveal that each stream activates a different area of auditory cortex, suggesting that the brain may distinguish acoustic stimuli on the basis of their frequency. However, the simple tones that are used in laboratory experiments bear little resemblance to the complex sounds we encounter in everyday life. These are often made up of multiple frequencies, and overlap—both in frequency and in time—with other sounds in the environment. Moreover, recent experiments have shown that if a subject hears two tones simultaneously, he or she perceives them as belonging to a single stream of sound even if they have different frequencies: models that assume that we distinguish stimuli from noise on the basis of frequency alone struggle to explain this observation. Now, Teki, Chait, et al. have used more complex sounds, in which frequency components of the target stimuli overlap with those of background signals, to obtain new insights into how the brain solves this problem. Subjects were extremely good at discriminating these complex target stimuli from background noise, and computational modelling confirmed that they did so via integration of both frequency and temporal information. The work of Teki, Chait, et al. thus offers the first explanation for our ability to home in on speech and other pertinent sounds, even amidst a sea of background noise. DOI:http://dx.doi.org/10.7554/eLife.00699.002
Collapse
Affiliation(s)
- Sundeep Teki
- Wellcome Trust Centre for Neuroimaging , University College London , London , United Kingdom
| | | | | | | | | |
Collapse
|
31
|
Millisecond stimulus onset-asynchrony enhances information about components in an odor mixture. J Neurosci 2013; 33:6060-9. [PMID: 23554487 DOI: 10.1523/jneurosci.5838-12.2013] [Citation(s) in RCA: 35] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
Airborne odorants rarely occur as pure, isolated stimuli. In a natural environment, odorants that intermingle from multiple sources create mixtures in which the onset and offset of odor components are asynchronous. Odor mixtures are known to elicit interactions in both behavioral and physiological responses, changing the perceptive quality of mixtures compared with the components. However, relevant odors need to be segregated from a distractive background. Honeybees (Apis mellifera) can use stimulus onset asynchrony of as little as 6 ms to segregate learned odor components within a mixture. Using in vivo calcium imaging of projection neurons in the honeybee, we studied neuronal mechanisms of odor-background segregation based on stimulus onset asynchrony in the antennal lobe. We found that asynchronous mixtures elicit response patterns that are different from their synchronous counterpart: the responses to asynchronous mixtures contain more information about the constituent components. With longer onset shifts, more features of the components were present in the mixture response patterns. Moreover, we found that the processing of asynchronous mixtures activated more inhibitory interactions than the processing of synchronous mixtures. This study provides evidence of neuronal mechanisms that underlie odor-object segregation on a timescale much faster than found for mammals.
Collapse
|