51
|
Mårup SH, Kleber BA, Møller C, Vuust P. When direction matters: Neural correlates of interlimb coordination of rhythm and beat. Cortex 2024; 172:86-108. [PMID: 38241757 DOI: 10.1016/j.cortex.2023.11.019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2022] [Revised: 04/11/2023] [Accepted: 11/09/2023] [Indexed: 01/21/2024]
Abstract
In a previous experiment, we found evidence for a bodily hierarchy governing interlimb coordination of rhythm and beat, using five effectors: 1) Left foot, 2) Right foot, 3) Left hand, 4) Right hand and 5) Voice. The hierarchy implies that, during simultaneous rhythm and beat performance and using combinations of two of these effectors, executing the task by performing the rhythm with an effector that has a higher number than the beat effector is significantly easier than vice versa. To investigate the neural underpinnings of this proposed bodily hierarchy, we here scanned 46 professional musicians using fMRI as they performed a rhythmic pattern with one effector while keeping the beat with another. The conditions combined the voice and the right hand (V + RH), the right hand and the left hand (RH + LH), and the left hand and the right foot (LH + RF). Each effector combination was performed with and against the bodily hierarchy. Going against the bodily hierarchy increased tapping errors significantly and also increased activity in key brain areas functionally associated with top-down sensorimotor control and bottom-up feedback processing, such as the cerebellum and SMA. Conversely, going with the bodily hierarchy engaged areas functionally associated with the default mode network and regions involved in emotion processing. Theories of general brain function that hold prediction as a key principle, propose that action and perception are governed by the brain's attempt to minimise prediction error at different levels in the brain. Following this viewpoint, our results indicate that going against the hierarchy induces stronger prediction errors, while going with the hierarchy allows for a higher degree of automatization. Our results also support the notion of a bodily hierarchy in motor control that prioritizes certain conductive and supportive tapping roles in specific effector combinations.
Collapse
|
52
|
Jertberg RM, Begeer S, Geurts HM, Chakrabarti B, Van der Burg E. Perception of temporal synchrony not a prerequisite for multisensory integration. Sci Rep 2024; 14:4982. [PMID: 38424118 PMCID: PMC10904801 DOI: 10.1038/s41598-024-55572-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2023] [Accepted: 02/25/2024] [Indexed: 03/02/2024] Open
Abstract
Temporal alignment is often viewed as the most essential cue the brain can use to integrate information from across sensory modalities. However, the importance of conscious perception of synchrony to multisensory integration is a controversial topic. Conversely, the influence of cross-modal incongruence of higher level stimulus features such as phonetics on temporal processing is poorly understood. To explore the nuances of this relationship between temporal processing and multisensory integration, we presented 101 participants (ranging from 19 to 73 years of age) with stimuli designed to elicit the McGurk/MacDonald illusion (either matched or mismatched pairs of phonemes and visemes) with varying degrees of stimulus onset asynchrony between the visual and auditory streams. We asked them to indicate which syllable they perceived and whether the video and audio were synchronized on each trial. We found that participants often experienced the illusion despite not perceiving the stimuli as synchronous, and the same phonetic incongruence that produced the illusion also led to significant interference in simultaneity judgments. These findings challenge the longstanding assumption that perception of synchrony is a prerequisite to multisensory integration, support a more flexible view of multisensory integration, and suggest a complex, reciprocal relationship between temporal and multisensory processing.
Collapse
|
53
|
Marjieh R, Harrison PMC, Lee H, Deligiannaki F, Jacoby N. Timbral effects on consonance disentangle psychoacoustic mechanisms and suggest perceptual origins for musical scales. Nat Commun 2024; 15:1482. [PMID: 38369535 DOI: 10.1038/s41467-024-45812-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2022] [Accepted: 12/11/2023] [Indexed: 02/20/2024] Open
Abstract
The phenomenon of musical consonance is an essential feature in diverse musical styles. The traditional belief, supported by centuries of Western music theory and psychological studies, is that consonance derives from simple (harmonic) frequency ratios between tones and is insensitive to timbre. Here we show through five large-scale behavioral studies, comprising 235,440 human judgments from US and South Korean populations, that harmonic consonance preferences can be reshaped by timbral manipulations, even as far as to induce preferences for inharmonic intervals. We show how such effects may suggest perceptual origins for diverse scale systems ranging from the gamelan's slendro scale to the tuning of Western mean-tone and equal-tempered scales. Through computational modeling we show that these timbral manipulations dissociate competing psychoacoustic mechanisms underlying consonance, and we derive an updated computational model combining liking of harmonicity, disliking of fast beats (roughness), and liking of slow beats. Altogether, this work showcases how large-scale behavioral experiments can inform classical questions in auditory perception.
Collapse
|
54
|
林 欢, 刘 潘, 孙 钰, 俞 欣, 钱 君, 池 霞, 洪 琴. [Association between auditory processing and problem behaviors in preschool children: the mediating role of executive function]. ZHONGGUO DANG DAI ER KE ZA ZHI = CHINESE JOURNAL OF CONTEMPORARY PEDIATRICS 2024; 26:174-180. [PMID: 38436316 PMCID: PMC10921876 DOI: 10.7499/j.issn.1008-8830.2309067] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Subscribe] [Scholar Register] [Received: 09/14/2023] [Accepted: 12/25/2023] [Indexed: 03/05/2024]
Abstract
OBJECTIVES To investigate the association between auditory processing and problem behaviors in preschool children, as well as the mediating role of executive function. METHODS A total of 2 342 preschool children were selected from 7 kindergartens in Nanjing, China from June to August 2021. They were evaluated using Preschool Auditory Processing Assessment Scale, Conners Parent Symptom Questionnaire, and Behavior Rating Inventory of Executive Functioning-Preschool version. Children with different demographic features were compared in the scores and the abnormality rates of auditory processing, problem behaviors, and executive function. The influencing factors of the total scores of auditory processing, problem behaviors, and executive function were evaluated using multiple linear regression analysis. Whether executive function was a mediating factor between auditory processing and executive function was examined. RESULTS Sex and grade were the main influencing factors for the total score of auditory processing (P<0.05), and sex, grade, parental education level, and family economic status were the main influencing factors for the total scores of problem behaviors and executive function (P<0.05). The auditory processing score (rs=0.458, P<0.05) and problem behavior score (rs=0.185, P<0.05) were significantly positively correlated with the executive function score, and the auditory processing score was significantly positively correlated with the problem behavior score (rs=0.423, P<0.05). Executive function played a partial mediating role between auditory processing and problem behaviors, and the mediating effect accounted for 33.44% of the total effect. CONCLUSIONS Auditory processing can directly affect the problem behaviors of preschool children and indirectly affect problem behaviors through executive function.
Collapse
|
55
|
Kondo HM, Hasegawa R, Ezaki T, Sakata H, Ho HT. Functional coupling between auditory memory and verbal transformations. Sci Rep 2024; 14:3480. [PMID: 38347058 PMCID: PMC10861569 DOI: 10.1038/s41598-024-54013-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2023] [Accepted: 02/07/2024] [Indexed: 02/15/2024] Open
Abstract
The ability to parse sound mixtures into coherent auditory objects is fundamental to cognitive functions, such as speech comprehension and language acquisition. Yet, we still lack a clear understanding of how auditory objects are formed. To address this question, we studied a speech-specific case of perceptual multistability, called verbal transformations (VTs), in which a variety of verbal forms is induced by continuous repetition of a physically unchanging word. Here, we investigated the degree to which auditory memory through sensory adaptation influences VTs. Specifically, we hypothesized that when memory persistence is longer, participants are able to retain the current verbal form longer, resulting in sensory adaptation, which in turn, affects auditory perception. Participants performed VT and auditory memory tasks on different days. In the VT task, Japanese participants continuously reported their perception while listening to a Japanese word (2- or 3-mora in length) played repeatedly for 5 min. In the auditory memory task, a different sequence of three morae, e.g., /ka/, /hi/, and /su/, was presented to each ear simultaneously. After some period (0-4 s), participants were visually cued to recall one of the sequences, i.e., in the left or right ear. We found that delayed recall accuracy was negatively correlated with the number of VTs, particularly under 2-mora conditions. This suggests that memory persistence is important for formation and selection of perceptual objects.
Collapse
|
56
|
Lee HH, Groves K, Ripollés P, Carrasco M. Audiovisual integration in the McGurk effect is impervious to music training. Sci Rep 2024; 14:3262. [PMID: 38332159 PMCID: PMC10853564 DOI: 10.1038/s41598-024-53593-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2023] [Accepted: 02/01/2024] [Indexed: 02/10/2024] Open
Abstract
The McGurk effect refers to an audiovisual speech illusion where the discrepant auditory and visual syllables produce a fused percept between the visual and auditory component. However, little is known about how individual differences contribute to the McGurk effect. Here, we examined whether music training experience-which involves audiovisual integration-can modulate the McGurk effect. Seventy-three participants completed the Goldsmiths Musical Sophistication Index (Gold-MSI) questionnaire to evaluate their music expertise on a continuous scale. Gold-MSI considers participants' daily-life exposure to music learning experiences (formal and informal), instead of merely classifying people into different groups according to how many years they have been trained in music. Participants were instructed to report, via a 3-alternative forced choice task, "what a person said": /Ba/, /Ga/ or /Da/. The experiment consisted of 96 audiovisual congruent trials and 96 audiovisual incongruent (McGurk) trials. We observed no significant correlations between the susceptibility of the McGurk effect and the different subscales of the Gold-MSI (active engagement, perceptual abilities, music training, singing abilities, emotion) or the general musical sophistication composite score. Together, these findings suggest that music training experience does not modulate audiovisual integration in speech as reflected by the McGurk effect.
Collapse
|
57
|
Rotaru I, Geirnaert S, Heintz N, Van de Ryck I, Bertrand A, Francart T. What are we reallydecoding? Unveiling biases in EEG-based decoding of the spatial focus of auditory attention. J Neural Eng 2024; 21:016017. [PMID: 38266281 DOI: 10.1088/1741-2552/ad2214] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/13/2023] [Accepted: 01/24/2024] [Indexed: 01/26/2024]
Abstract
Objective.Spatial auditory attention decoding (Sp-AAD) refers to the task of identifying the direction of the speaker to which a person is attending in a multi-talker setting, based on the listener's neural recordings, e.g. electroencephalography (EEG). The goal of this study is to thoroughly investigate potential biases when training such Sp-AAD decoders on EEG data, particularly eye-gaze biases and latent trial-dependent confounds, which may result in Sp-AAD models that decode eye-gaze or trial-specific fingerprints rather than spatial auditory attention.Approach.We designed a two-speaker audiovisual Sp-AAD protocol in which the spatial auditory and visual attention were enforced to be either congruent or incongruent, and we recorded EEG data from sixteen participants undergoing several trials recorded at distinct timepoints. We trained a simple linear model for Sp-AAD based on common spatial patterns filters in combination with either linear discriminant analysis (LDA) or k-means clustering, and evaluated them both across- and within-trial.Main results.We found that even a simple linear Sp-AAD model is susceptible to overfitting to confounding signal patterns such as eye-gaze and trial fingerprints (e.g. due to feature shifts across trials), resulting in artificially high decoding accuracies. Furthermore, we found that changes in the EEG signal statistics across trials deteriorate the trial generalization of the classifier, even when the latter is retrained on the test trial with an unsupervised algorithm.Significance.Collectively, our findings confirm that there exist subtle biases and confounds that can strongly interfere with the decoding of spatial auditory attention from EEG. It is expected that more complicated non-linear models based on deep neural networks, which are often used for Sp-AAD, are even more vulnerable to such biases. Future work should perform experiments and model evaluations that avoid and/or control for such biases in Sp-AAD tasks.
Collapse
|
58
|
Wang X, Burger B, Wöllner C. Tapping to drumbeats in an online experiment changes our perception of time and expressiveness. PSYCHOLOGICAL RESEARCH 2024; 88:127-140. [PMID: 37300705 PMCID: PMC10806225 DOI: 10.1007/s00426-023-01835-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2022] [Accepted: 05/15/2023] [Indexed: 06/12/2023]
Abstract
Bodily movements along with music, such as tapping, are not only very frequent, but may also have a profound impact on our perception of time and emotions. The current study adopted an online tapping paradigm to investigate participants' time experiences and expressiveness judgements when they tapped and did not tap to a series of drumming performances that varied in tempo and rhythmic complexity. Participants were asked to judge durations, passage of time (PoT), and the expressiveness of the performances in two conditions: (1) Observing only, (2) Observing and tapping regularly to the perceived beats. Results show that tapping trials passed subjectively faster and were partially (in slow- and medium-tempo conditions) perceived shorter compared to the observing-only trials. Increases in musical tempo (in tapping trials) and in complexity led to faster PoT, potentially due to distracted attentional resources for the timing task. Participants' musical training modulated the effects of complexity on the judgments of expressiveness. In addition, increases in tapping speed led to duration overestimation among the less musically trained participants. Taken together, tapping to music may have altered the internal clock speed, affecting the temporal units accumulated in the pacemaker-counter model.
Collapse
|
59
|
Hüppi PS, Filippa M. Multisensory stimuli and pain perception in the newborn. Pediatr Res 2024; 95:603-604. [PMID: 37833529 PMCID: PMC10899096 DOI: 10.1038/s41390-023-02833-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Accepted: 09/02/2023] [Indexed: 10/15/2023]
|
60
|
Dudarev V, Kai J, Brar N, Enns JT. Listening to trees in the forest: Attentional set influences how semantic and acoustic factors interact in auditory perception. Atten Percept Psychophys 2024; 86:381-391. [PMID: 38177945 DOI: 10.3758/s13414-023-02835-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/14/2023] [Indexed: 01/06/2024]
Abstract
Studies of auditory object perception claim that semantic properties dominate acoustic properties in determining identification accuracy. Yet the direction of the semantic effect is mixed, with some studies showing an advantage for detecting incongruent sounds and others reporting a congruent sound advantage. Here we examine the role of the participant's attentional set when identifying auditory objects in naturalistic soundscapes. We varied the acoustic and semantic properties of the sounds orthogonally in two experiments. In Experiment 1 participants tuned their attention broadly to detect any change between two successive soundscapes (e.g., two restaurant soundscapes, with and without a child coughing). In Experiment 2 they tuned attention more narrowly to a probe presented after a soundscape (e.g., a restaurant soundscape with a child coughing, followed by the coughing sound alone). In both experiments, semantic relations between the objects and backgrounds helped to disambiguate objects that blended acoustically with the background. When attending globally (Experiment 1), objects that were acoustically similar yet semantically incongruent tended to be missed (e.g., bouncing basketball on a construction site), as though camouflaged by the gist of the soundscape. When attending locally (Experiment 2), semantically congruent foil objects led to false positive reports under acoustically similar conditions (hammering sounds on a construction site), as though the gist of the soundscape contributed to their plausible inclusion. In summary, although attentional set had a strong influence on the specific kinds of errors made, both results pointed to participants using a semantically congruent high-level schema to report the sounds they heard.
Collapse
|
61
|
Breadmore HL, Halliday LF, Carroll JM. Variability in auditory processing performance is associated with reading difficulties rather than with history of otitis media. DYSLEXIA (CHICHESTER, ENGLAND) 2024; 30:e1760. [PMID: 38262626 DOI: 10.1002/dys.1760] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/23/2023] [Revised: 10/05/2023] [Accepted: 12/13/2023] [Indexed: 01/25/2024]
Abstract
The nature and cause of auditory processing deficits in dyslexic individuals have been debated for decades. Auditory processing deficits were argued to be the first step in a causal chain of difficulties, leading to difficulties in speech perception and thereby phonological processing and literacy difficulties. More recently, it has been argued that auditory processing difficulties may not be causally related to language and literacy difficulties. This study compares two groups who have phonological processing impairments for different reasons: dyslexia and a history of otitis media (OM). We compared their discrimination thresholds and response variability to chronological age- and reading age-matched controls, across three auditory processing tasks: frequency discrimination, rise-time discrimination and speech perception. Dyslexic children showed raised frequency discrimination thresholds in comparison with age-matched controls but did not differ from reading age-matched controls or individuals with a history of OM. There were no group differences on speech perception or rise-time tasks. For the dyslexic children, there was an association between phonological awareness and frequency discrimination response variability, but no association with thresholds. These findings are not consistent with a 'causal chain' explanation but could be accounted for within a multiple deficits view of literacy difficulties.
Collapse
|
62
|
Kiepe F, Kraus N, Hesselmann G. Self-initiation enhances perceptual processing of auditory stimuli in an online study. Atten Percept Psychophys 2024; 86:587-601. [PMID: 38148430 DOI: 10.3758/s13414-023-02827-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/07/2023] [Indexed: 12/28/2023]
Abstract
Understanding how the brain incorporates sensory and motor information will enable better theory building on human perception and behavior. In this study, we aimed to estimate the influence of predictive mechanisms on the magnitude and variability of sensory attenuation in two online samples. After the presentation of a visual cue stimulus, participants (Experiment 1: N = 224, Experiment 2: N = 84) compared the loudness of two consecutive tones in a two-alternative forced-choice task. In Experiment 1, the first tone was either self-initiated or not; in Experiment 2, the second tone was either self-initiated or not (active and passive condition, respectively). We further manipulated identity prediction (i.e., the congruence of pre-learned cue-sound combinations; congruent vs. incongruent), and the duration of the onset delay (to account for effects of attentional differences between the passive and active condition, 50 ms vs. 0 ms). We critically discuss our results within the framework of both classical (i.e., motor-based forward models) and contemporary approaches (i.e., predictive processing framework). Contrary to our preregistered hypothesis, we observed enhanced perceptual processing, instead of attenuation, for self-initiated auditory sensory input. Further, our results reveal an effect of fixed sound delays on the processing of motor and non-motor-based predictive information, and may point to according shifts in attention, leading to a perceptual bias. These results might best be captured by a hybrid explanatory model, combining predictions based on self-initiated motor action with a global predictive mechanism.
Collapse
|
63
|
Hu J, Vetter P. How the eyes respond to sounds. Ann N Y Acad Sci 2024; 1532:18-36. [PMID: 38152040 DOI: 10.1111/nyas.15093] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/29/2023]
Abstract
Eye movements have been extensively studied with respect to visual stimulation. However, we live in a multisensory world, and how the eyes are driven by other senses has been explored much less. Here, we review the evidence on how audition can trigger and drive different eye responses and which cortical and subcortical neural correlates are involved. We provide an overview on how different types of sounds, from simple tones and noise bursts to spatially localized sounds and complex linguistic stimuli, influence saccades, microsaccades, smooth pursuit, pupil dilation, and eye blinks. The reviewed evidence reveals how the auditory system interacts with the oculomotor system, both behaviorally and neurally, and how this differs from visually driven eye responses. Some evidence points to multisensory interaction, and potential multisensory integration, but the underlying computational and neural mechanisms are still unclear. While there are marked differences in how the eyes respond to auditory compared to visual stimuli, many aspects of auditory-evoked eye responses remain underexplored, and we summarize the key open questions for future research.
Collapse
|
64
|
Vathagavorakul R, Gonjo T, Homma M. The influence of sound waves and musical experiences on movement coordination with beats. Hum Mov Sci 2024; 93:103170. [PMID: 38043482 DOI: 10.1016/j.humov.2023.103170] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2023] [Revised: 11/12/2023] [Accepted: 11/21/2023] [Indexed: 12/05/2023]
Abstract
Synchronizing movement with external stimuli is important in musicians and athletes. This study investigated the effects of sound characteristics, including sound with harmonics (square wave) and without harmonics (sine wave) and levels of expertise in sports and music on rhythmic ability. Thirty-two university students participated in the study. The participants were divided into sixteen music education (ME) and sixteen physical education (PE) majors. They were asked to perform finger tapping tasks with 1,2 and 3 Hz beat rates, tapping in time with the sine and square wave beat produced by a metronome. The relative phase angle of finger tapping and the onset time of metronome sound were calculated using circular statistics. The results showed that type of wave and music experience affected the rhythmic ability of participants. Our study highlights the importance of types of waves on rhythmic ability, especially for participants with no background in music. The square wave is recommended for athletes to learn to synchronize their movement with beats.
Collapse
|
65
|
Wang K, Fang Y, Guo Q, Shen L, Chen Q. Superior Attentional Efficiency of Auditory Cue via the Ventral Auditory-thalamic Pathway. J Cogn Neurosci 2024; 36:303-326. [PMID: 38010315 DOI: 10.1162/jocn_a_02090] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2023]
Abstract
Auditory commands are often executed more efficiently than visual commands. However, empirical evidence on the underlying behavioral and neural mechanisms remains scarce. In two experiments, we manipulated the delivery modality of informative cues and the prediction violation effect and found consistently enhanced RT benefits for the matched auditory cues compared with the matched visual cues. At the neural level, when the bottom-up perceptual input matched the prior prediction induced by the auditory cue, the auditory-thalamic pathway was significantly activated. Moreover, the stronger the auditory-thalamic connectivity, the higher the behavioral benefits of the matched auditory cue. When the bottom-up input violated the prior prediction induced by the auditory cue, the ventral auditory pathway was specifically involved. Moreover, the stronger the ventral auditory-prefrontal connectivity, the larger the behavioral costs caused by the violation of the auditory cue. In addition, the dorsal frontoparietal network showed a supramodal function in reacting to the violation of informative cues irrespective of the delivery modality of the cue. Taken together, the results reveal novel behavioral and neural evidence that the superior efficiency of the auditory cue is twofold: The auditory-thalamic pathway is associated with improvements in task performance when the bottom-up input matches the auditory cue, whereas the ventral auditory-prefrontal pathway is involved when the auditory cue is violated.
Collapse
|
66
|
Lee T, Rieke C, Niemczak C, Kobrina A, Clavier O, Gonzalez J, Fuente A, Alvarez KM, Gonzalez-Quiroz M, Buckey J, Saunders J. Assessment of Central Auditory Processing in Children Using a Novel Tablet-Based Platform: Application for Low- and Middle-Income Countries. Otol Neurotol 2024; 45:176-183. [PMID: 38206066 PMCID: PMC10783812 DOI: 10.1097/mao.0000000000004085] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/12/2024]
Abstract
OBJECTIVE Evaluate whether a portable, tablet-based central auditory processing (CAP) test system using native language training videos and administered by minimally trained community health workers can produce CAP results comparable to previously published norms. Our secondary aim was to determine subject parameters that influence test results. STUDY DESIGN Cross-sectional study. SETTING Community-based settings in Chontales, Nicaragua, New Hampshire, and Florida. PATIENTS English- and/or Spanish-speaking children and adolescents (n = 245; average age, 12.20 yr; range, 6-18 yr). MAIN OUTCOME MEASURES Completion of the following tests with responses comparable to published norms: Pure-tone average (PTA), gap detection threshold (GDT), fixed-level frequency threshold, masking level difference (MLD), Hearing in Noise Test (HINT), Dichotic Digits Test (DDT), and Frequency Pattern Recognition (FPR) test. RESULTS GDT, HINT, and DDT had comparable results to previously published normative values. MLD and FPR results differed compared with previously published normative values. Most CAP tests (MLD, GDT, HINT) results were independent of age and PTA (p = 0.1-0.9). However, DDT was associated with age and PTA (p < 0.0001). CONCLUSIONS Pediatric CAP testing can be successfully completed in remote low- and middle- income country environments using a tablet-based platform without the presence of an audiologist. Performance on DDT improved with age but deteriorated with hearing loss. Further investigation is warranted to assess the variability of FPR.
Collapse
|
67
|
Liu H, Bai Y, Xu Z, Liu J, Ni G, Ming D. The scalp time-varying network of auditory spatial attention in "cocktail-party" situations. Hear Res 2024; 442:108946. [PMID: 38150794 DOI: 10.1016/j.heares.2023.108946] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/09/2023] [Revised: 12/21/2023] [Accepted: 12/22/2023] [Indexed: 12/29/2023]
Abstract
Sound source localization in "cocktail-party" situations is a remarkable ability of the human auditory system. However, the neural mechanisms underlying auditory spatial attention are still largely unknown. In this study, the "cocktail-party" situations are simulated through multiple sound sources and presented through head-related transfer functions and headphones. Furthermore, the scalp time-varying network of auditory spatial attention is constructed using the high-temporal resolution electroencephalogram, and its network properties are measured quantitatively using graph theory analysis. The results show that the time-varying network of auditory spatial attention in "cocktail-party" situations is more complex and partially different than in simple acoustic situations, especially in the early- and middle-latency periods. The network coupling strength increases continuously over time, and the network hub shifts from the posterior temporal lobe to the parietal lobe and then to the frontal lobe region. In addition, the right hemisphere has a stronger network strength for processing auditory spatial information in "cocktail-party" situations, i.e., the right hemisphere has higher clustering levels, higher transmission efficiency, and more node degrees during the early- and middle-latency periods, while this phenomenon disappears and appears symmetrically during the late-latency period. These findings reveal different network patterns and properties of auditory spatial attention in "cocktail-party" situations during different periods and demonstrate the dominance of the right hemisphere in the dynamic processing of auditory spatial information.
Collapse
|
68
|
Merz S, Frings C, Spence C. Motion perception in touch: resolving contradictory findings by varying probabilities of different trial types. PSYCHOLOGICAL RESEARCH 2024; 88:148-155. [PMID: 37369933 PMCID: PMC10805958 DOI: 10.1007/s00426-023-01849-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2023] [Accepted: 06/10/2023] [Indexed: 06/29/2023]
Abstract
Representational momentum describes the typical overestimation of the final location of a moving stimulus in the direction of stimulus motion. While systematically observed in different sensory modalities, especially vision and audition, in touch, empirical findings indicate a mixed pattern of results, with some published studies suggesting the existence of the phenomenon, while others do not. In the present study, one possible moderating variable, the relative probabilities of different trial types, was explored in an attempt to resolve the seemingly contradictory findings in the literature. In some studies, only consistently moving target stimuli were presented and no representational momentum was observed, while other studies have included inconsistently moving target stimuli in the same experimental block, and observed representational momentum. Therefore, the present study was designed to systematically compare the localization of consistent target motion stimuli across two experimental blocks, for which either only consistent motion trials were presented, or else mixed with inconsistent target motion trials. The results indicate a strong influence of variations in the probability of different trial types on the occurrence of representational momentum. That is, representational momentum only occurred when both trial types (inconsistent and consistent target motion) were presented within one experimental block. The results are discussed in light of recent theoretical advancements in the literature, namely the speed prior account of motion perception.
Collapse
|
69
|
Lee J, Park S. Multi-modal Representation of the Size of Space in the Human Brain. J Cogn Neurosci 2024; 36:340-361. [PMID: 38010320 DOI: 10.1162/jocn_a_02092] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2023]
Abstract
To estimate the size of an indoor space, we must analyze the visual boundaries that limit the spatial extent and acoustic cues from reflected interior surfaces. We used fMRI to examine how the brain processes the geometric size of indoor scenes when various types of sensory cues are presented individually or together. Specifically, we asked whether the size of space is represented in a modality-specific way or in an integrative way that combines multimodal cues. In a block-design study, images or sounds that depict small- and large-sized indoor spaces were presented. Visual stimuli were real-world pictures of empty spaces that were small or large. Auditory stimuli were sounds convolved with different reverberations. By using a multivoxel pattern classifier, we asked whether the two sizes of space can be classified in visual, auditory, and visual-auditory combined conditions. We identified both sensory-specific and multimodal representations of the size of space. To further investigate the nature of the multimodal region, we specifically examined whether it contained multimodal information in a coexistent or integrated form. We found that angular gyrus and the right medial frontal gyrus had modality-integrated representation, displaying sensitivity to the match in the spatial size information conveyed through image and sound. Background functional connectivity analysis further demonstrated that the connection between sensory-specific regions and modality-integrated regions increases in the multimodal condition compared with single modality conditions. Our results suggest that spatial size perception relies on both sensory-specific and multimodal representations, as well as their interplay during multimodal perception.
Collapse
|
70
|
Putkinen V, Zhou X, Gan X, Yang L, Becker B, Sams M, Nummenmaa L. Bodily maps of musical sensations across cultures. Proc Natl Acad Sci U S A 2024; 121:e2308859121. [PMID: 38271338 PMCID: PMC10835118 DOI: 10.1073/pnas.2308859121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2023] [Accepted: 12/01/2023] [Indexed: 01/27/2024] Open
Abstract
Emotions, bodily sensations and movement are integral parts of musical experiences. Yet, it remains unknown i) whether emotional connotations and structural features of music elicit discrete bodily sensations and ii) whether these sensations are culturally consistent. We addressed these questions in a cross-cultural study with Western (European and North American, n = 903) and East Asian (Chinese, n = 1035). We precented participants with silhouettes of human bodies and asked them to indicate the bodily regions whose activity they felt changing while listening to Western and Asian musical pieces with varying emotional and acoustic qualities. The resulting bodily sensation maps (BSMs) varied as a function of the emotional qualities of the songs, particularly in the limb, chest, and head regions. Music-induced emotions and corresponding BSMs were replicable across Western and East Asian subjects. The BSMs clustered similarly across cultures, and cluster structures were similar for BSMs and self-reports of emotional experience. The acoustic and structural features of music were consistently associated with the emotion ratings and music-induced bodily sensations across cultures. These results highlight the importance of subjective bodily experience in music-induced emotions and demonstrate consistent associations between musical features, music-induced emotions, and bodily sensations across distant cultures.
Collapse
|
71
|
Cheung VKM, Harrison PMC, Koelsch S, Pearce MT, Friederici AD, Meyer L. Cognitive and sensory expectations independently shape musical expectancy and pleasure. Philos Trans R Soc Lond B Biol Sci 2024; 379:20220420. [PMID: 38104601 PMCID: PMC10725761 DOI: 10.1098/rstb.2022.0420] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2023] [Accepted: 10/20/2023] [Indexed: 12/19/2023] Open
Abstract
Expectation is crucial for our enjoyment of music, yet the underlying generative mechanisms remain unclear. While sensory models derive predictions based on local acoustic information in the auditory signal, cognitive models assume abstract knowledge of music structure acquired over the long term. To evaluate these two contrasting mechanisms, we compared simulations from four computational models of musical expectancy against subjective expectancy and pleasantness ratings of over 1000 chords sampled from 739 US Billboard pop songs. Bayesian model comparison revealed that listeners' expectancy and pleasantness ratings were predicted by the independent, non-overlapping, contributions of cognitive and sensory expectations. Furthermore, cognitive expectations explained over twice the variance in listeners' perceived surprise compared to sensory expectations, suggesting a larger relative importance of long-term representations of music structure over short-term sensory-acoustic information in musical expectancy. Our results thus emphasize the distinct, albeit complementary, roles of cognitive and sensory expectations in shaping musical pleasure, and suggest that this expectancy-driven mechanism depends on musical information represented at different levels of abstraction along the neural hierarchy. This article is part of the theme issue 'Art, aesthetics and predictive processing: theoretical and empirical perspectives'.
Collapse
|
72
|
Barchet AV, Rimmele JM, Pelofi C. TenseMusic: An automatic prediction model for musical tension. PLoS One 2024; 19:e0296385. [PMID: 38241238 PMCID: PMC10798497 DOI: 10.1371/journal.pone.0296385] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2023] [Accepted: 12/12/2023] [Indexed: 01/21/2024] Open
Abstract
The perception of tension and release dynamics constitutes one of the essential aspects of music listening. However, modeling musical tension to predict perception of listeners has been a challenge to researchers. Seminal work demonstrated that tension is reported consistently by listeners and can be accurately predicted from a discrete set of musical features, combining them into a weighted sum of slopes reflecting their combined dynamics over time. However, previous modeling approaches lack an automatic pipeline for feature extraction that would make them widely accessible to researchers in the field. Here, we present TenseMusic: an open-source automatic predictive tension model that operates with a musical audio as the only input. Using state-of-the-art music information retrieval (MIR) methods, it automatically extracts a set of six features (i.e., loudness, pitch height, tonal tension, roughness, tempo, and onset frequency) to use as predictors for musical tension. The algorithm was optimized using Lasso regression to best predict behavioral tension ratings collected on 38 Western classical musical pieces. Its performance was then tested by assessing the correlation between the predicted tension and unseen continuous behavioral tension ratings yielding large mean correlations between ratings and predictions approximating r = .60 across all pieces. We hope that providing the research community with this well-validated open-source tool for predicting musical tension will motivate further work in music cognition and contribute to elucidate the neural and cognitive correlates of tension dynamics for various musical genres and cultures.
Collapse
|
73
|
Massaeli F, Power SD. EEG-based hierarchical classification of level of demand and modality of auditory and visual sensory processing. J Neural Eng 2024; 21:016008. [PMID: 38176028 DOI: 10.1088/1741-2552/ad1ac1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2023] [Accepted: 01/04/2024] [Indexed: 01/06/2024]
Abstract
Objective.To date, most research on electroencephalography (EEG)-based mental workload detection for passive brain-computer interface (pBCI) applications has focused on identifying the overall level of cognitive resources required, such as whether the workload is high or low. We propose, however, that being able to determine the specific type of cognitive resources being used, such as visual or auditory, would also be useful. This would enable the pBCI to take more appropriate action to reduce the overall level of cognitive demand on the user. For example, if a high level of workload was detected and it is determined that the user is primarily engaged in visual information processing, then the pBCI could cause some information to be presented aurally instead. In our previous work we showed that EEG could be used to differentiate visual from auditory processing tasks when the level of processing is high, but the two modalities could not be distinguished when the level of cognitive processing demand was very low. The current study aims to build on this work and move toward the overall objective of developing a pBCI that is capable of predicting both the level and the type of cognitive resources being used.Approach.Fifteen individuals undertook carefully designed visual and auditory tasks while their EEG data was being recorded. In this study, we incorporated a more diverse range of sensory processing conditions including not only single-modality conditions (i.e. those requiring one of either visual or auditory processing) as in our previous study, but also dual-modality conditions (i.e. those requiring both visual and auditory processing) and no-task/baseline conditions (i.e. when the individual is not engaged in either visual or auditory processing).Main results.Using regularized linear discriminant analysis within a hierarchical classification algorithm, the overall cognitive demand was predicted with an accuracy of more than 86%, while the presence or absence of visual and auditory sensory processing were each predicted with an accuracy of approximately 70%.Significance.The findings support the feasibility of establishing a pBCI that can determine both the level and type of attentional resources required by the user at any given moment. This pBCI could assist in enhancing safety in hazardous jobs by triggering the most effective and efficient adaptation strategies when high workload conditions are detected.
Collapse
|
74
|
Dalla Bella S, Janaqi S, Benoit CE, Farrugia N, Bégel V, Verga L, Harding EE, Kotz SA. Unravelling individual rhythmic abilities using machine learning. Sci Rep 2024; 14:1135. [PMID: 38212632 PMCID: PMC10784578 DOI: 10.1038/s41598-024-51257-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2023] [Accepted: 01/02/2024] [Indexed: 01/13/2024] Open
Abstract
Humans can easily extract the rhythm of a complex sound, like music, and move to its regular beat, like in dance. These abilities are modulated by musical training and vary significantly in untrained individuals. The causes of this variability are multidimensional and typically hard to grasp in single tasks. To date we lack a comprehensive model capturing the rhythmic fingerprints of both musicians and non-musicians. Here we harnessed machine learning to extract a parsimonious model of rhythmic abilities, based on behavioral testing (with perceptual and motor tasks) of individuals with and without formal musical training (n = 79). We demonstrate that variability in rhythmic abilities and their link with formal and informal music experience can be successfully captured by profiles including a minimal set of behavioral measures. These findings highlight that machine learning techniques can be employed successfully to distill profiles of rhythmic abilities, and ultimately shed light on individual variability and its relationship with both formal musical training and informal musical experiences.
Collapse
|
75
|
Ngodup T, Irie T, Elkins SP, Trussell LO. The Na + leak channel NALCN controls spontaneous activity and mediates synaptic modulation by α2-adrenergic receptors in auditory neurons. eLife 2024; 12:RP89520. [PMID: 38197879 PMCID: PMC10945507 DOI: 10.7554/elife.89520] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/11/2024] Open
Abstract
Cartwheel interneurons of the dorsal cochlear nucleus (DCN) potently suppress multisensory signals that converge with primary auditory afferent input, and thus regulate auditory processing. Noradrenergic fibers from locus coeruleus project to the DCN, and α2-adrenergic receptors inhibit spontaneous spike activity but simultaneously enhance synaptic strength in cartwheel cells, a dual effect leading to enhanced signal-to-noise for inhibition. However, the ionic mechanism of this striking modulation is unknown. We generated a glycinergic neuron-specific knockout of the Na+ leak channel NALCN in mice and found that its presence was required for spontaneous firing in cartwheel cells. Activation of α2-adrenergic receptors inhibited both NALCN and spike generation, and this modulation was absent in the NALCN knockout. Moreover, α2-dependent enhancement of synaptic strength was also absent in the knockout. GABAB receptors mediated inhibition through NALCN as well, acting on the same population of channels as α2 receptors, suggesting close apposition of both receptor subtypes with NALCN. Thus, multiple neuromodulatory systems determine the impact of synaptic inhibition by suppressing the excitatory leak channel, NALCN.
Collapse
|