76
|
Meiser A, Lena Knoll A, Bleichner MG. High-density ear-EEG for understanding ear-centered EEG. J Neural Eng 2024; 21:016001. [PMID: 38118173 DOI: 10.1088/1741-2552/ad1783] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2023] [Accepted: 12/20/2023] [Indexed: 12/22/2023]
Abstract
Background. Mobile ear-EEG provides the opportunity to record EEG unobtrusively in everyday life. However, in real-life, the EEG data quickly becomes difficult to interpret, as the neural signal is contaminated by other, non-neural signal contributions. Due to the small number of electrodes in ear-EEG devices, the interpretation of the EEG becomes even more difficult. For meaningful and reliable ear-EEG, it is crucial that the brain signals we wish to record in real life are well-understood and that we make optimal use of the available electrodes. Their placement should be guided by prior knowledge about the characteristics of the signal of interest.Objective.We want to understand the signal we record with ear-EEG and make recommendations on how to optimally place a limited number of electrodes.Approach.We built a high-density ear-EEG with 31 channels spaced densely around one ear. We used it to record four auditory event-related potentials (ERPs): the mismatch negativity, the P300, the N100 and the N400. With this data, we gain an understanding of how different stages of auditory processing are reflected in ear-EEG. We investigate the electrode configurations that carry the most information and use a mass univariate ERP analysis to identify the optimal channel configuration. We additionally use a multivariate approach to investigate the added value of multi-channel recordings.Main results.We find significant condition differences for all ERPs. The different ERPs vary considerably in their spatial extent and different electrode positions are necessary to optimally capture each component. In the multivariate analysis, we find that the investigation of the ERPs benefits strongly from multi-channel ear-EEG.Significance.Our work emphasizes the importance of a strong theoretical and practical background when building and using ear-EEG. We provide recommendations on finding the optimal electrode positions. These results will guide future research employing ear-EEG in real-life scenarios.
Collapse
|
77
|
Davidson A, Souza P. Relationships Between Auditory Processing and Cognitive Abilities in Adults: A Systematic Review. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:296-345. [PMID: 38147487 DOI: 10.1044/2023_jslhr-22-00716] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/28/2023]
Abstract
PURPOSE The contributions from the central auditory and cognitive systems play a major role in communication. Understanding the relationship between auditory and cognitive abilities has implications for auditory rehabilitation for clinical patients. The purpose of this systematic review is to address the question, "In adults, what is the relationship between central auditory processing abilities and cognitive abilities?" METHOD Preferred Reporting Items for Systematic Reviews and Meta-Analyses guidelines were followed to identify, screen, and determine eligibility for articles that addressed the research question of interest. Medical librarians and subject matter experts assisted in search strategy, keyword review, and structuring the systematic review process. To be included, articles needed to have an auditory measure (either behavioral or electrophysiologic), a cognitive measure that assessed individual ability, and the measures needed to be compared to one another. RESULTS Following two rounds of identification and screening, 126 articles were included for full analysis. Central auditory processing (CAP) measures were grouped into categories (behavioral: speech in noise, altered speech, temporal processing, binaural processing; electrophysiologic: mismatch negativity, P50, N200, P200, and P300). The most common CAP measures were sentence recognition in speech-shaped noise and the P300. Cognitive abilities were grouped into constructs, and the most common construct was working memory. The findings were mixed, encompassing both significant and nonsignificant relationships; therefore, the results do not conclusively establish a direct link between CAP and cognitive abilities. Nonetheless, several consistent relationships emerged across different domains. Distorted or noisy speech was related to working memory or processing speed. Auditory temporal order tasks showed significant relationships with working memory, fluid intelligence, or multidomain cognitive measures. For electrophysiology, relationships were observed between some cortical evoked potentials and working memory or executive/inhibitory processes. Significant results were consistent with the hypothesis that assessments of CAP and cognitive processing would be positively correlated. CONCLUSIONS Results from this systematic review summarize relationships between CAP and cognitive processing, but also underscore the complexity of these constructs, the importance of study design, and the need to select an appropriate measure. The relationship between auditory and cognitive abilities is complex but can provide informative context when creating clinical management plans. This review supports a need to develop guidelines and training for audiologists who wish to consider individual central auditory and cognitive abilities in patient care. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.24855174.
Collapse
|
78
|
Brima T, Freedman EG, Prinsloo KD, Augustine EF, Adams HR, Wang KH, Mink JW, Shaw LH, Mantel EP, Foxe JJ. Assessing the integrity of auditory sensory memory processing in CLN3 disease (Juvenile Neuronal Ceroid Lipofuscinosis (Batten disease)): an auditory evoked potential study of the duration-evoked mismatch negativity (MMN). J Neurodev Disord 2024; 16:3. [PMID: 38183037 PMCID: PMC10770910 DOI: 10.1186/s11689-023-09515-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/25/2023] [Accepted: 12/12/2023] [Indexed: 01/07/2024] Open
Abstract
BACKGROUND We interrogated auditory sensory memory capabilities in individuals with CLN3 disease (juvenile neuronal ceroid lipofuscinosis), specifically for the feature of "duration" processing. Given decrements in auditory processing abilities associated with later-stage CLN3 disease, we hypothesized that the duration-evoked mismatch negativity (MMN) of the event related potential (ERP) would be a marker of progressively atypical cortical processing in this population, with potential applicability as a brain-based biomarker in clinical trials. METHODS We employed three stimulation rates (fast: 450 ms, medium: 900 ms, slow: 1800 ms), allowing for assessment of the sustainability of the auditory sensory memory trace. The robustness of MMN directly relates to the rate at which the regularly occurring stimulus stream is presented. As presentation rate slows, robustness of the sensory memory trace diminishes. By manipulating presentation rate, the strength of the sensory memory trace is parametrically varied, providing greater sensitivity to detect auditory cortical dysfunction. A secondary hypothesis was that duration-evoked MMN abnormalities in CLN3 disease would be more severe at slower presentation rates, resulting from greater demand on the sensory memory system. RESULTS Data from individuals with CLN3 disease (N = 21; range 6-28 years of age) showed robust MMN responses (i.e., intact auditory sensory memory processes) at the medium stimulation rate. However, at the fastest rate, MMN was significantly reduced, and at the slowest rate, MMN was not detectable in CLN3 disease relative to neurotypical controls (N = 41; ages 6-26 years). CONCLUSIONS Results reveal emerging insufficiencies in this critical auditory perceptual system in individuals with CLN3 disease.
Collapse
|
79
|
Kathios N, Sachs ME, Zhang E, Ou Y, Loui P. Generating New Musical Preferences From Multilevel Mapping of Predictions to Reward. Psychol Sci 2024; 35:34-54. [PMID: 38019607 DOI: 10.1177/09567976231214185] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2023] Open
Abstract
Much of what we know and love about music hinges on our ability to make successful predictions, which appears to be an intrinsically rewarding process. Yet the exact process by which learned predictions become pleasurable is unclear. Here we created novel melodies in an alternative scale different from any established musical culture to show how musical preference is generated de novo. Across nine studies (n = 1,185), adult participants learned to like more frequently presented items that adhered to this rapidly learned structure, suggesting that exposure and prediction errors both affected self-report liking ratings. Learning trajectories varied by music-reward sensitivity but were similar for U.S. and Chinese participants. Furthermore, functional MRI activity in auditory areas reflected prediction errors, whereas functional connectivity between auditory and medial prefrontal regions reflected both exposure and prediction errors. Collectively, results support predictive coding as a cognitive mechanism by which new musical sounds become rewarding.
Collapse
|
80
|
Schubert J, Suess N, Weisz N. Individual prediction tendencies do not generalize across modalities. Psychophysiology 2024; 61:e14435. [PMID: 37691098 PMCID: PMC10909557 DOI: 10.1111/psyp.14435] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2023] [Revised: 08/04/2023] [Accepted: 08/22/2023] [Indexed: 09/12/2023]
Abstract
Predictive processing theories, which model the brain as a "prediction machine", explain a wide range of cognitive functions, including learning, perception and action. Furthermore, it is increasingly accepted that aberrant prediction tendencies play a crucial role in psychiatric disorders. Given this explanatory value for clinical psychiatry, prediction tendencies are often implicitly conceptualized as individual traits or as tendencies that generalize across situations. As this has not yet explicitly been shown, in the current study, we quantify to what extent the individual tendency to anticipate sensory features of high probability generalizes across modalities. Using magnetoencephalography (MEG), we recorded brain activity while participants were presented with a sequence of four different (either visual or auditory) stimuli, which changed according to predefined transitional probabilities of two entropy levels: ordered vs. random. Our results show that, on a group-level, under conditions of low entropy, stimulus features of high probability are preactivated in the auditory but not in the visual modality. Crucially, the magnitude of the individual tendency to predict sensory events seems not to correlate between the two modalities. Furthermore, reliability statistics indicate poor internal consistency, suggesting that the measures from the different modalities are unlikely to reflect a single, common cognitive process. In sum, our findings suggest that quantification and interpretation of individual prediction tendencies cannot be generalized across modalities.
Collapse
|
81
|
Alwashmi K, Meyer G, Rowe F, Ward R. Enhancing learning outcomes through multisensory integration: A fMRI study of audio-visual training in virtual reality. Neuroimage 2024; 285:120483. [PMID: 38048921 DOI: 10.1016/j.neuroimage.2023.120483] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2023] [Revised: 11/18/2023] [Accepted: 12/01/2023] [Indexed: 12/06/2023] Open
Abstract
The integration of information from different sensory modalities is a fundamental process that enhances perception and performance in real and virtual environments (VR). Understanding these mechanisms, especially during learning tasks that exploit novel multisensory cue combinations provides opportunities for the development of new rehabilitative interventions. This study aimed to investigate how functional brain changes support behavioural performance improvements during an audio-visual (AV) learning task. Twenty healthy participants underwent a 30 min daily VR training for four weeks. The task was an AV adaptation of a 'scanning training' paradigm that is commonly used in hemianopia rehabilitation. Functional magnetic resonance imaging (fMRI) and performance data were collected at baseline, after two and four weeks of training, and four weeks post-training. We show that behavioural performance, operationalised as mean reaction time reduction in VR, significantly improves. In separate tests in a controlled laboratory environment, we showed that the behavioural performance gains in the VR training environment transferred to a significant mean RT reduction for the trained AV voluntary task on a computer screen. Enhancements were observed in both the visual-only and AV conditions, with the latter demonstrating a faster response time supported by the presence of audio cues. The behavioural learning effect also transfers to two additional tasks that were tested: a visual search task and an involuntary visual task. Our fMRI results reveal an increase in functional activation (BOLD signal) in multisensory brain regions involved in early-stage AV processing: the thalamus, the caudal inferior parietal lobe and cerebellum. These functional changes were only observed for the trained, multisensory, task and not for unimodal visual stimulation. Functional activation changes in the thalamus were significantly correlated to behavioural performance improvements. This study demonstrates that incorporating spatial auditory cues to voluntary visual training in VR leads to augmented brain activation changes in multisensory integration, resulting in measurable performance gains across tasks. The findings highlight the potential of VR-based multisensory training as an effective method for enhancing cognitive function and as a potentially valuable tool in rehabilitative programmes.
Collapse
|
82
|
Mandal A, Liesefeld AM, Liesefeld HR. The surprising robustness of visual search against concurrent auditory distraction. J Exp Psychol Hum Percept Perform 2024; 50:99-118. [PMID: 38236258 DOI: 10.1037/xhp0001168] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/19/2024]
Abstract
People often complain about distraction by irrelevant sounds that reportedly hamper performance on concurrent visual tasks demanding the allocation of focused attention toward relevant stimuli, such as processing street signs during driving. To study this everyday issue experimentally, we devised a cross-modal distraction paradigm, inspired by a standard visual-distraction paradigm (additional-singleton paradigm) that is highly sensitive to measure interference on the allocation of attention. In a visual-search pop-out task, participants reported whether a salient target (a tilted bar) was present or absent, while a completely irrelevant, but salient auditory distractor accompanied some trials. To our surprise, the results revealed no notable distraction on visual-search performance (controlled for speed-accuracy tradeoffs). Reliable auditory distraction failed to occur even when the distractor was a (highly salient) auditory oddball or was additionally presented with a temporal advantage of 300 ms. However, when the auditory modality was made relevant globally while maintaining its irrelevance to the visual-search task, we finally observed the expected interference effect. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
Collapse
|
83
|
Niu Y, Chen N, Zhu H, Jin J, Li G. Music-oriented auditory attention detection from electroencephalogram. Neurosci Lett 2024; 818:137534. [PMID: 37871827 DOI: 10.1016/j.neulet.2023.137534] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2023] [Revised: 10/16/2023] [Accepted: 10/19/2023] [Indexed: 10/25/2023]
Abstract
Music-oriented auditory attention detection (AAD) aims at determining which instrument in polyphonic music a listener is paying attention to by analyzing the listener's electroencephalogram (EEG). However, the existing linear models cannot effectively mimic the nonlinearity of the human brain, resulting in limited performance. Thus, a nonlinear music-oriented AAD model is proposed in this paper. Firstly, an auditory feature and a musical feature are fused to represent musical sources precisely and comprehensively. Secondly, the EEG is enhanced if music stimuli are presented in stereo. Thirdly, a neural network architecture is constructed to capture nonlinear and dynamic interactions between the EEG and auditory stimuli. Finally, the musical source most similar to the EEG in the common embedding space is identified as the attended one. Experimental results demonstrate that the proposed model outperforms all baseline models. On 1-s decision windows, it reaches accuracies of 92.6% and 81.7% under mono duo and trio stimuli, respectively. Additionally, it can be easily extended to speech-oriented AAD. This work can open up new possibilities for studies on both brain neural activity decoding and music information retrieval.
Collapse
|
84
|
Salselas I, Pereira F, Sousa E. Inducing visual attention through audiovisual stimuli: Can synchronous sound be a salient event? Perception 2024; 53:31-43. [PMID: 37872670 PMCID: PMC10798022 DOI: 10.1177/03010066231208127] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2023] [Accepted: 09/29/2023] [Indexed: 10/25/2023]
Abstract
We present an experimental research aiming to explore how spatial attention may be biased through auditory stimuli. In particular, we investigate how synchronous sound and image may affect attention and increase the saliency of the audiovisual event. We have designed and implemented an experimental study where subjects, wearing an eye-tracking system, were examined regarding their gaze toward the audiovisual stimuli being displayed. The audiovisual stimuli were specifically tailored for this experiment, consisting of videos contrasting in terms of Synch Points (i.e., moments where a visual event is associated with a visible trigger movement, synchronous with its correspondent sound). While consistency across audiovisual sensory modalities revealed to be an attention-drawing feature, when combined with synchrony, it clearly emphasized the biasing, triggering orienting, that is, focal attention towards the particular scene that contains the Synch Point. Consequently, results revealed synchrony to be a saliency factor, contributing to the strengthening of the focal attention.
Collapse
|
85
|
Barros ACMPD, Lopes RV, Gil D, Carmo ACFD, Onishi ET, Branco-Barreiro FCA. Auditory training for tinnitus treatment: a scoping review. Braz J Otorhinolaryngol 2024; 90:101361. [PMID: 38006725 PMCID: PMC10709205 DOI: 10.1016/j.bjorl.2023.101361] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Accepted: 11/05/2023] [Indexed: 11/27/2023] Open
Abstract
OBJECTIVES Our study aimed to verify the evidence of auditory training employed in the audiological treatment of tinnitus in adults and older adults. METHODS Scoping review based on a search for articles in journals available in MEDLINE (PubMed), Embase (Elsevier), LILACS (BVS), and Cochrane Library. Titles and abstracts of the retrieved articles were assessed by peers, following the eligibility criteria; they were afterward read in full text, and the references were hand searched in the results found. Studies' level of evidence was classified into very high (Level A+), high (Level A), moderate (Level B), limited (Level C), low (Level D), or very low (Level D-) based on the Critically Appraised Topics. RESULTS 2160 records were identified in the searching stage and 15 studies were eligible for data extraction. Study design, sample characterization, auditory training tasks, sound stimuli, outcome measures, and results were extracted. Frequency discrimination training was the most frequent strategy, followed by auditory attentional skills training and multisensory training. Almost all studies with daily auditory training sessions reported significant benefits demonstrated in at least one outcome measure. Studies that used auditory discrimination training and attentional auditory skill stimulation to treat tinnitus obtained quality evidence levels ranging from limited to high (C‒A) and studies that applied multisensory training or attentional training combined with counseling and passive listening in tinnitus patients reached a high-quality evidence level (A). CONCLUSION Recent studies had higher levels of evidence and considered attentional factors and multisensory pathways in auditory training strategies.
Collapse
|
86
|
Nave K, Carrillo C, Jacoby N, Trainor L, Hannon E. The development of rhythmic categories as revealed through an iterative production task. Cognition 2024; 242:105634. [PMID: 37820488 DOI: 10.1016/j.cognition.2023.105634] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2023] [Revised: 10/03/2023] [Accepted: 10/04/2023] [Indexed: 10/13/2023]
Abstract
Both humans and non-humans (e.g. birds and primates) preferentially produce and perceive auditory rhythms with simple integer ratios. In addition, these preferences (biases) tend to reflect specific integer-ratio rhythms that are common to one's cultural listening experience. To better understand the developmental trajectory of these biases, we estimated children's rhythm biases across the entire rhythm production space of simple (e.g., ratios of 1, 2, and 3) three-interval rhythms. North American children aged 6-11 years completed an iterative rhythm production task, in which they attempted to tap in synchrony with repeating three-interval rhythms chosen randomly from the space. For each rhythm, the child's produced rhythm was presented back to them as the stimulus, and over the course of 5 such iterations we used their final reproductions to estimate their rhythmic biases or priors. Results suggest that regardless of the initial rhythm, after 5 iterations, children's tapping converged on rhythms with (nearly) simple integer ratios, indicating that, like adults, their rhythmic priors consist of rhythms with simple-integer ratios. Furthermore, the relative weights (or prominence of different rhythmic priors) observed in children were highly correlated with those of adults. However, we also observed some age-related changes, especially for the ratio types that vary most across cultures. In an additional rhythm perception task, children were better at detecting rhythmic disruptions to a culturally familiar rhythm (in 4/4 m with 2:1:1 ratio pattern) than to a culturally unfamiliar rhythm (7/8 m with 3:2:2 ratios), and performance in this task was correlated with tapping variability in the iterative task. Taken together, our findings provide evidence that children as young as 6-years-old exhibit simple integer-ratio categorical rhythm priors in their rhythm production that closely resemble those of adults in the same culture.
Collapse
|
87
|
Vickers DA, Moore BCJ. Editorial: Cochlear Implants and Music. Trends Hear 2024; 28:23312165241231685. [PMID: 38361376 PMCID: PMC10874149 DOI: 10.1177/23312165241231685] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2024] [Accepted: 01/24/2024] [Indexed: 02/17/2024] Open
|
88
|
Cogen T, Cetin Kara H, Kara E, Telci F, Yener HM. Investigation of the relationship between hyperacusis and auditory processing difficulties in individuals with normal hearing. Eur Arch Otorhinolaryngol 2024; 281:469-477. [PMID: 37819548 DOI: 10.1007/s00405-023-08269-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2023] [Accepted: 09/26/2023] [Indexed: 10/13/2023]
Abstract
PURPOSE This study aimed to investigate the prevalence of hyperacusis and auditory processing (listening) difficulties (APDiff) in normal-hearing young adults and to explore the relationship between these two conditions. METHODS A total of 333 participants aged 18-24 underwent pure-tone audiometry tests and completed the Khalfa Hyperacusis Questionnaire (KHQ) and the University of Cincinnati Auditory Processing Inventory (UCAPI) online. RESULTS Hyperacusis was identified in 26.43% of participants, and APDiff in 36.04%. A strong, positive correlation was found between KHQ and UCAPI scores (r = 0.603, p < 0.001). Logistic regression models indicated that hyperacusis significantly increased the likelihood of having APDiff (OR 7.688, p < 0.001). CONCLUSION Despite the high prevalence of hyperacusis and APDiff in young adults, few seek help from audiology clinics. Our findings highlight a significant link between hyperacusis and APDiff, emphasizing the need for screening protocols for auditory processing skills in young adults due to these conditions' potential social and academic impacts.
Collapse
|
89
|
van der Willigen RF, Versnel H, van Opstal AJ. Spectral-temporal processing of naturalistic sounds in monkeys and humans. J Neurophysiol 2024; 131:38-63. [PMID: 37965933 DOI: 10.1152/jn.00129.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2023] [Revised: 10/23/2023] [Accepted: 11/13/2023] [Indexed: 11/16/2023] Open
Abstract
Human speech and vocalizations in animals are rich in joint spectrotemporal (S-T) modulations, wherein acoustic changes in both frequency and time are functionally related. In principle, the primate auditory system could process these complex dynamic sounds based on either an inseparable representation of S-T features or, alternatively, a separable representation. The separability hypothesis implies an independent processing of spectral and temporal modulations. We collected comparative data on the S-T hearing sensitivity in humans and macaque monkeys to a wide range of broadband dynamic spectrotemporal ripple stimuli employing a yes-no signal-detection task. Ripples were systematically varied, as a function of density (spectral modulation frequency), velocity (temporal modulation frequency), or modulation depth, to cover a listener's full S-T modulation sensitivity, derived from a total of 87 psychometric ripple detection curves. Audiograms were measured to control for normal hearing. Determined were hearing thresholds, reaction time distributions, and S-T modulation transfer functions (MTFs), both at the ripple detection thresholds and at suprathreshold modulation depths. Our psychophysically derived MTFs are consistent with the hypothesis that both monkeys and humans employ analogous perceptual strategies: S-T acoustic information is primarily processed separable. Singular value decomposition (SVD), however, revealed a small, but consistent, inseparable spectral-temporal interaction. Finally, SVD analysis of the known visual spatiotemporal contrast sensitivity function (CSF) highlights that human vision is space-time inseparable to a much larger extent than is the case for S-T sensitivity in hearing. Thus, the specificity with which the primate brain encodes natural sounds appears to be less strict than is required to adequately deal with natural images.NEW & NOTEWORTHY We provide comparative data on primate audition of naturalistic sounds comprising hearing thresholds, reaction time distributions, and spectral-temporal modulation transfer functions. Our psychophysical experiments demonstrate that auditory information is primarily processed in a spectral-temporal-independent manner by both monkeys and humans. Singular value decomposition of known visual spatiotemporal contrast sensitivity, in comparison to our auditory spectral-temporal sensitivity, revealed a striking contrast in how the brain encodes natural sounds as opposed to natural images, as vision appears to be space-time inseparable.
Collapse
|
90
|
Torres NL, Castro SL, Silva S. Visual movement impairs duration discrimination at short intervals. Q J Exp Psychol (Hove) 2024; 77:57-69. [PMID: 36717537 PMCID: PMC10712207 DOI: 10.1177/17470218231156542] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2022] [Revised: 01/19/2023] [Accepted: 01/24/2023] [Indexed: 02/01/2023]
Abstract
The classic advantage of audition over vision in time processing has been recently challenged by studies using continuously moving visual stimuli such as bouncing balls. Bouncing balls drive beat-based synchronisation better than static visual stimuli (flashes) and as efficiently as auditory ones (beeps). It is yet unknown how bouncing balls modulate performance in duration perception. Our previous study addressing this was inconclusive: there were no differences among bouncing balls, flashes, and beeps, but this could have been due to the fact that intervals were too long to allow sensitivity to modality (visual vs auditory). In this study, we conducted a first experiment to determine whether shorter intervals elicit cross-stimulus differences. We found that short (mean 157 ms) but not medium (326 ms) intervals made duration perception worse for bouncing balls compared with flashes and beeps. In a second experiment, we investigated whether the lower efficiency of bouncing balls was due to experimental confounds, lack of realism, or movement. We ruled out the experimental confounds and found support for the hypothesis that visual movement-be it continuous or discontinuous-impairs duration perception at short interval lengths. Therefore, unlike beat-based synchronisation, duration perception does not benefit from continuous visual movement, which may even have a detrimental effect at short intervals.
Collapse
|
91
|
Yang X, Ying C, Zhu L, Wenjing W. The neural oscillations in delta- and theta-bands contribute to divided attention in audiovisual integration. Perception 2024; 53:44-60. [PMID: 37899595 DOI: 10.1177/03010066231208539] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/31/2023]
Abstract
One of key mechanisms implicated in multisensory processing is neural oscillations in distinct frequency band. Many studies explored the modulation of attention by recording the electroencephalography signals when subjects attended one modality, and ignored the other modality input. However, when attention is directed toward one modality, it may be not always possible to shut out completely inputs from a different modality. Since many situations require division of attention between audition and vision, it is imperative to investigate the neural mechanisms underlying processing of concurrent auditory and visual sensory streams. In the present study, we designed a task of audiovisual semantic discrimination, in which the subjects were asked to share attention to both auditory and visual stimuli. We explored the contribution of neural oscillations in lower-frequency to the modulation of divided attention on audiovisual integration. Our results implied that theta-band activity contributes to the early modulation of divided attention, and delta-band activity contributes to the late modulation of divided attention to audiovisual integration. Moreover, the fronto-central delta- and theta-bands activity is likely a marker of divided attention in audiovisual integration, and the neural oscillation on delta- and theta-bands is conducive to allocating attention resources to dual-tasking involving task-coordinating abilities.
Collapse
|
92
|
Sulfaro AA, Robinson AK, Carlson TA. Properties of imagined experience across visual, auditory, and other sensory modalities. Conscious Cogn 2024; 117:103598. [PMID: 38086154 DOI: 10.1016/j.concog.2023.103598] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2023] [Revised: 10/13/2023] [Accepted: 10/23/2023] [Indexed: 01/16/2024]
Abstract
Little is known about the perceptual characteristics of mental images nor how they vary across sensory modalities. We conducted an exhaustive survey into how mental images are experienced across modalities, mainly targeting visual and auditory imagery of a single stimulus, the letter "O", to facilitate direct comparisons. We investigated temporal properties of mental images (e.g. onset latency, duration), spatial properties (e.g. apparent location), effort (e.g. ease, spontaneity, control), movement requirements (e.g. eye movements), real-imagined interactions (e.g. inner speech while reading), beliefs about imagery norms and terminologies, as well as respondent confidence. Participants also reported on the five traditional senses and their prominence during thinking, imagining, and dreaming. Overall, visual and auditory experiences dominated mental events, although auditory mental images were superior to visual mental images on almost every metric tested except regarding spatial properties. Our findings suggest that modality-specific differences in mental imagery may parallel those of other sensory neural processes.
Collapse
|
93
|
Alves MDC, Mancini PC, Teixeira LC. Modifications of auditory feedback and its effects on the voice of adult subjects: a scoping review. Codas 2023; 36:e20220202. [PMID: 38126424 PMCID: PMC10750862 DOI: 10.1590/2317-1782/20232022202pt] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2022] [Accepted: 05/29/2023] [Indexed: 12/23/2023] Open
Abstract
INTRODUCTION The auditory perception of voice and its production involve auditory feedback, kinesthetic cues and the feedforward system that produce different effects for the voice. The Lombard, Sidetone and Pitch-Shift-Reflex effects are the most studied. The mapping of scientific experiments on changes in auditory feedback for voice motor control makes it possible to examine the existing literature on the phenomenon and may contribute to voice training or therapies. PURPOSE To map experiments and research results with manipulation of auditory feedback for voice motor control in adults. METHOD Scope review following the Checklist Preferred Reporting Items for Systematic reviews and Meta-Analyses extension (PRISMA-ScR) to answer the question: "What are the investigation methods and main research findings on the manipulation of auditory feedback in voice self-monitoring of adults?". The search protocol was based on the Population, Concept, and Context (PCC) mnemonic strategy, in which the population is adult individuals, the concept is the manipulation of auditory feedback and the context is on motor voice control. Articles were searched in the databases: BVS/Virtual Health Library, MEDLINE/Medical Literature Analysis and Retrieval System online, COCHRANE, CINAHL/Cumulative Index to Nursing and Allied Health Literature, SCOPUS and WEB OF SCIENCE. RESULTS 60 articles were found, 19 on the Lombard Effect, 25 on the Pitch-shift-reflex effect, 12 on the Sidetone effect and four on the Sidetone/Lombard effect. The studies are in agreement that the insertion of a noise that masks the auditory feedback causes an increase in the individual's speech intensity and that the amplification of the auditory feedback promotes the reduction of the sound pressure level in the voice production. A reflex response to the change in pitch is observed in the auditory feedback, however, with particular characteristics in each study. CONCLUSION The material and method of the experiments are different, there are no standardizations in the tasks, the samples are varied and often reduced. The methodological diversity makes it difficult to generalize the results. The main findings of research on auditory feedback on voice motor control confirm that in the suppression of auditory feedback, the individual tends to increase the intensity of the voice. In auditory feedback amplification, the individual decreases the intensity and has greater control over the fundamental frequency, and in frequency manipulations, the individual tends to correct the manipulation. The few studies with dysphonic individuals show that they behave differently from non-dysphonic individuals.
Collapse
|
94
|
Meyer L, Rachman L, Araiza-Illan G, Gaudrain E, Başkent D. Use of a humanoid robot for auditory psychophysical testing. PLoS One 2023; 18:e0294328. [PMID: 38091272 PMCID: PMC10718414 DOI: 10.1371/journal.pone.0294328] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2023] [Accepted: 10/31/2023] [Indexed: 12/18/2023] Open
Abstract
Tasks in psychophysical tests can at times be repetitive and cause individuals to lose engagement during the test. To facilitate engagement, we propose the use of a humanoid NAO robot, named Sam, as an alternative interface for conducting psychophysical tests. Specifically, we aim to evaluate the performance of Sam as an auditory testing interface, given its potential limitations and technical differences, in comparison to the current laptop interface. We examine the results and durations of two voice perception tests, voice cue sensitivity and voice gender categorisation, obtained from both the conventionally used laptop interface and Sam. Both tests investigate the perception and use of two speaker-specific voice cues, fundamental frequency (F0) and vocal tract length (VTL), important for characterising voice gender. Responses are logged on the laptop using a connected mouse, and on Sam using the tactile sensors. Comparison of test results from both interfaces shows functional similarity between the interfaces and replicates findings from previous studies with similar tests. Comparison of test durations shows longer testing times with Sam, primarily due to longer processing times in comparison to the laptop, as well as other design limitations due to the implementation of the test on the robot. Despite the inherent constraints of the NAO robot, such as in sound quality, relatively long processing and testing times, and different methods of response logging, the NAO interface appears to facilitate collecting similar data to the current laptop interface, confirming its potential as an alternative psychophysical test interface for auditory perception tests.
Collapse
|
95
|
Thibault N, Albouy P, Grondin S. Distinct brain dynamics and networks for processing short and long auditory time intervals. Sci Rep 2023; 13:22018. [PMID: 38086944 PMCID: PMC10716402 DOI: 10.1038/s41598-023-49562-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2023] [Accepted: 12/09/2023] [Indexed: 12/18/2023] Open
Abstract
Psychophysical studies suggest that time intervals above and below 1.2 s are processed differently in the human brain. However, the neural underpinnings of this dissociation remain unclear. Here, we investigate whether distinct or common brain networks and dynamics support the passive perception of short (below 1.2 s) and long (above 1.2 s) empty time intervals. Twenty participants underwent an EEG recording during an auditory oddball paradigm with .8- and 1.6-s standard time intervals and deviant intervals either shorter (early) or longer (delayed) than the standard interval. We computed the auditory ERPs for each condition at the sensor and source levels. We then performed whole brain cluster-based permutation statistics for the CNV, N1 and P2, components, testing deviants against standards. A CNV was found only for above 1.2 s intervals (delayed deviants), with generators in temporo-parietal, SMA, and motor regions. Deviance detection of above 1.2 s intervals occurred during the N1 period over fronto-central sensors for delayed deviants only, with generators in parietal and motor regions. Deviance detection of below 1.2 s intervals occurred during the P2 period over fronto-central sensors for delayed deviants only, with generators in primary auditory cortex, SMA, IFG, cingulate and parietal cortex. We then identified deviance related changes in directed connectivity using bivariate Granger causality to highlight the networks dynamics associated with interval processing above and below 1.2. These results suggest that distinct brain dynamics and networks support the perception of time intervals above and below 1.2 s.
Collapse
|
96
|
Park WJ, Fine I. The perception of auditory motion in sighted and early blind individuals. Proc Natl Acad Sci U S A 2023; 120:e2310156120. [PMID: 38015842 PMCID: PMC10710053 DOI: 10.1073/pnas.2310156120] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2023] [Accepted: 10/29/2023] [Indexed: 11/30/2023] Open
Abstract
Motion perception is a fundamental sensory task that plays a critical evolutionary role. In vision, motion processing is classically described using a motion energy model with spatiotemporally nonseparable filters suited for capturing the smooth continuous changes in spatial position over time afforded by moving objects. However, it is still not clear whether the filters underlying auditory motion discrimination are also continuous motion detectors or infer motion from comparing discrete sound locations over time (spatiotemporally separable). We used a psychophysical reverse correlation paradigm, where participants discriminated the direction of a motion signal in the presence of spatiotemporal noise, to determine whether the filters underlying auditory motion discrimination were spatiotemporally separable or nonseparable. We then examined whether these auditory motion filters were altered as a result of early blindness. We found that both sighted and early blind individuals have separable filters. However, early blind individuals show increased sensitivity to auditory motion, with reduced susceptibility to noise and filters that were more accurate in detecting motion onsets/offsets. Model simulations suggest that this reliance on separable filters is optimal given the limited spatial resolution of auditory input.
Collapse
|
97
|
Silva JDM, Silva BCS, Lopes NBF, Jacob RTDS, Moret ALM. Effectiveness of Computerized Auditory Training on Speech Perception in Children With Hearing Loss: A Systematic Review. Am J Audiol 2023; 32:990-1004. [PMID: 37983178 DOI: 10.1044/2023_aja-23-00078] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2023] Open
Abstract
PURPOSE The purpose of this study was to investigate the effectiveness of using computerized auditory training in the auditory perception of speech in children who use electronic devices through a systematic review of the literature. METHOD This is a systematic review based on the elaboration of the following guiding question: In hearing-impaired children who use electronic devices, how effective is the use of computerized auditory training in the auditory perception of speech? The Cochrane Library, Embase, LILACS, PubMed, Web of Science, and Scopus databases and gray literature (Google Scholar and ProQuest) were consulted. RESULTS Nine hundred and eighteen records were identified, with additional two studies in the citation search, totaling 920 studies. After excluding the 184 duplicate references, using the Rayyan platform, the titles and abstracts were screened (n = 736), with 20 studies selected and considered for full reading. Six of the 20 studies were included in the systematic review of the literature. CONCLUSIONS Positive and beneficial results regarding computerized auditory training are evidenced, which describe and discuss the proof of the effectiveness of the positive effects on the auditory perception of speech in children using hearing aids and cochlear implants with regard to auditory detection skills, discrimination, recognition, and auditory memory. The positive results are inspiring for the continuity and expansion of studies that further investigate the use of computerized auditory training with contemporary technologies and have an impact on the improvement of access to specialized intervention.
Collapse
|
98
|
Repetto C, Rodella C, Conca F, Santi GC, Catricalà E. The Italian Sensorimotor Norms: Perception and action strength measures for 959 words. Behav Res Methods 2023; 55:4035-4047. [PMID: 36307624 PMCID: PMC10700458 DOI: 10.3758/s13428-022-02004-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 10/06/2022] [Indexed: 11/08/2022]
Abstract
Neuroscience research has provided evidence that semantic information is stored in a distributed brain network involved in sensorimotor and linguistic processing. More specifically, according to the embodied cognition accounts, the representation of concepts is deemed as grounded in our bodily states. For these reasons, normative measures of words should provide relevant information about the extent to which each word embeds perceptual and action properties. In the present study, we collected ratings for 959 Italian nouns and verbs from 398 volunteers, recruited via an online platform. The words were mostly taken from the Italian adaptation of the Affective Norms for English Words (ANEW). A pool of 145 verbs was added to the original set. All the words were rated on 11 sensorimotor dimensions: six perceptual modalities (vision, audition, taste, smell, touch, and interoception) and five effectors (hand-arm, foot-leg, torso, mouth, head). The new verbs were also rated on the ANEW dimensions. Results showed good reliability and consistency with previous studies. Relations between perceptual and motor dimensions are described and interpreted, along with relations between the sensorimotor and the affective dimensions. The currently developed dataset represents an important novelty, as it includes different word classes, i.e., both nouns and verbs, and integrates ratings of both sensorimotor and affective dimensions, along with other psycholinguistic parameters; all features only partially accomplished in previous studies.
Collapse
|
99
|
Ozses M, Ozbal Batuk M, Cicek Cinar B. Evaluation of auditory brainstem implant (ABI) users' auditory behavior in everyday life. Eur Arch Otorhinolaryngol 2023; 280:5299-5305. [PMID: 37272952 DOI: 10.1007/s00405-023-08046-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2023] [Accepted: 05/26/2023] [Indexed: 06/06/2023]
Abstract
PURPOSE This study aims to evaluate the everyday listening status of pediatric ABI users using the Turkish ABEL questionnaire. METHODS The study included 33 parents of children with auditory brainstem implant, and 28 parents of children with cochlear implant were included as a control group. All implant users were between the ages of 4-14. Parents answered the ABEL questionnaire to assess their child's auditory behavior in their daily living environment. In addition, Categories of Auditory Performance (CAP) and Speech Intelligibility Rating (SIR) scales were used to evaluate speech perception and production. RESULTS Auditory-aural, auditory awareness and ABEL total score of ABI users were statistically significantly lower than the CI group (p < 0.05). There was no statistically significant difference between the groups in the Conversational/Social skills subgroups. It was found that as the duration of ABI use increased, auditory-verbal, social skills and total scores increased significantly. In addition, correlations were obtained between ABEL total and subscale scores and CAP and SIR scores. CONCLUSIONS Parents believe that their children adapt nicely to ABI and are aware of environmental sounds. This study reveals the auditory, aural, and social skills of children using ABI through the regards of their parents. This study showed that the ABEL questionnaire, which was used in previous studies to express parental views of children with hearing aids and cochlear implants, can also be used for parents of children using ABI.
Collapse
|
100
|
Watanabe K, Srinivasan R, Richards VM. Effects of sound segregation cues on multi-sound intensity discrimination. JASA EXPRESS LETTERS 2023; 3:124401. [PMID: 38038677 PMCID: PMC10694800 DOI: 10.1121/10.0022559] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/14/2023] [Accepted: 11/13/2023] [Indexed: 12/02/2023]
Abstract
The effects of sound segregation cues on the sensitivity to intensity increments were explored. Listeners indicated whether the second and fourth sounds (harmonic complexes) within a five-sound sequence were increased in intensity. The target sound had a fundamental frequency of 250 Hz. In different conditions, nontarget sounds had different fundamental frequencies, different spectral shapes, and unique frequency regions relative to the target. For targets more intense than nontargets, nontarget characteristics did not affect thresholds. For targets less intense than the nontargets, thresholds improved when the targets and nontargets had unique frequency regions.
Collapse
|