1
|
De Brigard F. Episodic memory without autonoetic consciousness. Philos Trans R Soc Lond B Biol Sci 2024; 379:20230410. [PMID: 39278243 PMCID: PMC11496718 DOI: 10.1098/rstb.2023.0410] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2024] [Revised: 07/08/2024] [Accepted: 07/19/2024] [Indexed: 09/18/2024] Open
Abstract
Ever since Tulving's influential 1985 article 'Memory and consciousness', it has become traditional to think of autonoetic consciousness as necessary for episodic memory. This paper questions this claim. Specifically, it argues that the construct of autonoetic consciousness lacks validity and that, even if it was valid, it would still not be necessary for episodic memory. The paper ends with a proposal to go back to a functional/computational characterization of episodic memory in which its characteristic phenomenology is a contingent feature of the retrieval process and, as a result, open to empirical scrutiny. The proposal also dovetails with recent taxonomies of memory that are independent of conscious awareness and suggests strategies to evaluate within- and between-individual variability in the conscious experience of episodic memories in human and non-human agents. This article is part of the theme issue 'Elements of episodic memory: lessons from 40 years of research'.
Collapse
|
2
|
Zhang H, Xie J, Tao Q, Ge Z, Xiong Y, Xu G, Li M, Han C. The effect of rhythmic stimuli with spatial information on sensorimotor synchronization: an EEG and EMG study. Front Neurosci 2024; 18:1448051. [PMID: 39429702 PMCID: PMC11486764 DOI: 10.3389/fnins.2024.1448051] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2024] [Accepted: 09/16/2024] [Indexed: 10/22/2024] Open
Abstract
Introduction Sensorimotor synchronization (SMS) is the human ability to align body movement rhythms with external rhythmic stimuli. While the effects of rhythmic stimuli containing only temporal information on SMS have been extensively studied, less is known about how spatial information affects SMS performance. This study investigates the neural mechanisms underlying SMS with rhythmic stimuli that include both temporal and spatial information, providing insights into the influence of these factors across different sensory modalities. Methods This study compared the effects temporal information and spatial information on SMS performance across different stimuli conditions. We simultaneously recorded the electroencephalogram (EEG), the electromyogram (EMG), and behavioral data as subjects performed synchronized tapping to rhythmic stimuli. The study analyzed SMS performance under conditions including auditory, visual, and auditory-visual motion stimuli (containing both temporal and spatial information), as well as auditory, visual, and auditory-visual non-motion stimuli (containing only temporal information). Specifically, the research examined behavioral data (i.e., mean asynchrony, absolute asynchrony, and variability), neural oscillations, cortico-muscular coherence (CMC), and brain connectivity. Results The results demonstrated that SMS performance was superior with rhythmic stimuli containing both temporal and spatial information compared to stimuli with only temporal information. Moreover, sensory-motor neural entrainment was stronger during SMS with rhythmic stimuli containing spatial information within the same sensory modality. SMS with both types of rhythmic stimuli was found to be dynamically modulated by neural oscillations and cortical-muscular coupling in the beta band (13-30 Hz). Discussion These findings provide deeper insights into the combined effects of temporal and spatial information, as well as sensory modality, on SMS performance. The study highlights the dynamic modulation of SMS by neural oscillations and CMC, particularly in the beta band, offering valuable contributions to understanding the neural basis of sensorimotor synchronization.
Collapse
Affiliation(s)
- Huanqing Zhang
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, China
| | - Jun Xie
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, China
- School of Mechanical Engineering, Xinjiang University, Ürümqi, China
| | - Qing Tao
- School of Mechanical Engineering, Xinjiang University, Ürümqi, China
| | - Zengle Ge
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, China
| | - Yu Xiong
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, China
| | - Guanghua Xu
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, China
| | - Min Li
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, China
| | - Chengcheng Han
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, China
| |
Collapse
|
3
|
Grassi M, Felline A, Orlandi N, Toffanin M, Goli GP, Senyuva HA, Migliardi M, Contemori G. PSYCHOACOUSTICS-WEB: A free online tool for the estimation of auditory thresholds. Behav Res Methods 2024; 56:7465-7481. [PMID: 38709452 PMCID: PMC11362506 DOI: 10.3758/s13428-024-02430-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/12/2024] [Indexed: 05/07/2024]
Abstract
PSYCHOACOUSTICS-WEB is an online tool written in JavaScript and PHP that enables the estimation of auditory sensory thresholds via adaptive threshold tracking. The toolbox implements the transformed up-down methods proposed by Levitt (Journal of the Acoustical Society of America, 49, 467-477, (1971) for a set of classic psychoacoustical tasks: frequency, intensity, and duration discrimination of pure tones; duration discrimination and gap detection of noise; and amplitude modulation detection with noise carriers. The toolbox can be used through a common web browser; it works with both fixed and mobile devices, and requires no programming skills. PSYCHOACOUSTICS-WEB is suitable for laboratory, classroom, and online testing and is designed for two main types of users: an occasional user and, above all, an experimenter using the toolbox for their own research. This latter user can create a personal account, customise existing experiments, and share them in the form of direct links to further users (e.g., the participants of a hypothetical experiment). Finally, because data storage is centralised, the toolbox offers the potential for creating a database of auditory skills.
Collapse
Affiliation(s)
- Massimo Grassi
- Deparment of General Psychology, University of Padua, Via Venezia 8, 35131, Padua, Italy.
| | - Andrea Felline
- Department of Information Engineering, University of Padua, Padua, Italy
| | - Niccolò Orlandi
- Department of Information Engineering, University of Padua, Padua, Italy
| | - Mattia Toffanin
- Department of Information Engineering, University of Padua, Padua, Italy
| | - Gnana Prakash Goli
- Department of Information Engineering, University of Padua, Padua, Italy
| | - Hurcan Andrei Senyuva
- Deparment of General Psychology, University of Padua, Via Venezia 8, 35131, Padua, Italy
| | - Mauro Migliardi
- Department of Information Engineering, University of Padua, Padua, Italy
| | - Giulio Contemori
- Deparment of General Psychology, University of Padua, Via Venezia 8, 35131, Padua, Italy
| |
Collapse
|
4
|
Pinto JD, Temp DA, Ferreira L, Souza AEHD, Garcia MV, Andrade AND, Biaggio EPV. Mismatch Negativity in Children with Deficits in Auditory Abilities. Int Arch Otorhinolaryngol 2024; 28:e561-e567. [PMID: 39464353 PMCID: PMC11511459 DOI: 10.1055/s-0044-1785458] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2023] [Accepted: 01/16/2024] [Indexed: 10/29/2024] Open
Abstract
Introduction Mismatch negativity (MMN) represents a negative component of event-related potentials, which is mentioned by guidelines as an important tool to provide measurable data regarding the functionality of the auditory system in acoustic processing. However, the literature still lacks reliable data that can support the clinical use of this potential in the complementary diagnosis of central auditory processing (CAP) disorder (CAPD). Objectives To analyze whether MMN assessment might be associated with the CAP behavioral test battery, as well as to assess the effects of auditory ability deficits on MMN responses in the pediatric population. Methods In total, 45 age-matched children participated in the study. They were submitted to the CAP behavior assessment and to MMN. The children were tested with a combination of speech contrast consisting of acoustic syllables [da] versus [ta], governed by the oddball paradigm. Results Mismatch negativity did not show a direct association with a single test but with the combination of the four tests used as a behavioral test battery to identify CAPD. The results also indicated that the auditory ability deficits influenced the measurement of MMN latency ( p = 0.003*), but not the amplitude ( p = 0.857) or the area ( p = 0.577). Conclusion Mismatch negativity was shown to be statistically associated with the battery of tests used to identify deficits in auditory abilities in the studied sample rather than with a single behavioral test. The deficits in auditory abilities were observed in the MMN latency. Mismatch negativity can be used to assess children with CAPD.
Collapse
Affiliation(s)
- Julia Dalcin Pinto
- Department of Speech Therapy, Centro de Ciências da Saúde, Universidade Federal de Santa Maria (UFSM), Santa Maria, RS, Brazil
| | - Déborah Aurélio Temp
- Department of Speech Therapy, Centro de Ciências da Saúde, Universidade Federal de Santa Maria (UFSM), Santa Maria, RS, Brazil
| | - Laís Ferreira
- Department of Speech Therapy, Centro de Ciências da Saúde, Universidade Federal de Santa Maria (UFSM), Santa Maria, RS, Brazil
| | - Amália El Hatal de Souza
- Department of Speech Therapy, Centro de Ciências da Saúde, Universidade Federal de Santa Maria (UFSM), Santa Maria, RS, Brazil
| | - Michele Vargas Garcia
- Department of Speech Therapy, Centro de Ciências da Saúde, Universidade Federal de Santa Maria (UFSM), Santa Maria, RS, Brazil
| | - Adriana Neves de Andrade
- Department of Speech Therapy, Escola Paulista de Medicina, Universidade Federal de São Paulo, São Paulo, SP, Brazil
| | - Eliara Pinto Vieira Biaggio
- Department of Speech Therapy, Centro de Ciências da Saúde, Universidade Federal de Santa Maria (UFSM), Santa Maria, RS, Brazil
| |
Collapse
|
5
|
Zhao X, Yang X. Aging affects auditory contributions to focus perception in Jianghuai Mandarina). THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 155:2990-3004. [PMID: 38717206 DOI: 10.1121/10.0025928] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/13/2023] [Accepted: 04/20/2024] [Indexed: 09/20/2024]
Abstract
Speakers can place their prosodic prominence on any locations within a sentence, generating focus prosody for listeners to perceive new information. This study aimed to investigate age-related changes in the bottom-up processing of focus perception in Jianghuai Mandarin by clarifying the perceptual cues and the auditory processing abilities involved in the identification of focus locations. Young, middle-aged, and older speakers of Jianghuai Mandarin completed a focus identification task and an auditory perception task. The results showed that increasing age led to a decrease in listeners' accuracy rate in identifying focus locations, with all participants performing the worst when dynamic pitch cues were inaccessible. Auditory processing abilities did not predict focus perception performance in young and middle-aged listeners but accounted significantly for the variance in older adults' performance. These findings suggest that age-related deteriorations in focus perception can be largely attributed to declined auditory processing of perceptual cues. Poor ability to extract frequency modulation cues may be the most important underlying psychoacoustic factor for older adults' difficulties in perceiving focus prosody in Jianghuai Mandarin. The results contribute to our understanding of the bottom-up mechanisms involved in linguistic prosody processing in aging adults, particularly in tonal languages.
Collapse
Affiliation(s)
- Xinxian Zhao
- School of Foreign Studies, Tongji University, Shanghai 200092, China
| | - Xiaohu Yang
- School of Foreign Studies, Tongji University, Shanghai 200092, China
| |
Collapse
|
6
|
Caprini F, Zhao S, Chait M, Agus T, Pomper U, Tierney A, Dick F. Generalization of auditory expertise in audio engineers and instrumental musicians. Cognition 2024; 244:105696. [PMID: 38160651 DOI: 10.1016/j.cognition.2023.105696] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2021] [Revised: 12/04/2023] [Accepted: 12/13/2023] [Indexed: 01/03/2024]
Abstract
From auditory perception to general cognition, the ability to play a musical instrument has been associated with skills both related and unrelated to music. However, it is unclear if these effects are bound to the specific characteristics of musical instrument training, as little attention has been paid to other populations such as audio engineers and designers whose auditory expertise may match or surpass that of musicians in specific auditory tasks or more naturalistic acoustic scenarios. We explored this possibility by comparing students of audio engineering (n = 20) to matched conservatory-trained instrumentalists (n = 24) and to naive controls (n = 20) on measures of auditory discrimination, auditory scene analysis, and speech in noise perception. We found that audio engineers and performing musicians had generally lower psychophysical thresholds than controls, with pitch perception showing the largest effect size. Compared to controls, audio engineers could better memorise and recall auditory scenes composed of non-musical sounds, whereas instrumental musicians performed best in a sustained selective attention task with two competing streams of tones. Finally, in a diotic speech-in-babble task, musicians showed lower signal-to-noise-ratio thresholds than both controls and engineers; however, a follow-up online study did not replicate this musician advantage. We also observed differences in personality that might account for group-based self-selection biases. Overall, we showed that investigating a wider range of forms of auditory expertise can help us corroborate (or challenge) the specificity of the advantages previously associated with musical instrument training.
Collapse
Affiliation(s)
- Francesco Caprini
- Department of Psychological Sciences, Birkbeck, University of London, UK.
| | - Sijia Zhao
- Department of Experimental Psychology, University of Oxford, UK
| | - Maria Chait
- University College London (UCL) Ear Institute, UK
| | - Trevor Agus
- School of Arts, English and Languages, Queen's University Belfast, UK
| | - Ulrich Pomper
- Department of Cognition, Emotion, and Methods in Psychology, Universität Wien, Austria
| | - Adam Tierney
- Department of Psychological Sciences, Birkbeck, University of London, UK
| | - Fred Dick
- Department of Experimental Psychology, University College London (UCL), UK
| |
Collapse
|
7
|
Symons A, Dick F, Tierney A. Salient sounds distort time perception and production. Psychon Bull Rev 2024; 31:137-147. [PMID: 37430179 PMCID: PMC10866776 DOI: 10.3758/s13423-023-02305-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 05/03/2023] [Indexed: 07/12/2023]
Abstract
The auditory world is often cacophonous, with some sounds capturing attention and distracting us from our goals. Despite the universality of this experience, many questions remain about how and why sound captures attention, how rapidly behavior is disrupted, and how long this interference lasts. Here, we use a novel measure of behavioral disruption to test predictions made by models of auditory salience. Models predict that goal-directed behavior is disrupted immediately after points in time that feature a high degree of spectrotemporal change. We find that behavioral disruption is precisely time-locked to the onset of distracting sound events: Participants who tap to a metronome temporarily increase their tapping speed 750 ms after the onset of distractors. Moreover, this response is greater for more salient sounds (larger amplitude) and sound changes (greater pitch shift). We find that the time course of behavioral disruption is highly similar after acoustically disparate sound events: Both sound onsets and pitch shifts of continuous background sounds speed responses at 750 ms, with these effects dying out by 1,750 ms. These temporal distortions can be observed using only data from the first trial across participants. A potential mechanism underlying these results is that arousal increases after distracting sound events, leading to an expansion of time perception, and causing participants to misjudge when their next movement should begin.
Collapse
Affiliation(s)
- Ashley Symons
- Department of Psychological Sciences, Birkbeck, University of London, Malet Street, London, WC1E 7HX, UK
| | - Fred Dick
- Experimental Psychology, Division of Psychology and Language Sciences, University College London, London, UK.
| | - Adam Tierney
- Department of Psychological Sciences, Birkbeck, University of London, Malet Street, London, WC1E 7HX, UK
| |
Collapse
|
8
|
Humes LE, Dhar S, Manchaiah V, Sharma A, Chisolm TH, Arnold ML, Sanchez VA. A Perspective on Auditory Wellness: What It Is, Why It Is Important, and How It Can Be Managed. Trends Hear 2024; 28:23312165241273342. [PMID: 39150412 PMCID: PMC11329910 DOI: 10.1177/23312165241273342] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2024] [Revised: 06/25/2024] [Accepted: 07/16/2024] [Indexed: 08/17/2024] Open
Abstract
During the last decade, there has been a move towards consumer-centric hearing healthcare. This is a direct result of technological advancements (e.g., merger of consumer grade hearing aids with consumer grade earphones creating a wide range of hearing devices) as well as policy changes (e.g., the U.S. Food and Drug Administration creating a new over-the-counter [OTC] hearing aid category). In addition to various direct-to-consumer (DTC) hearing devices available on the market, there are also several validated tools for the self-assessment of auditory function and the detection of ear disease, as well as tools for education about hearing loss, hearing devices, and communication strategies. Further, all can be made easily available to a wide range of people. This perspective provides a framework and identifies tools to improve and maintain optimal auditory wellness across the adult life course. A broadly available and accessible set of tools that can be made available on a digital platform to aid adults in the assessment and as needed, the improvement, of auditory wellness is discussed.
Collapse
Affiliation(s)
- Larry E. Humes
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, Indiana, USA
- Department of Communication Sciences and Disorders, Northwestern University, Evanston, Illinois, USA
| | - Sumitrajit Dhar
- Department of Communication Sciences and Disorders, Northwestern University, Evanston, Illinois, USA
| | - Vinaya Manchaiah
- Department of Otolaryngology-Head and Neck Surgery, University of Colorado School of Medicine, Aurora, Colorado, USA
- UCHealth Hearing and Balance, University of Colorado Hospital, Aurora, Colorado, USA
- Virtual Hearing Lab, Collaborative initiative between University of Colorado School of Medicine and University of Pretoria, Aurora, Colorado, USA
- Department of Speech-Language Pathology and Audiology, University of Pretoria, Pretoria, South Africa
- Department of Speech and Hearing, Manipal College of Health Professions, Manipal Academy of Higher Education, Manipal, Karnataka, India
| | - Anu Sharma
- Brain and Behavior Laboratory, Department of Speech Language and Hearing Sciences, Institute of Cognitive Science, Center for Neuroscience, University of Colorado Boulder, Boulder, Colorado, USA
| | - Theresa H. Chisolm
- Department of Communication Sciences & Disorders, University of South Florida, Tampa, Florida, USA
| | - Michelle L. Arnold
- Department of Communication Sciences & Disorders, University of South Florida, Tampa, Florida, USA
| | - Victoria A. Sanchez
- Department of Communication Sciences & Disorders, University of South Florida, Tampa, Florida, USA
- Department of Otolaryngology-Head & Neck Surgery, University of South Florida, Tampa, Florida, USA
| |
Collapse
|
9
|
Chow JK, Palmeri TJ, Pluck G, Gauthier I. Evidence for an amodal domain-general object recognition ability. Cognition 2023; 238:105542. [PMID: 37419065 DOI: 10.1016/j.cognition.2023.105542] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2022] [Revised: 06/26/2023] [Accepted: 06/27/2023] [Indexed: 07/09/2023]
Abstract
A general object recognition ability predicts performance across a variety of high-level visual tests, categories, and performance in haptic recognition. Does this ability extend to auditory recognition? Vision and haptics tap into similar representations of shape and texture. In contrast, features of auditory perception like pitch, timbre, or loudness do not readily translate into shape percepts related to edges, surfaces, or spatial arrangement of parts. We find that an auditory object recognition ability correlates highly with a visual object recognition ability after controlling for general intelligence, perceptual speed, low-level visual ability, and memory ability. Auditory object recognition was a stronger predictor of visual object recognition than all control measures across two experiments, even though those control variables were also tested visually. These results point towards a single high-level ability used in both vision and audition. Much work highlights how the integration of visual and auditory information is important in specific domains (e.g., speech, music), with evidence for some overlap of visual and auditory neural representations. Our results are the first to reveal a domain-general ability, o, that predicts object recognition performance in both visual and auditory tests. Because o is domain-general, it reveals mechanisms that apply across a wide range of situations, independent of experience and knowledge. As o is distinct from general intelligence, it is well positioned to potentially add predictive validity when explaining individual differences in a variety of tasks, above and beyond measures of common cognitive abilities like general intelligence and working memory.
Collapse
Affiliation(s)
- Jason K Chow
- Department of Psychology, Vanderbilt University, USA.
| | | | - Graham Pluck
- Faculty of Psychology, Chulalongkorn University, Thailand
| | | |
Collapse
|
10
|
Vinay, Moore BCJ. Exploiting individual differences to assess the role of place and phase locking cues in auditory frequency discrimination at 2 kHz. Sci Rep 2023; 13:13801. [PMID: 37612303 PMCID: PMC10447419 DOI: 10.1038/s41598-023-40571-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Accepted: 08/13/2023] [Indexed: 08/25/2023] Open
Abstract
The relative role of place and temporal mechanisms in auditory frequency discrimination was assessed for a centre frequency of 2 kHz. Four measures of frequency discrimination were obtained for 63 normal-hearing participants: detection of frequency modulation using modulation rates of 2 Hz (FM2) and 20 Hz (FM20); detection of a change in frequency across successive pure tones (difference limen for frequency, DLF); and detection of changes in the temporal fine structure of bandpass filtered complex tones centred at 2 kHz (TFS). Previous work has suggested that: FM2 depends on the use of both temporal and place cues; FM20 depends primarily on the use of place cues because the temporal mechanism cannot track rapid changes in frequency; DLF depends primarily on temporal cues; TFS depends exclusively on temporal cues. This led to the following predicted patterns of the correlations of scores across participants: DLF and TFS should be highly correlated; FM2 should be correlated with DLF and TFS; FM20 should not be correlated with DLF or TFS. The results were broadly consistent with these predictions and with the idea that frequency discrimination at 2 kHz depends partly or primarily on temporal cues except for frequency modulation detection at a high rate.
Collapse
Affiliation(s)
- Vinay
- Audiology Group, Department of Neuromedicine and Movement Science, Faculty of Medicine and Health Sciences, Norwegian University of Science and Technology (NTNU), Tungasletta 2, 7491, Trondheim, Norway.
| | - Brian C J Moore
- Cambridge Hearing Group, Department of Psychology, University of Cambridge, Cambridge, UK
| |
Collapse
|
11
|
Pearson DV, Shen Y, McAuley JD, Kidd GR. Differential sensitivity to speech rhythms in young and older adults. Front Psychol 2023; 14:1160236. [PMID: 37251054 PMCID: PMC10213510 DOI: 10.3389/fpsyg.2023.1160236] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2023] [Accepted: 04/19/2023] [Indexed: 05/31/2023] Open
Abstract
Sensitivity to the temporal properties of auditory patterns tends to be poorer in older listeners, and this has been hypothesized to be one factor contributing to their poorer speech understanding. This study examined sensitivity to speech rhythms in young and older normal-hearing subjects, using a task designed to measure the effect of speech rhythmic context on the detection of changes in the timing of word onsets in spoken sentences. A temporal-shift detection paradigm was used in which listeners were presented with an intact sentence followed by two versions of the sentence in which a portion of speech was replaced with a silent gap: one with correct gap timing (the same duration as the missing speech) and one with altered gap timing (shorter or longer than the duration of the missing speech), resulting in an early or late resumption of the sentence after the gap. The sentences were presented with either an intact rhythm or an altered rhythm preceding the silent gap. Listeners judged which sentence had the altered gap timing, and thresholds for the detection of deviations from the correct timing were calculated separately for shortened and lengthened gaps. Both young and older listeners demonstrated lower thresholds in the intact rhythm condition than in the altered rhythm conditions. However, shortened gaps led to lower thresholds than lengthened gaps for the young listeners, while older listeners were not sensitive to the direction of the change in timing. These results show that both young and older listeners rely on speech rhythms to generate temporal expectancies for upcoming speech events. However, the absence of lower thresholds for shortened gaps among the older listeners indicates a change in speech-timing expectancies with age. A further examination of individual differences within the older group revealed that those with better rhythm-discrimination abilities (from a separate study) tended to show the same heightened sensitivity to early events observed with the young listeners.
Collapse
Affiliation(s)
- Dylan V. Pearson
- Department of Speech, Language, and Hearing Sciences, Indiana University, Bloomington, IN, United States
| | - Yi Shen
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, United States
| | - J. Devin McAuley
- Department of Psychology, Michigan State University, East Lansing, MI, United States
| | - Gary R. Kidd
- Department of Speech, Language, and Hearing Sciences, Indiana University, Bloomington, IN, United States
| |
Collapse
|
12
|
Mathias SR, Knowles EEM, Mollon J, Rodrigue AL, Woolsey MK, Hernandez AM, Garret AS, Fox PT, Olvera RL, Peralta JM, Kumar S, Göring HHH, Duggirala R, Curran JE, Blangero J, Glahn DC. Cocktail-party listening and cognitive abilities show strong pleiotropy. Front Neurol 2023; 14:1071766. [PMID: 36970519 PMCID: PMC10035755 DOI: 10.3389/fneur.2023.1071766] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/16/2022] [Accepted: 02/21/2023] [Indexed: 03/11/2023] Open
Abstract
Introduction The cocktail-party problem refers to the difficulty listeners face when trying to attend to relevant sounds that are mixed with irrelevant ones. Previous studies have shown that solving these problems relies on perceptual as well as cognitive processes. Previously, we showed that speech-reception thresholds (SRTs) on a cocktail-party listening task were influenced by genetic factors. Here, we estimated the degree to which these genetic factors overlapped with those influencing cognitive abilities. Methods We measured SRTs and hearing thresholds (HTs) in 493 listeners, who ranged in age from 18 to 91 years old. The same individuals completed a cognitive test battery comprising 18 measures of various cognitive domains. Individuals belonged to large extended pedigrees, which allowed us to use variance component models to estimate the narrow-sense heritability of each trait, followed by phenotypic and genetic correlations between pairs of traits. Results All traits were heritable. The phenotypic and genetic correlations between SRTs and HTs were modest, and only the phenotypic correlation was significant. By contrast, all genetic SRT-cognition correlations were strong and significantly different from 0. For some of these genetic correlations, the hypothesis of complete pleiotropy could not be rejected. Discussion Overall, the results suggest that there was substantial genetic overlap between SRTs and a wide range of cognitive abilities, including abilities without a major auditory or verbal component. The findings highlight the important, yet sometimes overlooked, contribution of higher-order processes to solving the cocktail-party problem, raising an important caveat for future studies aiming to identify specific genetic factors that influence cocktail-party listening.
Collapse
Affiliation(s)
- Samuel R. Mathias
- Department of Psychiatry, Boston Children's Hospital, Boston, MA, United States
- Harvard Medical School, Boston, MA, United States
| | - Emma E. M. Knowles
- Department of Psychiatry, Boston Children's Hospital, Boston, MA, United States
- Harvard Medical School, Boston, MA, United States
| | - Josephine Mollon
- Department of Psychiatry, Boston Children's Hospital, Boston, MA, United States
- Harvard Medical School, Boston, MA, United States
| | - Amanda L. Rodrigue
- Department of Psychiatry, Boston Children's Hospital, Boston, MA, United States
- Harvard Medical School, Boston, MA, United States
| | - Mary K. Woolsey
- Research Imaging Institute, University of Texas Health Science Center, San Antonio, TX, United States
| | - Alyssa M. Hernandez
- Research Imaging Institute, University of Texas Health Science Center, San Antonio, TX, United States
| | - Amy S. Garret
- Research Imaging Institute, University of Texas Health Science Center, San Antonio, TX, United States
| | - Peter T. Fox
- Research Imaging Institute, University of Texas Health Science Center, San Antonio, TX, United States
- South Texas Veterans Health Care System, San Antonio, TX, United States
| | - Rene L. Olvera
- Research Imaging Institute, University of Texas Health Science Center, San Antonio, TX, United States
| | - Juan M. Peralta
- Department of Human Genetics, South Texas Diabetes and Obesity Institute, University of Texas Rio Grande Valley School of Medicine, Brownsville, TX, United States
| | - Satish Kumar
- Department of Human Genetics, South Texas Diabetes and Obesity Institute, University of Texas Rio Grande Valley School of Medicine, Brownsville, TX, United States
| | - Harald H. H. Göring
- Department of Human Genetics, South Texas Diabetes and Obesity Institute, University of Texas Rio Grande Valley School of Medicine, Brownsville, TX, United States
| | - Ravi Duggirala
- Department of Human Genetics, South Texas Diabetes and Obesity Institute, University of Texas Rio Grande Valley School of Medicine, Brownsville, TX, United States
| | - Joanne E. Curran
- Department of Human Genetics, South Texas Diabetes and Obesity Institute, University of Texas Rio Grande Valley School of Medicine, Brownsville, TX, United States
| | - John Blangero
- Department of Human Genetics, South Texas Diabetes and Obesity Institute, University of Texas Rio Grande Valley School of Medicine, Brownsville, TX, United States
| | - David C. Glahn
- Department of Psychiatry, Boston Children's Hospital, Boston, MA, United States
- Harvard Medical School, Boston, MA, United States
| |
Collapse
|
13
|
Jasmin K, Tierney A, Obasih C, Holt L. Short-term perceptual reweighting in suprasegmental categorization. Psychon Bull Rev 2023; 30:373-382. [PMID: 35915382 PMCID: PMC9971089 DOI: 10.3758/s13423-022-02146-5] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/05/2022] [Indexed: 11/08/2022]
Abstract
Segmental speech units such as phonemes are described as multidimensional categories whose perception involves contributions from multiple acoustic input dimensions, and the relative perceptual weights of these dimensions respond dynamically to context. For example, when speech is altered to create an "accent" in which two acoustic dimensions are correlated in a manner opposite that of long-term experience, the dimension that carries less perceptual weight is down-weighted to contribute less in category decisions. It remains unclear, however, whether this short-term reweighting extends to perception of suprasegmental features that span multiple phonemes, syllables, or words, in part because it has remained debatable whether suprasegmental features are perceived categorically. Here, we investigated the relative contribution of two acoustic dimensions to word emphasis. Participants categorized instances of a two-word phrase pronounced with typical covariation of fundamental frequency (F0) and duration, and in the context of an artificial "accent" in which F0 and duration (established in prior research on English speech as "primary" and "secondary" dimensions, respectively) covaried atypically. When categorizing "accented" speech, listeners rapidly down-weighted the secondary dimension (duration). This result indicates that listeners continually track short-term regularities across speech input and dynamically adjust the weight of acoustic evidence for suprasegmental decisions. Thus, dimension-based statistical learning appears to be a widespread phenomenon in speech perception extending to both segmental and suprasegmental categorization.
Collapse
Affiliation(s)
- Kyle Jasmin
- Department of Psychology, Wolfson Building, Royal Holloway, University of London, Egham, Surrey, TW20 0EX, UK.
| | | | | | - Lori Holt
- Carnegie Mellon University, Pittsburgh, PA, USA
| |
Collapse
|
14
|
Genetic Hearing Loss Affects Cochlear Processing. Genes (Basel) 2022; 13:genes13111923. [DOI: 10.3390/genes13111923] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2022] [Revised: 10/12/2022] [Accepted: 10/20/2022] [Indexed: 11/04/2022] Open
Abstract
The relationship between speech recognition and hereditary hearing loss is not straightforward. Underlying genetic defects might determine an impaired cochlear processing of sound. We obtained data from nine groups of patients with a specific type of genetic hearing loss. For each group, the affected cochlear site-of-lesion was determined based on previously published animal studies. Retrospectively obtained speech recognition scores in noise were related to several aspects of supra-threshold cochlear processing as assessed by psychophysical measurements. The differences in speech perception in noise between these patient groups could be explained by these factors and partially by the hypothesized affected structure of the cochlea, suggesting that speech recognition in noise was associated with a genetics-related malfunctioning of the cochlea. In particular, regression models indicate that loudness growth and spectral resolution best describe the cochlear distortions and are thus a good biomarker for speech understanding in noise.
Collapse
|
15
|
Stilp CE, Shorey AE, King CJ. Nonspeech sounds are not all equally good at being nonspeech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 152:1842. [PMID: 36182316 DOI: 10.1121/10.0014174] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/11/2022] [Accepted: 08/30/2022] [Indexed: 06/16/2023]
Abstract
Perception of speech sounds has a long history of being compared to perception of nonspeech sounds, with rich and enduring debates regarding how closely they share similar underlying processes. In many instances, perception of nonspeech sounds is directly compared to that of speech sounds without a clear explanation of how related these sounds are to the speech they are selected to mirror (or not mirror). While the extreme acoustic variability of speech sounds is well documented, this variability is bounded by the common source of a human vocal tract. Nonspeech sounds do not share a common source, and as such, exhibit even greater acoustic variability than that observed for speech. This increased variability raises important questions about how well perception of a given nonspeech sound might resemble or model perception of speech sounds. Here, we offer a brief review of extremely diverse nonspeech stimuli that have been used in the efforts to better understand perception of speech sounds. The review is organized according to increasing spectrotemporal complexity: random noise, pure tones, multitone complexes, environmental sounds, music, speech excerpts that are not recognized as speech, and sinewave speech. Considerations are offered for stimulus selection in nonspeech perception experiments moving forward.
Collapse
Affiliation(s)
- Christian E Stilp
- Department of Psychological and Brain Sciences, University of Louisville, Louisville, Kentucky 40292, USA
| | - Anya E Shorey
- Department of Psychological and Brain Sciences, University of Louisville, Louisville, Kentucky 40292, USA
| | - Caleb J King
- Department of Psychological and Brain Sciences, University of Louisville, Louisville, Kentucky 40292, USA
| |
Collapse
|
16
|
Saito K, Kachlicka M, Suzukida Y, Petrova K, Lee BJ, Tierney A. Auditory precision hypothesis-L2: Dimension-specific relationships between auditory processing and second language segmental learning. Cognition 2022; 229:105236. [PMID: 36027789 DOI: 10.1016/j.cognition.2022.105236] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2021] [Revised: 01/11/2022] [Accepted: 07/25/2022] [Indexed: 11/29/2022]
Abstract
Growing evidence suggests a broad relationship between individual differences in auditory processing ability and the rate and ultimate attainment of language acquisition throughout the lifespan, including post-pubertal second language (L2) speech learning. However, little is known about how the precision of processing of specific auditory dimensions relates to the acquisition of specific L2 segmental contrasts. In the context of 100 late Japanese-English bilinguals with diverse profiles of classroom and immersion experience, the current study set out to investigate the link between the perception of several auditory dimensions (F3 frequency, F2 frequency, and duration) in non-verbal sounds and English [r]-[l] perception and production proficiency. Whereas participants' biographical factors (the presence/absence of immersion) accounted for a large amount of variance in the success of learning this contrast, the outcomes were also tied to their acuity to the most reliable, new auditory cues (F3 variation) and the less reliable but already-familiar cues (F2 variation). This finding suggests that individuals can vary in terms of how they perceive, utilize, and make the most of information conveyed by specific acoustic dimensions. When perceiving more naturalistic spoken input, where speech contrasts can be distinguished via a combination of numerous cues, some can attain a high-level of L2 speech proficiency by using nativelike and/or non-nativelike strategies in a complementary fashion.
Collapse
|
17
|
The Intelligibility of Time-Compressed Speech Is Correlated with the Ability to Listen in Modulated Noise. J Assoc Res Otolaryngol 2022; 23:413-426. [DOI: 10.1007/s10162-021-00832-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2021] [Accepted: 12/15/2021] [Indexed: 10/18/2022] Open
|
18
|
Individualized Assays of Temporal Coding in the Ascending Human Auditory System. eNeuro 2022; 9:ENEURO.0378-21.2022. [PMID: 35193890 PMCID: PMC8925652 DOI: 10.1523/eneuro.0378-21.2022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2021] [Revised: 01/12/2022] [Accepted: 02/08/2022] [Indexed: 11/21/2022] Open
Abstract
Neural phase-locking to temporal fluctuations is a fundamental and unique mechanism by which acoustic information is encoded by the auditory system. The perceptual role of this metabolically expensive mechanism, the neural phase-locking to temporal fine structure (TFS) in particular, is debated. Although hypothesized, it is unclear whether auditory perceptual deficits in certain clinical populations are attributable to deficits in TFS coding. Efforts to uncover the role of TFS have been impeded by the fact that there are no established assays for quantifying the fidelity of TFS coding at the individual level. While many candidates have been proposed, for an assay to be useful, it should not only intrinsically depend on TFS coding, but should also have the property that individual differences in the assay reflect TFS coding per se over and beyond other sources of variance. Here, we evaluate a range of behavioral and electroencephalogram (EEG)-based measures as candidate individualized measures of TFS sensitivity. Our comparisons of behavioral and EEG-based metrics suggest that extraneous variables dominate both behavioral scores and EEG amplitude metrics, rendering them ineffective. After adjusting behavioral scores using lapse rates, and extracting latency or percent-growth metrics from EEG, interaural timing sensitivity measures exhibit robust behavior-EEG correlations. Together with the fact that unambiguous theoretical links can be made relating binaural measures and phase-locking to TFS, our results suggest that these "adjusted" binaural assays may be well suited for quantifying individual TFS processing.
Collapse
|
19
|
Goldsworthy RL, Bissmeyer SRS, Camarena A. Advantages of Pulse Rate Compared to Modulation Frequency for Temporal Pitch Perception in Cochlear Implant Users. J Assoc Res Otolaryngol 2022; 23:137-150. [PMID: 34981263 PMCID: PMC8782986 DOI: 10.1007/s10162-021-00828-w] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/11/2021] [Accepted: 12/01/2021] [Indexed: 02/03/2023] Open
Abstract
Most cochlear implants encode the fundamental frequency of periodic sounds by amplitude modulation of constant-rate pulsatile stimulation. Pitch perception provided by such stimulation strategies is markedly poor. Two experiments are reported here that consider potential advantages of pulse rate compared to modulation frequency for providing stimulation timing cues for pitch. The first experiment examines beat frequency distortion that occurs when modulating constant-rate pulsatile stimulation. This distortion has been reported on previously, but the results presented here indicate that distortion occurs for higher stimulation rates than previously reported. The second experiment examines pitch resolution as provided by pulse rate compared to modulation frequency. The results indicate that pitch discrimination is better with pulse rate than with modulation frequency. The advantage was large for rates near what has been suggested as the upper limit of temporal pitch perception conveyed by cochlear implants. The results are relevant to sound processing design for cochlear implants particularly for algorithms that encode fundamental frequency into deep envelope modulations or into precisely timed pulsatile stimulation.
Collapse
Affiliation(s)
- Raymond L Goldsworthy
- Auditory Research Center, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA.
| | - Susan R S Bissmeyer
- Auditory Research Center, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Department of Biomedical Engineering, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, USA
| | - Andres Camarena
- Auditory Research Center, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Neuroscience Graduate Program, University of Southern California, Los Angeles, CA, USA
| |
Collapse
|
20
|
Deng J, Zheng X, Zhu C, Lin Z, Ye Y. Auditory acuity and musical ability in young adults with high schizotypal traits. Compr Psychiatry 2022; 114:152297. [PMID: 35123176 DOI: 10.1016/j.comppsych.2022.152297] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/02/2021] [Revised: 12/07/2021] [Accepted: 01/10/2022] [Indexed: 11/30/2022] Open
Abstract
BACKGROUND Despite auditory cognition dysfunction being consistently found in people with schizophrenia, the evidence from non-clinical individuals with schizotypy is rare and inconsistent. No studies thus far have comprehensively assessed the association among auditory perception, musical cognition, and schizotypy in non-clinical samples. AIM We aimed to explore abnormalities in auditory skills, from basic perception to musical ability, among individuals with schizotypal traits. METHOD An extreme-group design was adopted. Sixty-six participants from the schizotypy and control groups were screened from 1093 young adults using the Schizotypal Personality Questionnaire (SPQ). Auditory acuity was assessed using four auditory discrimination threshold tests, and musical ability was evaluated through the Montreal Battery of Evaluation of Amusia (MBEA). Basic demographic information and musical backgrounds were assessed and matched, and depression, anxiety, and digit-span index were evaluated and controlled. RESULTS Elevated sensitivity in auditory perception and improved musical talent were found in young adults with high schizotypal traits. Auditory acuity and musical ability were positively correlated with schizotypy and its factors among participants across groups. A regression analysis in the control group showed that cognitive perceptual scores of SPQ positively predicted auditory temporal sensitivity. The mediation analysis revealed an indirect effect of pure tone duration discrimination between musical rhythmic ability and positive factor of schizotypy. DISCUSSION Elevated sensitivity in auditory temporal perception and improved musical talent in young adults with high schizotypy may contribute to explaining the variation of auditory process in the development of schizophrenia-spectrum disorders. It can also help elucidate the association between psychopathology and creativity in auditory modality.
Collapse
Affiliation(s)
- Jue Deng
- School of Psychology, Fujian Normal University, No. 1 Technology Road, Higher Education Mega-Center, Fuzhou City, Fujian Province 350117, China
| | - Xiaoqian Zheng
- School of Psychology, Fujian Normal University, No. 1 Technology Road, Higher Education Mega-Center, Fuzhou City, Fujian Province 350117, China
| | - Chenxin Zhu
- School of Psychology, Fujian Normal University, No. 1 Technology Road, Higher Education Mega-Center, Fuzhou City, Fujian Province 350117, China
| | - Ziyue Lin
- Mental Health Center, Fujian Medical University, No. 1 Xuefu North Road, Higher Education Mega-Center, Fuzhou City, Fujian Province 350122, China
| | - Yiduo Ye
- School of Psychology, Fujian Normal University, No. 1 Technology Road, Higher Education Mega-Center, Fuzhou City, Fujian Province 350117, China.
| |
Collapse
|
21
|
Humes LE, Kidd GR, Lentz JJ. Differences Between Young and Older Adults in Working Memory and Performance on the Test of Basic Auditory Capabilities †. Front Psychol 2022; 12:804891. [PMID: 35095690 PMCID: PMC8790016 DOI: 10.3389/fpsyg.2021.804891] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/29/2021] [Accepted: 12/13/2021] [Indexed: 11/29/2022] Open
Abstract
The Test of Basic Auditory Capabilities (TBAC) is a battery of auditory-discrimination tasks and speech-identification tasks that has been normed on several hundred young normal-hearing adults. Previous research with the TBAC suggested that cognitive function may impact the performance of older adults. Here, we examined differences in performance on several TBAC tasks between a group of 34 young adults with a mean age of 22.5 years (SD = 3.1 years) and a group of 115 older adults with a mean age of 69.2 years (SD = 6.2 years) recruited from the local community. Performance of the young adults was consistent with prior norms for this age group. Not surprisingly, the two groups differed significantly in hearing loss and working memory with the older adults having more hearing loss and poorer working memory than the young adults. The two age groups also differed significantly in performance on six of the nine measures extracted from the TBAC (eight test scores and one average test score) with the older adults consistently performing worse than the young adults. However, when these age-group comparisons were repeated with working memory and hearing loss as covariates, the groups differed in performance on only one of the nine auditory measures from the TBAC. For eight of the nine TBAC measures, working memory was a significant covariate and hearing loss never emerged as a significant factor. Thus, the age-group deficits observed initially on the TBAC most often appeared to be mediated by age-related differences in working memory rather than deficits in auditory processing. The results of these analyses of age-group differences were supported further by linear-regression analyses with each of the 9 TBAC scores serving as the dependent measure and age, hearing loss, and working memory as the predictors. Regression analyses were conducted for the full set of 149 adults and for just the 115 older adults. Working memory again emerged as the predominant factor impacting TBAC performance. It is concluded that working memory should be considered when comparing the performance of young and older adults on auditory tasks, including the TBAC.
Collapse
Affiliation(s)
- Larry E. Humes
- Department of Speech, Language, and Hearing Sciences, Indiana University, Bloomington, IN, United States
| | | | | |
Collapse
|
22
|
Miceli G, Caccia A. Cortical disorders of speech processing: Pure word deafness and auditory agnosia. HANDBOOK OF CLINICAL NEUROLOGY 2022; 187:69-87. [PMID: 35964993 DOI: 10.1016/b978-0-12-823493-8.00005-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Selective disorders of auditory speech processing due to brain lesions are reviewed. Over 120 years after the first anatomic report (Dejerine and Sérieux, 1898), fewer than 80 cumulative cases of generalized auditory agnosia and pure word deafness with documented brain lesions are on record. Most patients (approximately 70%) had vascular lesions. Damage is very frequently bilateral in generalized auditory agnosia, and more frequently unilateral in pure word deafness. In unilateral cases, anatomical disconnection is not a prerequisite, and disorders may be due to functional disconnection. Regardless of whether lesions are unilateral or bilateral, speech processing difficulties emerge in the presence of damage to the superior temporal regions of the language-dominant hemisphere, suggesting that speech input is processed asymmetrically at early stages already. Extant evidence does not allow establishing whether processing asymmetry originates in the primary auditory cortex or in higher associative cortices, nor whether auditory processing in the brainstem is entirely symmetric. Results are consistent with the view that the difficulty in processing auditory input characterized by quick spectral and/or temporal changes is one of the critical dimensions of the disorder. Forthcoming studies should focus on detailed audiologic, neurolinguistic, and neuroanatomic descriptions of each case.
Collapse
Affiliation(s)
- Gabriele Miceli
- Center for Mind/Brain Sciences - CIMeC, University of Trento, Rovereto, Italy; Centro Interdisciplinare Linceo 'Beniamino Segre'-Accademia dei Lincei, Rome, Italy.
| | - Antea Caccia
- Center for Mind/Brain Sciences - CIMeC, University of Trento, Rovereto, Italy; Department of Psychology, University of Milano-Bicocca, Milan, Italy
| |
Collapse
|
23
|
Carcagno S, Plack CJ. Relations between speech-reception, psychophysical temporal processing, and subcortical electrophysiological measures of auditory function in humans. Hear Res 2022; 417:108456. [PMID: 35149333 PMCID: PMC8935383 DOI: 10.1016/j.heares.2022.108456] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/07/2021] [Revised: 01/05/2022] [Accepted: 01/27/2022] [Indexed: 11/04/2022]
|
24
|
Lentz JJ, Humes LE, Kidd GR. Differences in Auditory Perception Between Young and Older Adults When Controlling for Differences in Hearing Loss and Cognition. Trends Hear 2022; 26:23312165211066180. [PMID: 34989641 PMCID: PMC8753078 DOI: 10.1177/23312165211066180] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
This study was designed to examine age effects on various auditory perceptual skills using a large group of listeners (155 adults, 121 aged 60-88 years and 34 aged 18-30 years), while controlling for the factors of hearing loss and working memory (WM). All subjects completed 3 measures of WM, 7 psychoacoustic tasks (24 conditions) and a hearing assessment. Psychophysical measures were selected to tap phenomena thought to be mediated by higher-level auditory function and included modulation detection, modulation detection interference, informational masking (IM), masking level difference (MLD), anisochrony detection, harmonic mistuning, and stream segregation. Principal-components analysis (PCA) was applied to each psychoacoustic test. For 6 of the 7 tasks, a single component represented performance across the multiple stimulus conditions well, whereas the modulation-detection interference (MDI) task required two components to do so. The effect of age was analyzed using a general linear model applied to each psychoacoustic component. Once hearing loss and WM were accounted for as covariates in the analyses, estimated marginal mean thresholds were lower for older adults on tasks based on temporal processing. When evaluated separately, hearing loss led to poorer performance on roughly 1/2 the tasks and declines in WM accounted for poorer performance on 6 of the 8 psychoacoustic components. These results make clear the need to interpret age-group differences in performance on psychoacoustic tasks in light of cognitive declines commonly associated with aging, and point to hearing loss and cognitive declines as negatively influencing auditory perceptual skills.
Collapse
Affiliation(s)
- Jennifer J. Lentz
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
| | - Larry E. Humes
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
| | - Gary R. Kidd
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
| |
Collapse
|
25
|
Jasmin K, Dick F, Tierney AT. The Multidimensional Battery of Prosody Perception (MBOPP). Wellcome Open Res 2021; 5:4. [PMID: 35282675 PMCID: PMC8881696 DOI: 10.12688/wellcomeopenres.15607.2] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/20/2021] [Indexed: 11/20/2022] Open
Abstract
Prosody can be defined as the rhythm and intonation patterns spanning words, phrases and sentences. Accurate perception of prosody is an important component of many aspects of language processing, such as parsing grammatical structures, recognizing words, and determining where emphasis may be placed. Prosody perception is important for language acquisition and can be impaired in language-related developmental disorders. However, existing assessments of prosodic perception suffer from some shortcomings. These include being unsuitable for use with typically developing adults due to ceiling effects and failing to allow the investigator to distinguish the unique contributions of individual acoustic features such as pitch and temporal cues. Here we present the Multi-Dimensional Battery of Prosody Perception (MBOPP), a novel tool for the assessment of prosody perception. It consists of two subtests: Linguistic Focus, which measures the ability to hear emphasis or sentential stress, and Phrase Boundaries, which measures the ability to hear where in a compound sentence one phrase ends, and another begins. Perception of individual acoustic dimensions (Pitch and Duration) can be examined separately, and test difficulty can be precisely calibrated by the experimenter because stimuli were created using a continuous voice morph space. We present validation analyses from a sample of 59 individuals and discuss how the battery might be deployed to examine perception of prosody in various populations.
Collapse
Affiliation(s)
- Kyle Jasmin
- Department of Psychology, Royal Holloway, University of London, Ehgam, TW20 0EX, UK
| | - Frederic Dick
- Psychological Sciences, Birkbeck University of London, London, WC1E 7HX, UK
| | | |
Collapse
|
26
|
Jasmin K, Dick F, Tierney AT. The Multidimensional Battery of Prosody Perception (MBOPP). Wellcome Open Res 2021; 5:4. [PMID: 35282675 PMCID: PMC8881696 DOI: 10.12688/wellcomeopenres.15607.1] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/20/2021] [Indexed: 09/03/2023] Open
Abstract
Prosody can be defined as the rhythm and intonation patterns spanning words, phrases and sentences. Accurate perception of prosody is an important component of many aspects of language processing, such as parsing grammatical structures, recognizing words, and determining where emphasis may be placed. Prosody perception is important for language acquisition and can be impaired in language-related developmental disorders. However, existing assessments of prosodic perception suffer from some shortcomings. These include being unsuitable for use with typically developing adults due to ceiling effects and failing to allow the investigator to distinguish the unique contributions of individual acoustic features such as pitch and temporal cues. Here we present the Multi-Dimensional Battery of Prosody Perception (MBOPP), a novel tool for the assessment of prosody perception. It consists of two subtests: Linguistic Focus, which measures the ability to hear emphasis or sentential stress, and Phrase Boundaries, which measures the ability to hear where in a compound sentence one phrase ends, and another begins. Perception of individual acoustic dimensions (Pitch and Duration) can be examined separately, and test difficulty can be precisely calibrated by the experimenter because stimuli were created using a continuous voice morph space. We present validation analyses from a sample of 59 individuals and discuss how the battery might be deployed to examine perception of prosody in various populations.
Collapse
Affiliation(s)
- Kyle Jasmin
- Department of Psychology, Royal Holloway, University of London, Ehgam, TW20 0EX, UK
| | - Frederic Dick
- Psychological Sciences, Birkbeck University of London, London, WC1E 7HX, UK
| | | |
Collapse
|
27
|
Skerritt-Davis B, Elhilali M. Neural Encoding of Auditory Statistics. J Neurosci 2021; 41:6726-6739. [PMID: 34193552 PMCID: PMC8336711 DOI: 10.1523/jneurosci.1887-20.2021] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2020] [Revised: 05/19/2021] [Accepted: 05/26/2021] [Indexed: 11/21/2022] Open
Abstract
The human brain extracts statistical regularities embedded in real-world scenes to sift through the complexity stemming from changing dynamics and entwined uncertainty along multiple perceptual dimensions (e.g., pitch, timbre, location). While there is evidence that sensory dynamics along different auditory dimensions are tracked independently by separate cortical networks, how these statistics are integrated to give rise to unified objects remains unknown, particularly in dynamic scenes that lack conspicuous coupling between features. Using tone sequences with stochastic regularities along spectral and spatial dimensions, this study examines behavioral and electrophysiological responses from human listeners (male and female) to changing statistics in auditory sequences and uses a computational model of predictive Bayesian inference to formulate multiple hypotheses for statistical integration across features. Neural responses reveal multiplexed brain responses reflecting both local statistics along individual features in frontocentral networks, together with global (object-level) processing in centroparietal networks. Independent tracking of local surprisal along each acoustic feature reveals linear modulation of neural responses, while global melody-level statistics follow a nonlinear integration of statistical beliefs across features to guide perception. Near identical results are obtained in separate experiments along spectral and spatial acoustic dimensions, suggesting a common mechanism for statistical inference in the brain. Potential variations in statistical integration strategies and memory deployment shed light on individual variability between listeners in terms of behavioral efficacy and fidelity of neural encoding of stochastic change in acoustic sequences.SIGNIFICANCE STATEMENT The world around us is complex and ever changing: in everyday listening, sound sources evolve along multiple dimensions, such as pitch, timbre, and spatial location, and they exhibit emergent statistical properties that change over time. In the face of this complexity, the brain builds an internal representation of the external world by collecting statistics from the sensory input along multiple dimensions. Using a Bayesian predictive inference model, this work considers alternative hypotheses for how statistics are combined across sensory dimensions. Behavioral and neural responses from human listeners show the brain multiplexes two representations, where local statistics along each feature linearly affect neural responses, and global statistics nonlinearly combine statistical beliefs across dimensions to shape perception of stochastic auditory sequences.
Collapse
|
28
|
Pesnot Lerousseau J, Arnold G, Auvray M. Training-induced plasticity enables visualizing sounds with a visual-to-auditory conversion device. Sci Rep 2021; 11:14762. [PMID: 34285265 PMCID: PMC8292401 DOI: 10.1038/s41598-021-94133-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2021] [Accepted: 06/28/2021] [Indexed: 12/04/2022] Open
Abstract
Sensory substitution devices aim at restoring visual functions by converting visual information into auditory or tactile stimuli. Although these devices show promise in the range of behavioral abilities they allow, the processes underlying their use remain underspecified. In particular, while an initial debate focused on the visual versus auditory or tactile nature of sensory substitution, since over a decade, the idea that it reflects a mixture of both has emerged. In order to investigate behaviorally the extent to which visual and auditory processes are involved, participants completed a Stroop-like crossmodal interference paradigm before and after being trained with a conversion device which translates visual images into sounds. In addition, participants' auditory abilities and their phenomenologies were measured. Our study revealed that, after training, when asked to identify sounds, processes shared with vision were involved, as participants’ performance in sound identification was influenced by the simultaneously presented visual distractors. In addition, participants’ performance during training and their associated phenomenology depended on their auditory abilities, revealing that processing finds its roots in the input sensory modality. Our results pave the way for improving the design and learning of these devices by taking into account inter-individual differences in auditory and visual perceptual strategies.
Collapse
Affiliation(s)
| | | | - Malika Auvray
- Sorbonne Université, CNRS UMR 7222, Institut des Systèmes Intelligents et de Robotique (ISIR), 75005, Paris, France.
| |
Collapse
|
29
|
McAuley JD, Shen Y, Smith T, Kidd GR. Effects of speech-rhythm disruption on selective listening with a single background talker. Atten Percept Psychophys 2021; 83:2229-2240. [PMID: 33782913 PMCID: PMC10612531 DOI: 10.3758/s13414-021-02298-x] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 03/05/2021] [Indexed: 11/08/2022]
Abstract
Recent work by McAuley et al. (Attention, Perception, & Psychophysics, 82, 3222-3233, 2020) using the Coordinate Response Measure (CRM) paradigm with a multitalker background revealed that altering the natural rhythm of target speech amidst background speech worsens target recognition (a target-rhythm effect), while altering background speech rhythm improves target recognition (a background-rhythm effect). Here, we used a single-talker background to examine the role of specific properties of target and background sound patterns on selective listening without the complexity of multiple background stimuli. Experiment 1 manipulated the sex of the background talker, presented with a male target talker, to assess target and background-rhythm effects with and without a strong pitch cue to aid perceptual segregation. Experiment 2 used a vocoded single-talker background to examine target and background-rhythm effects with envelope-based speech rhythms preserved, but without semantic content or temporal fine structure. While a target-rhythm effect was present with all backgrounds, the background-rhythm effect was only observed for the same-sex background condition. Results provide additional support for a selective entrainment hypothesis, while also showing that the background-rhythm effect is not driven by envelope-based speech rhythm alone, and may be reduced or eliminated when pitch or other acoustic differences provide a strong basis for selective listening.
Collapse
Affiliation(s)
- J Devin McAuley
- Department of Psychology, Michigan State University, East Lansing, MI, 48824, USA.
| | - Yi Shen
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, USA
| | - Toni Smith
- Department of Psychology, Michigan State University, East Lansing, MI, 48824, USA
| | - Gary R Kidd
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
| |
Collapse
|
30
|
Zhang L, Schlaghecken F, Harte J, Roberts KL. The Influence of the Type of Background Noise on Perceptual Learning of Speech in Noise. Front Neurosci 2021; 15:646137. [PMID: 34012384 PMCID: PMC8126633 DOI: 10.3389/fnins.2021.646137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/25/2020] [Accepted: 04/06/2021] [Indexed: 11/13/2022] Open
Abstract
OBJECTIVES Auditory perceptual learning studies tend to focus on the nature of the target stimuli. However, features of the background noise can also have a significant impact on the amount of benefit that participants obtain from training. This study explores whether perceptual learning of speech in background babble noise generalizes to other, real-life environmental background noises (car and rain), and if the benefits are sustained over time. DESIGN Normal-hearing native English speakers were randomly assigned to a training (n = 12) or control group (n = 12). Both groups completed a pre- and post-test session in which they identified Bamford-Kowal-Bench (BKB) target words in babble, car, or rain noise. The training group completed speech-in-babble noise training on three consecutive days between the pre- and post-tests. A follow up session was conducted between 8 and 18 weeks after the post-test session (training group: n = 9; control group: n = 7). RESULTS Participants who received training had significantly higher post-test word identification accuracy than control participants for all three types of noise, although benefits were greatest for the babble noise condition and weaker for the car- and rain-noise conditions. Both training and control groups maintained their pre- to post-test improvement over a period of several weeks for speech in babble noise, but returned to pre-test accuracy for speech in car and rain noise. CONCLUSION The findings show that training benefits can show some generalization from speech-in-babble noise to speech in other types of environmental noise. Both groups sustained their learning over a period of several weeks for speech-in-babble noise. As the control group received equal exposure to all three noise types, the sustained learning with babble noise, but not other noises, implies that a structural feature of babble noise was conducive to the sustained improvement. These findings emphasize the importance of considering the background noise as well as the target stimuli in auditory perceptual learning studies.
Collapse
Affiliation(s)
- Liping Zhang
- Department of Otolaryngology-Head and Neck Surgery, Shandong Provincial ENT Hospital, Cheeloo College of Medicine, Shandong University, Jinan, China
- Warwick Manufacturing Group, University of Warwick, Coventry, United Kingdom
| | | | - James Harte
- Warwick Manufacturing Group, University of Warwick, Coventry, United Kingdom
- Interacoustics Research Unit, Technical University of Denmark, Lyngby, Denmark
| | - Katherine L. Roberts
- Department of Psychology, University of Warwick, Coventry, United Kingdom
- Department of Psychology, Nottingham Trent University, Nottingham, United Kingdom
| |
Collapse
|
31
|
Errors on a Speech-in-Babble Sentence Recognition Test Reveal Individual Differences in Acoustic Phonetic Perception and Babble Misallocations. Ear Hear 2021; 42:673-690. [PMID: 33928926 DOI: 10.1097/aud.0000000000001020] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/20/2022]
Abstract
OBJECTIVES The ability to recognize words in connected speech under noisy listening conditions is critical to everyday communication. Many processing levels contribute to the individual listener's ability to recognize words correctly against background speech, and there is clinical need for measures of individual differences at different levels. Typical listening tests of speech recognition in noise require a list of items to obtain a single threshold score. Diverse abilities measures could be obtained through mining various open-set recognition errors during multi-item tests. This study sought to demonstrate that an error mining approach using open-set responses from a clinical sentence-in-babble-noise test can be used to characterize abilities beyond signal-to-noise ratio (SNR) threshold. A stimulus-response phoneme-to-phoneme sequence alignment software system was used to achieve automatic, accurate quantitative error scores. The method was applied to a database of responses from normal-hearing (NH) adults. Relationships between two types of response errors and words correct scores were evaluated through use of mixed models regression. DESIGN Two hundred thirty-three NH adults completed three lists of the Quick Speech in Noise test. Their individual open-set speech recognition responses were automatically phonemically transcribed and submitted to a phoneme-to-phoneme stimulus-response sequence alignment system. The computed alignments were mined for a measure of acoustic phonetic perception, a measure of response text that could not be attributed to the stimulus, and a count of words correct. The mined data were statistically analyzed to determine whether the response errors were significant factors beyond stimulus SNR in accounting for the number of words correct per response from each participant. This study addressed two hypotheses: (1) Individuals whose perceptual errors are less severe recognize more words correctly under difficult listening conditions due to babble masking and (2) Listeners who are better able to exclude incorrect speech information such as from background babble and filling in recognize more stimulus words correctly. RESULTS Statistical analyses showed that acoustic phonetic accuracy and exclusion of babble background were significant factors, beyond the stimulus sentence SNR, in accounting for the number of words a participant recognized. There was also evidence that poorer acoustic phonetic accuracy could occur along with higher words correct scores. This paradoxical result came from a subset of listeners who had also performed subjective accuracy judgments. Their results suggested that they recognized more words while also misallocating acoustic cues from the background into the stimulus, without realizing their errors. Because the Quick Speech in Noise test stimuli are locked to their own babble sample, misallocations of whole words from babble into the responses could be investigated in detail. The high rate of common misallocation errors for some sentences supported the view that the functional stimulus was the combination of the target sentence and its babble. CONCLUSIONS Individual differences among NH listeners arise both in terms of words accurately identified and errors committed during open-set recognition of sentences in babble maskers. Error mining to characterize individual listeners can be done automatically at the levels of acoustic phonetic perception and the misallocation of background babble words into open-set responses. Error mining can increase test information and the efficiency and accuracy of characterizing individual listeners.
Collapse
|
32
|
Auditory and visual short-term memory: influence of material type, contour, and musical expertise. PSYCHOLOGICAL RESEARCH 2021; 86:421-442. [PMID: 33881610 PMCID: PMC8885540 DOI: 10.1007/s00426-021-01519-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2020] [Accepted: 04/08/2021] [Indexed: 01/14/2023]
Abstract
Short-term memory has mostly been investigated with verbal or visuospatial stimuli and less so with other categories of stimuli. Moreover, the influence of sensory modality has been explored almost solely in the verbal domain. The present study compared visual and auditory short-term memory for different types of materials, aiming to understand whether sensory modality and material type can influence short-term memory performance. Furthermore, we aimed to assess if music expertise can modulate memory performance, as previous research has reported better auditory memory (and to some extent, visual memory), and better auditory contour recognition for musicians than non-musicians. To do so, we adapted the same recognition paradigm (delayed-matching to sample) across different types of stimuli. In each trial, participants (musicians and non-musicians) were presented with two sequences of events, separated by a silent delay, and had to indicate whether the two sequences were identical or different. The performance was compared for auditory and visual materials belonging to three different categories: (1) verbal (i.e., syllables); (2) nonverbal (i.e., that could not be easily denominated) with contour (based on loudness or luminance variations); and (3) nonverbal without contour (pink noise sequences or kanji letters sequences). Contour and no-contour conditions referred to whether the sequence can entail (or not) a contour (i.e., a pattern of up and down changes) based on non-pitch features. Results revealed a selective advantage of musicians for auditory no-contour stimuli and for contour stimuli (both visual and auditory), suggesting that musical expertise is associated with specific short-term memory advantages in domains close to the trained domain, also extending cross-modally when stimuli have contour information. Moreover, our results suggest a role of encoding strategies (i.e., how the material is represented mentally during the task) for short-term-memory performance.
Collapse
|
33
|
Skerritt-Davis B, Elhilali M. Computational framework for investigating predictive processing in auditory perception. J Neurosci Methods 2021; 360:109177. [PMID: 33839191 DOI: 10.1016/j.jneumeth.2021.109177] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2020] [Revised: 03/07/2021] [Accepted: 03/25/2021] [Indexed: 11/24/2022]
Abstract
BACKGROUND The brain tracks sound sources as they evolve in time, collecting contextual information to predict future sensory inputs. Previous work in predictive coding typically focuses on the perception of predictable stimuli, leaving the implementation of these same neural processes in more complex, real-world environments containing randomness and uncertainty up for debate. NEW METHOD To facilitate investigation into the perception of less tightly-controlled listening scenarios, we present a computational model as a tool to ask targeted questions about the underlying predictive processes that connect complex sensory inputs to listener behavior and neural responses. In the modeling framework, observed sound features (e.g. pitch) are tracked sequentially using Bayesian inference. Sufficient statistics are inferred from past observations at multiple time scales and used to make predictions about future observation while tracking the statistical structure of the sensory input. RESULTS Facets of the model are discussed in terms of their application to perceptual research, and examples taken from real-world audio demonstrate the model's flexibility to capture a variety of statistical structures along various perceptual dimensions. COMPARISON WITH EXISTING METHODS Previous models are often targeted toward interpreting a particular experimental paradigm (e.g., oddball paradigm), perceptual dimension (e.g., pitch processing), or task (e.g., speech segregation), thus limiting their ability to generalize to other domains. The presented model is designed as a flexible and practical tool for broad application. CONCLUSION The model is presented as a general framework for generating new hypotheses and guiding investigation into the neural processes underlying predictive coding of complex scenes.
Collapse
Affiliation(s)
| | - Mounya Elhilali
- Johns Hopkins University, 3400 N Charles St, Baltimore, MD, USA.
| |
Collapse
|
34
|
Shen J. Older Listeners' Perception of Speech With Strengthened and Weakened Dynamic Pitch Cues in Background Noise. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2021; 64:348-358. [PMID: 33439741 PMCID: PMC8632513 DOI: 10.1044/2020_jslhr-20-00116] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/15/2020] [Revised: 07/28/2020] [Accepted: 09/21/2020] [Indexed: 06/12/2023]
Abstract
Purpose Dynamic pitch, which is defined as the variation in fundamental frequency, is an acoustic cue that aids speech perception in noise. This study examined the effects of strengthened and weakened dynamic pitch cues on older listeners' speech perception in noise, as well as how these effects were modulated by individual factors including spectral perception ability. Method The experiment measured speech reception thresholds in noise in both younger listeners with normal hearing and older listeners whose hearing status ranged from near-normal hearing to mild-to-moderate sensorineural hearing loss. The pitch contours of the target speech were manipulated to create four levels of dynamic pitch strength: weakened, original, mildly strengthened, and strengthened. Listeners' spectral perception ability was measured using tests of spectral ripple and frequency modulation discrimination. Results Both younger and older listeners performed worse with manipulated dynamic pitch cues than with original dynamic pitch. The effects of dynamic pitch on older listeners' speech recognition were associated with their age but not with their perception of spectral information. Those older listeners who were relatively younger were more negatively affected by dynamic pitch manipulations. Conclusions The findings suggest the current pitch manipulation strategy is detrimental for older listeners to perceive speech in noise, as compared to original dynamic pitch. While the influence of age on the effects of dynamic pitch is likely due to age-related declines in pitch perception, the spectral measures used in this study were not strong predictors for dynamic pitch effects. Taken together, these results indicate next steps in this line of work should be focused on how to manipulate acoustic cues in speech in order to improve speech perception in noise for older listeners.
Collapse
Affiliation(s)
- Jing Shen
- Department of Speech, Language and Hearing Sciences, Western Michigan University, Kalamazoo
| |
Collapse
|
35
|
Lutfi RA, Rodriguez B, Lee J. The Listener Effect in Multitalker Speech Segregation and Talker Identification. Trends Hear 2021; 25:23312165211051886. [PMID: 34693853 PMCID: PMC8544763 DOI: 10.1177/23312165211051886] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/02/2021] [Accepted: 09/20/2021] [Indexed: 12/04/2022] Open
Abstract
Over six decades ago, Cherry (1953) drew attention to what he called the "cocktail-party problem"; the challenge of segregating the speech of one talker from others speaking at the same time. The problem has been actively researched ever since but for all this time one observation has eluded explanation. It is the wide variation in performance of individual listeners. That variation was replicated here for four major experimental factors known to impact performance: differences in task (talker segregation vs. identification), differences in the voice features of talkers (pitch vs. location), differences in the voice similarity and uncertainty of talkers (informational masking), and the presence or absence of linguistic cues. The effect of these factors on the segregation of naturally spoken sentences and synthesized vowels was largely eliminated in psychometric functions relating the performance of individual listeners to that of an ideal observer, d'ideal. The effect of listeners remained as differences in the slopes of the functions (fixed effect) with little within-listener variability in the estimates of slope (random effect). The results make a case for considering the listener a factor in multitalker segregation and identification equal in status to any major experimental variable.
Collapse
Affiliation(s)
- Robert A. Lutfi
- Auditory Behavioral Research Lab, Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | - Briana Rodriguez
- Auditory Behavioral Research Lab, Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | - Jungmee Lee
- Auditory Behavioral Research Lab, Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| |
Collapse
|
36
|
Lutfi RA, Rodriguez B, Lee J, Pastore T. A test of model classes accounting for individual differences in the cocktail-party effect. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 148:4014. [PMID: 33379927 PMCID: PMC7775115 DOI: 10.1121/10.0002961] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/18/2020] [Revised: 11/06/2020] [Accepted: 12/03/2020] [Indexed: 06/12/2023]
Abstract
Listeners differ widely in the ability to follow the speech of a single talker in a noisy crowd-what is called the cocktail-party effect. Differences may arise for any one or a combination of factors associated with auditory sensitivity, selective attention, working memory, and decision making required for effective listening. The present study attempts to narrow the possibilities by grouping explanations into model classes based on model predictions for the types of errors that distinguish better from poorer performing listeners in a vowel segregation and talker identification task. Two model classes are considered: those for which the errors are predictably tied to the voice variation of talkers (decision weight models) and those for which the errors occur largely independently of this variation (internal noise models). Regression analyses of trial-by-trial responses, for different tasks and task demands, show overwhelmingly that the latter type of error is responsible for the performance differences among listeners. The results are inconsistent with models that attribute the performance differences to differences in the reliance listeners place on relevant voice features in this decision. The results are consistent instead with models for which largely stimulus-independent, stochastic processes cause information loss at different stages of auditory processing.
Collapse
Affiliation(s)
- Robert A Lutfi
- Auditory Behavioral Research Lab, Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida 33620, USA
| | - Briana Rodriguez
- Auditory Behavioral Research Lab, Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida 33620, USA
| | - Jungmee Lee
- Auditory Behavioral Research Lab, Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida 33620, USA
| | - Torben Pastore
- Spatial Hearing Lab, College of Health Solutions, Arizona State University, Tempe, Arizona 85281, USA
| |
Collapse
|
37
|
Cohen Priva U, Sanker C. Natural Leaders: Some Interlocutors Elicit Greater Convergence Across Conversations and Across Characteristics. Cogn Sci 2020; 44:e12897. [PMID: 33037640 DOI: 10.1111/cogs.12897] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2020] [Revised: 08/19/2020] [Accepted: 08/25/2020] [Indexed: 11/28/2022]
Abstract
Are there individual tendencies in convergence, such that some speakers consistently converge more than others? Similarly, are there natural "leaders," speakers with whom others converge more? Are such tendencies consistent across different linguistic characteristics? We use the Switchboard Corpus to perform a large-scale convergence study of speakers in multiple conversations with different interlocutors, across six linguistic characteristics. Because each speaker participated in several conversations, it is possible to look for individual differences in speakers' likelihood of converging and interlocutors' likelihood of eliciting convergence. We only find evidence for individual differences by interlocutor, not by speaker: There are natural leaders of convergence, who elicit more convergence than others across characteristics and across conversations. The lack of similar evidence for speakers who converge more than others suggests that social factors have a stronger effect in mediating convergence than putative individual tendencies in producing convergence, or that such tendencies are characteristic-specific.
Collapse
Affiliation(s)
- Uriel Cohen Priva
- Department of Cognitive, Linguistic, and Psychological Sciences, Brown University
| | | |
Collapse
|
38
|
Jaeger M, Mirkovic B, Bleichner MG, Debener S. Decoding the Attended Speaker From EEG Using Adaptive Evaluation Intervals Captures Fluctuations in Attentional Listening. Front Neurosci 2020; 14:603. [PMID: 32612507 PMCID: PMC7308709 DOI: 10.3389/fnins.2020.00603] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2019] [Accepted: 05/15/2020] [Indexed: 11/13/2022] Open
Abstract
Listeners differ in their ability to attend to a speech stream in the presence of a competing sound. Differences in speech intelligibility in noise cannot be fully explained by the hearing ability which suggests the involvement of additional cognitive factors. A better understanding of the temporal fluctuations in the ability to pay selective auditory attention to a desired speech stream may help in explaining these variabilities. In order to better understand the temporal dynamics of selective auditory attention, we developed an online auditory attention decoding (AAD) processing pipeline based on speech envelope tracking in the electroencephalogram (EEG). Participants had to attend to one audiobook story while a second one had to be ignored. Online AAD was applied to track the attention toward the target speech signal. Individual temporal attention profiles were computed by combining an established AAD method with an adaptive staircase procedure. The individual decoding performance over time was analyzed and linked to behavioral performance as well as subjective ratings of listening effort, motivation, and fatigue. The grand average attended speaker decoding profile derived in the online experiment indicated performance above chance level. Parameters describing the individual AAD performance in each testing block indicated significant differences in decoding performance over time to be closely related to the behavioral performance in the selective listening task. Further, an exploratory analysis indicated that subjects with poor decoding performance reported higher listening effort and fatigue compared to good performers. Taken together our results show that online EEG based AAD in a complex listening situation is feasible. Adaptive attended speaker decoding profiles over time could be used as an objective measure of behavioral performance and listening effort. The developed online processing pipeline could also serve as a basis for future EEG based near real-time auditory neurofeedback systems.
Collapse
Affiliation(s)
- Manuela Jaeger
- Neuropsychology Lab, Department of Psychology, University of Oldenburg, Oldenburg, Germany.,Fraunhofer Institute for Digital Media Technology IDMT, Division Hearing, Speech and Audio Technology, Oldenburg, Germany
| | - Bojana Mirkovic
- Neuropsychology Lab, Department of Psychology, University of Oldenburg, Oldenburg, Germany.,Cluster of Excellence Hearing4all, University of Oldenburg, Oldenburg, Germany
| | - Martin G Bleichner
- Neuropsychology Lab, Department of Psychology, University of Oldenburg, Oldenburg, Germany.,Neurophysiology of Everyday Life Lab, Department of Psychology, University of Oldenburg, Oldenburg, Germany
| | - Stefan Debener
- Neuropsychology Lab, Department of Psychology, University of Oldenburg, Oldenburg, Germany.,Cluster of Excellence Hearing4all, University of Oldenburg, Oldenburg, Germany.,Research Center for Neurosensory Science, University of Oldenburg, Oldenburg, Germany
| |
Collapse
|
39
|
The Timbre Perception Test (TPT): A new interactive musical assessment tool to measure timbre perception ability. Atten Percept Psychophys 2020; 82:3658-3675. [PMID: 32529570 PMCID: PMC7536169 DOI: 10.3758/s13414-020-02058-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
To date, tests that measure individual differences in the ability to perceive musical timbre are scarce in the published literature. The lack of such tool limits research on how timbre, a primary attribute of sound, is perceived and processed among individuals. The current paper describes the development of the Timbre Perception Test (TPT), in which participants use a slider to reproduce heard auditory stimuli that vary along three important dimensions of timbre: envelope, spectral flux, and spectral centroid. With a sample of 95 participants, the TPT was calibrated and validated against measures of related abilities and examined for its reliability. The results indicate that a short-version (8 minutes) of the TPT has good explanatory support from a factor analysis model, acceptable internal reliability (α = .69, ωt = .70), good test–retest reliability (r = .79) and substantial correlations with self-reported general musical sophistication (ρ = .63) and pitch discrimination (ρ = .56), as well as somewhat lower correlations with duration discrimination (ρ = .27), and musical instrument discrimination abilities (ρ = .33). Overall, the TPT represents a robust tool to measure an individual’s timbre perception ability. Furthermore, the use of sliders to perform a reproductive task has shown to be an effective approach in threshold testing. The current version of the TPT is openly available for research purposes.
Collapse
|
40
|
Jasmin K, Dick F, Holt LL, Tierney A. Tailored perception: Individuals' speech and music perception strategies fit their perceptual abilities. J Exp Psychol Gen 2020; 149:914-934. [PMID: 31589067 PMCID: PMC7133494 DOI: 10.1037/xge0000688] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2018] [Revised: 08/09/2019] [Accepted: 08/12/2019] [Indexed: 01/09/2023]
Abstract
Perception involves integration of multiple dimensions that often serve overlapping, redundant functions, for example, pitch, duration, and amplitude in speech. Individuals tend to prioritize these dimensions differently (stable, individualized perceptual strategies), but the reason for this has remained unclear. Here we show that perceptual strategies relate to perceptual abilities. In a speech cue weighting experiment (trial N = 990), we first demonstrate that individuals with a severe deficit for pitch perception (congenital amusics; N = 11) categorize linguistic stimuli similarly to controls (N = 11) when the main distinguishing cue is duration, which they perceive normally. In contrast, in a prosodic task where pitch cues are the main distinguishing factor, we show that amusics place less importance on pitch and instead rely more on duration cues-even when pitch differences in the stimuli are large enough for amusics to discern. In a second experiment testing musical and prosodic phrase interpretation (N = 16 amusics; 15 controls), we found that relying on duration allowed amusics to overcome their pitch deficits to perceive speech and music successfully. We conclude that auditory signals, because of their redundant nature, are robust to impairments for specific dimensions, and that optimal speech and music perception strategies depend not only on invariant acoustic dimensions (the physical signal), but on perceptual dimensions whose precision varies across individuals. Computational models of speech perception (indeed, all types of perception involving redundant cues e.g., vision and touch) should therefore aim to account for the precision of perceptual dimensions and characterize individuals as well as groups. (PsycInfo Database Record (c) 2020 APA, all rights reserved).
Collapse
Affiliation(s)
| | - Fred Dick
- Department of Psychological Sciences
| | | | | |
Collapse
|
41
|
Online sonification for golf putting gesture: reduced variability of motor behaviour and perceptual judgement. Exp Brain Res 2020; 238:883-895. [DOI: 10.1007/s00221-020-05757-3] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2019] [Accepted: 02/19/2020] [Indexed: 11/26/2022]
|
42
|
Toon J, Kukona A. Activating Semantic Knowledge During Spoken Words and Environmental Sounds: Evidence From the Visual World Paradigm. Cogn Sci 2020; 44:e12810. [PMID: 31960505 DOI: 10.1111/cogs.12810] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2018] [Revised: 10/09/2019] [Accepted: 11/11/2019] [Indexed: 11/30/2022]
Abstract
Two visual world experiments investigated the activation of semantically related concepts during the processing of environmental sounds and spoken words. Participants heard environmental sounds such as barking or spoken words such as "puppy" while viewing visual arrays with objects such as a bone (semantically related competitor) and candle (unrelated distractor). In Experiment 1, a puppy (target) was also included in the visual array; in Experiment 2, it was not. During both types of auditory stimuli, competitors were fixated significantly more than distractors, supporting the coactivation of semantically related concepts in both cases; comparisons of the two types of auditory stimuli also revealed significantly larger effects with environmental sounds than spoken words. We discuss implications of these results for theories of semantic knowledge.
Collapse
Affiliation(s)
- Josef Toon
- Division of Psychology, De Montfort University
| | | |
Collapse
|
43
|
Otsuka S, Nakagawa S, Furukawa S. Relationship between cochlear mechanics and speech-in-noise reception performance. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:EL265. [PMID: 31590549 DOI: 10.1121/1.5125008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/17/2019] [Accepted: 08/17/2019] [Indexed: 06/10/2023]
Abstract
Some normal-hearing listeners report difficulties in speech perception in noisy environments, and the cause is not well understood. The present study explores the correlation between speech-in-noise reception performance and cochlear mechanical characteristics, which were evaluated using a principal component analysis of the otoacoustic emission (OAE) spectra. A principal component, specifically a characteristic dip at around 2-2.5 kHz in OAE spectra, correlated with speech reception thresholds in noise but not in quiet. The results suggest that subclinical cochlear dysfunction specifically contributes to difficulties in speech perception in noisy environments, which is possibly a new form of "hidden hearing deficits."
Collapse
Affiliation(s)
- Sho Otsuka
- Center for Frontier Medical Engineering, Chiba University, 1-33 Yayoicho, Inageku, Chiba-shi, Chiba Pref., 263-8522 Japan
| | - Seiji Nakagawa
- Center for Frontier Medical Engineering, Chiba University, 1-33 Yayoicho, Inageku, Chiba-shi, Chiba Pref., 263-8522 Japan
| | - Shigeto Furukawa
- NTT Communication Science Laboratories, NTT Corporation, 3-1, Morinosato Wakamiya, Atsugi-shi, Kanagawa Pref., 243-0198, , ,
| |
Collapse
|
44
|
Anderson SR, Kan A, Litovsky RY. Asymmetric temporal envelope encoding: Implications for within- and across-ear envelope comparison. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:1189. [PMID: 31472559 PMCID: PMC7051005 DOI: 10.1121/1.5121423] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/25/2018] [Revised: 07/24/2019] [Accepted: 07/24/2019] [Indexed: 05/17/2023]
Abstract
Separating sound sources in acoustic environments relies on making ongoing, highly accurate spectro-temporal comparisons. However, listeners with hearing impairment may have varying quality of temporal encoding within or across ears, which may limit the listeners' ability to make spectro-temporal comparisons between places-of-stimulation. In this study in normal hearing listeners, depth of amplitude modulation (AM) for sinusoidally amplitude modulated (SAM) tones was manipulated in an effort to reduce the coding of periodicity in the auditory nerve. The ability to judge differences in AM rates was studied for stimuli presented to different cochlear places-of-stimulation, within- or across-ears. It was hypothesized that if temporal encoding was poorer for one tone in a pair, then sensitivity to differences in AM rate of the pair would decrease. Results indicated that when the depth of AM was reduced from 50% to 20% for one SAM tone in a pair, sensitivity to differences in AM rate decreased. Sensitivity was greatest for AM rates near 90 Hz and depended upon the places-of-stimulation being compared. These results suggest that degraded temporal representations in the auditory nerve for one place-of-stimulation could lead to deficits comparing that temporal information with other places-of-stimulation.
Collapse
Affiliation(s)
- Sean R Anderson
- Waisman Center, University of Wisconsin-Madison, Madison, Wisconsin 53705, USA
| | - Alan Kan
- Waisman Center, University of Wisconsin-Madison, Madison, Wisconsin 53705, USA
| | - Ruth Y Litovsky
- Waisman Center, University of Wisconsin-Madison, Madison, Wisconsin 53705, USA
| |
Collapse
|
45
|
Bharadwaj HM, Mai AR, Simpson JM, Choi I, Heinz MG, Shinn-Cunningham BG. Non-Invasive Assays of Cochlear Synaptopathy - Candidates and Considerations. Neuroscience 2019; 407:53-66. [PMID: 30853540 DOI: 10.1016/j.neuroscience.2019.02.031] [Citation(s) in RCA: 73] [Impact Index Per Article: 14.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2018] [Revised: 02/21/2019] [Accepted: 02/25/2019] [Indexed: 12/31/2022]
Abstract
Studies in multiple species, including in post-mortem human tissue, have shown that normal aging and/or acoustic overexposure can lead to a significant loss of afferent synapses innervating the cochlea. Hypothetically, this cochlear synaptopathy can lead to perceptual deficits in challenging environments and can contribute to central neural effects such as tinnitus. However, because cochlear synaptopathy can occur without any measurable changes in audiometric thresholds, synaptopathy can remain hidden from standard clinical diagnostics. To understand the perceptual sequelae of synaptopathy and to evaluate the efficacy of emerging therapies, sensitive and specific non-invasive measures at the individual patient level need to be established. Pioneering experiments in specific mice strains have helped identify many candidate assays. These include auditory brainstem responses, the middle-ear muscle reflex, envelope-following responses, and extended high-frequency audiograms. Unfortunately, because these non-invasive measures can be also affected by extraneous factors other than synaptopathy, their application and interpretation in humans is not straightforward. Here, we systematically examine six extraneous factors through a series of interrelated human experiments aimed at understanding their effects. Using strategies that may help mitigate the effects of such extraneous factors, we then show that these suprathreshold physiological assays exhibit across-individual correlations with each other indicative of contributions from a common physiological source consistent with cochlear synaptopathy. Finally, we discuss the application of these assays to two key outstanding questions, and discuss some barriers that still remain. This article is part of a Special Issue entitled: Hearing Loss, Tinnitus, Hyperacusis, Central Gain.
Collapse
Affiliation(s)
- Hari M Bharadwaj
- Department of Speech, Language, and Hearing Sciences, Purdue University, West Lafayette, IN; Weldon School of Biomedical Engineering, Purdue University, West Lafayette, IN.
| | - Alexandra R Mai
- Department of Speech, Language, and Hearing Sciences, Purdue University, West Lafayette, IN
| | - Jennifer M Simpson
- Department of Speech, Language, and Hearing Sciences, Purdue University, West Lafayette, IN
| | - Inyong Choi
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City, IA
| | - Michael G Heinz
- Department of Speech, Language, and Hearing Sciences, Purdue University, West Lafayette, IN; Weldon School of Biomedical Engineering, Purdue University, West Lafayette, IN
| | | |
Collapse
|
46
|
Different Associations between Auditory Function and Cognition Depending on Type of Auditory Function and Type of Cognition. Ear Hear 2019; 40:1210-1219. [PMID: 30807540 PMCID: PMC6706331 DOI: 10.1097/aud.0000000000000700] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
Supplemental Digital Content is available in the text. Previous studies strongly suggest that declines in auditory threshold can lead to impaired cognition. The aim of this study was to expand that picture by investigating how the relationships between age, auditory function, and cognitive function vary with the types of auditory and cognitive function considered.
Collapse
|
47
|
Kraus MS, Walker TM, Jarskog LF, Millet RA, Keefe RSE. Basic auditory processing deficits and their association with auditory emotion recognition in schizophrenia. Schizophr Res 2019; 204:155-161. [PMID: 30268821 DOI: 10.1016/j.schres.2018.08.031] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/25/2018] [Revised: 08/20/2018] [Accepted: 08/21/2018] [Indexed: 11/16/2022]
Abstract
BACKGROUND Individuals with schizophrenia are impaired in their ability to recognize emotions based on vocal cues and these impairments are associated with poor global outcome. Basic perceptual processes, such as auditory pitch processing, are impaired in schizophrenia and contribute to difficulty identifying emotions. However, previous work has focused on a relatively narrow assessment of auditory deficits and their relation to emotion recognition impairment in schizophrenia. METHODS We have assessed 87 patients with schizophrenia and 73 healthy controls on a comprehensive battery of tasks spanning the five empirically derived domains of auditory function. We also explored the relationship between basic auditory processing and auditory emotion recognition within the patient group using correlational analysis. RESULTS Patients exhibited widespread auditory impairments across multiple domains of auditory function, with mostly medium effect sizes. Performance on all of the basic auditory tests correlated with auditory emotion recognition at the p < .01 level in the patient group, with 9 out of 13 tests correlating with emotion recognition at r = 0.40 or greater. After controlling for cognition, many of the largest correlations involved spectral processing within the phase-locking range and discrimination of vocally based stimuli. CONCLUSIONS While many auditory skills contribute to this impairment, deficient formant discrimination appears to be a key skill contributing to impaired emotion recognition as this was the only basic auditory skill to enter a step-wise multiple regression after first entering a measure of cognitive impairment, and formant discrimination accounted for significant unique variance in emotion recognition performance after accounting for deficits in pitch processing.
Collapse
Affiliation(s)
- Michael S Kraus
- Department of Psychiatry & Behavioral Sciences, Duke University Medical Center, 10 Duke Medicine Circle, Durham, NC 27710, USA.
| | - Trina M Walker
- Department of Psychiatry & Behavioral Sciences, Duke University Medical Center, 10 Duke Medicine Circle, Durham, NC 27710, USA
| | - L Fredrik Jarskog
- North Carolina Psychiatric Research Center, Department of Psychiatry, University of North Carolina at Chapel Hill, 101 Manning Dr # 1, Chapel Hill, NC 27514, USA
| | - Robert A Millet
- Carolina Behavioral Care, 4102 Ben Franklin Blvd Durham, NC 27704, USA
| | - Richard S E Keefe
- Department of Psychiatry & Behavioral Sciences, Duke University Medical Center, 10 Duke Medicine Circle, Durham, NC 27710, USA
| |
Collapse
|
48
|
Modular reconfiguration of an auditory control brain network supports adaptive listening behavior. Proc Natl Acad Sci U S A 2018; 116:660-669. [PMID: 30587584 PMCID: PMC6329957 DOI: 10.1073/pnas.1815321116] [Citation(s) in RCA: 34] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
How do brain networks shape our listening behavior? We here develop and test the hypothesis that, during challenging listening situations, intrinsic brain networks are reconfigured to adapt to the listening demands and thus, to enable successful listening. We find that, relative to a task-free resting state, networks of the listening brain show higher segregation of temporal auditory, ventral attention, and frontal control regions known to be involved in speech processing, sound localization, and effortful listening. Importantly, the relative change in modularity of this auditory control network predicts individuals’ listening success. Our findings shed light on how cortical communication dynamics tune selection and comprehension of speech in challenging listening situations and suggest modularity as the network principle of auditory attention. Speech comprehension in noisy, multitalker situations poses a challenge. Successful behavioral adaptation to a listening challenge often requires stronger engagement of auditory spatial attention and context-dependent semantic predictions. Human listeners differ substantially in the degree to which they adapt behaviorally and can listen successfully under such circumstances. How cortical networks embody this adaptation, particularly at the individual level, is currently unknown. We here explain this adaptation from reconfiguration of brain networks for a challenging listening task (i.e., a linguistic variant of the Posner paradigm with concurrent speech) in an age-varying sample of n = 49 healthy adults undergoing resting-state and task fMRI. We here provide evidence for the hypothesis that more successful listeners exhibit stronger task-specific reconfiguration (hence, better adaptation) of brain networks. From rest to task, brain networks become reconfigured toward more localized cortical processing characterized by higher topological segregation. This reconfiguration is dominated by the functional division of an auditory and a cingulo-opercular module and the emergence of a conjoined auditory and ventral attention module along bilateral middle and posterior temporal cortices. Supporting our hypothesis, the degree to which modularity of this frontotemporal auditory control network is increased relative to resting state predicts individuals’ listening success in states of divided and selective attention. Our findings elucidate how fine-tuned cortical communication dynamics shape selection and comprehension of speech. Our results highlight modularity of the auditory control network as a key organizational principle in cortical implementation of auditory spatial attention in challenging listening situations.
Collapse
|
49
|
Mednicoff S, Mejia S, Rashid JA, Chubb C. Many listeners cannot discriminate major vs minor tone-scrambles regardless of presentation rate. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 144:2242. [PMID: 30404491 PMCID: PMC6192795 DOI: 10.1121/1.5055990] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/26/2018] [Revised: 07/10/2018] [Accepted: 09/10/2018] [Indexed: 05/19/2023]
Abstract
A tone-scramble is a random sequence of pure tones. Previous studies have found that most listeners (≈ 70%) perform near chance in classifying rapid tone-scrambles composed of multiple copies of notes in G-major vs G-minor triads; the remaining listeners perform nearly perfectly [Chubb, Dickson, Dean, Fagan, Mann, Wright, Guan, Silva, Gregersen, and Kowalski (2013). J. Acoust. Soc. Am. 134(4), 3067-3078; Dean and Chubb (2017). J. Acoust. Soc. Am. 142(3), 1432-1440]. This study tested whether low-performing listeners might improve with slower stimuli. In separate tasks, stimuli were tone-scrambles presented at 115, 231, 462, and 923 notes per min. In each task, the listener classified (with feedback) stimuli as major vs minor. Listeners who performed poorly in any of these tasks performed poorly in all of them. Strikingly, performance was worst in the task with the slowest stimuli. In all tasks, most listeners were biased to respond "major" ("minor") if the stimulus ended on a note high (low) in pitch. Dean and Chubb introduced the name "scale-sensitivity" for the cognitive resource that separates high- from low-performing listeners in tone-scramble classification tasks, suggesting that this resource confers sensitivity to the full gamut of qualities that music can attain by being in a scale. In ruling out the possibility that performance in these tasks depends on speed of presentation, the current results bolster this interpretation.
Collapse
Affiliation(s)
- Solena Mednicoff
- Department of Cognitive Sciences, University of California at Irvine, Irvine, California 92697-5100, USA
| | - Stephanie Mejia
- Department of Cognitive Sciences, University of California at Irvine, Irvine, California 92697-5100, USA
| | - Jordan Ali Rashid
- Department of Cognitive Sciences, University of California at Irvine, Irvine, California 92697-5100, USA
| | - Charles Chubb
- Department of Cognitive Sciences, University of California at Irvine, Irvine, California 92697-5100, USA
| |
Collapse
|
50
|
Schnakenberg Martin AM, Bartolomeo L, Howell J, Hetrick WP, Bolbecker AR, Breier A, Kidd G, O'Donnell BF. Auditory feature perception and auditory hallucinatory experiences in schizophrenia spectrum disorder. Eur Arch Psychiatry Clin Neurosci 2018; 268:653-661. [PMID: 28936548 PMCID: PMC7126185 DOI: 10.1007/s00406-017-0839-1] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/25/2017] [Accepted: 09/14/2017] [Indexed: 01/20/2023]
Abstract
Schizophrenia spectrum disorder (SZ) is associated with deficits in auditory perception as well as auditory verbal hallucinations (AVH). However, the relationship between auditory feature perception and auditory verbal hallucinations (AVH), one of the most commonly occurring symptoms in psychosis, has not been well characterized. This study evaluated perception of a broad range of auditory features in SZ and determined whether current AVHs relate to auditory feature perception. Auditory perception, including frequency, intensity, duration, pulse-train and temporal order discrimination, as well as an embedded tone task, was assessed in both AVH (n = 20) and non-AVH (n = 24) SZ individuals and in healthy controls (n = 29) with the Test of Basic Auditory Capabilities (TBAC). The Hamilton Program for Schizophrenia Voices Questionnaire (HPSVQ) was used to assess the experience of auditory hallucinations in patients with SZ. Findings suggest that compared to controls, the SZ group had greater deficits on an array of auditory features, with non-AVH SZ individuals showing the most severe degree of abnormality. IQ and measures of cognitive processing were positively associated with performance on the TBAC for all SZ individuals, but not with the HPSVQ scores. These findings indicate that persons with SZ demonstrate impaired auditory perception for a broad range of features. It does not appear that impaired auditory perception is associated with recent auditory verbal hallucinations, but instead associated with the degree of intellectual impairment in SZ.
Collapse
Affiliation(s)
- Ashley M Schnakenberg Martin
- Department of Psychological and Brain Sciences, Indiana University-Bloomington, 1101 E 10th Street, Bloomington, IN, 47401, USA.
- Larue D. Carter Memorial Hospital, Indianapolis, IN, USA.
| | - Lisa Bartolomeo
- Department of Psychological and Brain Sciences, Indiana University-Bloomington, 1101 E 10th Street, Bloomington, IN, 47401, USA
- Larue D. Carter Memorial Hospital, Indianapolis, IN, USA
| | - Josselyn Howell
- Department of Psychological and Brain Sciences, Indiana University-Bloomington, 1101 E 10th Street, Bloomington, IN, 47401, USA
- Larue D. Carter Memorial Hospital, Indianapolis, IN, USA
| | - William P Hetrick
- Department of Psychological and Brain Sciences, Indiana University-Bloomington, 1101 E 10th Street, Bloomington, IN, 47401, USA
- Larue D. Carter Memorial Hospital, Indianapolis, IN, USA
- Department of Psychiatry, Indiana University School of Medicine, Indianapolis, IN, USA
| | - Amanda R Bolbecker
- Department of Psychological and Brain Sciences, Indiana University-Bloomington, 1101 E 10th Street, Bloomington, IN, 47401, USA
- Larue D. Carter Memorial Hospital, Indianapolis, IN, USA
| | - Alan Breier
- Larue D. Carter Memorial Hospital, Indianapolis, IN, USA
- Department of Psychiatry, Indiana University School of Medicine, Indianapolis, IN, USA
| | - Gary Kidd
- Department of Speech and Hearing Sciences, Indiana University-Bloomington, Bloomington, IN, USA
| | - Brian F O'Donnell
- Department of Psychological and Brain Sciences, Indiana University-Bloomington, 1101 E 10th Street, Bloomington, IN, 47401, USA
- Larue D. Carter Memorial Hospital, Indianapolis, IN, USA
- Department of Psychiatry, Indiana University School of Medicine, Indianapolis, IN, USA
| |
Collapse
|