1
|
Suri H, Rothschild G. Enhanced stability of complex sound representations relative to simple sounds in the auditory cortex. eNeuro 2022; 9:ENEURO.0031-22.2022. [PMID: 35868858 PMCID: PMC9347310 DOI: 10.1523/eneuro.0031-22.2022] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/21/2022] [Revised: 06/29/2022] [Accepted: 06/30/2022] [Indexed: 11/29/2022] Open
Abstract
Typical everyday sounds, such as those of speech or running water, are spectrotemporally complex. The ability to recognize complex sounds (CxS) and their associated meaning is presumed to rely on their stable neural representations across time. The auditory cortex is critical for processing of CxS, yet little is known of the degree of stability of auditory cortical representations of CxS across days. Previous studies have shown that the auditory cortex represents CxS identity with a substantial degree of invariance to basic sound attributes such as frequency. We therefore hypothesized that auditory cortical representations of CxS are more stable across days than those of sounds that lack spectrotemporal structure such as pure tones (PTs). To test this hypothesis, we recorded responses of identified L2/3 auditory cortical excitatory neurons to both PTs and CxS across days using two-photon calcium imaging in awake mice. Auditory cortical neurons showed significant daily changes of responses to both types of sounds, yet responses to CxS exhibited significantly lower rates of daily change than those of PTs. Furthermore, daily changes in response profiles to PTs tended to be more stimulus-specific, reflecting changes in sound selectivity, as compared to changes of CxS responses. Lastly, the enhanced stability of responses to CxS was evident across longer time intervals as well. Together, these results suggest that spectrotemporally CxS are more stably represented in the auditory cortex across time than PTs. These findings support a role of the auditory cortex in representing CxS identity across time.Significance statementThe ability to recognize everyday complex sounds such as those of speech or running water is presumed to rely on their stable neural representations. Yet, little is known of the degree of stability of single-neuron sound responses across days. As the auditory cortex is critical for complex sound perception, we hypothesized that the auditory cortical representations of complex sounds are relatively stable across days. To test this, we recorded sound responses of identified auditory cortical neurons across days in awake mice. We found that auditory cortical responses to complex sounds are significantly more stable across days as compared to those of simple pure tones. These findings support a role of the auditory cortex in representing complex sound identity across time.
Collapse
Affiliation(s)
- Harini Suri
- Department of Psychology, University of Michigan, Ann Arbor, MI, 48109, USA
| | - Gideon Rothschild
- Department of Psychology, University of Michigan, Ann Arbor, MI, 48109, USA
- Kresge Hearing Research Institute and Department of Otolaryngology - Head and Neck Surgery, University of Michigan, Ann Arbor, MI 48109, USA
| |
Collapse
|
2
|
Li H, Liang F, Zhong W, Yan L, Mesik L, Xiao Z, Tao HW, Zhang LI. Synaptic Mechanisms for Bandwidth Tuning in Awake Mouse Primary Auditory Cortex. Cereb Cortex 2020; 29:2998-3009. [PMID: 30010857 DOI: 10.1093/cercor/bhy165] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2018] [Revised: 06/08/2018] [Indexed: 11/12/2022] Open
Abstract
Spatial size tuning in the visual cortex has been considered as an important neuronal functional property for sensory perception. However, an analogous mechanism in the auditory system has remained controversial. In the present study, cell-attached recordings in the primary auditory cortex (A1) of awake mice revealed that excitatory neurons can be categorized into three types according to their bandwidth tuning profiles in response to band-passed noise (BPN) stimuli: nonmonotonic (NM), flat, and monotonic, with the latter two considered as non-tuned for bandwidth. The prevalence of bandwidth-tuned (i.e., NM) neurons increases significantly from layer 4 to layer 2/3. With sequential cell-attached and whole-cell voltage-clamp recordings from the same neurons, we found that the bandwidth preference of excitatory neurons is largely determined by the excitatory synaptic input they receive, and that the bandwidth selectivity is further enhanced by flatly tuned inhibition observed in all cells. The latter can be attributed at least partially to the flat tuning of parvalbumin inhibitory neurons. The tuning of auditory cortical neurons for bandwidth of BPN may contribute to the processing of complex sounds.
Collapse
Affiliation(s)
- Haifu Li
- Department of Physiology, School of Basic Medical Sciences, Southern Medical University, Guangzhou, China.,Zilkha Neurogenetic Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| | - Feixue Liang
- Zilkha Neurogenetic Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA.,Department of Medical Engineering, School of Biomedical Engineering, Southern Medical University, Guangzhou, China
| | - Wen Zhong
- Department of Physiology, School of Basic Medical Sciences, Southern Medical University, Guangzhou, China.,Zilkha Neurogenetic Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| | - Linqing Yan
- Department of Physiology, School of Basic Medical Sciences, Southern Medical University, Guangzhou, China.,Zilkha Neurogenetic Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| | - Lucas Mesik
- Zilkha Neurogenetic Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA.,Neuroscience Graduate Program, University of Southern California, Los Angeles, CA, USA
| | - Zhongju Xiao
- Department of Physiology, School of Basic Medical Sciences, Southern Medical University, Guangzhou, China
| | - Huizhong W Tao
- Zilkha Neurogenetic Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA.,Department of Physiology and Neuroscience, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| | - Li I Zhang
- Zilkha Neurogenetic Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA.,Department of Physiology and Neuroscience, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| |
Collapse
|
3
|
Sound identity is represented robustly in auditory cortex during perceptual constancy. Nat Commun 2018; 9:4786. [PMID: 30429465 PMCID: PMC6235866 DOI: 10.1038/s41467-018-07237-3] [Citation(s) in RCA: 20] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2018] [Accepted: 10/23/2018] [Indexed: 12/02/2022] Open
Abstract
Perceptual constancy requires neural representations that are selective for object identity, but also tolerant across identity-preserving transformations. How such representations arise in the brain and support perception remains unclear. Here, we study tolerant representation of sound identity in the auditory system by recording neural activity in auditory cortex of ferrets during perceptual constancy. Ferrets generalize vowel identity across variations in fundamental frequency, sound level and location, while neurons represent sound identity robustly across acoustic variations. Stimulus features are encoded with distinct time-courses in all conditions, however encoding of sound identity is delayed when animals fail to generalize and during passive listening. Neurons also encode information about task-irrelevant sound features, as well as animals’ choices and accuracy, while population decoding out-performs animals’ behavior. Our results show that during perceptual constancy, sound identity is represented robustly in auditory cortex across widely varying conditions, and behavioral generalization requires conserved timing of identity information. Perceptual constancy requires neural representations selective for object identity, yet tolerant of identity-preserving transformations. Here, the authors show that sound identity is represented robustly in auditory cortex and that behavioral generalization requires precise timing of identity information.
Collapse
|
4
|
Affiliation(s)
- Daniel J. Levitin
- Department of Psychology, McGill University, Montreal, QC H3A 1G1, Canada
| | - Jessica A. Grahn
- Department of Psychology and Brain and Mind Institute, Western University, London, Ontario N6A 5B7, Canada
| | - Justin London
- Departments of Music and Cognitive Science, Carleton College, Northfield, Minnesota 55057
| |
Collapse
|
5
|
Abstract
Goal-directed behavior can be characterized as a dynamic link between a sensory stimulus and a motor act. Neural correlates of many of the intermediate events of goal-directed behavior are found in the posterior parietal cortex. Although the parietal cortex’s role in guiding visual behaviors has received considerable attention, relatively little is known about its role in mediating auditory behaviors. Here, the authors review recent studies that have focused on how neurons in the lateral intraparietal area (area LIP) differentially process auditory and visual stimuli. These studies suggest that area LIP contains a modality-dependent representation that is highly dependent on behavioral context.
Collapse
Affiliation(s)
- Yale E Cohen
- Department of Psychological and Brain Sciences, Center for Cognitive Neuroscience, Dartmouth College, Hanover, NH
| | | | | |
Collapse
|
6
|
Hendrickson K, Walenski M, Friend M, Love T. The organization of words and environmental sounds in memory. Neuropsychologia 2015; 69:67-76. [PMID: 25624059 DOI: 10.1016/j.neuropsychologia.2015.01.035] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2014] [Revised: 11/21/2014] [Accepted: 01/23/2015] [Indexed: 11/24/2022]
Abstract
In the present study we used event-related potentials to compare the organization of linguistic and meaningful nonlinguistic sounds in memory. We examined N400 amplitudes as adults viewed pictures presented with words or environmental sounds that matched the picture (Match), that shared semantic features with the expected match (Near Violation), and that shared relatively few semantic features with the expected match (Far Violation). Words demonstrated incremental N400 amplitudes based on featural similarity from 300-700ms, such that both Near and Far Violations exhibited significant N400 effects, however Far Violations exhibited greater N400 effects than Near Violations. For environmental sounds, Far Violations but not Near Violations elicited significant N400 effects, in both early (300-400ms) and late (500-700ms) time windows, though a graded pattern similar to that of words was seen in the mid-latency time window (400-500ms). These results indicate that the organization of words and environmental sounds in memory is differentially influenced by featural similarity, with a consistently fine-grained graded structure for words but not sounds.
Collapse
Affiliation(s)
- Kristi Hendrickson
- Center for Research in Language, University of California, San Diego, USA; School of Speech, Language, and Hearing Sciences, San Diego State University, USA; Joint Doctoral Program in Language and Communicative Disorders, San Diego State University, USA.
| | - Matthew Walenski
- School of Speech, Language, and Hearing Sciences, San Diego State University, USA.
| | | | - Tracy Love
- Center for Research in Language, University of California, San Diego, USA; School of Speech, Language, and Hearing Sciences, San Diego State University, USA.
| |
Collapse
|
7
|
Christison-Lagay KL, Bennur S, Blackwell J, Lee JH, Schroeder T, Cohen YE. Natural variability in species-specific vocalizations constrains behavior and neural activity. Hear Res 2014; 312:128-42. [PMID: 24721001 PMCID: PMC4057037 DOI: 10.1016/j.heares.2014.03.007] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/31/2014] [Revised: 03/07/2014] [Accepted: 03/13/2014] [Indexed: 11/30/2022]
Abstract
A listener's capacity to discriminate between sounds is related to the amount of acoustic variability that exists between these sounds. However, a full understanding of how this natural variability impacts neural activity and behavior is lacking. Here, we tested monkeys' ability to discriminate between different utterances of vocalizations from the same acoustic class (i.e., coos and grunts), while neural activity was simultaneously recorded in the anterolateral belt region (AL) of the auditory cortex, a brain region that is a part of a pathway that mediates auditory perception. Monkeys could discriminate between coos better than they could discriminate between grunts. We also found AL activity was more informative about different coos than different grunts. This difference could be attributed, in part, to our finding that coos had more acoustic variability than grunts. Thus, intrinsic acoustic variability constrained the discriminability of AL spike trains and the ability of rhesus monkeys to discriminate between vocalizations.
Collapse
Affiliation(s)
| | - Sharath Bennur
- Dept. Otorhinolaryngology, U. Pennsylvania, Philadelphia, PA 19104, USA
| | - Jennifer Blackwell
- Neuroscience Graduate Group, University of Pennsylvania, Philadelphia, PA 19104, USA
| | - Jung H Lee
- Dept. Otorhinolaryngology, U. Pennsylvania, Philadelphia, PA 19104, USA
| | - Tim Schroeder
- Dept. Otorhinolaryngology, U. Pennsylvania, Philadelphia, PA 19104, USA
| | - Yale E Cohen
- Dept. Otorhinolaryngology, U. Pennsylvania, Philadelphia, PA 19104, USA; Neuroscience, U. Pennsylvania, Philadelphia, PA 19104, USA; Bioengineering, U. Pennsylvania, Philadelphia, PA 19104, USA.
| |
Collapse
|
8
|
Massoudi R, Van Wanrooij MM, Van Wetter SMCI, Versnel H, Van Opstal AJ. Stable bottom-up processing during dynamic top-down modulations in monkey auditory cortex. Eur J Neurosci 2013; 37:1830-42. [PMID: 23510187 DOI: 10.1111/ejn.12180] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2013] [Accepted: 02/11/2013] [Indexed: 11/26/2022]
Abstract
It is unclear whether top-down processing in the auditory cortex (AC) interferes with its bottom-up analysis of sound. Recent studies indicated non-acoustic modulations of AC responses, and that attention changes a neuron's spectrotemporal tuning. As a result, the AC would seem ill-suited to represent a stable acoustic environment, which is deemed crucial for auditory perception. To assess whether top-down signals influence acoustic tuning in tasks without directed attention, we compared monkey single-unit AC responses to dynamic spectrotemporal sounds under different behavioral conditions. Recordings were mostly made from neurons located in primary fields (primary AC and area R of the AC) that were well tuned to pure tones, with short onset latencies. We demonstrated that responses in the AC were substantially modulated during an auditory detection task and that these modulations were systematically related to top-down processes. Importantly, despite these significant modulations, the spectrotemporal receptive fields of all neurons remained remarkably stable. Our results suggest multiplexed encoding of bottom-up acoustic and top-down task-related signals at single AC neurons. This mechanism preserves a stable representation of the acoustic environment despite strong non-acoustic modulations.
Collapse
Affiliation(s)
- Roohollah Massoudi
- Department of Biophysics, Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands
| | | | | | | | | |
Collapse
|
9
|
Bizley J, King R. What Can Multisensory Processing Tell Us about the Functional Organization of Auditory Cortex? Front Neurosci 2011. [DOI: 10.1201/b11092-5] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022] Open
|
10
|
Bizley J, King R. What Can Multisensory Processing Tell Us about the Functional Organization of Auditory Cortex? Front Neurosci 2011. [DOI: 10.1201/9781439812174-5] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022] Open
|
11
|
Perrodin C, Kayser C, Logothetis NK, Petkov CI. Voice cells in the primate temporal lobe. Curr Biol 2011; 21:1408-15. [PMID: 21835625 PMCID: PMC3398143 DOI: 10.1016/j.cub.2011.07.028] [Citation(s) in RCA: 118] [Impact Index Per Article: 9.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2011] [Revised: 06/20/2011] [Accepted: 07/20/2011] [Indexed: 11/28/2022]
Abstract
Communication signals are important for social interactions and survival and are thought to receive specialized processing in the visual and auditory systems. Whereas the neural processing of faces by face clusters and face cells has been repeatedly studied [1-5], less is known about the neural representation of voice content. Recent functional magnetic resonance imaging (fMRI) studies have localized voice-preferring regions in the primate temporal lobe [6, 7], but the hemodynamic response cannot directly assess neurophysiological properties. We investigated the responses of neurons in an fMRI-identified voice cluster in awake monkeys, and here we provide the first systematic evidence for voice cells. "Voice cells" were identified, in analogy to "face cells," as neurons responding at least 2-fold stronger to conspecific voices than to "nonvoice" sounds or heterospecific voices. Importantly, whereas face clusters are thought to contain high proportions of face cells [4] responding broadly to many faces [1, 2, 4, 5, 8-10], we found that voice clusters contain moderate proportions of voice cells. Furthermore, individual voice cells exhibit high stimulus selectivity. The results reveal the neurophysiological bases for fMRI-defined voice clusters in the primate brain and highlight potential differences in how the auditory and visual systems generate selective representations of communication signals.
Collapse
Affiliation(s)
- Catherine Perrodin
- Department of Physiology of Cognitive Processes, Max Planck Institute for Biological Cybernetics, 72076 Tübingen, Germany
| | | | | | | |
Collapse
|
12
|
Goycoolea M, Mena I, Neubauer S. Functional studies (NeuroSPECT) of the human auditory pathway after stimulating binaurally with pure tones. Acta Otolaryngol 2011; 131:371-6. [PMID: 21314243 DOI: 10.3109/00016489.2010.545076] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
Abstract
CONCLUSIONS Our observations support the concept of bilateral cortical activation with monaural and binaural auditory stimulation. The observation that most of the significantly activated areas were the same with monaural or binaural stimulation suggests that the differences in auditory perception with binaural stimulation are not due to the involvement of significantly different processing centers but, more likely, to the type of information that reaches these centers for processing. The observation that the degree of stimulation was less intense in binaural than in monaural stimulation supports the concept that a richer binaural auditory stimulation compared with monaural stimulation does not mean summation of stimuli but integration and better processing of the information. For normal bilateral hearing subjects, a monaural stimulus is an uncommon event and may thus explain the more intense response. The repeatability of the results for monaural and binaural stimulation with pure tones in the same subjects confirms the consistency of the testing method. OBJECTIVES (1) To determine which areas of the cerebral cortex and basal ganglia are activated by binaural stimulation with pure tones (left and right ear simultaneously) and what type of response occurs (e.g. excitatory or inhibitory) in these different areas. (2) To determine the degree of ipsilateral and/or contralateral cortical activation and/or inhibition. (3) To compare the data with our previous reports of monaural stimulation using the same technique and the same subjects. (4) To evaluate the consistency of our testing method. METHODS Brain perfusion single photon emission computed tomography (SPECT) evaluation was conducted using auditory binaural stimulation with pure tones in six normal volunteers. Both ears were stimulated simultaneously. Tc99m HMPAO was injected while pure tones were delivered and the SPECT imaging was done 1 h later. RESULTS After delivering pure tones there was bilateral activation in Brodmann areas 7 (somatosensory association cortex), 9 and 10 (executive frontal areas), 17, 18, and 19 (associative visual cortex). There was also activation in temporal areas 21, 22 (auditory association areas), and parietal areas 39 and 40 (Wernicke). There was also marked activation in both thalami. These activated areas were the same as those in our previous reports with monaural stimulation in the same subjects. However, except for areas 17, 18, 23, 31, and 32 (which remained over 4 SD above the normal maximum), the degree of activation was less intense in binaural compared with monaural stimulation. Inhibition was also less intense in binaural stimulation.
Collapse
Affiliation(s)
- Marcos Goycoolea
- Department of Otorhinolaryngology, Clínica Las Condes, Santiago, Chile.
| | | | | |
Collapse
|
13
|
Recanzone GH. The biological basis of audition. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2010; 2:408-418. [PMID: 26302200 DOI: 10.1002/wcs.118] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Audition is one of the fundamental extrasensory percepts in mammals. Two of the primary objectives of audition are to determine where sounds originate from in space and what those sounds are. Neural processing of acoustic signals, which are commonly quite complex under natural conditions, is extensive in the brainstem, midbrain, and thalamus. This processing extracts multiple salient features that are then transmitted to the cerebral cortex. The cerebral cortex is a necessary neural structure for audition, or the perception of acoustic auditory objects and/or events. This entry will review the early processing along the ascending auditory central nervous system from the cochlea to the cerebral cortex. The neural mechanisms of audition will then be explored for spatial and non-spatial perception, drawing largely on examples from non-human primates, but insights gained from other mammalian species will also be covered. How these models relate to current studies in human subjects, using both functional imaging and invasive techniques, will also be explored as well as the types of future studies that will enable us to better understand the neural mechanisms of audition. WIREs Cogni Sci 2011 2 408-418 DOI: 10.1002/wcs.118 For further resources related to this article, please visit the WIREs website.
Collapse
Affiliation(s)
- Gregg H Recanzone
- Center for Neuroscience and Department of Neurobiology, Physiology and Behavior, University of California at Davis, One Shields Avenue, Davis, CA, USA
| |
Collapse
|
14
|
Krishnan A, Gandour JT, Bidelman GM. The effects of tone language experience on pitch processing in the brainstem. JOURNAL OF NEUROLINGUISTICS 2010; 23:81-95. [PMID: 20161561 PMCID: PMC2805250 DOI: 10.1016/j.jneuroling.2009.09.001] [Citation(s) in RCA: 86] [Impact Index Per Article: 6.1] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
Neural encoding of pitch in the auditory brainstem is shaped by long-term experience with language. The aim herein was to determine to what extent this experience-dependent effect is specific to a particular language. Analysis of variance of brainstem responses to Mandarin and Thai tones revealed that regardless of language identity, pitch-tracking accuracy of whole tones was higher in the two tone language groups (Chinese, Thai) compared to the non-tone language group (English), and that pitch strength of 40-ms tonal sections was generally more robust in tone relative to non-tone languages. Discriminant analysis of tonal sections, as defined by variation in direction and degree of slope, showed that moderate rising pitch was the most important variable for classifying English, Chinese, and Thai participants into their respective groups. We conclude that language-dependent enhancement of pitch representation transfers to other languages with similar phonological systems. From a neurobiological perspective, these findings suggest that neural mechanisms local to the brainstem are tuned for processing pitch dimensions that are perceptually salient depending upon the melodic patterns of a language.
Collapse
Affiliation(s)
- Ananthanarayan Krishnan
- Department of Speech Language Hearing Sciences, Purdue University, West Lafayette, Indiana, USA
| | - Jackson T. Gandour
- Department of Speech Language Hearing Sciences, Purdue University, West Lafayette, Indiana, USA
| | - Gavin M. Bidelman
- Department of Speech Language Hearing Sciences, Purdue University, West Lafayette, Indiana, USA
| |
Collapse
|
15
|
A functional role for the ventrolateral prefrontal cortex in non-spatial auditory cognition. Proc Natl Acad Sci U S A 2009; 106:20045-50. [PMID: 19897723 DOI: 10.1073/pnas.0907248106] [Citation(s) in RCA: 49] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Spatial and non-spatial sensory information is hypothesized to be evaluated in parallel pathways. In this study, we tested the spatial and non-spatial sensitivity of auditory neurons in the ventrolateral prefrontal cortex (vPFC), a cortical area in the non-spatial pathway. Activity was tested while non-human primates reported changes in an auditory stimulus' spatial or non-spatial features. We found that vPFC neurons were reliably modulated during a non-spatial auditory task but were not modulated during a spatial auditory task. The degree of modulation during the non-spatial task correlated positively with the monkeys' behavioral performance. These results are consistent with the hypotheses that the vPFC is part of a circuit involved in non-spatial auditory processing and that the vPFC plays a functional role in non-spatial auditory cognition.
Collapse
|
16
|
Krishnan A, Gandour JT. The role of the auditory brainstem in processing linguistically-relevant pitch patterns. BRAIN AND LANGUAGE 2009; 110:135-48. [PMID: 19366639 PMCID: PMC2731823 DOI: 10.1016/j.bandl.2009.03.005] [Citation(s) in RCA: 103] [Impact Index Per Article: 6.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/28/2008] [Revised: 01/12/2009] [Accepted: 03/15/2009] [Indexed: 05/15/2023]
Abstract
Historically, the brainstem has been neglected as a part of the brain involved in language processing. We review recent evidence of language-dependent effects in pitch processing based on comparisons of native vs. nonnative speakers of a tonal language from electrophysiological recordings in the auditory brainstem. We argue that there is enhancing of linguistically-relevant pitch dimensions or features well before the auditory signal reaches the cerebral cortex. We propose that long-term experience with a tone language sharpens the tuning characteristics of neurons along the pitch axis with enhanced sensitivity to linguistically-relevant, rapidly changing sections of pitch contours. Though not specific to a speech context, experience-dependent brainstem mechanisms for pitch representation are clearly sensitive to particular aspects of pitch contours that native speakers of a tone language have been exposed to. Such experience-dependent effects on lower-level sensory processing are compatible with more integrated, hierarchically organized pathways to language and the brain.
Collapse
|
17
|
Recanzone GH, Cohen YE. Serial and parallel processing in the primate auditory cortex revisited. Behav Brain Res 2009; 206:1-7. [PMID: 19686779 DOI: 10.1016/j.bbr.2009.08.015] [Citation(s) in RCA: 66] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2009] [Accepted: 08/12/2009] [Indexed: 11/24/2022]
Abstract
Over a decade ago it was proposed that the primate auditory cortex is organized in a serial and parallel manner in which there is a dorsal stream processing spatial information and a ventral stream processing non-spatial information. This organization is similar to the "what"/"where" processing of the primate visual cortex. This review will examine several key studies, primarily electrophysiological, that have tested this hypothesis. We also review several human-imaging studies that have attempted to define these processing streams in the human auditory cortex. While there is good evidence that spatial information is processed along a particular series of cortical areas, the support for a non-spatial processing stream is not as strong. Why this should be the case and how to better test this hypothesis is also discussed.
Collapse
Affiliation(s)
- Gregg H Recanzone
- Center for Neuroscience and Department of Neurobiology, Physiology and Behavior, University of California at Davis, One Shields Avenue, Davis, CA 95616, USA.
| | | |
Collapse
|
18
|
Petkov CI, Logothetis NK, Obleser J. Where are the human speech and voice regions, and do other animals have anything like them? Neuroscientist 2009; 15:419-29. [PMID: 19516047 DOI: 10.1177/1073858408326430] [Citation(s) in RCA: 48] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Modern lesion and imaging work in humans has been clarifying which brain regions are involved in the processing of speech and language. Concurrently, some of this work has aimed to bridge the gap to the seemingly incompatible evidence for multiple brain-processing pathways that first accumulated in nonhuman primates. For instance, the idea of a posterior temporal-parietal "Wernicke's" territory, which is thought to be instrumental for speech comprehension, conflicts with this region of the brain belonging to a spatial "where" pathway. At the same time a posterior speech-comprehension region ignores the anterior temporal lobe and its "what" pathway for evaluating the complex features of sensory input. Recent language models confirm that the posterior or dorsal stream has an important role in human communication, by a reconceptualization of the "where" into a "how-to" pathway with a connection to the motor system for speech comprehension. Others have tried to directly implicate the "what" pathway for speech comprehension, relying on the growing evidence in humans for anterior-temporal involvement in speech and voice processing. Coming full circle, we find that the recent imaging of vocalization and voice preferring regions in nonhuman primates allows us to make direct links to the human imaging data involving the anterior-temporal regions. The authors describe how comparison of the structure and function of the vocal communication system of humans and other animals is clarifying evolutionary relationships and the extent to which different species can model human brain function.
Collapse
Affiliation(s)
- Christopher I Petkov
- Max-Planck Institute for Biological Cybernetics, Tübingen, Germany, Institute of Neuroscience, Newcastle University, Newcastle upon Tyne, UK. chris.petkov@ ncl.ac.uk
| | | | | |
Collapse
|
19
|
Goycoolea M, Mena I, Neubauer S. Is there a difference in activation or in inhibition of cortical auditory centers depending on the ear that is stimulated? Acta Otolaryngol 2009; 129:348-53. [PMID: 18985461 DOI: 10.1080/00016480802495420] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
Abstract
CONCLUSIONS 1.With auditory stimuli cortical activation of Brodmann's areas 39 and 40 and inhibition of area 38 is bilateral. Inhibitory and excitatory relays play a role in the auditory pathways. 2. A statistically significant increased activation on the left side in areas 39 and 40, regardless of the stimulated ear, is suggestive that pure tones are preferably processed in the left hemisphere. 3. The significant difference in central inhibition depending on which ear is stimulated is supportive of the idea of a leading ear. OBJECTIVES The objectives were to determine cortical activation/inhibition, ipsi/contralateral in response to monaural stimulation with pure tones, and if the response differs for right/left ear stimulation. SUBJECTS AND METHODS Tc99m-HMPAO brain perfusion SPECT was done during monaural stimulation with pure tones in 10 volunteers. Ears were tested independently. RESULTS During auditory stimulation perfusion increased in both hemispheres in Brodmann's areas 39-40 and decreased in area 38,>2 SD above and below the normal mean respectively, in both hemispheres, regardless of which side was stimulated. A significantly more intense response was seen in left versus right in areas 39 and 40. In area 38 there was bilateral inhibition, significantly more intense in response to left than right ear stimulation.
Collapse
Affiliation(s)
- Marcos Goycoolea
- Department of Otorhinolaringology, Clínica Las Condes, Santiago, Chile.
| | | | | |
Collapse
|
20
|
Doehrmann O, Naumer MJ, Volz S, Kaiser J, Altmann CF. Probing category selectivity for environmental sounds in the human auditory brain. Neuropsychologia 2008; 46:2776-86. [PMID: 18597794 DOI: 10.1016/j.neuropsychologia.2008.05.011] [Citation(s) in RCA: 47] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/13/2008] [Revised: 05/02/2008] [Accepted: 05/12/2008] [Indexed: 11/29/2022]
Abstract
Earlier studies reported evidence suggesting distinct category-related auditory representations for environmental sounds such as animal vocalizations and tool sounds in superior and middle temporal regions of the temporal lobe. However, the degree of selectivity of these representations remains to be determined. The present study combined functional magnetic resonance imaging (fMRI) adaptation with a silent acquisition protocol to further investigate category-related auditory processing of environmental sounds. To this end, we consecutively presented pairs of sounds taken from the categories 'tool sounds' or 'animal vocalizations' with either the same or different identity/category. We examined the degree of selectivity as evidenced by adaptation effects to both or only one sound category in the course of whole-brain as well as functionally and anatomically constrained region of interest analyses. While most regions predominately in the temporal cortex showed an adaptation to both sound categories, particularly the left superior temporal gyrus (STG) and the left posterior middle temporal gyrus (pMTG) selectively adapted to animal vocalizations and tool sounds, respectively. However, the activation profiles of these regions differed with respect to the general responsiveness to sounds. While tool sounds still produced fMRI signals significantly different from fixation baseline in the STG, this was not the case for animal vocalizations in pMTG. Consistent with the interpretation of STG as an intermediate auditory processing stage, this region might differentiate auditory stimuli into categories based on variations of physical stimulus properties. However, processing in left pMTG seems to be even more restricted to action-related sounds of man-made objects.
Collapse
Affiliation(s)
- Oliver Doehrmann
- Institute of Medical Psychology, Johann Wolfgang Goethe-University, Heinrich-Hoffmann-Strasse 10 (Haus 93C), D-60528 Frankfurt am Main, Germany.
| | | | | | | | | |
Collapse
|
21
|
Brugge JF, Volkov IO, Oya H, Kawasaki H, Reale RA, Fenoy A, Steinschneider M, Howard MA. Functional localization of auditory cortical fields of human: click-train stimulation. Hear Res 2007; 238:12-24. [PMID: 18207680 DOI: 10.1016/j.heares.2007.11.012] [Citation(s) in RCA: 59] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/29/2007] [Revised: 11/28/2007] [Accepted: 11/30/2007] [Indexed: 11/16/2022]
Abstract
Averaged auditory evoked potentials (AEPs) to bilaterally presented 100 Hz click trains were recorded from multiple sites simultaneously within Heschl's gyrus (HG) and on the posterolateral surface of the superior temporal gyrus (STG) in epilepsy-surgery patients. Three auditory fields were identified based on AEP waveforms and their distribution. Primary (core) auditory cortex was localized to posteromedial HG. Here the AEP was characterized by a robust polyphasic low-frequency field potential having a short onset latency and on which was superimposed a smaller frequency-following response to the click train. Core AEPs exhibited the lowest response threshold and highest response amplitude at one HG site with threshold rising and amplitude declining systematically on either side of it. The AEPs recorded anterolateral to the core, if present, were typically of low amplitude, with little or no evidence of short-latency waves or the frequency-following response that characterized core AEPs. We suggest that this area is part of a lateral auditory belt system. Robust AEPs, with waveforms demonstrably different from those of the core or lateral belt, were localized to the posterolateral surface of the STG and conform to previously described field PLST.
Collapse
Affiliation(s)
- John F Brugge
- Department of Neurosurgery, University of Iowa College of Medicine, Iowa City, IA 52242, USA.
| | | | | | | | | | | | | | | |
Collapse
|
22
|
Raizada RDS, Poldrack RA. Selective Amplification of Stimulus Differences during Categorical Processing of Speech. Neuron 2007; 56:726-40. [PMID: 18031688 DOI: 10.1016/j.neuron.2007.11.001] [Citation(s) in RCA: 90] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2006] [Revised: 07/13/2007] [Accepted: 11/02/2007] [Indexed: 11/28/2022]
Affiliation(s)
- Rajeev D S Raizada
- Institute for Learning and Brain Sciences, University of Washington, Box 357988, Seattle, WA 98195, USA.
| | | |
Collapse
|
23
|
Deouell LY, Heller AS, Malach R, D'Esposito M, Knight RT. Cerebral responses to change in spatial location of unattended sounds. Neuron 2007; 55:985-96. [PMID: 17880900 DOI: 10.1016/j.neuron.2007.08.019] [Citation(s) in RCA: 89] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2006] [Revised: 07/01/2007] [Accepted: 08/20/2007] [Indexed: 11/20/2022]
Abstract
The neural basis of spatial processing in the auditory cortex has been controversial. Human fMRI studies suggest that a part of the planum temporale (PT) is involved in auditory spatial processing, but it was recently argued that this region is active only when the task requires voluntary spatial localization. If this is the case, then this region cannot harbor an ongoing spatial representation of the acoustic environment. In contrast, we show in three fMRI experiments that a region in the human medial PT is sensitive to background auditory spatial changes, even when subjects are not engaged in a spatial localization task, and in fact attend the visual modality. During such times, this area responded to rare location shifts, and even more so when spatial variation increased, consistent with spatially selective adaptation. Thus, acoustic space is represented in the human PT even when sound processing is not required by the ongoing task.
Collapse
Affiliation(s)
- Leon Y Deouell
- Department of Psychology and the Interdisciplinary Center for Neural Computation, The Hebrew University of Jerusalem, Jerusalem 91905, Israel.
| | | | | | | | | |
Collapse
|
24
|
Boumans T, Theunissen FE, Poirier C, Van Der Linden A. Neural representation of spectral and temporal features of song in the auditory forebrain of zebra finches as revealed by functional MRI. Eur J Neurosci 2007; 26:2613-26. [PMID: 17970728 PMCID: PMC2228391 DOI: 10.1111/j.1460-9568.2007.05865.x] [Citation(s) in RCA: 44] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Abstract
Song perception in songbirds, just as music and speech perception in humans, requires processing the spectral and temporal structure found in the succession of song-syllables. Using functional magnetic resonance imaging and synthetic songs that preserved exclusively either the temporal or the spectral structure of natural song, we investigated how vocalizations are processed in the avian forebrain. We found bilateral and equal activation of the primary auditory region, field L. The more ventral regions of field L showed depressed responses to the synthetic songs that lacked spectral structure. These ventral regions included subarea L3, medial-ventral subarea L and potentially the secondary auditory region caudal medial nidopallium. In addition, field L as a whole showed unexpected increased responses to the temporally filtered songs and this increase was the largest in the dorsal regions. These dorsal regions included L1 and the dorsal subareas L and L2b. Therefore, the ventral region of field L appears to be more sensitive to the preservation of both spectral and temporal information in the context of song processing. We did not find any differences in responses to playback of the bird's own song vs other familiar conspecific songs. We also investigated the effect of three commonly used anaesthetics on the blood oxygen level-dependent response: medetomidine, urethane and isoflurane. The extent of the area activated and the stimulus selectivity depended on the type of anaesthetic. We discuss these results in the context of what is known about the locus of action of the anaesthetics, and reports of neural activity measured in electrophysiological experiments.
Collapse
Affiliation(s)
- Tiny Boumans
- Bio-Imaging Laboratory, University of Antwerp, Belgium
| | | | | | | |
Collapse
|
25
|
Luo H, Poeppel D. Phase patterns of neuronal responses reliably discriminate speech in human auditory cortex. Neuron 2007; 54:1001-10. [PMID: 17582338 PMCID: PMC2703451 DOI: 10.1016/j.neuron.2007.06.004] [Citation(s) in RCA: 601] [Impact Index Per Article: 35.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2007] [Revised: 05/11/2007] [Accepted: 06/04/2007] [Indexed: 10/23/2022]
Abstract
How natural speech is represented in the auditory cortex constitutes a major challenge for cognitive neuroscience. Although many single-unit and neuroimaging studies have yielded valuable insights about the processing of speech and matched complex sounds, the mechanisms underlying the analysis of speech dynamics in human auditory cortex remain largely unknown. Here, we show that the phase pattern of theta band (4-8 Hz) responses recorded from human auditory cortex with magnetoencephalography (MEG) reliably tracks and discriminates spoken sentences and that this discrimination ability is correlated with speech intelligibility. The findings suggest that an approximately 200 ms temporal window (period of theta oscillation) segments the incoming speech signal, resetting and sliding to track speech dynamics. This hypothesized mechanism for cortical speech analysis is based on the stimulus-induced modulation of inherent cortical rhythms and provides further evidence implicating the syllable as a computational primitive for the representation of spoken language.
Collapse
Affiliation(s)
- Huan Luo
- Neuroscience and Cognitive Science Program, University of Maryland College Park, College Park MD 20742
- Department of Biology, University of Maryland College Park, College Park MD 20742
| | - David Poeppel
- Neuroscience and Cognitive Science Program, University of Maryland College Park, College Park MD 20742
- Department of Biology, University of Maryland College Park, College Park MD 20742
- Department of Linguistics, University of Maryland College Park, College Park MD 20742
| |
Collapse
|
26
|
Ischebeck AK, Friederici AD, Alter K. Processing Prosodic Boundaries in Natural and Hummed Speech: An fMRI Study. Cereb Cortex 2007; 18:541-52. [PMID: 17591598 DOI: 10.1093/cercor/bhm083] [Citation(s) in RCA: 30] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022] Open
Abstract
Speech contains prosodic cues such as pauses between different phrases of a sentence. These intonational phrase boundaries (IPBs) elicit a specific component in event-related brain potential studies, the so-called closure positive shift. The aim of the present functional magnetic resonance imaging study is to identify the neural correlates of this prosody-related component in sentences containing segmental and prosodic information (natural speech) and hummed sentences only containing prosodic information. Sentences with 2 IPBs both in normal and hummed speech activated the middle superior temporal gyrus, the rolandic operculum, and the gyrus of Heschl more strongly than sentences with 1 IPB. The results from a region of interest analysis of auditory cortex and auditory association areas suggest that the posterior rolandic operculum, in particular, supports the processing of prosodic information. A comparison of natural speech and hummed sentences revealed a number of left-hemispheric areas within the temporal lobe as well as in the frontal and parietal lobe that were activated more strongly for natural speech than for hummed sentences. These areas constitute the neural network for the processing of natural speech. The finding that no area was activated more strongly for hummed sentences compared with natural speech suggests that prosody is an integrated part of natural speech.
Collapse
Affiliation(s)
- Anja K Ischebeck
- Clinical Department of Neurology, Innsbruck Medical University, 6020 Innsbruck, Austria.
| | | | | |
Collapse
|
27
|
Affiliation(s)
- Andrew J King
- Department of Physiology, Anatomy and Genetics, Sherrington Building, University of Oxford, Parks Road, Oxford, UK.
| | | |
Collapse
|
28
|
Lenz D, Schadow J, Thaerig S, Busch NA, Herrmann CS. What's that sound? Matches with auditory long-term memory induce gamma activity in human EEG. Int J Psychophysiol 2007; 64:31-8. [PMID: 16959348 DOI: 10.1016/j.ijpsycho.2006.07.008] [Citation(s) in RCA: 42] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2006] [Revised: 06/24/2006] [Accepted: 07/13/2006] [Indexed: 10/24/2022]
Abstract
In recent years the cognitive functions of human gamma-band activity (30-100 Hz) advanced continuously into scientific focus. Not only bottom-up driven influences on 40 Hz activity have been observed, but also top-down processes seem to modulate responses in this frequency band. Among the various functions that have been related to gamma activity a pivotal role has been assigned to memory processes. Visual experiments suggested that gamma activity is involved in matching visual input to memory representations. Based on these findings we hypothesized that such memory related modulations of gamma activity exist in the auditory modality, as well. Thus, we chose environmental sounds for which subjects already had a long-term memory (LTM) representation and compared them to unknown, but physically similar sounds. 21 subjects had to classify sounds as 'recognized' or 'unrecognized', while EEG was recorded. Our data show significantly stronger activity in the induced gamma-band for recognized sounds in the time window between 300 and 500 ms after stimulus onset with a central topography. The results suggest that induced gamma-band activity reflects the matches between sounds and their representations in auditory LTM.
Collapse
Affiliation(s)
- Daniel Lenz
- Department of Biological Psychology, Otto-von-Guericke University of Magdeburg, P.O. Box 4120, 39016 Magdeburg, Germany
| | | | | | | | | |
Collapse
|
29
|
Saeb S, Gharibzadeh S, Towhidkhah F, Farajidavar A. Modeling the primary auditory cortex using dynamic synapses: can synaptic plasticity explain the temporal tuning? J Theor Biol 2007; 248:1-9. [PMID: 17559885 DOI: 10.1016/j.jtbi.2007.03.025] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/11/2006] [Revised: 03/20/2007] [Accepted: 03/20/2007] [Indexed: 11/17/2022]
Abstract
The molecular mechanisms underlying the temporal plasticity (temporal tuning) of cortical cells remain controversial. Experimental observations indicate that the neuronal responses at the primary auditory cortex are affected by behavioral learning. In this paper, we present a minimal feed-forward model of the primary auditory cortex, based on the dynamic synapse and the leaky integrate-and-fire neuron models, in order to search for the origin of the observed plasticity. We demonstrate that the frequency response of the model is markedly modified by regulating the contribution of synaptic facilitation to the short-term dynamics of synapses (U(1)). Consequently, we propose that the variation of this parameter may be responsible for primary auditory cortex enhancement achieved by behavioral training. Based on our model, we assume that the contribution of facilitation arises from the amount of Ca(2+) influx each time an action potential arrives at the nerve terminal. Regardless of what really leads to the long-term variation of Ca(2+) influx, we suggest that this process is responsible for the temporal tuning of responses observed in experimental studies. We believe that measurement of the long-term variation of Ca(2+) influx at the pre-synaptic area of the cortical cells in auditory learning trials would be the first step to validate our hypothesis.
Collapse
Affiliation(s)
- Sohrab Saeb
- Cognitive Neural Engineering Lab, Faculty of Biomedical Engineering, Amirkabir University of Technology (Tehran Polytechnic), Tehran, Iran
| | | | | | | |
Collapse
|
30
|
Griffiths TD, Kumar S, Warren JD, Stewart L, Stephan KE, Friston KJ. Approaches to the cortical analysis of auditory objects. Hear Res 2007; 229:46-53. [PMID: 17321704 DOI: 10.1016/j.heares.2007.01.010] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/21/2006] [Revised: 12/21/2006] [Accepted: 01/03/2007] [Indexed: 10/23/2022]
Abstract
We describe work that addresses the cortical basis for the analysis of auditory objects using 'generic' sounds that do not correspond to any particular events or sources (like vowels or voices) that have semantic association. The experiments involve the manipulation of synthetic sounds to produce systematic changes of stimulus features, such as spectral envelope. Conventional analyses of normal functional imaging data demonstrate that the analysis of spectral envelope and perceived timbral change involves a network consisting of planum temporale (PT) bilaterally and the right superior temporal sulcus (STS). Further analysis of imaging data using dynamic causal modelling (DCM) and Bayesian model selection was carried out in the right hemisphere areas to determine the effective connectivity between these auditory areas. Specifically, the objective was to determine if the analysis of spectral envelope in the network is done in a serial fashion (that is from HG to PT to STS) or parallel fashion (that is PT and STS receives input from HG simultaneously). Two families of models, serial and parallel (16 in total) that represent different hypotheses about the connectivity between HG, PT and STS were selected. The models within a family differ with respect to the pathway that is modulated by the analysis of spectral envelope. After the models are identified, Bayesian model selection procedure is then used to select the 'optimal' model from the specified models. The data strongly support a particular serial model containing modulation of the HG to PT effective connectivity during spectral envelope variation. Parallel work in neurological subjects addresses the effect of lesions to different parts of this network. We have recently studied in detail subjects with 'dystimbria': an alteration in the perceived quality of auditory objects distinct from pitch or loudness change. The subjects have lesions of the normal network described above with normal perception of pitch strength but abnormal perception of the analysis of spectral envelope change.
Collapse
Affiliation(s)
- Timothy D Griffiths
- Auditory Group, Medical School, Framlington Place, University of Newcastle, Newcastle upon Tyne, UK.
| | | | | | | | | | | |
Collapse
|
31
|
Chang JE, Bai JY, Zeng FG. Unintelligible low-frequency sound enhances simulated cochlear-implant speech recognition in noise. IEEE Trans Biomed Eng 2007; 53:2598-601. [PMID: 17152439 DOI: 10.1109/tbme.2006.883793] [Citation(s) in RCA: 75] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Speech can be recognized by multiple acoustic cues in both frequency and time domains. These acoustic cues are often thought to be redundant. One example is the low-frequency sound component below 300 Hz, which is not even transmitted by the majority of communication devices including telephones. Here, we showed that this low-frequency sound component, although unintelligible when presented alone, could improve the functional signal-to-noise ratio (SNR) by 10-15 dB for speech recognition in noise when presented in combination with a cochlear-implant simulation. A similar low-frequency enhancement effect could be obtained by presenting the low-frequency sound component to one ear and the cochlear-implant simulation to the other ear. However, a high-frequency sound could not produce a similar speech enhancement in noise. We argue that this low-frequency enhancement effect cannot be due to linear addition of intelligibility between low- and high-frequency components or an increase in the physical SNR. We suggest a brain-based mechanism that uses the voice pitch cue in the low-frequency sound to first segregate the target voice from the competing voice and then to group appropriate temporal envelope cues in the target voice for robust speech recognition under realistic listening situations.
Collapse
Affiliation(s)
- Janice E Chang
- Bioengineering Dept., University of California, Berkeley, CA 94720, USA
| | | | | |
Collapse
|
32
|
Lindenberg R, Scheef L. Supramodal language comprehension: Role of the left temporal lobe for listening and reading. Neuropsychologia 2007; 45:2407-15. [PMID: 17451759 DOI: 10.1016/j.neuropsychologia.2007.02.008] [Citation(s) in RCA: 49] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2006] [Revised: 02/07/2007] [Accepted: 02/11/2007] [Indexed: 11/26/2022]
Abstract
In this fMRI study, we aimed at identifying the cortical areas engaged in supramodal processing of language comprehension. BOLD changes were recorded in 19 healthy right-handed subjects reading or listening to a story. During the visual control tasks the volunteers attended to a series of continuous letterstrings or a fixation cross, while during the acoustic control tasks either a reversed text or white noise were presented. The conjunction of the visual and acoustic story processing yielded left-dominant activations which in comparison to language-like stimuli focused to the left middle temporal gyrus as well as to the supramarginal gyrus. We conclude that the core structure representing supramodal language comprehension is the left temporal lobe at both banks of the superior temporal sulcus.
Collapse
Affiliation(s)
- Robert Lindenberg
- Department of Radiology, Rheinische Friedrich-Wilhelms-University Bonn, Germany.
| | | |
Collapse
|
33
|
Meyer M, Baumann S, Jancke L. Electrical brain imaging reveals spatio-temporal dynamics of timbre perception in humans. Neuroimage 2006; 32:1510-23. [PMID: 16798014 DOI: 10.1016/j.neuroimage.2006.04.193] [Citation(s) in RCA: 56] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2005] [Revised: 03/29/2006] [Accepted: 04/10/2006] [Indexed: 11/27/2022] Open
Abstract
Timbre is a major attribute of sound perception and a key feature for the identification of sound quality. Here, we present event-related brain potentials (ERPs) obtained from sixteen healthy individuals while they discriminated complex instrumental tones (piano, trumpet, and violin) or simple sine wave tones that lack the principal features of timbre. Data analysis yielded enhanced N1 and P2 responses to instrumental tones relative to sine wave tones. Furthermore, we applied an electrical brain imaging approach using low-resolution electromagnetic tomography (LORETA) to estimate the neural sources of N1/P2 responses. Separate significance tests of instrumental vs. sine wave tones for N1 and P2 revealed distinct regions as principally governing timbre perception. In an initial stage (N1), timbre perception recruits left and right (peri-)auditory fields with an activity maximum over the right posterior Sylvian fissure (SF) and the posterior cingulate (PCC) territory. In the subsequent stage (P2), we uncovered enhanced activity in the vicinity of the entire cingulate gyrus. The involvement of extra-auditory areas in timbre perception may imply the presence of a highly associative processing level which might be generally related to musical sensations and integrates widespread medial areas of the human cortex. In summary, our results demonstrate spatio-temporally distinct stages in timbre perception which not only involve bilateral parts of the peri-auditory cortex but also medially situated regions of the human brain associated with emotional and auditory imagery functions.
Collapse
Affiliation(s)
- Martin Meyer
- Department of Neuropsychology, University of Zurich, Treichlerstrasse 10, CH-8032 Zurich, Switzerland.
| | | | | |
Collapse
|
34
|
Theunissen FE, Shaevitz SS. Auditory processing of vocal sounds in birds. Curr Opin Neurobiol 2006; 16:400-7. [PMID: 16842993 DOI: 10.1016/j.conb.2006.07.003] [Citation(s) in RCA: 54] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2006] [Accepted: 07/03/2006] [Indexed: 11/24/2022]
Abstract
The avian auditory system has become a model system to investigate how vocalizations are memorized and processed by the brain in order to mediate behavioral discrimination and recognition. Recent studies have shown that most of the avian auditory system responds preferentially and efficiently to sounds that have natural spectro-temporal statistics. In addition, neurons in secondary auditory forebrain areas have plastic response properties and are the most active when processing behaviorally relevant vocalizations. Physiological measurements show differential responses for vocalizations that were recently learned in discrimination tasks, and for the tutor song, a longer-term auditory memory that is used to guide vocal learning in male songbirds.
Collapse
Affiliation(s)
- Frédéric E Theunissen
- Department of Psychology and Neurosciences Institute, University of California, Berkeley, USA.
| | | |
Collapse
|
35
|
Warren JE, Wise RJS, Warren JD. Sounds do-able: auditory-motor transformations and the posterior temporal plane. Trends Neurosci 2005; 28:636-43. [PMID: 16216346 DOI: 10.1016/j.tins.2005.09.010] [Citation(s) in RCA: 181] [Impact Index Per Article: 9.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2005] [Revised: 08/22/2005] [Accepted: 09/26/2005] [Indexed: 10/25/2022]
Abstract
Accumulating evidence in humans and non-human primates implicates the posterior superior temporal plane (STP) in the processing of both auditory spatial information and vocal sounds. Such evidence is difficult to reconcile with existing accounts of the primate auditory brain. We propose that the posteromedial STP generates sequenced auditory representations by matching incoming auditory information with stored templates. These sequenced auditory representations are subsequently used to constrain motor responses. We argue for a re-assessment of the much-debated dorsal auditory pathway in terms of its generic behavioral role as an auditory "do" pathway.
Collapse
Affiliation(s)
- Jane E Warren
- Division of Neuroscience and Mental Health, Imperial College London, London W12 0NN, UK.
| | | | | |
Collapse
|
36
|
Cohen YE, Russ BE, Gifford GW, Kiringoda R, MacLean KA. Selectivity for the spatial and nonspatial attributes of auditory stimuli in the ventrolateral prefrontal cortex. J Neurosci 2005; 24:11307-16. [PMID: 15601937 PMCID: PMC6730358 DOI: 10.1523/jneurosci.3935-04.2004] [Citation(s) in RCA: 60] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
Spatial and nonspatial auditory processing is hypothesized to occur in parallel dorsal and ventral pathways, respectively. In this study, we tested the spatial and nonspatial sensitivity of auditory neurons in the ventrolateral prefrontal cortex (vPFC), a cortical area in the hypothetical nonspatial pathway. We found that vPFC neurons were modulated significantly by both the spatial and nonspatial attributes of an auditory stimulus. When comparing these responses with those in anterolateral belt region of the auditory cortex, which is hypothesized to be specialized for processing the nonspatial attributes of auditory stimuli, we found that the nonspatial sensitivity of vPFC neurons was poorer, whereas the spatial selectivity was better than anterolateral neurons. Also, the spatial and nonspatial sensitivity of vPFC neurons was comparable with that seen in the lateral intraparietal area, a cortical area that is a part of the dorsal pathway. These data suggest that substantial spatial and nonspatial processing occurs in both the dorsal and ventral pathways.
Collapse
Affiliation(s)
- Yale E Cohen
- Department of Psychological and Brain Sciences, Dartmouth College, Hanover, New Hampshire 03755, USA.
| | | | | | | | | |
Collapse
|
37
|
Abstract
There have been recent developments in our understanding of the auditory neuroscience of non-human primates that, to a certain extent, can be integrated with findings from human functional neuroimaging studies. This framework can be used to consider the cortical basis of complex sound processing in humans, including implications for speech perception, spatial auditory processing and auditory scene segregation.
Collapse
Affiliation(s)
- Sophie K Scott
- Institute of Cognitive Neuroscience, University College London, 17 Queen Square, London, WC1N 3AR, UK.
| |
Collapse
|
38
|
Coath M, Brader JM, Fusi S, Denham SL. Multiple views of the response of an ensemble of spectro-temporal features support concurrent classification of utterance, prosody, sex and speaker identity. NETWORK (BRISTOL, ENGLAND) 2005; 16:285-300. [PMID: 16411500 DOI: 10.1080/09548980500290120] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/06/2023]
Abstract
Models of auditory processing, particularly of speech, face many difficulties. These difficulties include variability among speakers, variability in speech rate and robustness to moderate distortions such as time compression. In contrast to the 'invariance of percept' (across different speakers, of different sexes, using different intonation, and so on) is the observation that we are sensitive to the identity, sex and intonation of the speaker. In previous work we have reported that a model based on ensembles of spectro-temporal feature detectors, derived from onset sensitive pre-processing of a limited class of stimuli, preserves significant information about the stimulus class. We have also shown that this is robust with respect to the exact choice of feature set, moderate time compression in the stimulus and speaker variation. Here we extend these results to show a) that by using a classifier based on a network of spiking neurons with spike-driven plasticity, the output of the ensemble constitutes an effective rate coding representation of complex sounds; and b) that the same set of spectro-temporal features concurrently preserve information about a range of qualitatively different classes into which the stimulus might fall. We show that it is possible for multiple views of the same pattern of responses to generate different percepts. This is consistent with suggestions that multiple parallel processes exist within the auditory 'what' pathway with attentional modulation enhancing the task-relevant classification type. We also show that the responses of the ensemble are sparse in the sense that a small number of features respond for each stimulus type. This has implications for the ensembles' ability to generalise, and to respond differentially to a wide variety of stimulus classes.
Collapse
Affiliation(s)
- M Coath
- Centre for Theoretical and Computational Neuroscience, University of Plymouth, Plymouth, UK.
| | | | | | | |
Collapse
|
39
|
Bizley JK, Nodal FR, Nelken I, King AJ. Functional organization of ferret auditory cortex. ACTA ACUST UNITED AC 2005; 15:1637-53. [PMID: 15703254 DOI: 10.1093/cercor/bhi042] [Citation(s) in RCA: 142] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
Abstract
We characterized the functional organization of different fields within the auditory cortex of anaesthetized ferrets. As previously reported, the primary auditory cortex, A1, and the anterior auditory field, AAF, are located on the middle ectosylvian gyrus. These areas exhibited a similar tonotopic organization, with high frequencies represented at the dorsal tip of the gyrus and low frequencies more ventrally, but differed in that AAF neurons had shorter response latencies than those in A1. On the basis of differences in frequency selectivity, temporal response properties and thresholds, we identified four more, previously undescribed fields. Two of these are located on the posterior ectosylvian gyrus and were tonotopically organized. Neurons in these areas responded robustly to tones, but had longer latencies, more sustained responses and a higher incidence of non-monotonic rate-level functions than those in the primary fields. Two further auditory fields, which were not tonotopically organized, were found on the anterior ectosylvian gyrus. Neurons in the more dorsal anterior area gave short-latency, transient responses to tones and were generally broadly tuned with a preference for high (>8 kHz) frequencies. Neurons in the other anterior area were frequently unresponsive to tones, but often responded vigorously to broadband noise. The presence of both tonotopic and non-tonotopic auditory cortical fields indicates that the organization of ferret auditory cortex is comparable to that seen in other mammals.
Collapse
|
40
|
Mrsic-Flogel TD, King AJ, Schnupp JWH. Encoding of virtual acoustic space stimuli by neurons in ferret primary auditory cortex. J Neurophysiol 2005; 93:3489-503. [PMID: 15659534 DOI: 10.1152/jn.00748.2004] [Citation(s) in RCA: 59] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Recent studies from our laboratory have indicated that the spatial response fields (SRFs) of neurons in the ferret primary auditory cortex (A1) with best frequencies > or =4 kHz may arise from a largely linear processing of binaural level and spectral localization cues. Here we extend this analysis to investigate how well the linear model can predict the SRFs of neurons with different binaural response properties and the manner in which SRFs change with increases in sound level. We also consider whether temporal features of the response (e.g., response latency) vary with sound direction and whether such variations can be explained by linear processing. In keeping with previous studies, we show that A1 SRFs, which we measured with individualized virtual acoustic space stimuli, expand and shift in direction with increasing sound level. We found that these changes are, in most cases, in good agreement with predictions from a linear threshold model. However, changes in spatial tuning with increasing sound level were generally less well predicted for neurons whose binaural frequency-time receptive field (FTRF) exhibited strong excitatory inputs from both ears than for those in which the binaural FTRF revealed either a predominantly inhibitory effect or no clear contribution from the ipsilateral ear. Finally, we found (in agreement with other authors) that many A1 neurons exhibit systematic response latency shifts as a function of sound-source direction, although these temporal details could usually not be predicted from the neuron's binaural FTRF.
Collapse
|
41
|
Sweet RA, Dorph-Petersen KA, Lewis DA. Mapping auditory core, lateral belt, and parabelt cortices in the human superior temporal gyrus. J Comp Neurol 2005; 491:270-89. [PMID: 16134138 DOI: 10.1002/cne.20702] [Citation(s) in RCA: 111] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
The goal of the present study was to determine whether the architectonic criteria used to identify the core, lateral belt, and parabelt auditory cortices in macaque monkeys (Macaca fascicularis) could be used to identify homologous regions in humans (Homo sapiens). Current evidence indicates that auditory cortex in humans, as in monkeys, is located on the superior temporal gyrus (STG), and is functionally and structurally altered in illnesses such as schizophrenia and Alzheimer's disease. In this study, we used serial sets of adjacent sections processed for Nissl substance, acetylcholinesterase, and parvalbumin to identify the distinguishing cyto- and chemoarchitectonic features of the core, lateral belt, and parabelt in monkey. These criteria were evaluated in postmortem tissue from a human subject, leading to the identification of additional criteria specific to human. The criteria were validated in an additional set of eight human subjects. Regions were delineated and their volumes estimated using the Cavalieri method in these subjects, and the sources of methodologic contribution to variability of the estimates was assessed. Serial reconstructions of the auditory cortex in humans were made showing the location of the lateral belt and parabelt with respect to gross anatomical landmarks. Architectonic criteria for the core, lateral belt, and parabelt were readily adapted from monkey to human. Additionally, we found evidence for an architectonic subdivision within the parabelt, present in both species. Variability of regional volume estimates was readily constrained using a multifaceted approach to reduce potential sources of variability in regional delineation.
Collapse
Affiliation(s)
- Robert A Sweet
- Department of Psychiatry, University of Pittsburgh, Pittsburgh, Pennsylvania 15213, USA.
| | | | | |
Collapse
|
42
|
Turner RP. The acute effect of music on interictal epileptiform discharges. Epilepsy Behav 2004; 5:662-8. [PMID: 15380117 DOI: 10.1016/j.yebeh.2004.07.003] [Citation(s) in RCA: 31] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/01/2004] [Accepted: 07/07/2004] [Indexed: 11/21/2022]
Abstract
This study was a prospective, randomized, single-blinded, crossover, placebo-controlled, pilot clinical trial investigating the effect of Mozart's Sonata for Two Pianos (K448) on the frequency of interictal epileptiform discharges (IEDs) from the EEGs of children with benign childhood epilepsy with centrotemporal spikes, or "rolandic" epilepsy. The goal was to demonstrate decreased frequency of IEDs with exposure to K448. Four subjects were recruited and 4-hour awake EEG recordings performed. IED frequency per minute was averaged over each of three epochs per hour. Mean IED count per epoch, standard deviations, and variance were calculated. Only complete waking epochs were analyzed. Two subjects demonstrated sufficient waking IEDs for statistical analysis, consisting of three epochs of K448-related effects. Significant decreases in IEDs per minute (33.7, 50.6, and 33.9%) were demonstrated comparing baseline with exposure to K448, but not to control music (Beethoven's Für Elise).
Collapse
Affiliation(s)
- Robert P Turner
- Department of Neurology, Medical University of South Carolina, Charleston, SC, USA.
| |
Collapse
|