1
|
Undurraga JA, Luke R, Van Yper L, Monaghan JJM, McAlpine D. The neural representation of an auditory spatial cue in the primate cortex. Curr Biol 2024; 34:2162-2174.e5. [PMID: 38718798 DOI: 10.1016/j.cub.2024.04.034] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2023] [Revised: 02/14/2024] [Accepted: 04/12/2024] [Indexed: 05/23/2024]
Abstract
Humans make use of small differences in the timing of sounds at the two ears-interaural time differences (ITDs)-to locate their sources. Despite extensive investigation, however, the neural representation of ITDs in the human brain is contentious, particularly the range of ITDs explicitly represented by dedicated neural detectors. Here, using magneto- and electro-encephalography (MEG and EEG), we demonstrate evidence of a sparse neural representation of ITDs in the human cortex. The magnitude of cortical activity to sounds presented via insert earphones oscillated as a function of increasing ITD-within and beyond auditory cortical regions-and listeners rated the perceptual quality of these sounds according to the same oscillating pattern. This pattern was accurately described by a population of model neurons with preferred ITDs constrained to the narrow, sound-frequency-dependent range evident in other mammalian species. When scaled for head size, the distribution of ITD detectors in the human cortex is remarkably like that recorded in vivo from the cortex of rhesus monkeys, another large primate that uses ITDs for source localization. The data solve a long-standing issue concerning the neural representation of ITDs in humans and suggest a representation that scales for head size and sound frequency in an optimal manner.
Collapse
Affiliation(s)
- Jaime A Undurraga
- Department of Linguistics, Macquarie University, 16 University Avenue, Sydney, NSW 2109, Australia; Interacoustics Research Unit, Technical University of Denmark, Ørsteds Plads, Building 352, 2800 Kgs. Lyngby, Denmark.
| | - Robert Luke
- Department of Linguistics, Macquarie University, 16 University Avenue, Sydney, NSW 2109, Australia; The Bionics Institute, 384-388 Albert St., East Melbourne, VIC 3002, Australia
| | - Lindsey Van Yper
- Department of Linguistics, Macquarie University, 16 University Avenue, Sydney, NSW 2109, Australia; Institute of Clinical Research, University of Southern Denmark, 5230 Odense, Denmark; Research Unit for ORL, Head & Neck Surgery and Audiology, Odense University Hospital & University of Southern Denmark, 5230 Odense, Denmark
| | - Jessica J M Monaghan
- Department of Linguistics, Macquarie University, 16 University Avenue, Sydney, NSW 2109, Australia; National Acoustic Laboratories, Australian Hearing Hub, 16 University Avenue, Sydney, NSW 2109, Australia
| | - David McAlpine
- Department of Linguistics, Macquarie University, 16 University Avenue, Sydney, NSW 2109, Australia; Macquarie University Hearing and the Australian Hearing Hub, Macquarie University, 16 University Avenue, Sydney, NSW 2109, Australia.
| |
Collapse
|
2
|
Hancock KE, Delgutte B. Neural coding of dichotic pitches in auditory midbrain. J Neurophysiol 2023; 129:872-893. [PMID: 36921210 PMCID: PMC10085564 DOI: 10.1152/jn.00511.2022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2022] [Revised: 02/24/2023] [Accepted: 03/13/2023] [Indexed: 03/17/2023] Open
Abstract
Dichotic pitches such as the Huggins pitch (HP) and the binaural edge pitch (BEP) are perceptual illusions whereby binaural noise that exhibits abrupt changes in interaural phase differences (IPDs) across frequency creates a tonelike pitch percept when presented to both ears, even though it does not produce a pitch when presented monaurally. At the perceptual and cortical levels, dichotic pitches behave as if an actual tone had been presented to the ears, yet investigations of neural correlates of dichotic pitch in single-unit responses at subcortical levels are lacking. We tested for cues to HP and BEP in the responses of binaural neurons in the auditory midbrain of anesthetized cats by varying the expected pitch frequency around each neuron's best frequency (BF). Neuronal firing rates showed specific features (peaks, troughs, or edges) when the pitch frequency crossed the BF, and the type of feature was consistent with a well-established model of binaural processing comprising frequency tuning, internal delays, and firing rates sensitive to interaural correlation. A Jeffress-like neural population model in which the behavior of individual neurons was governed by the cross-correlation model and the neurons were independently distributed along BF and best IPD predicted trends in human psychophysical HP detection but only when the model incorporated physiological BF and best IPD distributions. These results demonstrate the existence of a rate-place code for HP and BEP in the auditory midbrain and provide a firm physiological basis for models of dichotic pitches.NEW & NOTEWORTHY Dichotic pitches are perceptual illusions created centrally through binaural interactions that offer an opportunity to test theories of pitch and binaural hearing. Here we show that binaural neurons in auditory midbrain encode the frequency of two salient types of dichotic pitches via specific features in the pattern of firing rates along the tonotopic axis. This is the first combined single-unit and modeling study of responses of auditory neurons to stimuli evoking a dichotic pitch.
Collapse
Affiliation(s)
- Kenneth E Hancock
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, Massachusetts, United States
- Department of Otolaryngology, Head and Neck Surgery, Harvard Medical School, Boston, Massachusetts, United States
| | - Bertrand Delgutte
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, Massachusetts, United States
- Department of Otolaryngology, Head and Neck Surgery, Harvard Medical School, Boston, Massachusetts, United States
| |
Collapse
|
3
|
Are frog calls relatively difficult to locate by mammalian predators? J Comp Physiol A Neuroethol Sens Neural Behav Physiol 2023; 209:11-30. [PMID: 36508005 DOI: 10.1007/s00359-022-01594-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2022] [Revised: 10/26/2022] [Accepted: 11/08/2022] [Indexed: 12/14/2022]
Abstract
Frogs call in acoustically dense choruses to attract conspecific females. Their calls can potentially reveal their location to predators, many of which are mammals. However, frogs and mammals have very different acoustic receivers and mechanisms for determining sound source direction. We argue that frog calls may have been selected so that they are harder to locate with the direction-finding mechanisms of mammals. We focus on interaural time delay (ITD) estimation using delay-line coincidence detection (place code), and a binaural excitatory/inhibitory (E/I) ITD mechanism found in mammals with small heads (population code). We identify four "strategies" which frogs may employ to exploit the weaknesses of either mechanism. The first two strategies used by the frog confound delay estimation to increase direction ambiguity using highly periodic calls or narrowband calls. The third strategy relies on using short pulses. The E/I mechanism is susceptible to noise with sounds being pulled to the medial plane when signal-to-noise ratio is low. Together, these three strategies compromise both ongoing and onset determination of location using either mechanism. Finally, frogs call in dense choruses using various means for controlling synchrony, maintaining chorus tenure, and abruptly switching off calling, all of which serve to confound location finding. Of these strategies, only chorusing adversely impacts the localization performance of frogs' acoustic receivers. We illustrate these strategies with an analysis of calls from three different frog species.
Collapse
|
4
|
Luke R, Innes-Brown H, Undurraga JA, McAlpine D. Human cortical processing of interaural coherence. iScience 2022; 25:104181. [PMID: 35494228 PMCID: PMC9051632 DOI: 10.1016/j.isci.2022.104181] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2021] [Revised: 11/29/2021] [Accepted: 03/29/2022] [Indexed: 11/17/2022] Open
Abstract
Sounds reach the ears as a mixture of energy generated by different sources. Listeners extract cues that distinguish different sources from one another, including how similar sounds arrive at the two ears, the interaural coherence (IAC). Here, we find listeners cannot reliably distinguish two completely interaurally coherent sounds from a single sound with reduced IAC. Pairs of sounds heard toward the front were readily confused with single sounds with high IAC, whereas those heard to the sides were confused with single sounds with low IAC. Sounds that hold supra-ethological spatial cues are perceived as more diffuse than can be accounted for by their IAC, and this is accounted for by a computational model comprising a restricted, and sound-frequency dependent, distribution of auditory-spatial detectors. We observed elevated cortical hemodynamic responses for sounds with low IAC, suggesting that the ambiguity elicited by sounds with low interaural similarity imposes elevated cortical load.
Collapse
Affiliation(s)
- Robert Luke
- Macquarie University, Sydney, NSW, Australia
- The Bionics Institute, Melbourne, VIC, Australia
| | | | | | | |
Collapse
|
5
|
Exploring the distribution of statistical feature parameters for natural sound textures. PLoS One 2021; 16:e0238960. [PMID: 34161323 PMCID: PMC8221478 DOI: 10.1371/journal.pone.0238960] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2020] [Accepted: 06/03/2021] [Indexed: 11/19/2022] Open
Abstract
Sounds like “running water” and “buzzing bees” are classes of sounds which are a collective result of many similar acoustic events and are known as “sound textures”. A recent psychoacoustic study using sound textures has reported that natural sounding textures can be synthesized from white noise by imposing statistical features such as marginals and correlations computed from the outputs of cochlear models responding to the textures. The outputs being the envelopes of bandpass filter responses, the ‘cochlear envelope’. This suggests that the perceptual qualities of many natural sounds derive directly from such statistical features, and raises the question of how these statistical features are distributed in the acoustic environment. To address this question, we collected a corpus of 200 sound textures from public online sources and analyzed the distributions of the textures’ marginal statistics (mean, variance, skew, and kurtosis), cross-frequency correlations and modulation power statistics. A principal component analysis of these parameters revealed a great deal of redundancy in the texture parameters. For example, just two marginal principal components, which can be thought of as measuring the sparseness or burstiness of a texture, capture as much as 64% of the variance of the 128 dimensional marginal parameter space, while the first two principal components of cochlear correlations capture as much as 88% of the variance in the 496 correlation parameters. Knowledge of the statistical distributions documented here may help guide the choice of acoustic stimuli with high ecological validity in future research.
Collapse
|
6
|
Abstract
Humans and other animals use spatial hearing to rapidly localize events in the environment. However, neural encoding of sound location is a complex process involving the computation and integration of multiple spatial cues that are not represented directly in the sensory organ (the cochlea). Our understanding of these mechanisms has increased enormously in the past few years. Current research is focused on the contribution of animal models for understanding human spatial audition, the effects of behavioural demands on neural sound location encoding, the emergence of a cue-independent location representation in the auditory cortex, and the relationship between single-source and concurrent location encoding in complex auditory scenes. Furthermore, computational modelling seeks to unravel how neural representations of sound source locations are derived from the complex binaural waveforms of real-life sounds. In this article, we review and integrate the latest insights from neurophysiological, neuroimaging and computational modelling studies of mammalian spatial hearing. We propose that the cortical representation of sound location emerges from recurrent processing taking place in a dynamic, adaptive network of early (primary) and higher-order (posterior-dorsal and dorsolateral prefrontal) auditory regions. This cortical network accommodates changing behavioural requirements and is especially relevant for processing the location of real-life, complex sounds and complex auditory scenes.
Collapse
|
7
|
Gleiss H, Encke J, Lingner A, Jennings TR, Brosel S, Kunz L, Grothe B, Pecka M. Cooperative population coding facilitates efficient sound-source separability by adaptation to input statistics. PLoS Biol 2019; 17:e3000150. [PMID: 31356637 PMCID: PMC6687189 DOI: 10.1371/journal.pbio.3000150] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2019] [Revised: 08/08/2019] [Accepted: 07/11/2019] [Indexed: 01/31/2023] Open
Abstract
Our sensory environment changes constantly. Accordingly, neural systems continually adapt to the concurrent stimulus statistics to remain sensitive over a wide range of conditions. Such dynamic range adaptation (DRA) is assumed to increase both the effectiveness of the neuronal code and perceptual sensitivity. However, direct demonstrations of DRA-based efficient neuronal processing that also produces perceptual benefits are lacking. Here, we investigated the impact of DRA on spatial coding in the rodent brain and the perception of human listeners. Complex spatial stimulation with dynamically changing source locations elicited prominent DRA already on the initial spatial processing stage, the Lateral Superior Olive (LSO) of gerbils. Surprisingly, on the level of individual neurons, DRA diminished spatial tuning because of large response variability across trials. However, when considering single-trial population averages of multiple neurons, DRA enhanced the coding efficiency specifically for the concurrently most probable source locations. Intrinsic LSO population imaging of energy consumption combined with pharmacology revealed that a slow-acting LSO gain-control mechanism distributes activity across a group of neurons during DRA, thereby enhancing population coding efficiency. Strikingly, such “efficient cooperative coding” also improved neuronal source separability specifically for the locations that were most likely to occur. These location-specific enhancements in neuronal coding were paralleled by human listeners exhibiting a selective improvement in spatial resolution. We conclude that, contrary to canonical models of sensory encoding, the primary motive of early spatial processing is efficiency optimization of neural populations for enhanced source separability in the concurrent environment. The efficient coding hypothesis suggests that sensory processing adapts to the stimulus statistics to maximize information while minimizing energetic costs. This study finds that an auditory spatial processing circuit distributes activity across neurons to enhance processing efficiency, focally improving spatial resolution both in neurons and in human listeners.
Collapse
Affiliation(s)
- Helge Gleiss
- Division of Neurobiology, Department of Biology II, Ludwig-Maximilians-Universitaet Muenchen, Martinsried, Germany
| | - Jörg Encke
- Chair of Bio-Inspired Information Processing, Department of Electrical and Computer Engineering, Technical University of Munich, Garching, Germany
| | - Andrea Lingner
- Division of Neurobiology, Department of Biology II, Ludwig-Maximilians-Universitaet Muenchen, Martinsried, Germany
| | - Todd R. Jennings
- Division of Neurobiology, Department of Biology II, Ludwig-Maximilians-Universitaet Muenchen, Martinsried, Germany
| | - Sonja Brosel
- Division of Neurobiology, Department of Biology II, Ludwig-Maximilians-Universitaet Muenchen, Martinsried, Germany
| | - Lars Kunz
- Division of Neurobiology, Department of Biology II, Ludwig-Maximilians-Universitaet Muenchen, Martinsried, Germany
| | - Benedikt Grothe
- Division of Neurobiology, Department of Biology II, Ludwig-Maximilians-Universitaet Muenchen, Martinsried, Germany
| | - Michael Pecka
- Division of Neurobiology, Department of Biology II, Ludwig-Maximilians-Universitaet Muenchen, Martinsried, Germany
- * E-mail:
| |
Collapse
|
8
|
Rankin J, Rinzel J. Computational models of auditory perception from feature extraction to stream segregation and behavior. Curr Opin Neurobiol 2019; 58:46-53. [PMID: 31326723 DOI: 10.1016/j.conb.2019.06.009] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2019] [Accepted: 06/22/2019] [Indexed: 10/26/2022]
Abstract
Audition is by nature dynamic, from brainstem processing on sub-millisecond time scales, to segregating and tracking sound sources with changing features, to the pleasure of listening to music and the satisfaction of getting the beat. We review recent advances from computational models of sound localization, of auditory stream segregation and of beat perception/generation. A wealth of behavioral, electrophysiological and imaging studies shed light on these processes, typically with synthesized sounds having regular temporal structure. Computational models integrate knowledge from different experimental fields and at different levels of description. We advocate a neuromechanistic modeling approach that incorporates knowledge of the auditory system from various fields, that utilizes plausible neural mechanisms, and that bridges our understanding across disciplines.
Collapse
Affiliation(s)
- James Rankin
- College of Engineering, Mathematics and Physical Sciences, University of Exeter, Harrison Building, North Park Rd, Exeter EX4 4QF, UK.
| | - John Rinzel
- Center for Neural Science, New York University, 4 Washington Place, 10003 New York, NY, United States; Courant Institute of Mathematical Sciences, New York University, 251 Mercer St, 10012 New York, NY, United States
| |
Collapse
|
9
|
Kettler L, Carr CE. Neural Maps of Interaural Time Difference in the American Alligator: A Stable Feature in Modern Archosaurs. J Neurosci 2019; 39:3882-3896. [PMID: 30886018 PMCID: PMC6520516 DOI: 10.1523/jneurosci.2989-18.2019] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/26/2018] [Revised: 02/20/2019] [Accepted: 02/23/2019] [Indexed: 11/21/2022] Open
Abstract
Detection of interaural time differences (ITDs) is crucial for sound localization in most vertebrates. The current view is that optimal computational strategies of ITD detection depend mainly on head size and available frequencies, although evolutionary history should also be taken into consideration. In archosaurs, which include birds and crocodiles, the brainstem nucleus laminaris (NL) developed into the critical structure for ITD detection. In birds, ITDs are mapped in an orderly array or place code, whereas in the mammalian medial superior olive, the analog of NL, maps are not found. As yet, in crocodilians, topographical representations have not been identified. However, nontopographic representations of ITD cannot be excluded due to different anatomical and ethological features of birds and crocodiles. Therefore, we measured ITD-dependent responses in the NL of anesthetized American alligators of either sex and identified the location of the recording sites by lesions made after recording. The measured extracellular field potentials, or neurophonics, were strongly ITD tuned, and their preferred ITDs correlated with the position in NL. As in birds, delay lines, which compensate for external time differences, formed maps of ITD. The broad distributions of best ITDs within narrow frequency bands were not consistent with an optimal coding model. We conclude that the available acoustic cues and the architecture of the acoustic system in early archosaurs led to a stable and similar organization in today's birds and crocodiles, although physical features, such as internally coupled ears, head size, or shape, and audible frequency range, vary among the two groups.SIGNIFICANCE STATEMENT Interaural time difference (ITD) is an important cue for sound localization, and the optimal strategies for encoding ITD in neuronal populations are the subject of ongoing debate. We show that alligators form maps of ITD very similar to birds, suggesting that their common archosaur ancestor reached a stable coding solution different from mammals. Mammals and diapsids evolved tympanic hearing independently, and local optima can be reached in evolution that are not considered by global optimal coding models. Thus, the presence of ITD maps in the brainstem may reflect a local optimum in evolutionary development. Our results underline the importance of comparative animal studies and show that optimal models must be viewed in the light of evolutionary processes.
Collapse
Affiliation(s)
- Lutz Kettler
- Lehrstuhl für Zoologie, Technische Universität München, 85354 Freising, Germany, and
| | - Catherine E Carr
- Department of Biology, University of Maryland, College Park, Maryland 20742
| |
Collapse
|
10
|
Lingner A, Pecka M, Leibold C, Grothe B. A novel concept for dynamic adjustment of auditory space. Sci Rep 2018; 8:8335. [PMID: 29844516 PMCID: PMC5974081 DOI: 10.1038/s41598-018-26690-0] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2017] [Accepted: 05/15/2018] [Indexed: 11/22/2022] Open
Abstract
Traditionally, the auditory system is thought to serve reliable sound localization. Stimulus-history driven feedback circuits in the early binaural pathway, however, contradict this canonical concept and raise questions about their functional significance. Here we show that stimulus-history dependent changes in absolute space perception are poorly captured by the traditional labeled-line and hemispheric-difference models of auditory space coding. We therefore developed a new decoding model incorporating recent electrophysiological findings in which sound location is initially computed in both brain hemispheres independently and combined to yield a hemispherically balanced code. This model closely captures the observed absolute localization errors caused by stimulus history, and furthermore predicts a selective dilation and compression of perceptional space. These model predictions are confirmed by improvement and degradation of spatial resolution in human listeners. Thus, dynamic perception of auditory space facilitates focal sound source segregation at the expense of absolute sound localization, questioning existing concepts of spatial hearing.
Collapse
Affiliation(s)
- A Lingner
- Division of Neurobiology, Department Biology II, Ludwig-Maximilians-Universitaet Muenchen, Großhaderner Str. 2-4, D-82152, Martinsried, Planegg, Germany
| | - M Pecka
- Division of Neurobiology, Department Biology II, Ludwig-Maximilians-Universitaet Muenchen, Großhaderner Str. 2-4, D-82152, Martinsried, Planegg, Germany
| | - C Leibold
- Bernstein Center for Computational Neuroscience Munich, Großhaderner Straße 2-4, D-82152, Martinsried, Germany
| | - B Grothe
- Division of Neurobiology, Department Biology II, Ludwig-Maximilians-Universitaet Muenchen, Großhaderner Str. 2-4, D-82152, Martinsried, Planegg, Germany.
| |
Collapse
|
11
|
Encke J, Hemmert W. Extraction of Inter-Aural Time Differences Using a Spiking Neuron Network Model of the Medial Superior Olive. Front Neurosci 2018; 12:140. [PMID: 29559886 PMCID: PMC5845713 DOI: 10.3389/fnins.2018.00140] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2017] [Accepted: 02/21/2018] [Indexed: 11/13/2022] Open
Abstract
The mammalian auditory system is able to extract temporal and spectral features from sound signals at the two ears. One important cue for localization of low-frequency sound sources in the horizontal plane are inter-aural time differences (ITDs) which are first analyzed in the medial superior olive (MSO) in the brainstem. Neural recordings of ITD tuning curves at various stages along the auditory pathway suggest that ITDs in the mammalian brainstem are not represented in form of a Jeffress-type place code. An alternative is the hemispheric opponent-channel code, according to which ITDs are encoded as the difference in the responses of the MSO nuclei in the two hemispheres. In this study, we present a physiologically-plausible, spiking neuron network model of the mammalian MSO circuit and apply two different methods of extracting ITDs from arbitrary sound signals. The network model is driven by a functional model of the auditory periphery and physiological models of the cochlear nucleus and the MSO. Using a linear opponent-channel decoder, we show that the network is able to detect changes in ITD with a precision down to 10 μs and that the sensitivity of the decoder depends on the slope of the ITD-rate functions. A second approach uses an artificial neuronal network to predict ITDs directly from the spiking output of the MSO and ANF model. Using this predictor, we show that the MSO-network is able to reliably encode static and time-dependent ITDs over a large frequency range, also for complex signals like speech.
Collapse
Affiliation(s)
- Jörg Encke
- Bioanaloge-Informationsverarbeitung, Department of Electrical and Computer Engineering, Technical University Munich, Munich, Germany
| | - Werner Hemmert
- Bioanaloge-Informationsverarbeitung, Department of Electrical and Computer Engineering, Technical University Munich, Munich, Germany
| |
Collapse
|
12
|
Salminen NH, Jones SJ, Christianson GB, Marquardt T, McAlpine D. A common periodic representation of interaural time differences in mammalian cortex. Neuroimage 2018; 167:95-103. [PMID: 29122721 PMCID: PMC5854251 DOI: 10.1016/j.neuroimage.2017.11.012] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2017] [Revised: 10/01/2017] [Accepted: 11/04/2017] [Indexed: 11/16/2022] Open
Abstract
Binaural hearing, the ability to detect small differences in the timing and level of sounds at the two ears, underpins the ability to localize sound sources along the horizontal plane, and is important for decoding complex spatial listening environments into separate objects – a critical factor in ‘cocktail-party listening’. For human listeners, the most important spatial cue is the interaural time difference (ITD). Despite many decades of neurophysiological investigations of ITD sensitivity in small mammals, and computational models aimed at accounting for human perception, a lack of concordance between these studies has hampered our understanding of how the human brain represents and processes ITDs. Further, neural coding of spatial cues might depend on factors such as head-size or hearing range, which differ considerably between humans and commonly used experimental animals. Here, using magnetoencephalography (MEG) in human listeners, and electro-corticography (ECoG) recordings in guinea pig—a small mammal representative of a range of animals in which ITD coding has been assessed at the level of single-neuron recordings—we tested whether processing of ITDs in human auditory cortex accords with a frequency-dependent periodic code of ITD reported in small mammals, or whether alternative or additional processing stages implemented in psychoacoustic models of human binaural hearing must be assumed. Our data were well accounted for by a model consisting of periodically tuned ITD-detectors, and were highly consistent across the two species. The results suggest that the representation of ITD in human auditory cortex is similar to that found in other mammalian species, a representation in which neural responses to ITD are determined by phase differences relative to sound frequency rather than, for instance, the range of ITDs permitted by head size or the absolute magnitude or direction of ITD. ITD tuning is studied in human MEG and guinea pig ECoG with identical stimuli. Auditory cortical tuning to ITD is highly consistent across species. Results are consistent with a periodic, frequency-dependent code.
Collapse
Affiliation(s)
- Nelli H Salminen
- Brain and Mind Laboratory, Dept. of Neuroscience and Biomedical Engineering, MEG Core, Aalto NeuroImaging, Aalto University School of Science, Espoo, Finland.
| | - Simon J Jones
- UCL Ear Institute, 332 Gray's Inn Road, London, WC1X 8EE, UK
| | | | | | - David McAlpine
- UCL Ear Institute, 332 Gray's Inn Road, London, WC1X 8EE, UK; Dept of Linguistics, Australian Hearing Hub, Macquarie University, Sydney, NSW 2109, Australia
| |
Collapse
|
13
|
Dietz M, Lestang JH, Majdak P, Stern RM, Marquardt T, Ewert SD, Hartmann WM, Goodman DFM. A framework for testing and comparing binaural models. Hear Res 2017; 360:92-106. [PMID: 29208336 DOI: 10.1016/j.heares.2017.11.010] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/30/2017] [Revised: 11/03/2017] [Accepted: 11/24/2017] [Indexed: 11/19/2022]
Abstract
Auditory research has a rich history of combining experimental evidence with computational simulations of auditory processing in order to deepen our theoretical understanding of how sound is processed in the ears and in the brain. Despite significant progress in the amount of detail and breadth covered by auditory models, for many components of the auditory pathway there are still different model approaches that are often not equivalent but rather in conflict with each other. Similarly, some experimental studies yield conflicting results which has led to controversies. This can be best resolved by a systematic comparison of multiple experimental data sets and model approaches. Binaural processing is a prominent example of how the development of quantitative theories can advance our understanding of the phenomena, but there remain several unresolved questions for which competing model approaches exist. This article discusses a number of current unresolved or disputed issues in binaural modelling, as well as some of the significant challenges in comparing binaural models with each other and with the experimental data. We introduce an auditory model framework, which we believe can become a useful infrastructure for resolving some of the current controversies. It operates models over the same paradigms that are used experimentally. The core of the proposed framework is an interface that connects three components irrespective of their underlying programming language: The experiment software, an auditory pathway model, and task-dependent decision stages called artificial observers that provide the same output format as the test subject.
Collapse
Affiliation(s)
- Mathias Dietz
- National Centre for Audiology, Western University, London, ON, Canada.
| | - Jean-Hugues Lestang
- Department of Electrical and Electronic Engineering, Imperial College London, London, United Kingdom
| | - Piotr Majdak
- Institut für Schallforschung, Österreichische Akademie der Wissenschaften, Wien, Austria
| | | | | | - Stephan D Ewert
- Medizinische Physik, Universität Oldenburg, Oldenburg, Germany
| | | | - Dan F M Goodman
- Department of Electrical and Electronic Engineering, Imperial College London, London, United Kingdom
| |
Collapse
|
14
|
Signatures of Somatic Inhibition and Dendritic Excitation in Auditory Brainstem Field Potentials. J Neurosci 2017; 37:10451-10467. [PMID: 28947575 DOI: 10.1523/jneurosci.0600-17.2017] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2017] [Revised: 09/12/2017] [Accepted: 09/14/2017] [Indexed: 01/20/2023] Open
Abstract
Extracellular voltage recordings (Ve ; field potentials) provide an accessible view of in vivo neural activity, but proper interpretation of field potentials is a long-standing challenge. Computational modeling can aid in identifying neural generators of field potentials. In the auditory brainstem of cats, spatial patterns of sound-evoked Ve can resemble, strikingly, Ve generated by current dipoles. Previously, we developed a biophysically-based model of a binaural brainstem nucleus, the medial superior olive (MSO), that accounts qualitatively for observed dipole-like Ve patterns in sustained responses to monaural tones with frequencies >∼1000 Hz (Goldwyn et al., 2014). We have observed, however, that Ve patterns in cats of both sexes appear more monopole-like for lower-frequency tones. Here, we enhance our theory to accurately reproduce dipole and non-dipole features of Ve responses to monaural tones with frequencies ranging from 600 to 1800 Hz. By applying our model to data, we estimate time courses of paired input currents to MSO neurons. We interpret these inputs as dendrite-targeting excitation and soma-targeting inhibition (the latter contributes non-dipole-like features to Ve responses). Aspects of inferred inputs are consistent with synaptic inputs to MSO neurons including the tendencies of inhibitory inputs to attenuate in response to high-frequency tones and to precede excitatory inputs. Importantly, our updated theory can be tested experimentally by blocking synaptic inputs. MSO neurons perform a critical role in sound localization and binaural hearing. By solving an inverse problem to uncover synaptic inputs from Ve patterns we provide a new perspective on MSO physiology.SIGNIFICANCE STATEMENT Extracellular voltages (field potentials) are a common measure of brain activity. Ideally, one could infer from these data the activity of neurons and synapses that generate field potentials, but this "inverse problem" is not easily solved. We study brainstem field potentials in the region of the medial superior olive (MSO); a critical center in the auditory pathway. These field potentials exhibit distinctive spatial and temporal patterns in response to pure tone sounds. We use mathematical modeling in combination with physiological and anatomical knowledge of MSO neurons to plausibly explain how dendrite-targeting excitation and soma-targeting inhibition generate these field potentials. Inferring putative synaptic currents from field potentials advances our ability to study neural processing of sound in the MSO.
Collapse
|
15
|
Carr CE, Christensen-Dalsgaard J, Bierman H. Coupled ears in lizards and crocodilians. BIOLOGICAL CYBERNETICS 2016; 110:291-302. [PMID: 27734148 PMCID: PMC6003244 DOI: 10.1007/s00422-016-0698-2] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/10/2016] [Accepted: 09/17/2016] [Indexed: 05/22/2023]
Abstract
Lizard ears are coupled across the pharynx, and are very directional. In consequence all auditory responses should be directional, without a requirement for computation of sound source location. Crocodilian ears are connected through sinuses, and thus less tightly coupled. Coupling may improve the processing of low-frequency directional signals, while higher frequency signals appear to be progressively uncoupled. In both lizards and crocodilians, the increased directionality of the coupled ears leads to an effectively larger head and larger physiological range of ITDs. This increased physiological range is reviewed in the light of current theories of sound localization.
Collapse
Affiliation(s)
- Catherine E Carr
- Department of Biology, University of Maryland College Park, College Park, MD, 20742, USA.
| | | | - Hilary Bierman
- Department of Biology, University of Maryland College Park, College Park, MD, 20742, USA
| |
Collapse
|
16
|
Evolutionary trends in directional hearing. Curr Opin Neurobiol 2016; 40:111-117. [PMID: 27448850 DOI: 10.1016/j.conb.2016.07.001] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2016] [Revised: 06/30/2016] [Accepted: 07/06/2016] [Indexed: 01/08/2023]
Abstract
Tympanic hearing is a true evolutionary novelty that arose in parallel within early tetrapods. We propose that in these tetrapods, selection for sound localization in air acted upon pre-existing directionally sensitive brainstem circuits, similar to those in fishes. Auditory circuits in birds and lizards resemble this ancestral, directionally sensitive framework. Despite this anatomically similarity, coding of sound source location differs between birds and lizards, although all show mechanisms for enhancing sound source directionality. Comparisons with mammals reveal similarly complex interactions between coding strategies and evolutionary history.
Collapse
|
17
|
Physiological Evidence for a Midline Spatial Channel in Human Auditory Cortex. J Assoc Res Otolaryngol 2016; 17:331-40. [PMID: 27164943 PMCID: PMC4940291 DOI: 10.1007/s10162-016-0571-y] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2015] [Accepted: 04/26/2016] [Indexed: 12/23/2022] Open
Abstract
Studies with humans and other mammals have provided support for a two-channel representation of horizontal (“azimuthal”) space in the auditory system. In this representation, location-sensitive neurons contribute activity to one of two broadly tuned channels whose responses are compared to derive an estimate of sound-source location. One channel is maximally responsive to sounds towards the left and the other to sounds towards the right. However, recent psychophysical studies of humans, and physiological studies of other mammals, point to the presence of an additional channel, maximally responsive to the midline. In this study, we used electroencephalography to seek physiological evidence for such a midline channel in humans. We measured neural responses to probe stimuli presented from straight ahead (0 °) or towards the right (+30 ° or +90 °). Probes were preceded by adapter stimuli to temporarily suppress channel activity. Adapters came from 0 ° or alternated between left and right (−30 ° and +30 ° or −90 ° and +90 °). For the +90 ° probe, to which the right-tuned channel would respond most strongly, both accounts predict greatest adaptation when the adapters are at ±90 °. For the 0 ° probe, the two-channel account predicts greatest adaptation from the ±90 ° adapters, while the three-channel account predicts greatest adaptation when the adapters are at 0 ° because these adapters stimulate the midline-tuned channel which responds most strongly to the 0 ° probe. The results were consistent with the three-channel account. In addition, a computational implementation of the three-channel account fitted the probe response sizes well, explaining 93 % of the variance about the mean, whereas a two-channel implementation produced a poor fit and explained only 61 % of the variance.
Collapse
|
18
|
Carr CE, Christensen-Dalsgaard J. Sound Localization Strategies in Three Predators. BRAIN, BEHAVIOR AND EVOLUTION 2015; 86:17-27. [PMID: 26398572 DOI: 10.1159/000435946] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
Abstract
In this paper, we compare some of the neural strategies for sound localization and encoding interaural time differences (ITDs) in three predatory species of Reptilia, alligators, barn owls and geckos. Birds and crocodilians are sister groups among the extant archosaurs, while geckos are lepidosaurs. Despite the similar organization of their auditory systems, archosaurs and lizards use different strategies for encoding the ITDs that underlie localization of sound in azimuth. Barn owls encode ITD information using a place map, which is composed of neurons serving as labeled lines tuned for preferred spatial locations, while geckos may use a meter strategy or population code composed of broadly sensitive neurons that represent ITD via changes in the firing rate.
Collapse
Affiliation(s)
- Catherine E Carr
- Department of Biology, University of Maryland Center for the Comparative and Evolutionary Biology of Hearing, College Park, Md., USA
| | | |
Collapse
|
19
|
Palanca-Castan N, Köppl C. Change in the coding of interaural time difference along the tonotopic axis of the chicken nucleus laminaris. Front Neural Circuits 2015; 9:43. [PMID: 26347616 PMCID: PMC4542463 DOI: 10.3389/fncir.2015.00043] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2015] [Accepted: 08/05/2015] [Indexed: 11/23/2022] Open
Abstract
Interaural time differences (ITDs) are an important cue for the localization of sounds in azimuthal space. Both birds and mammals have specialized, tonotopically organized nuclei in the brain stem for the processing of ITD: medial superior olive in mammals and nucleus laminaris (NL) in birds. The specific way in which ITDs are derived was long assumed to conform to a delay-line model in which arrays of systematically arranged cells create a representation of auditory space with different cells responding maximally to specific ITDs. This model was supported by data from barn owl NL taken from regions above 3 kHz and from chicken above 1 kHz. However, data from mammals often do not show defining features of the Jeffress model such as a systematic topographic representation of best ITDs or the presence of axonal delay lines, and an alternative has been proposed in which neurons are not topographically arranged with respect to ITD and coding occurs through the assessment of the overall response of two large neuron populations, one in each hemisphere. Modeling studies have suggested that the presence of different coding systems could be related to the animal’s head size and frequency range rather than their phylogenetic group. Testing this hypothesis requires data from across the tonotopic range of both birds and mammals. The aim of this study was to obtain in vivo recordings from neurons in the low-frequency range (<1000 Hz) of chicken NL. Our data argues for the presence of a modified Jeffress system that uses the slopes of ITD-selective response functions instead of their peaks to topographically represent ITD at mid- to high frequencies. At low frequencies, below several 100 Hz, the data did not support any current model of ITD coding. This is different to what was previously shown in the barn owl and suggests that constraints in optimal ITD processing may be associated with the particular demands on sound localization determined by the animal’s ecological niche in the same way as other perceptual systems such as field of best vision.
Collapse
Affiliation(s)
- Nicolas Palanca-Castan
- Cluster of Excellence "Hearing4all" and Research Center Neurosensory Science and Department of Neuroscience, School of Medicine and Health Sciences, Carl von Ossietzky University of Oldenburg Oldenburg, Germany
| | - Christine Köppl
- Cluster of Excellence "Hearing4all" and Research Center Neurosensory Science and Department of Neuroscience, School of Medicine and Health Sciences, Carl von Ossietzky University of Oldenburg Oldenburg, Germany
| |
Collapse
|
20
|
Schnupp JWH, Garcia-Lazaro JA, Lesica NA. Periodotopy in the gerbil inferior colliculus: local clustering rather than a gradient map. Front Neural Circuits 2015; 9:37. [PMID: 26379508 PMCID: PMC4550179 DOI: 10.3389/fncir.2015.00037] [Citation(s) in RCA: 34] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2015] [Accepted: 07/07/2015] [Indexed: 11/13/2022] Open
Abstract
Periodicities in sound waveforms are widespread, and shape important perceptual attributes of sound including rhythm and pitch. Previous studies have indicated that, in the inferior colliculus (IC), a key processing stage in the auditory midbrain, neurons tuned to different periodicities might be arranged along a periodotopic axis which runs approximately orthogonal to the tonotopic axis. Here we map out the topography of frequency and periodicity tuning in the IC of gerbils in unprecedented detail, using pure tones and different periodic sounds, including click trains, sinusoidally amplitude modulated (SAM) noise and iterated rippled noise. We found that while the tonotopic map exhibited a clear and highly reproducible gradient across all animals, periodotopic maps varied greatly across different types of periodic sound and from animal to animal. Furthermore, periodotopic gradients typically explained only about 10% of the variance in modulation tuning between recording sites. However, there was a strong local clustering of periodicity tuning at a spatial scale of ca. 0.5 mm, which also differed from animal to animal.
Collapse
Affiliation(s)
- Jan W H Schnupp
- Department of Physiology, Anatomy and Genetics, University of Oxford Oxford, UK
| | | | | |
Collapse
|
21
|
Wood KC, Bizley JK. Relative sound localisation abilities in human listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 138:674-686. [PMID: 26328685 PMCID: PMC4610194 DOI: 10.1121/1.4923452] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/12/2015] [Revised: 06/16/2015] [Accepted: 06/22/2015] [Indexed: 06/05/2023]
Abstract
Spatial acuity varies with sound-source azimuth, signal-to-noise ratio, and the spectral characteristics of the sound source. Here, the spatial localisation abilities of listeners were assessed using a relative localisation task. This task tested localisation ability at fixed angular separations throughout space using a two-alternative forced-choice design across a variety of listening conditions. Subjects were required to determine whether a target sound originated to the left or right of a preceding reference in the presence of a multi-source noise background. Experiment 1 demonstrated that subjects' ability to determine the relative location of two sources declined with less favourable signal-to-noise ratios and at peripheral locations. Experiment 2 assessed performance with both broadband and spectrally restricted stimuli designed to limit localisation cues to predominantly interaural level differences or interaural timing differences (ITDs). Predictions generated from topographic, modified topographic, and two-channel models of sound localisation suggest that for low-pass stimuli, where ITD cues were dominant, the two-channel model provides an adequate description of the experimental data, whereas for broadband and high frequency bandpass stimuli none of the models was able to fully account for performance. Experiment 3 demonstrated that relative localisation performance was uninfluenced by shifts in gaze direction.
Collapse
Affiliation(s)
- Katherine C Wood
- University College London Ear Institute, 332 Grays Inn Road, London, WC1X 8EE, United Kingdom
| | - Jennifer K Bizley
- University College London Ear Institute, 332 Grays Inn Road, London, WC1X 8EE, United Kingdom
| |
Collapse
|
22
|
Palanca-Castan N, Köppl C. In vivo Recordings from Low-Frequency Nucleus Laminaris in the Barn Owl. BRAIN, BEHAVIOR AND EVOLUTION 2015; 85:271-86. [PMID: 26182962 DOI: 10.1159/000433513] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/08/2014] [Accepted: 01/20/2015] [Indexed: 11/19/2022]
Abstract
Localization of sound sources relies on 2 main binaural cues: interaural time differences (ITD) and interaural level differences. ITD computing is first carried out in tonotopically organized areas of the brainstem nucleus laminaris (NL) in birds and the medial superior olive (MSO) in mammals. The specific way in which ITD are derived was long assumed to conform to a delay line model in which arrays of systematically arranged cells create a representation of auditory space, with different cells responding maximally to specific ITD. This model conforms in many details to the particular case of the high-frequency regions (above 3 kHz) in the barn owl NL. However, data from recent studies in mammals are not consistent with a delay line model. A new model has been suggested in which neurons are not topographically arranged with respect to ITD and coding occurs through assessment of the overall response of 2 large neuron populations – 1 in each brainstem hemisphere. Currently available data comprise mainly low-frequency (<1,500 Hz) recordings in the case of mammals and higher-frequency recordings in the case of birds. This makes it impossible to distinguish between group-related adaptations and frequency-related adaptations. Here we report the first comprehensive data set from low-frequency NL in the barn owl and compare it to data from other avian and mammalian studies. Our data are consistent with a delay line model, so differences between ITD processing systems are more likely to have originated through divergent evolution of different vertebrate groups.
Collapse
Affiliation(s)
- Nicolas Palanca-Castan
- Cluster of Excellence Hearing4all, Research Center Neurosensory Science and Department of Neuroscience, School of Medicine and Health Sciences, Carl von Ossietzky University, Oldenburg, Germany
| | | |
Collapse
|