1
|
Mitchell PW, Carney LH. A Computational Model of Auditory Chirp-Velocity Sensitivity and Amplitude-Modulation Tuning in Inferior Colliculus Neurons. RESEARCH SQUARE 2024:rs.3.rs-4450943. [PMID: 38883707 PMCID: PMC11177976 DOI: 10.21203/rs.3.rs-4450943/v1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/18/2024]
Abstract
We demonstrate a model of chirp-velocity sensitivity in the inferior colliculus (IC) that retains the tuning to amplitude modulation (AM) that was established in earlier models. The mechanism of velocity sensitivity is sequence detection by octopus cells of the posteroventral cochlear nucleus, which have been proposed in physiological studies to respond preferentially to the order of arrival of cross-frequency inputs of different amplitudes. Model architecture is based on coincidence detection of a combination of excitatory and inhibitory inputs. Chirp-sensitivity of the IC output is largely controlled by the strength and timing of the chirp-sensitive octopus-cell inhibitory input. AM tuning is controlled by inhibition and excitation that are tuned to the same frequency. We present several example neurons that demonstrate the feasibility of the model in simulating realistic chirp-sensitivity and AM tuning for a wide range of characteristic frequencies. Additionally, we explore the systematic impact of varying parameters on model responses. The proposed model can be used to assess the contribution of IC chirp-velocity sensitivity to responses to complex sounds, such as speech.
Collapse
Affiliation(s)
- Paul W. Mitchell
- Department of Biomedical Engineering, University of Rochester, 601 Elmwood Ave, Rochester, NY, 14642, USA
| | - Laurel H. Carney
- Department of Biomedical Engineering, University of Rochester, 601 Elmwood Ave, Rochester, NY, 14642, USA
- Department of Neuroscience, University of Rochester, 601 Elmwood Ave, Rochester, NY, 14642, USA
| |
Collapse
|
2
|
Saddler MR, McDermott JH. Models optimized for real-world tasks reveal the necessity of precise temporal coding in hearing. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.04.21.590435. [PMID: 38712054 PMCID: PMC11071365 DOI: 10.1101/2024.04.21.590435] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/08/2024]
Abstract
Neurons encode information in the timing of their spikes in addition to their firing rates. Spike timing is particularly precise in the auditory nerve, where action potentials phase lock to sound with sub-millisecond precision, but its behavioral relevance is uncertain. To investigate the role of this temporal coding, we optimized machine learning models to perform real-world hearing tasks with simulated cochlear input. We asked how precise auditory nerve spike timing needed to be to reproduce human behavior. Models with high-fidelity phase locking exhibited more human-like sound localization and speech perception than models without, consistent with an essential role in human hearing. Degrading phase locking produced task-dependent effects, revealing how the use of fine-grained temporal information reflects both ecological task demands and neural implementation constraints. The results link neural coding to perception and clarify conditions in which prostheses that fail to restore high-fidelity temporal coding could in principle restore near-normal hearing.
Collapse
Affiliation(s)
- Mark R Saddler
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Center for Brains, Minds, and Machines, MIT, Cambridge, MA, USA
| | - Josh H McDermott
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Center for Brains, Minds, and Machines, MIT, Cambridge, MA, USA
- Program in Speech and Hearing Biosciences and Technology, Harvard, Cambridge, MA, USA
| |
Collapse
|
3
|
Carney LH. Neural Fluctuation Contrast as a Code for Complex Sounds: The Role and Control of Peripheral Nonlinearities. Hear Res 2024; 443:108966. [PMID: 38310710 PMCID: PMC10923127 DOI: 10.1016/j.heares.2024.108966] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/30/2023] [Revised: 01/14/2024] [Accepted: 01/26/2024] [Indexed: 02/06/2024]
Abstract
The nonlinearities of the inner ear are often considered to be obstacles that the central nervous system has to overcome to decode neural responses to sounds. This review describes how peripheral nonlinearities, such as saturation of the inner-hair-cell response and of the IHC-auditory-nerve synapse, are instead beneficial to the neural encoding of complex sounds such as speech. These nonlinearities set up contrast in the depth of neural-fluctuations in auditory-nerve responses along the tonotopic axis, referred to here as neural fluctuation contrast (NFC). Physiological support for the NFC coding hypothesis is reviewed, and predictions of several psychophysical phenomena, including masked detection and speech intelligibility, are presented. Lastly, a framework based on the NFC code for understanding how the medial olivocochlear (MOC) efferent system contributes to the coding of complex sounds is presented. By modulating cochlear gain control in response to both sound energy and fluctuations in neural responses, the MOC system is hypothesized to function not as a simple feedback gain-control device, but rather as a mechanism for enhancing NFC along the tonotopic axis, enabling robust encoding of complex sounds across a wide range of sound levels and in the presence of background noise. Effects of sensorineural hearing loss on the NFC code and on the MOC feedback system are presented and discussed.
Collapse
Affiliation(s)
- Laurel H Carney
- Depts. of Biomedical Engineering, Neuroscience, and Electrical & Computer Engineering University of Rochester, Rochester, NY, USA.
| |
Collapse
|
4
|
Wei L, Verschooten E, Joris PX. Enhancement of phase-locking in rodents. II. An axonal recording study in chinchilla. J Neurophysiol 2023; 130:751-767. [PMID: 37609701 DOI: 10.1152/jn.00474.2022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2022] [Revised: 08/07/2023] [Accepted: 08/15/2023] [Indexed: 08/24/2023] Open
Abstract
The trapezoid body (TB) contains axons of neurons residing in the anteroventral cochlear nucleus (AVCN) that provide excitatory and inhibitory inputs to the main monaural and binaural nuclei in the superior olivary complex (SOC). To understand the monaural and binaural response properties of neurons in the medial and lateral superior olive (MSO and LSO), it is important to characterize the temporal firing properties of these inputs. Because of its exceptional low-frequency hearing, the chinchilla (Chinchilla lanigera) is one of the widely used small animal models for studies of hearing. However, the characterization of the output of its ventral cochlear nucleus to the nuclei of the SOC is fragmentary. We obtained responses of TB axons to stimuli typically used in binaural studies and compared these responses to those of auditory nerve (AN) fibers, with a focus on temporal coding. We found enhancement of phase-locking and entrainment, i.e., the ability of a neuron to fire action potentials at a certain stimulus phase for nearly every stimulus period, in TB axons relative to AN fibers. Enhancement in phase-locking and entrainment are quantitatively more modest than in the cat but greater than in the gerbil. As in these species, these phenomena occur not only in low-frequency neurons stimulated at their characteristic frequency but also in neurons tuned to higher frequencies when stimulated with low-frequency tones, to which complex phase-locking behavior with multiple modes of firing per stimulus cycle is frequently observed.NEW & NOTEWORTHY The sensitivity of neurons to small time differences in sustained sounds to both ears is important for binaural hearing, and this sensitivity is critically dependent on phase-locking in the monaural pathways. Although studies in cat showed a marked improvement in phase-locking from the peripheral to the central auditory nervous system, the evidence in rodents is mixed. Here, we recorded from AN and TB of chinchilla and found temporal enhancement, though more limited than in cat.
Collapse
Affiliation(s)
- Liting Wei
- Laboratory of Auditory Neurophysiology, KU Leuven, Leuven, Belgium
| | - Eric Verschooten
- Laboratory of Auditory Neurophysiology, KU Leuven, Leuven, Belgium
| | - Philip X Joris
- Laboratory of Auditory Neurophysiology, KU Leuven, Leuven, Belgium
| |
Collapse
|
5
|
Vinay, Moore BCJ. Exploiting individual differences to assess the role of place and phase locking cues in auditory frequency discrimination at 2 kHz. Sci Rep 2023; 13:13801. [PMID: 37612303 PMCID: PMC10447419 DOI: 10.1038/s41598-023-40571-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Accepted: 08/13/2023] [Indexed: 08/25/2023] Open
Abstract
The relative role of place and temporal mechanisms in auditory frequency discrimination was assessed for a centre frequency of 2 kHz. Four measures of frequency discrimination were obtained for 63 normal-hearing participants: detection of frequency modulation using modulation rates of 2 Hz (FM2) and 20 Hz (FM20); detection of a change in frequency across successive pure tones (difference limen for frequency, DLF); and detection of changes in the temporal fine structure of bandpass filtered complex tones centred at 2 kHz (TFS). Previous work has suggested that: FM2 depends on the use of both temporal and place cues; FM20 depends primarily on the use of place cues because the temporal mechanism cannot track rapid changes in frequency; DLF depends primarily on temporal cues; TFS depends exclusively on temporal cues. This led to the following predicted patterns of the correlations of scores across participants: DLF and TFS should be highly correlated; FM2 should be correlated with DLF and TFS; FM20 should not be correlated with DLF or TFS. The results were broadly consistent with these predictions and with the idea that frequency discrimination at 2 kHz depends partly or primarily on temporal cues except for frequency modulation detection at a high rate.
Collapse
Affiliation(s)
- Vinay
- Audiology Group, Department of Neuromedicine and Movement Science, Faculty of Medicine and Health Sciences, Norwegian University of Science and Technology (NTNU), Tungasletta 2, 7491, Trondheim, Norway.
| | - Brian C J Moore
- Cambridge Hearing Group, Department of Psychology, University of Cambridge, Cambridge, UK
| |
Collapse
|
6
|
Moore BCJ, Vinay. Assessing mechanisms of frequency discrimination by comparison of different measures over a wide frequency range. Sci Rep 2023; 13:11379. [PMID: 37452119 PMCID: PMC10349105 DOI: 10.1038/s41598-023-38600-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2023] [Accepted: 07/11/2023] [Indexed: 07/18/2023] Open
Abstract
It has been hypothesized that auditory detection of frequency modulation (FM) for low FM rates depends on the use of both temporal (phase locking) and place cues, depending on the carrier frequency, while detection of FM at high rates depends primarily on the use of place cues. To test this, FM detection for 2 and 20 Hz rates was measured over a wide frequency range, 1-10 kHz, including high frequencies for which temporal cues are assumed to be very weak. Performance was measured over the same frequency range for a task involving detection of changes in the temporal fine structure (TFS) of bandpass filtered complex tones, for which performance is assumed to depend primarily on the use of temporal cues. FM thresholds were better for the 2- than for the 20-Hz rate for center frequencies up to 4 kHz, while the reverse was true for higher center frequencies. For both FM rates, the thresholds, expressed as a proportion of the center frequency, were roughly constant for center frequencies from 6 to 10 Hz, consistent with the use of place cues. For the TFS task, thresholds worsened progressively with increasing frequency above 4 kHz, consistent with the weakening of temporal cues.
Collapse
Affiliation(s)
- Brian C J Moore
- Cambridge Hearing Group, Department of Psychology, University of Cambridge, Cambridge, UK.
| | - Vinay
- Audiology Group, Department of Neuromedicine and Movement Science, Faculty of Medicine and Health Sciences, Norwegian University of Science and Technology (NTNU), Tungasletta 2, 7491, Trondheim, Norway
| |
Collapse
|
7
|
Klug J, Encke J, Dietz M. Characterization of the decline in auditory nerve phase locking at high frequencies. JASA EXPRESS LETTERS 2023; 3:074403. [PMID: 37477582 DOI: 10.1121/10.0020267] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/30/2023] [Accepted: 07/04/2023] [Indexed: 07/22/2023]
Abstract
The frequency dependence of phase locking in the auditory nerve influences various auditory coding mechanisms. The decline of phase locking with increasing frequency is commonly described by a low-pass filter. This study compares fitted low-pass filter parameters with the actual rate of phase locking decline. The decline is similar across studies and only 40 dB per decade, corresponding to the asymptotic decline of a second order filter.
Collapse
Affiliation(s)
- Jonas Klug
- Department of Medical Physics and Acoustics, University of Oldenburg, 26129 Oldenburg, , ,
| | - Jörg Encke
- Department of Medical Physics and Acoustics, University of Oldenburg, 26129 Oldenburg, , ,
| | - Mathias Dietz
- Department of Medical Physics and Acoustics, University of Oldenburg, 26129 Oldenburg, , ,
| |
Collapse
|
8
|
Faran M, Furst M. Inner-hair-cell induced hearing loss: A biophysical modeling perspective. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 153:1776. [PMID: 37002110 DOI: 10.1121/10.0017627] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/20/2022] [Accepted: 02/28/2023] [Indexed: 05/18/2023]
Abstract
In recent years, experimental studies have demonstrated that malfunction of the inner-hair cells and their synapse to the auditory nerve is a significant hearing loss (HL) contributor. This study presents a detailed biophysical model of the inner-hair cells embedded in an end-to-end computational model of the auditory pathway with an acoustic signal as an input and prediction of human audiometric thresholds as an output. The contribution of the outer hair cells is included in the mechanical model of the cochlea. Different types of HL were simulated by changing mechanical and biochemical parameters of the inner and outer hair cells. The predicted thresholds yielded common audiograms of hearing impairment. Outer hair cell damage could only introduce threshold shifts at mid-high frequencies up to 40 dB. Inner hair cell damage affects low and high frequencies differently. All types of inner hair cell deficits yielded a maximum of 40 dB HL at low frequencies. Only a significant reduction in the number of cilia of the inner-hair cells yielded HL of up to 120 dB HL at high frequencies. Sloping audiograms can be explained by a combination of gradual change in the number of cilia of inner and outer hair cells along the cochlear partition from apex to base.
Collapse
Affiliation(s)
- Michael Faran
- School of Electrical Engineering, Faculty of Engineering, Tel-Aviv University, Tel-Aviv, Israel
| | - Miriam Furst
- School of Electrical Engineering, Faculty of Engineering, Tel-Aviv University, Tel-Aviv, Israel
| |
Collapse
|
9
|
Sadagopan S, Kar M, Parida S. Quantitative models of auditory cortical processing. Hear Res 2023; 429:108697. [PMID: 36696724 PMCID: PMC9928778 DOI: 10.1016/j.heares.2023.108697] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/18/2022] [Revised: 12/17/2022] [Accepted: 01/12/2023] [Indexed: 01/15/2023]
Abstract
To generate insight from experimental data, it is critical to understand the inter-relationships between individual data points and place them in context within a structured framework. Quantitative modeling can provide the scaffolding for such an endeavor. Our main objective in this review is to provide a primer on the range of quantitative tools available to experimental auditory neuroscientists. Quantitative modeling is advantageous because it can provide a compact summary of observed data, make underlying assumptions explicit, and generate predictions for future experiments. Quantitative models may be developed to characterize or fit observed data, to test theories of how a task may be solved by neural circuits, to determine how observed biophysical details might contribute to measured activity patterns, or to predict how an experimental manipulation would affect neural activity. In complexity, quantitative models can range from those that are highly biophysically realistic and that include detailed simulations at the level of individual synapses, to those that use abstract and simplified neuron models to simulate entire networks. Here, we survey the landscape of recently developed models of auditory cortical processing, highlighting a small selection of models to demonstrate how they help generate insight into the mechanisms of auditory processing. We discuss examples ranging from models that use details of synaptic properties to explain the temporal pattern of cortical responses to those that use modern deep neural networks to gain insight into human fMRI data. We conclude by discussing a biologically realistic and interpretable model that our laboratory has developed to explore aspects of vocalization categorization in the auditory pathway.
Collapse
Affiliation(s)
- Srivatsun Sadagopan
- Department of Neurobiology, University of Pittsburgh, Pittsburgh, PA, USA; Center for Neuroscience, University of Pittsburgh, Pittsburgh, PA, USA; Center for the Neural Basis of Cognition, University of Pittsburgh, Pittsburgh, PA, USA; Department of Bioengineering, University of Pittsburgh, Pittsburgh, PA, USA; Department of Communication Science and Disorders, University of Pittsburgh, Pittsburgh, PA, USA.
| | - Manaswini Kar
- Department of Neurobiology, University of Pittsburgh, Pittsburgh, PA, USA; Center for Neuroscience, University of Pittsburgh, Pittsburgh, PA, USA; Center for the Neural Basis of Cognition, University of Pittsburgh, Pittsburgh, PA, USA
| | - Satyabrata Parida
- Department of Neurobiology, University of Pittsburgh, Pittsburgh, PA, USA; Center for Neuroscience, University of Pittsburgh, Pittsburgh, PA, USA
| |
Collapse
|
10
|
Oxenham AJ. Questions and controversies surrounding the perception and neural coding of pitch. Front Neurosci 2023; 16:1074752. [PMID: 36699531 PMCID: PMC9868815 DOI: 10.3389/fnins.2022.1074752] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2022] [Accepted: 12/16/2022] [Indexed: 01/12/2023] Open
Abstract
Pitch is a fundamental aspect of auditory perception that plays an important role in our ability to understand speech, appreciate music, and attend to one sound while ignoring others. The questions surrounding how pitch is represented in the auditory system, and how our percept relates to the underlying acoustic waveform, have been a topic of inquiry and debate for well over a century. New findings and technological innovations have led to challenges of some long-standing assumptions and have raised new questions. This article reviews some recent developments in the study of pitch coding and perception and focuses on the topic of how pitch information is extracted from peripheral representations based on frequency-to-place mapping (tonotopy), stimulus-driven auditory-nerve spike timing (phase locking), or a combination of both. Although a definitive resolution has proved elusive, the answers to these questions have potentially important implications for mitigating the effects of hearing loss via devices such as cochlear implants.
Collapse
Affiliation(s)
- Andrew J. Oxenham
- Center for Applied and Translational Sensory Science, University of Minnesota Twin Cities, Minneapolis, MN, United States
- Department of Psychology, University of Minnesota Twin Cities, Minneapolis, MN, United States
| |
Collapse
|
11
|
Bissmeyer SRS, Goldsworthy RL. Combining Place and Rate of Stimulation Improves Frequency Discrimination in Cochlear Implant Users. Hear Res 2022; 424:108583. [PMID: 35930901 PMCID: PMC10849775 DOI: 10.1016/j.heares.2022.108583] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/25/2022] [Revised: 06/20/2022] [Accepted: 07/21/2022] [Indexed: 11/04/2022]
Abstract
In the auditory system, frequency is represented as tonotopic and temporal response properties of the auditory nerve. While these response properties are inextricably linked in normal hearing, cochlear implants can separately excite tonotopic location and temporal synchrony using different electrodes and stimulation rates, respectively. This separation allows for the investigation of the contributions of tonotopic and temporal cues for frequency discrimination. The present study examines frequency discrimination in adult cochlear implant users as conveyed by electrode position and stimulation rate, separately and combined. The working hypothesis is that frequency discrimination is better provided by place and rate cues combined compared to either cue alone. This hypothesis was tested in two experiments. In the first experiment, frequency discrimination needed for melodic contour identification was measured for frequencies near 100, 200, and 400 Hz using frequency allocation modeled after clinical processors. In the second experiment, frequency discrimination for pitch ranking was measured for frequencies between 100 and 1600 Hz using an experimental frequency allocation designed to provide better access to place cues. The results of both experiments indicate that frequency discrimination is better with place and rate cues combined than with either cue alone. These results clarify how signal processing for cochlear implants could better encode frequency into place and rate of electrical stimulation. Further, the results provide insight into the contributions of place and rate cues for pitch.
Collapse
Affiliation(s)
- Susan R S Bissmeyer
- Department of Biomedical Engineering, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States; Auditory Research Center, Health Research Association, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, 1640 Marengo Street Suite 326, Los Angeles, CA 90033, United States.
| | - Raymond L Goldsworthy
- Department of Biomedical Engineering, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States; Auditory Research Center, Health Research Association, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, 1640 Marengo Street Suite 326, Los Angeles, CA 90033, United States
| |
Collapse
|
12
|
Mehta AH, Oxenham AJ. Role of perceptual integration in pitch discrimination at high frequenciesa). JASA EXPRESS LETTERS 2022; 2:084402. [PMID: 37311192 PMCID: PMC10264831 DOI: 10.1121/10.0013429] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/27/2022] [Accepted: 07/26/2022] [Indexed: 06/15/2023]
Abstract
At very high frequencies, fundamental-frequency difference limens (F0DLs) for five-component harmonic complex tones can be better than predicted by optimal integration of information, assuming performance is limited by noise at the peripheral level, but are in line with predictions based on more central sources of noise. This study investigates whether there is a minimum number of harmonic components needed for such super-optimal integration effects and if harmonic range or inharmonicity affects this super-optimal integration. Results show super-optimal integration, even with two harmonic components and for most combinations of consecutive harmonic, but not inharmonic, components.
Collapse
Affiliation(s)
- Anahita H Mehta
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455, USA ,
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455, USA ,
| |
Collapse
|
13
|
Joris PX. In praise of adventitious sounds. Hear Res 2022; 425:108592. [DOI: 10.1016/j.heares.2022.108592] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/04/2021] [Revised: 07/13/2022] [Accepted: 07/26/2022] [Indexed: 11/04/2022]
|
14
|
The burst gap is a peripheral temporal code for pitch perception that is shared across audition and touch. Sci Rep 2022; 12:11014. [PMID: 35773321 PMCID: PMC9246943 DOI: 10.1038/s41598-022-15269-5] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2022] [Accepted: 06/21/2022] [Indexed: 11/08/2022] Open
Abstract
When tactile afferents were manipulated to fire in periodic bursts of spikes, we discovered that the perceived pitch corresponded to the inter-burst interval (burst gap) in a spike train, rather than the spike rate or burst periodicity as previously thought. Given that tactile frequency mechanisms have many analogies to audition, and indications that temporal frequency channels are linked across the two modalities, we investigated whether there is burst gap temporal encoding in the auditory system. To link this putative neural code to perception, human subjects (n = 13, 6 females) assessed pitch elicited by trains of temporally-structured acoustic pulses in psychophysical experiments. Each pulse was designed to excite a fixed population of cochlear neurons, precluding place of excitation cues, and to elicit desired temporal spike trains in activated afferents. We tested periodicities up to 150 Hz using a variety of burst patterns and found striking deviations from periodicity-predicted pitch. Like the tactile system, the duration of the silent gap between successive bursts of neural activity best predicted perceived pitch, emphasising the role of peripheral temporal coding in shaping pitch. This suggests that temporal patterning of stimulus pulses in cochlear implant users might improve pitch perception.
Collapse
|
15
|
Cariani P, Baker JM. Time Is of the Essence: Neural Codes, Synchronies, Oscillations, Architectures. Front Comput Neurosci 2022; 16:898829. [PMID: 35814343 PMCID: PMC9262106 DOI: 10.3389/fncom.2022.898829] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2022] [Accepted: 05/04/2022] [Indexed: 11/25/2022] Open
Abstract
Time is of the essence in how neural codes, synchronies, and oscillations might function in encoding, representation, transmission, integration, storage, and retrieval of information in brains. This Hypothesis and Theory article examines observed and possible relations between codes, synchronies, oscillations, and types of neural networks they require. Toward reverse-engineering informational functions in brains, prospective, alternative neural architectures incorporating principles from radio modulation and demodulation, active reverberant circuits, distributed content-addressable memory, signal-signal time-domain correlation and convolution operations, spike-correlation-based holography, and self-organizing, autoencoding anticipatory systems are outlined. Synchronies and oscillations are thought to subserve many possible functions: sensation, perception, action, cognition, motivation, affect, memory, attention, anticipation, and imagination. These include direct involvement in coding attributes of events and objects through phase-locking as well as characteristic patterns of spike latency and oscillatory response. They are thought to be involved in segmentation and binding, working memory, attention, gating and routing of signals, temporal reset mechanisms, inter-regional coordination, time discretization, time-warping transformations, and support for temporal wave-interference based operations. A high level, partial taxonomy of neural codes consists of channel, temporal pattern, and spike latency codes. The functional roles of synchronies and oscillations in candidate neural codes, including oscillatory phase-offset codes, are outlined. Various forms of multiplexing neural signals are considered: time-division, frequency-division, code-division, oscillatory-phase, synchronized channels, oscillatory hierarchies, polychronous ensembles. An expandable, annotative neural spike train framework for encoding low- and high-level attributes of events and objects is proposed. Coding schemes require appropriate neural architectures for their interpretation. Time-delay, oscillatory, wave-interference, synfire chain, polychronous, and neural timing networks are discussed. Some novel concepts for formulating an alternative, more time-centric theory of brain function are discussed. As in radio communication systems, brains can be regarded as networks of dynamic, adaptive transceivers that broadcast and selectively receive multiplexed temporally-patterned pulse signals. These signals enable complex signal interactions that select, reinforce, and bind common subpatterns and create emergent lower dimensional signals that propagate through spreading activation interference networks. If memory traces share the same kind of temporal pattern forms as do active neuronal representations, then distributed, holograph-like content-addressable memories are made possible via temporal pattern resonances.
Collapse
Affiliation(s)
- Peter Cariani
- Hearing Research Center, Boston University, Boston, MA, United States
- Department of Otolaryngology-Head and Neck Surgery, Harvard Medical School, Boston, MA, United States
| | | |
Collapse
|
16
|
Stalnov O, Faran M, Koral Y, Furst M. Auditory detection probability of propeller noise in hover flight in presence of ambient soundscape. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 151:3719. [PMID: 35778181 DOI: 10.1121/10.0011546] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/03/2022] [Accepted: 05/13/2022] [Indexed: 06/15/2023]
Abstract
Unmanned aerial vehicles are rapidly advancing and becoming ubiquitous in an unlimited number of applications, from parcel delivery to people transportation. As unmanned aerial vehicle (UAV) markets expand, the increased acoustic nuisance on population becomes a more acute problem. Previous aircraft noise assessments have highlighted the necessity of a psychoacoustic metric for quantification of human audio perception. This study presents a framework for estimating propeller-based UAV auditory detection probability on the ground for a listener in a real-life scenario. The detection probability is derived by using its free-field measured acoustic background and estimating the UAV threshold according to a physiological model of the auditory pathway. The method is presented via results of an exemplar measurement in an anechoic environment with a single two- and five-bladed propeller. It was found that the auditory detection probability is primarily affected by the background noise level, whereas the number of blades is a less significant parameter. The significance of the proposed method lies in providing a quantitative evaluation of auditory detection probability of the UAV on the ground in the presence of a given soundscape. The results of this work are of practical significance since the method can aid anyone who plans a hovering flight mode.
Collapse
Affiliation(s)
- Oksana Stalnov
- Faculty of Aerospace Engineering, Technion - Israel Institute of Technology, Haifa, 32000, Israel
| | - Michael Faran
- Department of Bio-Medical Engineering, Faculty of Engineering, Tel-Aviv University, Tel-Aviv 69978, Israel
| | - Yonatan Koral
- School of Electrical Engineering, Faculty of Engineering, Tel-Aviv University, Tel-Aviv 69978, Israel
| | - Miriam Furst
- School of Electrical Engineering, Faculty of Engineering, Tel-Aviv University, Tel-Aviv 69978, Israel
| |
Collapse
|
17
|
Ripley S, Xia L, Zhang Z, Aiken SJ, Wang J. Animal-to-Human Translation Difficulties and Problems With Proposed Coding-in-Noise Deficits in Noise-Induced Synaptopathy and Hidden Hearing Loss. Front Neurosci 2022; 16:893542. [PMID: 35720689 PMCID: PMC9199355 DOI: 10.3389/fnins.2022.893542] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2022] [Accepted: 04/22/2022] [Indexed: 12/26/2022] Open
Abstract
Noise induced synaptopathy (NIS) and hidden hearing loss (NIHHL) have been hot topic in hearing research since a massive synaptic loss was identified in CBA mice after a brief noise exposure that did not cause permanent threshold shift (PTS) in 2009. Based upon the amount of synaptic loss and the bias of it to synapses with a group of auditory nerve fibers (ANFs) with low spontaneous rate (LSR), coding-in-noise deficit (CIND) has been speculated as the major difficult of hearing in subjects with NIS and NIHHL. This speculation is based upon the idea that the coding of sound at high level against background noise relies mainly on the LSR ANFs. However, the translation from animal data to humans for NIS remains to be justified due to the difference in noise exposure between laboratory animals and human subjects in real life, the lack of morphological data and reliable functional methods to quantify or estimate the loss of the afferent synapses by noise. Moreover, there is no clear, robust data revealing the CIND even in animals with the synaptic loss but no PTS. In humans, both positive and negative reports are available. The difficulty in verifying CINDs has led a re-examination of the hypothesis that CIND is the major deficit associated with NIS and NIHHL, and the theoretical basis of this idea on the role of LSR ANFs. This review summarized the current status of research in NIS and NIHHL, with focus on the translational difficulty from animal data to human clinicals, the technical difficulties in quantifying NIS in humans, and the problems with the SR theory on signal coding. Temporal fluctuation profile model was discussed as a potential alternative for signal coding at high sound level against background noise, in association with the mechanisms of efferent control on the cochlea gain.
Collapse
Affiliation(s)
- Sara Ripley
- School of Communication Sciences and Disorders, Dalhousie University, Halifax, NS, Canada
| | - Li Xia
- Department of Otolaryngology-Head and Neck Surgery, Mianyang Central Hospital, School of Medicine, University of Electronic Science and Technology of China, Mianyang, China
| | - Zhen Zhang
- Department of Otolaryngology-Head and Neck Surgery, Shanghai Jiao Tong University Affiliated Sixth People’s Hospital, Shanghai, China
- Otolaryngology Institute of Shanghai Jiao Tong University, Shanghai, China
| | - Steve J. Aiken
- School of Communication Sciences and Disorders, Dalhousie University, Halifax, NS, Canada
| | - Jian Wang
- School of Communication Sciences and Disorders, Dalhousie University, Halifax, NS, Canada
- Department of Otolaryngology-Head and Neck Surgery, Mianyang Central Hospital, School of Medicine, University of Electronic Science and Technology of China, Mianyang, China
| |
Collapse
|
18
|
Guest DR, Oxenham AJ. Human discrimination and modeling of high-frequency complex tones shed light on the neural codes for pitch. PLoS Comput Biol 2022; 18:e1009889. [PMID: 35239639 PMCID: PMC8923464 DOI: 10.1371/journal.pcbi.1009889] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2021] [Revised: 03/15/2022] [Accepted: 02/02/2022] [Indexed: 11/24/2022] Open
Abstract
Accurate pitch perception of harmonic complex tones is widely believed to rely on temporal fine structure information conveyed by the precise phase-locked responses of auditory-nerve fibers. However, accurate pitch perception remains possible even when spectrally resolved harmonics are presented at frequencies beyond the putative limits of neural phase locking, and it is unclear whether residual temporal information, or a coarser rate-place code, underlies this ability. We addressed this question by measuring human pitch discrimination at low and high frequencies for harmonic complex tones, presented either in isolation or in the presence of concurrent complex-tone maskers. We found that concurrent complex-tone maskers impaired performance at both low and high frequencies, although the impairment introduced by adding maskers at high frequencies relative to low frequencies differed between the tested masker types. We then combined simulated auditory-nerve responses to our stimuli with ideal-observer analysis to quantify the extent to which performance was limited by peripheral factors. We found that the worsening of both frequency discrimination and F0 discrimination at high frequencies could be well accounted for (in relative terms) by optimal decoding of all available information at the level of the auditory nerve. A Python package is provided to reproduce these results, and to simulate responses to acoustic stimuli from the three previously published models of the human auditory nerve used in our analyses.
Collapse
Affiliation(s)
- Daniel R. Guest
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota, United States of America
| | - Andrew J. Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota, United States of America
| |
Collapse
|
19
|
Moore BCJ. Listening to Music Through Hearing Aids: Potential Lessons for Cochlear Implants. Trends Hear 2022; 26:23312165211072969. [PMID: 35179052 PMCID: PMC8859663 DOI: 10.1177/23312165211072969] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022] Open
Abstract
Some of the problems experienced by users of hearing aids (HAs) when listening to music are relevant to cochlear implants (CIs). One problem is related to the high peak levels (up to 120 dB SPL) that occur in live music. Some HAs and CIs overload at such levels, because of the limited dynamic range of the microphones and analogue-to-digital converters (ADCs), leading to perceived distortion. Potential solutions are to use 24-bit ADCs or to include an adjustable gain between the microphones and the ADCs. A related problem is how to squeeze the wide dynamic range of music into the limited dynamic range of the user, which can be only 6-20 dB for CI users. In HAs, this is usually done via multi-channel amplitude compression (automatic gain control, AGC). In CIs, a single-channel front-end AGC is applied to the broadband input signal or a control signal derived from a running average of the broadband signal level is used to control the mapping of the channel envelope magnitude to an electrical signal. This introduces several problems: (1) an intense narrowband signal (e.g. a strong bass sound) reduces the level for all frequency components, making some parts of the music harder to hear; (2) the AGC introduces cross-modulation effects that can make a steady sound (e.g. sustained strings or a sung note) appear to fluctuate in level. Potential solutions are to use several frequency channels to create slowly varying gain-control signals and to use slow-acting (or dual time-constant) AGC rather than fast-acting AGC.
Collapse
Affiliation(s)
- Brian C J Moore
- Cambridge Hearing Group, Department of Psychology, 2152University of Cambridge, Cambridge, England
| |
Collapse
|
20
|
Deep neural network models of sound localization reveal how perception is adapted to real-world environments. Nat Hum Behav 2022; 6:111-133. [PMID: 35087192 PMCID: PMC8830739 DOI: 10.1038/s41562-021-01244-z] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2020] [Accepted: 10/29/2021] [Indexed: 11/15/2022]
Abstract
Mammals localize sounds using information from their two ears.
Localization in real-world conditions is challenging, as echoes provide
erroneous information, and noises mask parts of target sounds. To better
understand real-world localization we equipped a deep neural network with human
ears and trained it to localize sounds in a virtual environment. The resulting
model localized accurately in realistic conditions with noise and reverberation.
In simulated experiments, the model exhibited many features of human spatial
hearing: sensitivity to monaural spectral cues and interaural time and level
differences, integration across frequency, biases for sound onsets, and limits
on localization of concurrent sources. But when trained in unnatural
environments without either reverberation, noise, or natural sounds, these
performance characteristics deviated from those of humans. The results show how
biological hearing is adapted to the challenges of real-world environments and
illustrate how artificial neural networks can reveal the real-world constraints
that shape perception.
Collapse
|
21
|
Saddler MR, Gonzalez R, McDermott JH. Deep neural network models reveal interplay of peripheral coding and stimulus statistics in pitch perception. Nat Commun 2021; 12:7278. [PMID: 34907158 PMCID: PMC8671597 DOI: 10.1038/s41467-021-27366-6] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2020] [Accepted: 11/12/2021] [Indexed: 11/15/2022] Open
Abstract
Perception is thought to be shaped by the environments for which organisms are optimized. These influences are difficult to test in biological organisms but may be revealed by machine perceptual systems optimized under different conditions. We investigated environmental and physiological influences on pitch perception, whose properties are commonly linked to peripheral neural coding limits. We first trained artificial neural networks to estimate fundamental frequency from biologically faithful cochlear representations of natural sounds. The best-performing networks replicated many characteristics of human pitch judgments. To probe the origins of these characteristics, we then optimized networks given altered cochleae or sound statistics. Human-like behavior emerged only when cochleae had high temporal fidelity and when models were optimized for naturalistic sounds. The results suggest pitch perception is critically shaped by the constraints of natural environments in addition to those of the cochlea, illustrating the use of artificial neural networks to reveal underpinnings of behavior.
Collapse
Affiliation(s)
- Mark R Saddler
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA.
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA.
- Center for Brains, Minds and Machines, MIT, Cambridge, MA, USA.
| | - Ray Gonzalez
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Center for Brains, Minds and Machines, MIT, Cambridge, MA, USA
| | - Josh H McDermott
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA.
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA.
- Center for Brains, Minds and Machines, MIT, Cambridge, MA, USA.
- Program in Speech and Hearing Biosciences and Technology, Harvard University, Cambridge, MA, USA.
| |
Collapse
|
22
|
Gockel HE, Carlyon RP. On musical interval perception for complex tones at very high frequencies. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:2644. [PMID: 33940917 PMCID: PMC7612123 DOI: 10.1121/10.0004222] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/16/2020] [Accepted: 03/17/2021] [Indexed: 06/12/2023]
Abstract
Listeners appear able to extract a residue pitch from high-frequency harmonics for which phase locking to the temporal fine structure is weak or absent. The present study investigated musical interval perception for high-frequency harmonic complex tones using the same stimuli as Lau, Mehta, and Oxenham [J. Neurosci. 37, 9013-9021 (2017)]. Nine young musically trained listeners with especially good high-frequency hearing adjusted various musical intervals using harmonic complex tones containing harmonics 6-10. The reference notes had fundamental frequencies (F0s) of 280 or 1400 Hz. Interval matches were possible, albeit markedly worse, even when all harmonic frequencies were above the presumed limit of phase locking. Matches showed significantly larger systematic errors and higher variability, and subjects required more trials to finish a match for the high than for the low F0. Additional absolute pitch judgments from one subject with absolute pitch, for complex tones containing harmonics 1-5 or 6-10 with a wide range of F0s, were perfect when the lowest frequency component was below about 7 kHz, but at least 50% of responses were incorrect when it was 8 kHz or higher. The results are discussed in terms of the possible effects of phase-locking information and familiarity with high-frequency stimuli on pitch.
Collapse
|
23
|
de Cheveigné A. Harmonic Cancellation-A Fundamental of Auditory Scene Analysis. Trends Hear 2021; 25:23312165211041422. [PMID: 34698574 PMCID: PMC8552394 DOI: 10.1177/23312165211041422] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2021] [Revised: 07/23/2021] [Accepted: 07/09/2021] [Indexed: 11/16/2022] Open
Abstract
This paper reviews the hypothesis of harmonic cancellation according to which an interfering sound is suppressed or canceled on the basis of its harmonicity (or periodicity in the time domain) for the purpose of Auditory Scene Analysis. It defines the concept, discusses theoretical arguments in its favor, and reviews experimental results that support it, or not. If correct, the hypothesis may draw on time-domain processing of temporally accurate neural representations within the brainstem, as required also by the classic equalization-cancellation model of binaural unmasking. The hypothesis predicts that a target sound corrupted by interference will be easier to hear if the interference is harmonic than inharmonic, all else being equal. This prediction is borne out in a number of behavioral studies, but not all. The paper reviews those results, with the aim to understand the inconsistencies and come up with a reliable conclusion for, or against, the hypothesis of harmonic cancellation within the auditory system.
Collapse
Affiliation(s)
- Alain de Cheveigné
- Laboratoire des systèmes perceptifs, CNRS, Paris, France
- Département d’études cognitives, École normale supérieure, PSL
University, Paris, France
- UCL Ear Institute, London, UK
| |
Collapse
|
24
|
Bissmeyer SRS, Hossain S, Goldsworthy RL. Perceptual learning of pitch provided by cochlear implant stimulation rate. PLoS One 2020; 15:e0242842. [PMID: 33270735 PMCID: PMC7714175 DOI: 10.1371/journal.pone.0242842] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2019] [Accepted: 11/10/2020] [Indexed: 11/19/2022] Open
Abstract
Cochlear implant users hear pitch evoked by stimulation rate, but discrimination diminishes for rates above 300 Hz. This upper limit on rate pitch is surprising given the remarkable and specialized ability of the auditory nerve to respond synchronously to stimulation rates at least as high as 3 kHz and arguably as high as 10 kHz. Sensitivity to stimulation rate as a pitch cue varies widely across cochlear implant users and can be improved with training. The present study examines individual differences and perceptual learning of stimulation rate as a cue for pitch ranking. Adult cochlear implant users participated in electrode psychophysics that involved testing once per week for three weeks. Stimulation pulse rate discrimination was measured in bipolar and monopolar configurations for apical and basal electrodes. Base stimulation rates between 100 and 800 Hz were examined. Individual differences were quantified using psychophysically derived metrics of spatial tuning and temporal integration. This study examined distribution of measures across subjects, predictive power of psychophysically derived metrics of spatial tuning and temporal integration, and the effect of training on rate discrimination thresholds. Psychophysical metrics of spatial tuning and temporal integration were not predictive of stimulation rate discrimination, but discrimination thresholds improved at lower frequencies with training. Since most clinical devices do not use variable stimulation rates, it is unknown to what extent recipients may learn to use stimulation rate cues if provided in a clear and consistent manner.
Collapse
Affiliation(s)
- Susan R. S. Bissmeyer
- Department of Biomedical Engineering, Viterbi School of Engineering, University of Southern California, Los Angeles, California, United States of America
- Auditory Research Center, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, Los Angeles, California, United States of America
| | - Shaikat Hossain
- Auditory Research Center, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, Los Angeles, California, United States of America
| | - Raymond L. Goldsworthy
- Department of Biomedical Engineering, Viterbi School of Engineering, University of Southern California, Los Angeles, California, United States of America
- Auditory Research Center, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, Los Angeles, California, United States of America
| |
Collapse
|
25
|
Whiteford KL, Kreft HA, Oxenham AJ. The role of cochlear place coding in the perception of frequency modulation. eLife 2020; 9:58468. [PMID: 32996463 PMCID: PMC7556860 DOI: 10.7554/elife.58468] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2020] [Accepted: 09/29/2020] [Indexed: 12/17/2022] Open
Abstract
Natural sounds convey information via frequency and amplitude modulations (FM and AM). Humans are acutely sensitive to the slow rates of FM that are crucial for speech and music. This sensitivity has long been thought to rely on precise stimulus-driven auditory-nerve spike timing (time code), whereas a coarser code, based on variations in the cochlear place of stimulation (place code), represents faster FM rates. We tested this theory in listeners with normal and impaired hearing, spanning a wide range of place-coding fidelity. Contrary to predictions, sensitivity to both slow and fast FM correlated with place-coding fidelity. We also used incoherent AM on two carriers to simulate place coding of FM and observed poorer sensitivity at high carrier frequencies and fast rates, two properties of FM detection previously ascribed to the limits of time coding. The results suggest a unitary place-based neural code for FM across all rates and carrier frequencies.
Collapse
Affiliation(s)
- Kelly L Whiteford
- Department of Psychology, University of Minnesota, Minneapolis, United States
| | - Heather A Kreft
- Department of Psychology, University of Minnesota, Minneapolis, United States
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, United States
| |
Collapse
|
26
|
Mehta AH, Oxenham AJ. Effect of lowest harmonic rank on fundamental-frequency difference limens varies with fundamental frequency. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:2314. [PMID: 32359332 PMCID: PMC7166120 DOI: 10.1121/10.0001092] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/19/2019] [Revised: 03/25/2020] [Accepted: 03/27/2020] [Indexed: 06/11/2023]
Abstract
This study investigated the relationship between fundamental frequency difference limens (F0DLs) and the lowest harmonic number present over a wide range of F0s (30-2000 Hz) for 12-component harmonic complex tones that were presented in either sine or random phase. For fundamental frequencies (F0s) between 100 and 400 Hz, a transition from low (∼1%) to high (∼5%) F0DLs occurred as the lowest harmonic number increased from about seven to ten, in line with earlier studies. At lower and higher F0s, the transition between low and high F0DLs occurred at lower harmonic numbers. The worsening performance at low F0s was reasonably well predicted by the expected decrease in spectral resolution below about 500 Hz. At higher F0s, the degradation in performance at lower harmonic numbers could not be predicted by changes in spectral resolution but remained relatively good (<2%-3%) in some conditions, even when all harmonics were above 8 kHz, confirming that F0 can be extracted from harmonics even when temporal envelope or fine-structure cues are weak or absent.
Collapse
Affiliation(s)
- Anahita H Mehta
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| |
Collapse
|
27
|
Ewert SD, Paraouty N, Lorenzi C. A two‐path model of auditory modulation detection using temporal fine structure and envelope cues. Eur J Neurosci 2020; 51:1265-1278. [DOI: 10.1111/ejn.13846] [Citation(s) in RCA: 20] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2017] [Revised: 01/18/2018] [Accepted: 01/18/2018] [Indexed: 11/30/2022]
Affiliation(s)
- Stephan D. Ewert
- Medizinische Physik and Cluster of Excellence Hearing4All Universität Oldenburg 26111 Oldenburg Germany
| | - Nihaad Paraouty
- Laboratoire des systèmes perceptifs Département d’études cognitives, École normale supérieure CNRS PSL Research University Paris France
| | - Christian Lorenzi
- Laboratoire des systèmes perceptifs Département d’études cognitives, École normale supérieure CNRS PSL Research University Paris France
| |
Collapse
|
28
|
Henry KS, Amburgey KN, Abrams KS, Carney LH. Identifying cues for tone-in-noise detection using decision variable correlation in the budgerigar (Melopsittacus undulatus). THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:984. [PMID: 32113293 PMCID: PMC7010520 DOI: 10.1121/10.0000621] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/06/2019] [Revised: 01/02/2020] [Accepted: 01/07/2020] [Indexed: 06/05/2023]
Abstract
Previous studies evaluated cues for masked tone detection using reproducible noise waveforms. Human results founded on this approach suggest that tone detection is based on combined energy and envelope (ENV) cues, but detection cues in nonhuman species are less clear. Decision variable correlation (DVC) was used to evaluate tone-in-noise detection cues in the budgerigar, an avian species with human-like behavioral sensitivity to many complex sounds. DVC quantifies a model's ability to predict trial-by-trial variance in behavioral responses. Budgerigars were behaviorally conditioned to detect 500-Hz tones in wideband (WB; 100-3000 Hz) and narrowband (NB; 452-552 Hz) noise. Behavioral responses were obtained using a single-interval, two-alternative discrimination task and two-down, one-up adaptive tracking procedures. Tone-detection thresholds in WB noise were higher than human thresholds, putatively due to broader peripheral frequency tuning, whereas NB thresholds were within ∼1 dB of human results. Budgerigar average hit and false-alarm rates across noise waveforms were consistent, highly correlated across subjects, and correlated to human results. Trial-by-trial behavioral results in NB noise were best explained by a model combining energy and ENV cues. In contrast, WB results were better predicted by ENV-based or multiple-channel energy detector models. These results suggest that budgerigars and humans use similar cues for tone-in-noise detection.
Collapse
Affiliation(s)
- Kenneth S Henry
- Department of Otolaryngology, University of Rochester, Rochester, New York 14642, USA
| | - Kassidy N Amburgey
- Department of Otolaryngology, University of Rochester, Rochester, New York 14642, USA
| | - Kristina S Abrams
- Department of Neuroscience, University of Rochester, Rochester, New York 14642, USA
| | - Laurel H Carney
- Department of Biomedical Engineering, University of Rochester, Rochester, New York 14642, USA
| |
Collapse
|
29
|
Carcagno S, Lakhani S, Plack CJ. Consonance perception beyond the traditional existence region of pitch. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:2279. [PMID: 31671967 DOI: 10.1121/1.5127845] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/20/2019] [Accepted: 09/12/2019] [Indexed: 06/10/2023]
Abstract
Some theories posit that the perception of consonance is based on neural periodicity detection, which is dependent on accurate phase locking of auditory nerve fibers to features of the stimulus waveform. In the current study, 15 listeners were asked to rate the pleasantness of complex tone dyads (2 note chords) forming various harmonic intervals and bandpass filtered in a high-frequency region (all components >5.8 kHz), where phase locking to the rapid stimulus fine structure is thought to be severely degraded or absent. The two notes were presented to opposite ears. Consonant intervals (minor third and perfect fifth) received higher ratings than dissonant intervals (minor second and tritone). The results could not be explained in terms of phase locking to the slower waveform envelope because the preference for consonant intervals was higher when the stimuli were harmonic, compared to a condition in which they were made inharmonic by shifting their component frequencies by a constant offset, so as to preserve their envelope periodicity. Overall the results indicate that, if phase locking is indeed absent at frequencies greater than ∼5 kHz, neural periodicity detection is not necessary for the perception of consonance.
Collapse
Affiliation(s)
- Samuele Carcagno
- Department of Psychology, Lancaster University, Lancaster, LA1 4YF, United Kingdom
| | - Saday Lakhani
- Department of Psychology, Lancaster University, Lancaster, LA1 4YF, United Kingdom
| | - Christopher J Plack
- Department of Psychology, Lancaster University, Lancaster, LA1 4YF, United Kingdom
| |
Collapse
|
30
|
Divergent Auditory Nerve Encoding Deficits Between Two Common Etiologies of Sensorineural Hearing Loss. J Neurosci 2019; 39:6879-6887. [PMID: 31285299 DOI: 10.1523/jneurosci.0038-19.2019] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2019] [Revised: 05/08/2019] [Accepted: 06/09/2019] [Indexed: 11/21/2022] Open
Abstract
Speech intelligibility can vary dramatically between individuals with similar clinically defined severity of hearing loss based on the audiogram. These perceptual differences, despite equal audiometric-threshold elevation, are often assumed to reflect central-processing variations. Here, we compared peripheral-processing in auditory nerve (AN) fibers of male chinchillas between two prevalent hearing loss etiologies: metabolic hearing loss (MHL) and noise-induced hearing loss (NIHL). MHL results from age-related reduction of the endocochlear potential due to atrophy of the stria vascularis. MHL in the present study was induced using furosemide, which provides a validated model of age-related MHL in young animals by reversibly inhibiting the endocochlear potential. Effects of MHL on peripheral processing were assessed using Wiener-kernel (system identification) analyses of single AN fiber responses to broadband noise, for direct comparison to previously published AN responses from animals with NIHL. Wiener-kernel analyses show that even mild NIHL causes grossly abnormal coding of low-frequency stimulus components. In contrast, for MHL the same abnormal coding was only observed with moderate to severe loss. For equal sensitivity loss, coding impairment was substantially less severe with MHL than with NIHL, probably due to greater preservation of the tip-to-tail ratio of cochlear frequency tuning with MHL compared with NIHL rather than different intrinsic AN properties. Differences in peripheral neural coding between these two pathologies-the more severe of which, NIHL, is preventable-likely contribute to individual speech perception differences. Our results underscore the need to minimize noise overexposure and for strategies to personalize diagnosis and treatment for individuals with sensorineural hearing loss.SIGNIFICANCE STATEMENT Differences in speech perception ability between individuals with similar clinically defined severity of hearing loss are often assumed to reflect central neural-processing differences. Here, we demonstrate for the first time that peripheral neural processing of complex sounds differs dramatically between the two most common etiologies of hearing loss. Greater processing impairment with noise-induced compared with an age-related (metabolic) hearing loss etiology may explain heightened speech perception difficulties in people overexposed to loud environments. These results highlight the need for public policies to prevent noise-induced hearing loss, an entirely avoidable hearing loss etiology, and for personalized strategies to diagnose and treat sensorineural hearing loss.
Collapse
|
31
|
Mino H. Modeling of spike trains in auditory nerves with self-exciting point processes of the von Mises type. BIOLOGICAL CYBERNETICS 2019; 113:347-356. [PMID: 31004189 DOI: 10.1007/s00422-019-00799-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/25/2018] [Accepted: 04/08/2019] [Indexed: 06/09/2023]
Abstract
This article presents the modeling of spike trains in auditory nerve fiber (ANF) models with a one-memory self-exciting point process (SEPP) of the von Mises type. The ANF models were acoustically stimulated by a synaptic current of inner hair cells, or electrically stimulated by sinusoidally amplitude-modulated pulsatile waveforms. It has been shown that the parameters of one-memory SEPP of the von Mises type could be estimated by numerically maximizing the likelihood function from sample realizations of the spike trains in response to acoustic or electric stimulus. Furthermore, it was found that period histograms of the one-memory SEPP generated artificially on the basis of the estimated von Mises parameters agreed well with those of acoustic or electric stimulus, by performing the uniform-scores test. It implies that the waveforms of pulsatile electric stimuli should be selected such that the spike trains can be represented by one-memory SEPP of the von Mises type with appropriate parameters, efficiently carrying information to the cochlear implant user's brain, like that in acoustic stimulation of the healthy ear. The findings presented in this paper may play an important role in determining optimal parameters of pulsatile electric stimuli by using one-memory SEPP of the von Mises type, and further in the design of better cochlear prostheses.
Collapse
Affiliation(s)
- Hiroyuki Mino
- Department of Electrical Engineering, Kanto Gakuin University, 1-50-1 Mutsuura E., Kanazawa-ku, Yokohama, 236-8501, Japan.
| |
Collapse
|
32
|
The upper frequency limit for the use of phase locking to code temporal fine structure in humans: A compilation of viewpoints. Hear Res 2019; 377:109-121. [PMID: 30927686 PMCID: PMC6524635 DOI: 10.1016/j.heares.2019.03.011] [Citation(s) in RCA: 56] [Impact Index Per Article: 11.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/30/2018] [Revised: 02/09/2019] [Accepted: 03/13/2019] [Indexed: 11/27/2022]
Abstract
The relative importance of neural temporal and place coding in auditory perception is still a matter of much debate. The current article is a compilation of viewpoints from leading auditory psychophysicists and physiologists regarding the upper frequency limit for the use of neural phase locking to code temporal fine structure in humans. While phase locking is used for binaural processing up to about 1500 Hz, there is disagreement regarding the use of monaural phase-locking information at higher frequencies. Estimates of the general upper limit proposed by the contributors range from 1500 to 10000 Hz. The arguments depend on whether or not phase locking is needed to explain psychophysical discrimination performance at frequencies above 1500 Hz, and whether or not the phase-locked neural representation is sufficiently robust at these frequencies to provide useable information. The contributors suggest key experiments that may help to resolve this issue, and experimental findings that may cause them to change their minds. This issue is of crucial importance to our understanding of the neural basis of auditory perception in general, and of pitch perception in particular. Phase locking is used in binaural processing for frequencies up to ∼1500 Hz. Estimates of the general upper limit (inc. monaural processing) vary from 1500 to 10000 Hz. Direct recordings from human auditory nerve would determine peripheral limitation. Understanding of the central processing of temporal and place cues is needed to establish an upper limit.
Collapse
|
33
|
Verschooten E, Desloovere C, Joris PX. High-resolution frequency tuning but not temporal coding in the human cochlea. PLoS Biol 2018; 16:e2005164. [PMID: 30321166 PMCID: PMC6201958 DOI: 10.1371/journal.pbio.2005164] [Citation(s) in RCA: 33] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2017] [Revised: 10/25/2018] [Accepted: 09/25/2018] [Indexed: 11/24/2022] Open
Abstract
Frequency tuning and phase-locking are two fundamental properties generated in the cochlea, enabling but also limiting the coding of sounds by the auditory nerve (AN). In humans, these limits are unknown, but high resolution has been postulated for both properties. Electrophysiological recordings from the AN of normal-hearing volunteers indicate that human frequency tuning, but not phase-locking, exceeds the resolution observed in animal models. The coding of sounds by the cochlea depends on two primary properties: frequency selectivity, which refers to the ability to separate sounds into their different frequency components, and phase-locking, which refers to the neural coding of the temporal waveform of these components. These properties have been well characterized in animals using neurophysiological recordings from single neurons of the auditory nerve (AN), but this approach is not feasible in humans. As a result, there is considerable controversy as to how these two properties may differ between humans and the small animals typically used in neurophysiological studies. It has been proposed that humans excel both in frequency selectivity and in the range of frequencies over which they have phase-locking. We developed a technique to quantify these properties using mass potentials from the AN, recorded via the middle ear in human volunteers with normal hearing. We find that humans have unusually sharp frequency tuning but that the upper frequency limit of phase-locking is at best similar to—and more likely lower than—that of the nonhuman animals conventionally used in experiments.
Collapse
Affiliation(s)
- Eric Verschooten
- Laboratory of Auditory Neurophysiology, KU Leuven, Leuven, Belgium
| | - Christian Desloovere
- Department of Otorhinolaryngology, Head and Neck Surgery, KU Leuven, Leuven, Belgium
| | - Philip X. Joris
- Laboratory of Auditory Neurophysiology, KU Leuven, Leuven, Belgium
- * E-mail:
| |
Collapse
|
34
|
Horst JW, McGee J, Walsh EJ. Input-output curves of low and high spontaneous rate auditory nerve fibers are exponential near threshold. Hear Res 2018; 367:195-206. [DOI: 10.1016/j.heares.2018.06.007] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/30/2017] [Revised: 06/12/2018] [Accepted: 06/12/2018] [Indexed: 10/28/2022]
|
35
|
Carney LH. Supra-Threshold Hearing and Fluctuation Profiles: Implications for Sensorineural and Hidden Hearing Loss. J Assoc Res Otolaryngol 2018; 19:331-352. [PMID: 29744729 PMCID: PMC6081887 DOI: 10.1007/s10162-018-0669-5] [Citation(s) in RCA: 93] [Impact Index Per Article: 15.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2017] [Accepted: 04/19/2018] [Indexed: 12/22/2022] Open
Abstract
An important topic in contemporary auditory science is supra-threshold hearing. Difficulty hearing at conversational speech levels in background noise has long been recognized as a problem of sensorineural hearing loss, including that associated with aging (presbyacusis). Such difficulty in listeners with normal thresholds has received more attention recently, especially associated with descriptions of synaptopathy, the loss of auditory nerve (AN) fibers as a result of noise exposure or aging. Synaptopathy has been reported to cause a disproportionate loss of low- and medium-spontaneous rate (L/MSR) AN fibers. Several studies of synaptopathy have assumed that the wide dynamic ranges of L/MSR AN fiber rates are critical for coding supra-threshold sounds. First, this review will present data from the literature that argues against a direct role for average discharge rates of L/MSR AN fibers in coding sounds at moderate to high sound levels. Second, the encoding of sounds at supra-threshold levels is examined. A key assumption in many studies is that saturation of AN fiber discharge rates limits neural encoding, even though the majority of AN fibers, high-spontaneous rate (HSR) fibers, have saturated average rates at conversational sound levels. It is argued here that the cross-frequency profile of low-frequency neural fluctuation amplitudes, not average rates, encodes complex sounds. As described below, this fluctuation-profile coding mechanism benefits from both saturation of inner hair cell (IHC) transduction and average rate saturation associated with the IHC-AN synapse. Third, the role of the auditory efferent system, which receives inputs from L/MSR fibers, is revisited in the context of fluctuation-profile coding. The auditory efferent system is hypothesized to maintain and enhance neural fluctuation profiles. Lastly, central mechanisms sensitive to neural fluctuations are reviewed. Low-frequency fluctuations in AN responses are accentuated by cochlear nucleus neurons which, either directly or via other brainstem nuclei, relay fluctuation profiles to the inferior colliculus (IC). IC neurons are sensitive to the frequency and amplitude of low-frequency fluctuations and convert fluctuation profiles from the periphery into a phase-locked rate profile that is robust across a wide range of sound levels and in background noise. The descending projection from the midbrain (IC) to the efferent system completes a functional loop that, combined with inputs from the L/MSR pathway, is hypothesized to maintain "sharp" supra-threshold hearing, reminiscent of visual mechanisms that regulate optical accommodation. Examples from speech coding and detection in noise are reviewed. Implications for the effects of synaptopathy on control mechanisms hypothesized to influence supra-threshold hearing are discussed. This framework for understanding neural coding and control mechanisms for supra-threshold hearing suggests strategies for the design of novel hearing aid signal-processing and electrical stimulation patterns for cochlear implants.
Collapse
Affiliation(s)
- Laurel H Carney
- Departments of Biomedical Engineering, Neuroscience, and Electrical & Computer Engineering, Del Monte Institute for Neuroscience, University of Rochester, 601 Elmwood Ave., Box 603, Rochester, NY, 14642, USA.
| |
Collapse
|
36
|
The Interplay Between Spike-Time and Spike-Rate Modes in the Auditory Nerve Encodes Tone-In-Noise Threshold. J Neurosci 2018; 38:5727-5738. [PMID: 29793977 DOI: 10.1523/jneurosci.3103-17.2018] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2017] [Revised: 05/04/2018] [Accepted: 05/10/2018] [Indexed: 11/21/2022] Open
Abstract
Auditory nerve fibers (ANFs) encode pure tones through two modes of coding, spike time and spike rate, depending on the tone frequency. In response to a low-frequency tone, ANF firing is phase locked to the sinusoidal waveform. Because time coding vanishes with an increase in the tone frequency, high-frequency tone coding relies on the spike rate of the ANFs. Adding a continuous broadband noise to a tone compresses the rate intensity function of ANFs and shifts its dynamic range toward higher intensities. Therefore, the ANFs with high-threshold/low-spontaneous rate (SR) are thought to contribute to behavioral tone detection in noise. However, this theory relies on the discharge rate of the ANFs. The direct comparison with the masking threshold through spike timing, irrespective of the spontaneous rate, has not so far been investigated. Taking advantage of a unique proxy to quantify the spike synchrony (i.e., the shuffle autocorrelogram), we show in female gerbils that high-SR ANFs are more adapted to encode low-frequency thresholds through temporal code, giving them a strong robustness in noise. By comparing behavioral thresholds measured using prepulse inhibition of the acoustical startle reflex with population thresholds calculated from ANFs pooled per octave band, we show that threshold-based spike timing provides a better estimate of behavioral thresholds in the low-frequency range, whereas the high-frequency behavioral thresholds rely on the spiking rate, particularly in noise. This emphasizes the complementarity of temporal and rate modes to code tone-in-noise thresholds over a large range of frequencies.SIGNIFICANCE STATEMENT There is a general agreement that high-threshold/low-spontaneous rate (SR) auditory nerve fibers (ANFs) are of prime importance for tone detection in noise. However, this theory is based on the discharge rate of the fibers. Comparing the behavioral thresholds and single ANF thresholds shows that this is only true in the high-frequency range of tone stimulations. In the low-frequency range of tones (up to 2.7 kHz in the gerbil), the most sensitive ANFs (high-SR fibers) carry neural information through a spike-timing mode, even for noise in which tones do not induce a noticeable increment in the spike rate. This emphasizes the interplay between spike-time and spike-rate modes in the auditory nerve to encode tone-in-noise threshold over a large range of tone frequencies.
Collapse
|
37
|
Arzounian D, de Kerangal M, de Cheveigné A. Sequential dependencies in pitch judgments. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:3047. [PMID: 29195443 DOI: 10.1121/1.5009938] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Studies that measure pitch discrimination relate a subject's response on each trial to the stimuli presented on that trial, but there is evidence that behavior depends also on earlier stimulation. Here, listeners heard a sequence of tones and reported after each tone whether it was higher or lower in pitch than the previous tone. Frequencies were determined by an adaptive staircase targeting 75% correct, with interleaved tracks to ensure independence between consecutive frequency changes. Responses for this specific task were predicted by a model that took into account the frequency interval on the current trial, as well as the interval and response on the previous trial. This model was superior to simpler models. The dependence on the previous interval was positive (assimilative) for all subjects, consistent with persistence of the sensory trace. The dependence on the previous response was either positive or negative, depending on the subject, consistent with a subject-specific suboptimal response strategy. It is argued that a full stimulus + response model is necessary to account for effects of stimulus history and obtain an accurate estimate of sensory noise.
Collapse
Affiliation(s)
- Dorothée Arzounian
- Laboratoire des Systèmes Perceptifs, Département d'Etudes Cognitives, Ecole normale supérieure, PSL Research University, CNRS, 29 rue d'Ulm, Paris, 75005, France
| | - Mathilde de Kerangal
- The Ear Institute, University College London, 332 Grays Inn Road, Kings Cross, London, WC1X 8EE, United Kingdom
| | - Alain de Cheveigné
- Laboratoire des Systèmes Perceptifs, Département d'Etudes Cognitives, Ecole normale supérieure, PSL Research University, CNRS, 29 rue d'Ulm, Paris, 75005, France
| |
Collapse
|
38
|
Abstract
Auditory perception is our main gateway to communication with others via speech and music, and it also plays an important role in alerting and orienting us to new events. This review provides an overview of selected topics pertaining to the perception and neural coding of sound, starting with the first stage of filtering in the cochlea and its profound impact on perception. The next topic, pitch, has been debated for millennia, but recent technical and theoretical developments continue to provide us with new insights. Cochlear filtering and pitch both play key roles in our ability to parse the auditory scene, enabling us to attend to one auditory object or stream while ignoring others. An improved understanding of the basic mechanisms of auditory perception will aid us in the quest to tackle the increasingly important problem of hearing loss in our aging population.
Collapse
Affiliation(s)
- Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455;
| |
Collapse
|
39
|
Lau BK, Mehta AH, Oxenham AJ. Superoptimal Perceptual Integration Suggests a Place-Based Representation of Pitch at High Frequencies. J Neurosci 2017; 37:9013-9021. [PMID: 28821642 PMCID: PMC5597982 DOI: 10.1523/jneurosci.1507-17.2017] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2017] [Revised: 07/29/2017] [Accepted: 08/05/2017] [Indexed: 11/21/2022] Open
Abstract
Pitch, the perceptual correlate of sound repetition rate or frequency, plays an important role in speech perception, music perception, and listening in complex acoustic environments. Despite the perceptual importance of pitch, the neural mechanisms that underlie it remain poorly understood. Although cortical regions responsive to pitch have been identified, little is known about how pitch information is extracted from the inner ear itself. The two primary theories of peripheral pitch coding involve stimulus-driven spike timing, or phase locking, in the auditory nerve (time code), and the spatial distribution of responses along the length of the cochlear partition (place code). To rule out the use of timing information, we tested pitch discrimination of very high-frequency tones (>8 kHz), well beyond the putative limit of phase locking. We found that high-frequency pure-tone discrimination was poor, but when the tones were combined into a harmonic complex, a dramatic improvement in discrimination ability was observed that exceeded performance predicted by the optimal integration of peripheral information from each of the component frequencies. The results are consistent with the existence of pitch-sensitive neurons that rely only on place-based information from multiple harmonically related components. The results also provide evidence against the common assumption that poor high-frequency pure-tone pitch perception is the result of peripheral neural-coding constraints. The finding that place-based spectral coding is sufficient to elicit complex pitch at high frequencies has important implications for the design of future neural prostheses to restore hearing to deaf individuals.SIGNIFICANCE STATEMENT The question of how pitch is represented in the ear has been debated for over a century. Two competing theories involve timing information from neural spikes in the auditory nerve (time code) and the spatial distribution of neural activity along the length of the cochlear partition (place code). By using very high-frequency tones unlikely to be coded via time information, we discovered that information from the individual harmonics is combined so efficiently that performance exceeds theoretical predictions based on the optimal integration of information from each harmonic. The findings have important implications for the design of auditory prostheses because they suggest that enhanced spatial resolution alone may be sufficient to restore pitch via such implants.
Collapse
Affiliation(s)
- Bonnie K Lau
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455
| | - Anahita H Mehta
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455
| |
Collapse
|
40
|
Zuk N, Delgutte B. Neural coding of time-varying interaural time differences and time-varying amplitude in the inferior colliculus. J Neurophysiol 2017; 118:544-563. [PMID: 28381487 DOI: 10.1152/jn.00797.2016] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2016] [Revised: 03/29/2017] [Accepted: 03/31/2017] [Indexed: 11/22/2022] Open
Abstract
Binaural cues occurring in natural environments are frequently time varying, either from the motion of a sound source or through interactions between the cues produced by multiple sources. Yet, a broad understanding of how the auditory system processes dynamic binaural cues is still lacking. In the current study, we directly compared neural responses in the inferior colliculus (IC) of unanesthetized rabbits to broadband noise with time-varying interaural time differences (ITD) with responses to noise with sinusoidal amplitude modulation (SAM) over a wide range of modulation frequencies. On the basis of prior research, we hypothesized that the IC, one of the first stages to exhibit tuning of firing rate to modulation frequency, might use a common mechanism to encode time-varying information in general. Instead, we found weaker temporal coding for dynamic ITD compared with amplitude modulation and stronger effects of adaptation for amplitude modulation. The differences in temporal coding of dynamic ITD compared with SAM at the single-neuron level could be a neural correlate of "binaural sluggishness," the inability to perceive fluctuations in time-varying binaural cues at high modulation frequencies, for which a physiological explanation has so far remained elusive. At ITD-variation frequencies of 64 Hz and above, where a temporal code was less effective, noise with a dynamic ITD could still be distinguished from noise with a constant ITD through differences in average firing rate in many neurons, suggesting a frequency-dependent tradeoff between rate and temporal coding of time-varying binaural information.NEW & NOTEWORTHY Humans use time-varying binaural cues to parse auditory scenes comprising multiple sound sources and reverberation. However, the neural mechanisms for doing so are poorly understood. Our results demonstrate a potential neural correlate for the reduced detectability of fluctuations in time-varying binaural information at high speeds, as occurs in reverberation. The results also suggest that the neural mechanisms for processing time-varying binaural and monaural cues are largely distinct.
Collapse
Affiliation(s)
- Nathaniel Zuk
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, Massachusetts.,Speech and Hearing Bioscience and Technology Program, Harvard-MIT Division of Health Sciences and Technology, Cambridge, Massachusetts; and
| | - Bertrand Delgutte
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, Massachusetts; .,Speech and Hearing Bioscience and Technology Program, Harvard-MIT Division of Health Sciences and Technology, Cambridge, Massachusetts; and.,Department of Otolaryngology, Harvard Medical School, Boston, Massachusetts
| |
Collapse
|
41
|
Bianchi F, Fereczkowski M, Zaar J, Santurette S, Dau T. Complex-Tone Pitch Discrimination in Listeners With Sensorineural Hearing Loss. Trends Hear 2016; 20:20/0/2331216516655793. [PMID: 27604780 PMCID: PMC5017569 DOI: 10.1177/2331216516655793] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
Physiological studies have shown that noise-induced sensorineural hearing loss (SNHL) enhances the amplitude of envelope coding in auditory-nerve fibers. As pitch coding of unresolved complex tones is assumed to rely on temporal envelope coding mechanisms, this study investigated pitch-discrimination performance in listeners with SNHL. Pitch-discrimination thresholds were obtained for 14 normal-hearing (NH) and 10 hearing-impaired (HI) listeners for sine-phase (SP) and random-phase (RP) complex tones. When all harmonics were unresolved, the HI listeners performed, on average, worse than NH listeners in the RP condition but similarly to NH listeners in the SP condition. The increase in pitch-discrimination performance for the SP relative to the RP condition (F0DL ratio) was significantly larger in the HI as compared with the NH listeners. Cochlear compression and auditory-filter bandwidths were estimated in the same listeners. The estimated reduction of cochlear compression was significantly correlated with the increase in the F0DL ratio, while no correlation was found with filter bandwidth. The effects of degraded frequency selectivity and loss of compression were considered in a simplified peripheral model as potential factors in envelope enhancement. The model revealed that reducing cochlear compression significantly enhanced the envelope of an unresolved SP complex tone, while not affecting the envelope of a RP complex tone. This envelope enhancement in the SP condition was significantly correlated with the increased pitch-discrimination performance for the SP relative to the RP condition in the HI listeners.
Collapse
Affiliation(s)
- Federica Bianchi
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Michal Fereczkowski
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Johannes Zaar
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Sébastien Santurette
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Torsten Dau
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Kgs. Lyngby, Denmark
| |
Collapse
|
42
|
Heil P, Peterson AJ. Spike timing in auditory-nerve fibers during spontaneous activity and phase locking. Synapse 2016; 71:5-36. [DOI: 10.1002/syn.21925] [Citation(s) in RCA: 28] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2015] [Revised: 07/20/2016] [Accepted: 07/24/2016] [Indexed: 12/22/2022]
Affiliation(s)
- Peter Heil
- Department of Systems Physiology of Learning; Leibniz Institute for Neurobiology; Magdeburg 39118 Germany
- Center for Behavioral Brain Sciences; Magdeburg Germany
| | - Adam J. Peterson
- Department of Systems Physiology of Learning; Leibniz Institute for Neurobiology; Magdeburg 39118 Germany
| |
Collapse
|
43
|
Paraouty N, Ewert SD, Wallaert N, Lorenzi C. Interactions between amplitude modulation and frequency modulation processing: Effects of age and hearing loss. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 140:121. [PMID: 27475138 DOI: 10.1121/1.4955078] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Frequency modulation (FM) and amplitude modulation (AM) detection thresholds were measured for a 500-Hz carrier frequency and a 5-Hz modulation rate. For AM detection, FM at the same rate as the AM was superimposed with varying FM depth. For FM detection, AM at the same rate was superimposed with varying AM depth. The target stimuli always contained both amplitude and frequency modulations, while the standard stimuli only contained the interfering modulation. Young and older normal-hearing listeners, as well as older listeners with mild-to-moderate sensorineural hearing loss were tested. For all groups, AM and FM detection thresholds were degraded in the presence of the interfering modulation. AM detection with and without interfering FM was hardly affected by either age or hearing loss. While aging had an overall detrimental effect on FM detection with and without interfering AM, there was a trend that hearing loss further impaired FM detection in the presence of AM. Several models using optimal combination of temporal-envelope cues at the outputs of off-frequency filters were tested. The interfering effects could only be predicted for hearing-impaired listeners. This indirectly supports the idea that, in addition to envelope cues resulting from FM-to-AM conversion, normal-hearing listeners use temporal fine-structure cues for FM detection.
Collapse
Affiliation(s)
- Nihaad Paraouty
- Laboratoire des Systèmes Perceptifs (CNRS UMR 8248), Institut d'Etude de la Cognition, Ecole normale supérieure, Paris Sciences et Lettres Research University, 29 rue d'Ulm, 75005 Paris, France
| | - Stephan D Ewert
- Medizinische Physik and Cluster of Excellence Hearing4All, Universität Oldenburg, 26111 Oldenburg, Germany
| | - Nicolas Wallaert
- Laboratoire des Systèmes Perceptifs (CNRS UMR 8248), Institut d'Etude de la Cognition, Ecole normale supérieure, Paris Sciences et Lettres Research University, 29 rue d'Ulm, 75005 Paris, France
| | - Christian Lorenzi
- Laboratoire des Systèmes Perceptifs (CNRS UMR 8248), Institut d'Etude de la Cognition, Ecole normale supérieure, Paris Sciences et Lettres Research University, 29 rue d'Ulm, 75005 Paris, France
| |
Collapse
|
44
|
Distorted Tonotopic Coding of Temporal Envelope and Fine Structure with Noise-Induced Hearing Loss. J Neurosci 2016; 36:2227-37. [PMID: 26888932 DOI: 10.1523/jneurosci.3944-15.2016] [Citation(s) in RCA: 28] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
UNLABELLED People with cochlear hearing loss have substantial difficulty understanding speech in real-world listening environments (e.g., restaurants), even with amplification from a modern digital hearing aid. Unfortunately, a disconnect remains between human perceptual studies implicating diminished sensitivity to fast acoustic temporal fine structure (TFS) and animal studies showing minimal changes in neural coding of TFS or slower envelope (ENV) structure. Here, we used general system-identification (Wiener kernel) analyses of chinchilla auditory nerve fiber responses to Gaussian noise to reveal pronounced distortions in tonotopic coding of TFS and ENV following permanent, noise-induced hearing loss. In basal fibers with characteristic frequencies (CFs) >1.5 kHz, hearing loss introduced robust nontonotopic coding (i.e., at the wrong cochlear place) of low-frequency TFS, while ENV responses typically remained at CF. As a consequence, the highest dominant frequency of TFS coding in response to Gaussian noise was 2.4 kHz in noise-overexposed fibers compared with 4.5 kHz in control fibers. Coding of ENV also became nontonotopic in more pronounced cases of cochlear damage. In apical fibers, more classical hearing-loss effects were observed, i.e., broadened tuning without a significant shift in best frequency. Because these distortions and dissociations of TFS/ENV disrupt tonotopicity, a fundamental principle of auditory processing necessary for robust signal coding in background noise, these results have important implications for understanding communication difficulties faced by people with hearing loss. Further, hearing aids may benefit from distinct amplification strategies for apical and basal cochlear regions to address fundamentally different coding deficits. SIGNIFICANCE STATEMENT Speech-perception problems associated with noise overexposure are pervasive in today's society, even with modern digital hearing aids. Unfortunately, the underlying physiological deficits in neural coding remain unclear. Here, we used innovative system-identification analyses of auditory nerve fiber responses to Gaussian noise to uncover pronounced distortions in coding of rapidly varying acoustic temporal fine structure and slower envelope cues following noise trauma. Because these distortions degrade and diminish the tonotopic representation of temporal acoustic features, a fundamental principle of auditory processing, the results represent a critical advancement in our understanding of the physiological bases of communication disorders. The detailed knowledge provided by this work will help guide the design of signal-processing strategies aimed at alleviating everyday communication problems for people with hearing loss.
Collapse
|
45
|
Wallaert N, Moore BCJ, Lorenzi C. Comparing the effects of age on amplitude modulation and frequency modulation detection. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 139:3088. [PMID: 27369130 DOI: 10.1121/1.4953019] [Citation(s) in RCA: 31] [Impact Index Per Article: 3.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/22/2023]
Abstract
Frequency modulation (FM) and amplitude modulation (AM) detection thresholds were measured at 40 dB sensation level for young (22-28 yrs) and older (44-66 yrs) listeners with normal audiograms for a carrier frequency of 500 Hz and modulation rates of 2 and 20 Hz. The number of modulation cycles, N, varied between 2 and 9. For FM detection, uninformative AM at the same rate as the FM was superimposed to disrupt excitation-pattern cues. For both groups, AM and FM detection thresholds were lower for the 2-Hz than for the 20-Hz rate, and AM and FM detection thresholds decreased with increasing N. Thresholds were higher for older than for younger listeners, especially for FM detection at 2 Hz, possibly reflecting the effect of age on the use of temporal-fine-structure cues for 2-Hz FM detection. The effect of increasing N was similar across groups for both AM and FM. However, at 20 Hz, older listeners showed a greater effect of increasing N than younger listeners for both AM and FM. The results suggest that ageing reduces sensitivity to both excitation-pattern and temporal-fine-structure cues for modulation detection, but more so for the latter, while sparing temporal integration of these cues at low modulation rates.
Collapse
Affiliation(s)
- Nicolas Wallaert
- UMR CNRS LSP 8248, Institut d'Etude de la Cognition, Ecole normale supérieure, Paris Sciences et Lettres, 29 rue d'Ulm, 75005 Paris, France
| | - Brian C J Moore
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, United Kingdom
| | - Christian Lorenzi
- UMR CNRS LSP 8248, Institut d'Etude de la Cognition, Ecole normale supérieure, Paris Sciences et Lettres, 29 rue d'Ulm, 75005 Paris, France
| |
Collapse
|
46
|
O'Brien GE, Imennov NS, Rubinstein JT. Simulating electrical modulation detection thresholds using a biophysical model of the auditory nerve. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 139:2448. [PMID: 27250141 DOI: 10.1121/1.4947430] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Modulation detection thresholds (MDTs) assess listeners' sensitivity to changes in the temporal envelope of a signal and have been shown to strongly correlate with speech perception in cochlear implant users. MDTs are simulated with a stochastic model of a population of auditory nerve fibers that has been verified to accurately simulate a number of physiologically important temporal response properties. The procedure to estimate detection thresholds has previously been applied to stimulus discrimination tasks. The population model simulates the MDT-stimulus intensity relationship measured in cochlear implant users. The model also recreates the shape of the modulation transfer function and the relationship between MDTs and carrier rate. Discrimination based on fluctuations in synchronous firing activity predicts better performance at low carrier rates, but quantitative measures of modulation coding predict better neural representation of high carrier rate stimuli. Manipulating the number of fibers and a temporal integration parameter, the width of a sliding temporal integration window, varies properties of the MDTs, such as cutoff frequency and peak threshold. These results demonstrate the importance of using a multi-diameter fiber population in modeling the MDTs and demonstrate a wider applicability of this model to simulating behavioral performance in cochlear implant listeners.
Collapse
Affiliation(s)
- Gabrielle E O'Brien
- Department of Otolaryngology, V. M. Bloedel Hearing Research Center, University of Washington, Box 3657923, CHDD building, CD 176, Seattle, Washington 98196, USA
| | - Nikita S Imennov
- Department of Otolaryngology, V. M. Bloedel Hearing Research Center, University of Washington, Box 3657923, CHDD building, CD 176, Seattle, Washington 98196, USA
| | - Jay T Rubinstein
- Department of Otolaryngology, V. M. Bloedel Hearing Research Center, University of Washington, Box 3657923, CHDD building, CD 176, Seattle, Washington 98196, USA
| |
Collapse
|
47
|
Carlyon RP, Deeks JM. Combined neural and behavioural measures of temporal pitch perception in cochlear implant users. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 138:2885-2905. [PMID: 26627764 DOI: 10.1121/1.4934275] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Four experiments measured the perceptual and neural correlates of the temporal pattern of electrical stimulation applied to one cochlear-implant (CI) electrode, for several subjects. Neural effects were estimated from the electrically evoked compound action potential (ECAP) to each pulse. Experiment 1 attenuated every second pulse of a 200-pps pulse train. Increasing attenuation caused pitch to drop and the ECAP to become amplitude modulated, thereby providing an estimate of the relationship between neural modulation and pitch. Experiment 2 showed that the pitch of a 200-pps pulse train can be reduced by delaying every second pulse, so that the inter-pulse-intervals alternate between longer and shorter intervals. This caused the ECAP to become amplitude modulated, but not by enough to account for the change in pitch. Experiment 3 replicated the finding that rate discrimination deteriorates with increases in baseline rate. This was accompanied by an increase in ECAP modulation, but by an amount that produced only a small effect on pitch in experiment 1. Experiment 4 showed that preceding a pulse train with a carefully selected "pre-pulse" could reduce ECAP modulation, but did not improve rate discrimination. Implications for theories of pitch and for limitations of pitch perception in CI users are discussed.
Collapse
Affiliation(s)
- Robert P Carlyon
- MRC Cognition & Brain Sciences Unit, 15 Chaucer Road, Cambridge CB2 7EF, England
| | - John M Deeks
- MRC Cognition & Brain Sciences Unit, 15 Chaucer Road, Cambridge CB2 7EF, England
| |
Collapse
|
48
|
Marmel F, Plack CJ, Hopkins K. The role of excitation-pattern cues in the detection of frequency shifts in bandpass-filtered complex tones. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 137:2687-97. [PMID: 25994700 PMCID: PMC5044982 DOI: 10.1121/1.4919315] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
One task intended to measure sensitivity to temporal fine structure (TFS) involves the discrimination of a harmonic complex tone from a tone in which all harmonics are shifted upwards by the same amount in hertz. Both tones are passed through a fixed bandpass filter centered on the high harmonics to reduce the availability of excitation-pattern cues and a background noise is used to mask combination tones. The role of frequency selectivity in this "TFS1" task was investigated by varying level. Experiment 1 showed that listeners performed more poorly at a high level than at a low level. Experiment 2 included intermediate levels and showed that performance deteriorated for levels above about 57 dB sound pressure level. Experiment 3 estimated the magnitude of excitation-pattern cues from the variation in forward masking of a pure tone as a function of frequency shift in the complex tones. There was negligible variation, except for the lowest level used. The results indicate that the changes in excitation level at threshold for the TFS1 task would be too small to be usable. The results are consistent with the TFS1 task being performed using TFS cues, and with frequency selectivity having an indirect effect on performance via its influence on TFS cues.
Collapse
Affiliation(s)
- Frederic Marmel
- School of Psychological Sciences, Manchester Academic Health Science Centre, The University of Manchester, Oxford Road, Manchester M13 9PL, United Kingdom
| | - Christopher J. Plack
- School of Psychological Sciences, Manchester Academic Health Science Centre, The University of Manchester, Oxford Road, Manchester M13 9PL, United Kingdom
| | - Kathryn Hopkins
- School of Psychological Sciences, Manchester Academic Health Science Centre, The University of Manchester, Oxford Road, Manchester M13 9PL, United Kingdom
| |
Collapse
|
49
|
Abstract
In the diverse mechanosensory systems that animals evolved, the waveform of stimuli can be encoded by phase locking in spike trains of primary afferents. Coding of the fine structure of sounds via phase locking is thought to be critical for hearing. The upper frequency limit of phase locking varies across species and is unknown in humans. We applied a method developed previously, which is based on neural adaptation evoked by forward masking, to analyze mass potentials recorded on the cochlea and auditory nerve in the cat. The method allows us to separate neural phase locking from receptor potentials. We find that the frequency limit of neural phase locking obtained from mass potentials was very similar to that reported for individual auditory nerve fibers. The results suggest that this is a promising approach to examine neural phase locking in humans with normal or impaired hearing or in other species for which direct recordings from primary afferents are not feasible.
Collapse
|
50
|
Implications of within-fiber temporal coding for perceptual studies of F0 discrimination and discrimination of harmonic and inharmonic tone complexes. J Assoc Res Otolaryngol 2015; 15:465-82. [PMID: 24658856 DOI: 10.1007/s10162-014-0451-2] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2013] [Accepted: 02/17/2014] [Indexed: 10/25/2022] Open
Abstract
Recent psychophysical studies suggest that normal-hearing (NH) listeners can use acoustic temporal-fine-structure (TFS) cues for accurately discriminating shifts in the fundamental frequency (F0) of complex tones, or equal shifts in all component frequencies, even when the components are peripherally unresolved. The present study quantified both envelope (ENV) and TFS cues in single auditory-nerve (AN) fiber responses (henceforth referred to as neural ENV and TFS cues) from NH chinchillas in response to harmonic and inharmonic complex tones similar to those used in recent psychophysical studies. The lowest component in the tone complex (i.e., harmonic rank N) was systematically varied from 2 to 20 to produce various resolvability conditions in chinchillas (partially resolved to completely unresolved). Neural responses to different pairs of TEST (F0 or frequency shifted) and standard or reference (REF) stimuli were used to compute shuffled cross-correlograms, from which cross-correlation coefficients representing the degree of similarity between responses were derived separately for TFS and ENV. For a given F0 shift, the dissimilarity (TEST vs. REF) was greater for neural TFS than ENV. However, this difference was stimulus-based; the sensitivities of the neural TFS and ENV metrics were equivalent for equal absolute shifts of their relevant frequencies (center component and F0, respectively). For the F0-discrimination task, both ENV and TFS cues were available and could in principle be used for task performance. However, in contrast to human performance, neural TFS cues quantified with our cross-correlation coefficients were unaffected by phase randomization, suggesting that F0 discrimination for unresolved harmonics does not depend solely on TFS cues. For the frequency-shift (harmonic-versus-inharmonic) discrimination task, neural ENV cues were not available. Neural TFS cues were available and could in principle support performance in this task; however, in contrast to human-listeners' performance, these TFS cues showed no dependence on N. We conclude that while AN-fiber responses contain TFS-related cues, which can in principle be used to discriminate changes in F0 or equal shifts in component frequencies of peripherally unresolved harmonics, performance in these two psychophysical tasks appears to be limited by other factors (e.g., central processing noise).
Collapse
|