1
|
van den Wildenberg MF, Bremen P. Heterogeneous spatial tuning in the auditory pathway of the Mongolian Gerbil (Meriones unguiculatus). Eur J Neurosci 2024; 60:4954-4981. [PMID: 39085952 DOI: 10.1111/ejn.16472] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2021] [Revised: 06/17/2024] [Accepted: 07/09/2024] [Indexed: 08/02/2024]
Abstract
Sound-source localization is based on spatial cues arising due to interactions of sound waves with the torso, head and ears. Here, we evaluated neural responses to free-field sound sources in the central nucleus of the inferior colliculus (CIC), the medial geniculate body (MGB) and the primary auditory cortex (A1) of Mongolian gerbils. Using silicon probes we recorded from anaesthetized gerbils positioned in the centre of a sound-attenuating, anechoic chamber. We measured rate-azimuth functions (RAFs) with broad-band noise of varying levels presented from loudspeakers spanning 210° in azimuth and characterized RAFs by calculating spatial centroids, Equivalent Rectangular Receptive Fields (ERRFs), steepest slope locations and spatial-separation thresholds. To compare neuronal responses with behavioural discrimination thresholds from the literature we performed a neurometric analysis based on signal-detection theory. All structures demonstrated heterogeneous spatial tuning with a clear dominance of contralateral tuning. However, the relative amount of contralateral tuning decreased from the CIC to A1. In all three structures spatial tuning broadened with increasing sound-level. This effect was strongest in CIC and weakest in A1. Neurometric spatial-separation thresholds compared well with behavioural discrimination thresholds for locations directly in front of the animal. Our findings contrast with those reported for another rodent, the rat, which exhibits homogenous and sharply delimited contralateral spatial tuning. Spatial tuning in gerbils resembles more closely the tuning reported in A1 of cats, ferrets and non-human primates. Interestingly, gerbils, in contrast to rats, share good low-frequency hearing with carnivores and non-human primates, which may account for the observed spatial tuning properties.
Collapse
Affiliation(s)
| | - Peter Bremen
- Department of Neuroscience, Erasmus MC, Rotterdam, The Netherlands
| |
Collapse
|
2
|
Ihlefeld A, Alamatsaz N, Shapley RM. Population rate-coding predicts correctly that human sound localization depends on sound intensity. eLife 2019; 8:47027. [PMID: 31633481 PMCID: PMC6802950 DOI: 10.7554/elife.47027] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2019] [Accepted: 09/20/2019] [Indexed: 12/02/2022] Open
Abstract
Human sound localization is an important computation performed by the brain. Models of sound localization commonly assume that sound lateralization from interaural time differences is level invariant. Here we observe that two prevalent theories of sound localization make opposing predictions. The labelled-line model encodes location through tuned representations of spatial location and predicts that perceived direction is level invariant. In contrast, the hemispheric-difference model encodes location through spike-rate and predicts that perceived direction becomes medially biased at low sound levels. Here, behavioral experiments find that softer sounds are perceived closer to midline than louder sounds, favoring rate-coding models of human sound localization. Analogously, visual depth perception, which is based on interocular disparity, depends on the contrast of the target. The similar results in hearing and vision suggest that the brain may use a canonical computation of location: encoding perceived location through population spike rate relative to baseline. Being able to localize sounds helps us make sense of the world around us. The brain works out sound direction by comparing the times of when sound reaches the left versus the right ear. This cue is known as interaural time difference, or ITD for short. But how exactly the brain decodes this information is still unknown. The brain contains nerve cells that each show maximum activity in response to one particular ITD. One idea is that these nerve cells are arranged in the brain like a map from left to right, and that the brain then uses this map to estimate sound direction. This is known as the Jeffress model, after the scientist who first proposed it. There is some evidence that birds and alligators actually use a system like this to localize sounds, but no such map of nerve cells has yet been identified in mammals. An alternative possibility is that the brain compares activity across groups of ITD-sensitive nerve cells. One of the oldest and simplest ways to measure this is to compare nerve activity in the left and right hemispheres of the brain. This readout is known as the hemispheric difference model. By analyzing data from published studies, Ihlefeld, Alamatsaz, and Shapley discovered that these two models make opposing predictions about the effects of volume. The Jeffress model predicts that the volume of a sound will not affect a person’s ability to localize it. By contrast, the hemispheric difference model predicts that very soft sounds will lead to systematic errors, so that for the same ITD, softer sounds are perceived closer towards the front than louder sounds. To investigate this further, Ihlefeld, Alamatsaz, and Shapley asked healthy volunteers to localize sounds of different volumes. The volunteers tended to mis-localize quieter sounds, believing them to be closer to the body’s midline than they actually were, which is inconsistent with the predictions of the Jeffress model. These new findings also reveal key parallels to processing in the visual system. Visual areas of the brain estimate how far away an object is by comparing the input that reaches the two eyes. But these estimates are also systematically less accurate for low-contrast stimuli than for high-contrast ones, just as sound localization is less accurate for softer sounds than for louder ones. The idea that the brain uses the same basic strategy to localize both sights and sounds generates a number of predictions for future studies to test.
Collapse
Affiliation(s)
- Antje Ihlefeld
- New Jersey Institute of Technology, Newark, United States
| | - Nima Alamatsaz
- New Jersey Institute of Technology, Newark, United States.,Rutgers University, Newark, United States
| | | |
Collapse
|
3
|
Yin TC, Smith PH, Joris PX. Neural Mechanisms of Binaural Processing in the Auditory Brainstem. Compr Physiol 2019; 9:1503-1575. [DOI: 10.1002/cphy.c180036] [Citation(s) in RCA: 26] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
|
4
|
Montagne C, Zhou Y. Audiovisual Interactions in Front and Rear Space. Front Psychol 2018; 9:713. [PMID: 29867678 PMCID: PMC5962672 DOI: 10.3389/fpsyg.2018.00713] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2017] [Accepted: 04/23/2018] [Indexed: 11/13/2022] Open
Abstract
The human visual and auditory systems do not encode an entirely overlapped space when static head and body position are maintained. While visual capture of sound source location in the frontal field is known to be immediate and direct, visual influence in the rear auditory space behind the subject remains under-studied. In this study we investigated the influence of presenting frontal LED flashes on the perceived location of a phantom sound source generated using time-delay-based stereophony. Our results show that frontal visual stimuli affected auditory localization in two different ways - (1) auditory responses were laterally shifted (left or right) toward the location of the light stimulus and (2) auditory responses were more often in the frontal field. The observed visual effects do not adhere to the spatial rule of multisensory interaction with regard to the physical proximity of cues. Instead, the influence of visual cues interacted closely with front-back confusions in auditory localization. In particular, visually induced shift along the left-right direction occurred most often when an auditory stimulus was localized in the same (frontal) field as the light stimulus, even when the actual sound sources were presented from behind a subject. Increasing stimulus duration (from 15-ms to 50-ms) significantly mitigated the rates of front-back confusion and the associated effects of visual stimuli. These findings suggest that concurrent visual stimulation elicits a strong frontal bias in auditory localization and confirm that temporal integration plays an important role in decreasing front-back errors under conditions requiring multisensory spatial processing.
Collapse
Affiliation(s)
- Christopher Montagne
- Laboratory of Auditory Computation & Neurophysiology, Department of Speech and Hearing Science, College of Health Solutions, Arizona State University, Tempe, AZ, United States
| | - Yi Zhou
- Laboratory of Auditory Computation & Neurophysiology, Department of Speech and Hearing Science, College of Health Solutions, Arizona State University, Tempe, AZ, United States
| |
Collapse
|
5
|
Acoustic Perturbation of Breathing: A Newly Discovered Response to Soft Sounds in Rats Using an Approach of Image Analysis. J Med Biol Eng 2018. [DOI: 10.1007/s40846-018-0381-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
|
6
|
Representation of Multidimensional Stimuli: Quantifying the Most Informative Stimulus Dimension from Neural Responses. J Neurosci 2017; 37:7332-7346. [PMID: 28663198 DOI: 10.1523/jneurosci.0318-17.2017] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2017] [Revised: 06/09/2017] [Accepted: 06/17/2017] [Indexed: 11/21/2022] Open
Abstract
A common way to assess the function of sensory neurons is to measure the number of spikes produced by individual neurons while systematically varying a given dimension of the stimulus. Such measured tuning curves can then be used to quantify the accuracy of the neural representation of the stimulus dimension under study, which can in turn be related to behavioral performance. However, tuning curves often change shape when other dimensions of the stimulus are varied, reflecting the simultaneous sensitivity of neurons to multiple stimulus features. Here we illustrate how one-dimensional information analyses are misleading in this context, and propose a framework derived from Fisher information that allows the quantification of information carried by neurons in multidimensional stimulus spaces. We use this method to probe the representation of sound localization in auditory neurons of chinchillas and guinea pigs of both sexes, and show how heterogeneous tuning properties contribute to a representation of sound source position that is robust to changes in sound level.SIGNIFICANCE STATEMENT Sensory neurons' responses are typically modulated simultaneously by numerous stimulus properties, which can result in an overestimation of neural acuity with existing one-dimensional neural information transmission measures. To overcome this limitation, we develop new, compact expressions of Fisher information-derived measures that bound the robust encoding of separate stimulus dimensions in the context of multidimensional stimuli. We apply this method to the problem of the representation of sound source location in the face of changes in sound source level by neurons of the auditory midbrain.
Collapse
|
7
|
Siniscalchi M, Laddago S, Quaranta A. Auditory lateralization of conspecific and heterospecific vocalizations in cats. Laterality 2015; 21:215-27. [DOI: 10.1080/1357650x.2015.1116541] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
8
|
Day ML, Delgutte B. Neural population encoding and decoding of sound source location across sound level in the rabbit inferior colliculus. J Neurophysiol 2015; 115:193-207. [PMID: 26490292 DOI: 10.1152/jn.00643.2015] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2015] [Accepted: 10/17/2015] [Indexed: 11/22/2022] Open
Abstract
At lower levels of sensory processing, the representation of a stimulus feature in the response of a neural population can vary in complex ways across different stimulus intensities, potentially changing the amount of feature-relevant information in the response. How higher-level neural circuits could implement feature decoding computations that compensate for these intensity-dependent variations remains unclear. Here we focused on neurons in the inferior colliculus (IC) of unanesthetized rabbits, whose firing rates are sensitive to both the azimuthal position of a sound source and its sound level. We found that the azimuth tuning curves of an IC neuron at different sound levels tend to be linear transformations of each other. These transformations could either increase or decrease the mutual information between source azimuth and spike count with increasing level for individual neurons, yet population azimuthal information remained constant across the absolute sound levels tested (35, 50, and 65 dB SPL), as inferred from the performance of a maximum-likelihood neural population decoder. We harnessed evidence of level-dependent linear transformations to reduce the number of free parameters in the creation of an accurate cross-level population decoder of azimuth. Interestingly, this decoder predicts monotonic azimuth tuning curves, broadly sensitive to contralateral azimuths, in neurons at higher levels in the auditory pathway.
Collapse
Affiliation(s)
- Mitchell L Day
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, Massachusetts; and Department of Otology and Laryngology, Harvard Medical School, Boston, Massachusetts
| | - Bertrand Delgutte
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, Massachusetts; and Department of Otology and Laryngology, Harvard Medical School, Boston, Massachusetts
| |
Collapse
|
9
|
Gai Y, Ruhland JL, Yin TCT. Behavior and modeling of two-dimensional precedence effect in head-unrestrained cats. J Neurophysiol 2015; 114:1272-85. [PMID: 26133795 DOI: 10.1152/jn.00214.2015] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2015] [Accepted: 06/29/2015] [Indexed: 11/22/2022] Open
Abstract
The precedence effect (PE) is an auditory illusion that occurs when listeners localize nearly coincident and similar sounds from different spatial locations, such as a direct sound and its echo. It has mostly been studied in humans and animals with immobile heads in the horizontal plane; speaker pairs were often symmetrically located in the frontal hemifield. The present study examined the PE in head-unrestrained cats for a variety of paired-sound conditions along the horizontal, vertical, and diagonal axes. Cats were trained with operant conditioning to direct their gaze to the perceived sound location. Stereotypical PE-like behaviors were observed for speaker pairs placed in azimuth or diagonally in the frontal hemifield as the interstimulus delay was varied. For speaker pairs in the median sagittal plane, no clear PE-like behavior occurred. Interestingly, when speakers were placed diagonally in front of the cat, certain PE-like behavior emerged along the vertical dimension. However, PE-like behavior was not observed when both speakers were located in the left hemifield. A Hodgkin-Huxley model was used to simulate responses of neurons in the medial superior olive (MSO) to sound pairs in azimuth. The novel simulation incorporated a low-threshold potassium current and frequency mismatches to generate internal delays. The model exhibited distinct PE-like behavior, such as summing localization and localization dominance. The simulation indicated that certain encoding of the PE could have occurred before information reaches the inferior colliculus, and MSO neurons with binaural inputs having mismatched characteristic frequencies may play an important role.
Collapse
Affiliation(s)
- Yan Gai
- Department of Neuroscience, University of Wisconsin, Madison, Wisconsin; and Department of Biomedical Engineering, Saint Louis University, St. Louis, Missouri
| | - Janet L Ruhland
- Department of Neuroscience, University of Wisconsin, Madison, Wisconsin; and
| | - Tom C T Yin
- Department of Neuroscience, University of Wisconsin, Madison, Wisconsin; and
| |
Collapse
|
10
|
Johnson JS, O'Connor KN, Sutter ML. Segregating two simultaneous sounds in elevation using temporal envelope: Human psychophysics and a physiological model. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 138:33-43. [PMID: 26233004 PMCID: PMC4491017 DOI: 10.1121/1.4922224] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/18/2014] [Revised: 04/29/2015] [Accepted: 05/21/2015] [Indexed: 06/04/2023]
Abstract
The ability to segregate simultaneous sound sources based on their spatial locations is an important aspect of auditory scene analysis. While the role of sound azimuth in segregation is well studied, the contribution of sound elevation remains unknown. Although previous studies in humans suggest that elevation cues alone are not sufficient to segregate simultaneous broadband sources, the current study demonstrates they can suffice. Listeners segregating a temporally modulated noise target from a simultaneous unmodulated noise distracter differing in elevation fall into two statistically distinct groups: one that identifies target direction accurately across a wide range of modulation frequencies (MF) and one that cannot identify target direction accurately and, on average, reports the opposite direction of the target for low MF. A non-spiking model of inferior colliculus neurons that process single-source elevation cues suggests that the performance of both listener groups at the population level can be accounted for by the balance of excitatory and inhibitory inputs in the model. These results establish the potential for broadband elevation cues to contribute to the computations underlying sound source segregation and suggest a potential mechanism underlying this contribution.
Collapse
Affiliation(s)
- Jeffrey S Johnson
- Center for Neuroscience, University of California at Davis, 1544 Newton Court, Davis, California 95618, USA
| | - Kevin N O'Connor
- Center for Neuroscience, University of California at Davis, 1544 Newton Court, Davis, California 95618, USA
| | - Mitchell L Sutter
- Center for Neuroscience, University of California at Davis, 1544 Newton Court, Davis, California 95618, USA
| |
Collapse
|
11
|
Brown AD, Stecker GC, Tollin DJ. The precedence effect in sound localization. J Assoc Res Otolaryngol 2015; 16:1-28. [PMID: 25479823 PMCID: PMC4310855 DOI: 10.1007/s10162-014-0496-2] [Citation(s) in RCA: 64] [Impact Index Per Article: 7.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2014] [Accepted: 10/13/2014] [Indexed: 11/29/2022] Open
Abstract
In ordinary listening environments, acoustic signals reaching the ears directly from real sound sources are followed after a few milliseconds by early reflections arriving from nearby surfaces. Early reflections are spectrotemporally similar to their source signals but commonly carry spatial acoustic cues unrelated to the source location. Humans and many other animals, including nonmammalian and even invertebrate animals, are nonetheless able to effectively localize sound sources in such environments, even in the absence of disambiguating visual cues. Robust source localization despite concurrent or nearly concurrent spurious spatial acoustic information is commonly attributed to an assortment of perceptual phenomena collectively termed "the precedence effect," characterizing the perceptual dominance of spatial information carried by the first-arriving signal. Here, we highlight recent progress and changes in the understanding of the precedence effect and related phenomena.
Collapse
Affiliation(s)
- Andrew D. Brown
- />Department of Physiology and Biophysics, University of Colorado School of Medicine, Aurora, CO 80045 USA
| | - G. Christopher Stecker
- />Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN 37232 USA
| | - Daniel J. Tollin
- />Department of Physiology and Biophysics, University of Colorado School of Medicine, Aurora, CO 80045 USA
| |
Collapse
|
12
|
Koka K, Tollin DJ. Linear coding of complex sound spectra by discharge rate in neurons of the medial nucleus of the trapezoid body (MNTB) and its inputs. Front Neural Circuits 2014; 8:144. [PMID: 25565971 PMCID: PMC4267272 DOI: 10.3389/fncir.2014.00144] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2014] [Accepted: 11/25/2014] [Indexed: 11/25/2022] Open
Abstract
The interaural level difference (ILD) cue to sound location is first encoded in the lateral superior olive (LSO). ILD sensitivity results because the LSO receives excitatory input from the ipsilateral cochlear nucleus and inhibitory input indirectly from the contralateral cochlear nucleus via glycinergic neurons of the ipsilateral medial nucleus of the trapezoid body (MNTB). It is hypothesized that in order for LSO neurons to encode ILDs, the sound spectra at both ears must be accurately encoded via spike rate by their afferents. This spectral-coding hypothesis has not been directly tested in MNTB, likely because MNTB neurons have been mostly described and studied recently in regards to their abilities to encode temporal aspects of sounds, not spectral. Here, we test the hypothesis that MNTB neurons and their inputs from the cochlear nucleus and auditory nerve code sound spectra via discharge rate. The Random Spectral Shape (RSS) method was used to estimate how the levels of 100-ms duration spectrally stationary stimuli were weighted, both linearly and non-linearly, across a wide band of frequencies. In general, MNTB neurons, and their globular bushy cell inputs, were found to be well-modeled by a linear weighting of spectra demonstrating that the pathways through the MNTB can accurately encode sound spectra including those resulting from the acoustical cues to sound location provided by head-related directional transfer functions (DTFs). Together with the anatomical and biophysical specializations for timing in the MNTB-LSO complex, these mechanisms may allow ILDs to be computed for complex stimuli with rapid spectrotemporally-modulated envelopes such as speech and animal vocalizations and moving sound sources.
Collapse
Affiliation(s)
- Kanthaiah Koka
- Department of Physiology and Biophysics, University of Colorado School of Medicine Aurora, CO, USA
| | - Daniel J Tollin
- Department of Physiology and Biophysics, University of Colorado School of Medicine Aurora, CO, USA
| |
Collapse
|
13
|
Jones AE, Ruhland JL, Gai Y, Yin TCT. Simultaneous comparison of two sound localization measures. Hear Res 2014; 317:33-40. [PMID: 25261773 DOI: 10.1016/j.heares.2014.08.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/20/2014] [Revised: 08/26/2014] [Accepted: 08/30/2014] [Indexed: 10/24/2022]
Abstract
Almost all behavioral studies of sound localization have used either an approach-to-target or pointing/orienting task to assess absolute sound localization performance, yet there are very few direct comparisons of these measures. In an approach-to-target task, the subject is trained to walk to a sound source from a fixed location. In an orienting task, finger, head and/or eye movements are monitored while the subject's body is typically constrained. The fact that subjects may also initiate head and eye movements toward the target during the approach-to-target task allows us to measure the accuracy of the initial orienting response and compare it with subsequent target selection. To perform this comparison, we trained cats to localize a broadband noise presented randomly from one of four speakers located ± 30° and ± 60° in azimuth. The cat responded to each sound presentation by walking to and pressing a lever at the perceived location, and a food reward was delivered if the first attempt was correct. In tandem, we recorded initial head and eye orienting movements, via magnetic search coils, immediately following target onset and prior to the walking response. Reducing either stimulus duration or level resulted in a systematic decline in both measurements of localization performance. When the task was easy, localization performance was accurate for both measures. When the task was more difficult, the number of incorrect (i.e., wrong selection) and no-go (i.e., no selection) responses increased. Interestingly, for many of the incorrect trials, there was a dissociation between the orienting response and the target selected, and for many of the no-go trials, the gaze oriented towards the correct target even though the cat did not move to it. This suggests different neural systems governing walking to a target as compared to unconditioned gaze orienting.
Collapse
Affiliation(s)
- Amy E Jones
- Department of Neuroscience, University of Wisconsin, Madison, WI 53706, USA.
| | - Janet L Ruhland
- Department of Neuroscience, University of Wisconsin, Madison, WI 53706, USA.
| | - Yan Gai
- Department of Neuroscience, University of Wisconsin, Madison, WI 53706, USA.
| | - Tom C T Yin
- Department of Neuroscience, University of Wisconsin, Madison, WI 53706, USA.
| |
Collapse
|
14
|
Gai Y, Ruhland JL, Yin TCT. Localization of click trains and speech by cats: the negative level effect. J Assoc Res Otolaryngol 2014; 15:789-800. [PMID: 24942705 DOI: 10.1007/s10162-014-0469-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2014] [Accepted: 05/28/2014] [Indexed: 11/30/2022] Open
Abstract
Although localization of sound in elevation is believed to depend on spectral cues, it has been shown with human listeners that the temporal features of sound can also greatly affect localization performance. Of particular interest is a phenomenon known as the negative level effect, which describes the deterioration of localization ability in elevation with increasing sound level and is observed only with impulsive or short-duration sound. The present study uses the gaze positions of domestic cats as measures of perceived locations of sound targets varying in azimuth and elevation. The effects of sound level on localization in terms of accuracy, precision, and response latency were tested for sound with different temporal features, such as a click train, a single click, a continuous sound that had the same frequency spectrum of the click train, and speech segments. In agreement with previous human studies, negative level effects were only observed with click-like stimuli and only in elevation. In fact, localization of speech sounds in elevation benefited significantly when the sound level increased. Our findings indicate that the temporal continuity of a sound can affect the frequency analysis performed by the auditory system, and the variation in the frequency spectrum contained in speech sound does not interfere much with the spectral coding for its location in elevation.
Collapse
Affiliation(s)
- Yan Gai
- Department of Neuroscience, University of Wisconsin-Madison, Medical Sciences Building, 251 SMI, 1300 University Avenue, Madison, WI, 53706, USA,
| | | | | |
Collapse
|
15
|
Macpherson EA, Sabin AT. Vertical-plane sound localization with distorted spectral cues. Hear Res 2013; 306:76-92. [PMID: 24076423 DOI: 10.1016/j.heares.2013.09.007] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/16/2013] [Revised: 09/11/2013] [Accepted: 09/17/2013] [Indexed: 10/26/2022]
Abstract
For human listeners, the primary cues for localization in the vertical plane are provided by the direction-dependent filtering of the pinnae, head, and upper body. Vertical-plane localization generally is accurate for broadband sounds, but when such sounds are presented at near-threshold levels or at high levels with short durations (<20 ms), the apparent location is biased toward the horizontal plane (i.e., elevation gain <1). We tested the hypothesis that these effects result in part from distorted peripheral representations of sound spectra. Human listeners indicated the apparent position of 100-ms, 50-60 dB SPL, wideband noise-burst targets by orienting their heads. The targets were synthesized in virtual auditory space and presented over headphones. Faithfully synthesized targets were interleaved with targets for which the directional transfer function spectral notches were filled in, peaks were leveled off, or the spectral contrast of the entire profile was reduced or expanded. As notches were filled in progressively or peaks leveled progressively, elevation gain decreased in a graded manner similar to that observed as sensation level is reduced below 30 dB or, for brief sounds, increased above 45 dB. As spectral contrast was reduced, gain dropped only at the most extreme reduction (25% of normal). Spectral contrast expansion had little effect. The results are consistent with the hypothesis that loss of representation of spectral features contributes to reduced elevation gain at low and high sound levels. The results also suggest that perceived location depends on a correlation-like spectral matching process that is sensitive to the relative, rather than absolute, across-frequency shape of the spectral profile.
Collapse
Affiliation(s)
- Ewan A Macpherson
- Kresge Hearing Research Institute, University of Michigan Medical School, 1150 W. Medical Center Drive, Ann Arbor, MI 48109-5616, USA; National Centre for Audiology, Western University, 1201 Western Road, London, Ontario, Canada N6G 1H1.
| | | |
Collapse
|
16
|
Gai Y, Ruhland JL, Yin TCT. Effects of forward masking on sound localization in cats: basic findings with broadband maskers. J Neurophysiol 2013; 110:1600-10. [PMID: 23843432 DOI: 10.1152/jn.00255.2013] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Forward masking is traditionally measured with a detection task in which the addition of a preceding masking sound results in an increased signal-detection threshold. Little is known about the influence of forward masking on localization of free-field sound for human or animal subjects. Here we recorded gaze shifts of two head-unrestrained cats during localization using a search-coil technique. A broadband (BB) noise masker was presented straight ahead. A brief signal could come from 1 of the 17 speaker locations in the frontal hemifield. The signal was either a BB or a band-limited (BL) noise. For BB targets, the presence of the forward masker reduced localization accuracy at almost all target levels (20 to 80 dB SPL) along both horizontal and vertical dimensions. Temporal decay of masking was observed when a 15-ms interstimulus gap was added between the end of the masker and the beginning of the target. A large effect of forward masking was also observed for BL targets with low (0.2-2 kHz) and mid (2-7 kHz) frequencies, indicating that the interaural timing cue is susceptible to forward masking. Except at low sound levels, a small or little effect was observed for high-frequency (7-15 kHz) targets, indicating that the interaural level and the spectral cues in that frequency range remained relatively robust. Our findings suggest that different localization mechanisms can operate independently in a complex listening environment.
Collapse
Affiliation(s)
- Yan Gai
- Department of Neuroscience, University of Wisconsin, Madison, Wisconsin
| | | | | |
Collapse
|