51
|
Catic J, Santurette S, Dau T. The role of reverberation-related binaural cues in the externalization of speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 138:1154-1167. [PMID: 26328729 DOI: 10.1121/1.4928132] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
The perception of externalization of speech sounds was investigated with respect to the monaural and binaural cues available at the listeners' ears in a reverberant environment. Individualized binaural room impulse responses (BRIRs) were used to simulate externalized sound sources via headphones. The measured BRIRs were subsequently modified such that the proportion of the response containing binaural vs monaural information was varied. Normal-hearing listeners were presented with speech sounds convolved with such modified BRIRs. Monaural reverberation cues were found to be sufficient for the externalization of a lateral sound source. In contrast, for a frontal source, an increased amount of binaural cues from reflections was required in order to obtain well externalized sound images. It was demonstrated that the interaction between the interaural cues of the direct sound and the reverberation strongly affects the perception of externalization. An analysis of the short-term binaural cues showed that the amount of fluctuations of the binaural cues corresponded well to the externalization ratings obtained in the listening tests. The results further suggested that the precedence effect is involved in the auditory processing of the dynamic binaural cues that are utilized for externalization perception.
Collapse
Affiliation(s)
- Jasmina Catic
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, DTU Bygning 352, Ørsteds Plads, 2800 Kongens Lyngby, Denmark
| | - Sébastien Santurette
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, DTU Bygning 352, Ørsteds Plads, 2800 Kongens Lyngby, Denmark
| | - Torsten Dau
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, DTU Bygning 352, Ørsteds Plads, 2800 Kongens Lyngby, Denmark
| |
Collapse
|
52
|
Spiousas I, Etchemendy PE, Vergara RO, Calcagno ER, Eguia MC. An Auditory Illusion of Proximity of the Source Induced by Sonic Crystals. PLoS One 2015. [PMID: 26222281 PMCID: PMC4519286 DOI: 10.1371/journal.pone.0133271] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/03/2022] Open
Abstract
In this work we report an illusion of proximity of a sound source created by a sonic crystal placed between the source and a listener. This effect seems, at first, paradoxical to naïve listeners since the sonic crystal is an obstacle formed by almost densely packed cylindrical scatterers. Even when the singular acoustical properties of these periodic composite materials have been studied extensively (including band gaps, deaf bands, negative refraction, and birrefringence), the possible perceptual effects remain unexplored. The illusion reported here is studied through acoustical measurements and a psychophysical experiment. The results of the acoustical measurements showed that, for a certain frequency range and region in space where the focusing phenomenon takes place, the sonic crystal induces substantial increases in binaural intensity, direct-to-reverberant energy ratio and interaural cross-correlation values, all cues involved in the auditory perception of distance. Consistently, the results of the psychophysical experiment revealed that the presence of the sonic crystal between the sound source and the listener produces a significant reduction of the perceived relative distance to the sound source.
Collapse
Affiliation(s)
- Ignacio Spiousas
- Laboratorio de Acústica y Percepción Sonora, Universidad Nacional de Quilmes, Bernal, Buenos Aires, Argentina
| | - Pablo E. Etchemendy
- Laboratorio de Acústica y Percepción Sonora, Universidad Nacional de Quilmes, Bernal, Buenos Aires, Argentina
| | - Ramiro O. Vergara
- Laboratorio de Acústica y Percepción Sonora, Universidad Nacional de Quilmes, Bernal, Buenos Aires, Argentina
| | - Esteban R. Calcagno
- Laboratorio de Acústica y Percepción Sonora, Universidad Nacional de Quilmes, Bernal, Buenos Aires, Argentina
| | - Manuel C. Eguia
- Laboratorio de Acústica y Percepción Sonora, Universidad Nacional de Quilmes, Bernal, Buenos Aires, Argentina
- * E-mail:
| |
Collapse
|
53
|
Sunder K, Gan WS, Tan EL. Modeling distance-dependent individual head-related transfer functions in the horizontal plane using frontal projection headphones. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 138:150-171. [PMID: 26233016 DOI: 10.1121/1.4919347] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
The veracity of virtual audio is degraded by the use of non-individualized head-related transfer functions (HRTFs) due to the introduction of front-back, elevation confusions, and timbral coloration. Hence, an accurate reproduction of spatial sound demands the use of individualized HRTFs. Measuring distance-dependent individualized HRTFs can be extremely tedious, since it requires precise measurements at several distances in the proximal region (<1 m) for each individual. This paper proposes a technique to model distance-dependent individualized HRTFs in the horizontal plane using "frontal projection headphones playback" that does not require individualized measurements. The frontal projection headphones [Sunder, Tan, and Gan (2013). J. Audio Eng. Soc. 61, 989-1000] project the sound directly onto the pinnae from the front, and thus inherently create listener's idiosyncratic pinna cues at the eardrum. Perceptual experiments were conducted to investigate cues (auditory parallax and interaural level differences) that aid distance perception in anechoic conditions. Interaural level differences were identified as the prominent cue for distance perception and a spherical head model was used to model these distance-dependent features. Detailed psychophysical experiments revealed that the modeled distance-dependent individualized HRTFs exhibited localization performance close to the measured distance-dependent individualized HRTFs for all subjects.
Collapse
Affiliation(s)
- Kaushik Sunder
- Digital Signal Processing Lab, School of Electrical and Electronics Engineering, Nanyang Technological University, Singapore 639798, Singapore
| | - Woon-Seng Gan
- Digital Signal Processing Lab, School of Electrical and Electronics Engineering, Nanyang Technological University, Singapore 639798, Singapore
| | - Ee-Leng Tan
- Digital Signal Processing Lab, School of Electrical and Electronics Engineering, Nanyang Technological University, Singapore 639798, Singapore
| |
Collapse
|
54
|
Auditory distance coding in rabbit midbrain neurons and human perception: monaural amplitude modulation depth as a cue. J Neurosci 2015; 35:5360-72. [PMID: 25834060 DOI: 10.1523/jneurosci.3798-14.2015] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
Mechanisms underlying sound source distance localization are not well understood. Here we tested the hypothesis that a novel mechanism can create monaural distance sensitivity: a combination of auditory midbrain neurons' sensitivity to amplitude modulation (AM) depth and distance-dependent loss of AM in reverberation. We used virtual auditory space (VAS) methods for sounds at various distances in anechoic and reverberant environments. Stimulus level was constant across distance. With increasing modulation depth, some rabbit inferior colliculus neurons increased firing rates whereas others decreased. These neurons exhibited monotonic relationships between firing rates and distance for monaurally presented noise when two conditions were met: (1) the sound had AM, and (2) the environment was reverberant. The firing rates as a function of distance remained approximately constant without AM in either environment and, in an anechoic condition, even with AM. We corroborated this finding by reproducing the distance sensitivity using a neural model. We also conducted a human psychophysical study using similar methods. Normal-hearing listeners reported perceived distance in response to monaural 1 octave 4 kHz noise source sounds presented at distances of 35-200 cm. We found parallels between the rabbit neural and human responses. In both, sound distance could be discriminated only if the monaural sound in reverberation had AM. These observations support the hypothesis. When other cues are available (e.g., in binaural hearing), how much the auditory system actually uses the AM as a distance cue remains to be determined.
Collapse
|
55
|
Van der Stoep N, Nijboer T, Van der Stigchel S, Spence C. Multisensory interactions in the depth plane in front and rear space: A review. Neuropsychologia 2015; 70:335-49. [DOI: 10.1016/j.neuropsychologia.2014.12.007] [Citation(s) in RCA: 40] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2014] [Revised: 12/03/2014] [Accepted: 12/04/2014] [Indexed: 11/28/2022]
|
56
|
Preattentive processing of horizontal motion, radial motion, and intensity changes of sounds. Neuroreport 2014; 24:861-5. [PMID: 24022175 DOI: 10.1097/wnr.0000000000000006] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
Localization of sound sources is critical for an appropriate behavioral response. This is not only true for localization in the horizontal plane but also for localization in depth. Depth ranging of sound sources implicates various distance cues, among others sound intensity. In this study, we measured human electroencephalography and compared mismatch negativity (MMN) amplitudes and latencies for horizontal motion, radial motion, and pure intensity changes in the free field. We observed similar MMN latencies for horizontal and radial motion, whereas MMN responses to pure intensity changes were comparably delayed. MMN amplitudes and latencies were not different for approaching and receding sounds. Our data suggest similar fast processing for horizontal and radial motion, whereas pure intensity changes are possibly processed with less priority.
Collapse
|
57
|
Silva CC, Mendonça C, Mouta S, Silva R, Campos JC, Santos J. Depth cues and perceived audiovisual synchrony of biological motion. PLoS One 2013; 8:e80096. [PMID: 24244617 PMCID: PMC3828238 DOI: 10.1371/journal.pone.0080096] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2013] [Accepted: 09/30/2013] [Indexed: 11/18/2022] Open
Abstract
BACKGROUND Due to their different propagation times, visual and auditory signals from external events arrive at the human sensory receptors with a disparate delay. This delay consistently varies with distance, but, despite such variability, most events are perceived as synchronic. There is, however, contradictory data and claims regarding the existence of compensatory mechanisms for distance in simultaneity judgments. PRINCIPAL FINDINGS In this paper we have used familiar audiovisual events--a visual walker and footstep sounds--and manipulated the number of depth cues. In a simultaneity judgment task we presented a large range of stimulus onset asynchronies corresponding to distances of up to 35 meters. We found an effect of distance over the simultaneity estimates, with greater distances requiring larger stimulus onset asynchronies, and vision always leading. This effect was stronger when both visual and auditory cues were present but was interestingly not found when depth cues were impoverished. SIGNIFICANCE These findings reveal that there should be an internal mechanism to compensate for audiovisual delays, which critically depends on the depth information available.
Collapse
Affiliation(s)
- Carlos César Silva
- Department of Informatics, University of Minho, Braga, Braga, Portugal
- * E-mail:
| | - Catarina Mendonça
- Department of Psychology, School of Medicine and Health Sciences, Carl von Ossietzky University, Oldenburg, Lower Saxony, Germany
| | - Sandra Mouta
- Department de Psicologia Bàsica, University de Barcelona, Barcelona, Catalonia, Spain
- INESC Porto, Instituto de Engenharia de Sistemas e Computadores, Porto, Porto, Portugal
| | - Rosa Silva
- Department of Basic Psychology, School of Psychology, University of Minho, Braga, Braga, Portugal
| | - José Creissac Campos
- Department of Informatics, School of Engineering, University of Minho, Braga, Braga, Portugal
- HASLab (High-Assurance Software Laboratory), INESC TEC Porto, Porto, Portugal
| | - Jorge Santos
- Department of Basic Psychology, School of Psychology, University of Minho, Braga, Braga, Portugal
- Centro Algoritmi, University of Minho, Guimarães, Braga, Portugal
- Centro de Computação Gráfica, Guimarães, Braga, Portugal
| |
Collapse
|
58
|
Exogenous orienting of crossmodal attention in 3-D space: support for a depth-aware crossmodal attentional system. Psychon Bull Rev 2013; 21:708-14. [PMID: 24101573 DOI: 10.3758/s13423-013-0532-y] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
The aim of the present study was to investigate exogenous crossmodal orienting of attention in three-dimensional (3-D) space. Most studies in which the orienting of attention has been examined in 3-D space concerned either exogenous intramodal or endogenous crossmodal attention. Evidence for exogenous crossmodal orienting of attention in depth is lacking. Endogenous and exogenous attention are behaviorally different, suggesting that they are two different mechanisms. We used the orthogonal spatial-cueing paradigm and presented auditory exogenous cues at one of four possible locations in near or far space before the onset of a visual target. Cues could be presented at the same (valid) or at a different (invalid) depth from the target (radial validity), and on the same (valid) or on a different (invalid) side (horizontal validity), whereas we blocked the depth at which visual targets were presented. Next to an overall validity effect (valid RTs < invalid RTs) in horizontal space, we observed an interaction between the horizontal and radial validity of the cue: The horizontal validity effect was present only when the cue and the target were presented at the same depth. No horizontal validity effect was observed when the cue and the target were presented at different depths. These results suggest that exogenous crossmodal attention is "depth-aware," and they are discussed in the context of the supramodal hypothesis of attention.
Collapse
|
59
|
Altmann CF, Ono K, Callan A, Matsuhashi M, Mima T, Fukuyama H. Environmental reverberation affects processing of sound intensity in right temporal cortex. Eur J Neurosci 2013; 38:3210-20. [PMID: 23869792 DOI: 10.1111/ejn.12318] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2013] [Revised: 06/13/2013] [Accepted: 06/20/2013] [Indexed: 11/28/2022]
Abstract
Although sound reverberation is considered a nuisance variable in most studies investigating auditory processing, it can serve as a cue for loudness constancy, a phenomenon describing constant loudness perception in spite of changing sound source distance. In this study, we manipulated room reverberation characteristics to test their effect on psychophysical loudness constancy and we tested with magnetoencephalography on human subjects for neural responses reflecting loudness constancy. Psychophysically, we found that loudness constancy was present in strong, but not weak, reverberation conditions. In contrast, the dependence of sound distance judgment on actual distance was similar across conditions. We observed brain activity reflecting behavioral loudness constancy, i.e. inverse scaling of the evoked magnetic fields with distance for weak reverberation but constant responses across distance for strong reverberation from ~210 to 270 ms after stimulus onset. Distributed magnetoencephalography source reconstruction revealed underlying neural generators within the right middle temporal and right inferior anterior temporal lobe. Our data suggest a dissociation of loudness constancy and distance perception, implying a direct usage of reverberation cues for constructing constant loudness across distance. Furthermore, our magnetoencephalography data suggest involvement of auditory association areas in the right middle and right inferior anterior temporal cortex in this process.
Collapse
Affiliation(s)
- Christian F Altmann
- Graduate School of Medicine, Human Brain Research Center, Kyoto University, Kyoto, Japan; Career-Path Promotion Unit for Young Life Scientists, Kyoto University, Kyoto, Japan
| | | | | | | | | | | |
Collapse
|
60
|
Ahveninen J, Kopčo N, Jääskeläinen IP. Psychophysics and neuronal bases of sound localization in humans. Hear Res 2013; 307:86-97. [PMID: 23886698 DOI: 10.1016/j.heares.2013.07.008] [Citation(s) in RCA: 56] [Impact Index Per Article: 5.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/15/2013] [Revised: 07/02/2013] [Accepted: 07/10/2013] [Indexed: 10/26/2022]
Abstract
Localization of sound sources is a considerable computational challenge for the human brain. Whereas the visual system can process basic spatial information in parallel, the auditory system lacks a straightforward correspondence between external spatial locations and sensory receptive fields. Consequently, the question how different acoustic features supporting spatial hearing are represented in the central nervous system is still open. Functional neuroimaging studies in humans have provided evidence for a posterior auditory "where" pathway that encompasses non-primary auditory cortex areas, including the planum temporale (PT) and posterior superior temporal gyrus (STG), which are strongly activated by horizontal sound direction changes, distance changes, and movement. However, these areas are also activated by a wide variety of other stimulus features, posing a challenge for the interpretation that the underlying areas are purely spatial. This review discusses behavioral and neuroimaging studies on sound localization, and some of the competing models of representation of auditory space in humans. This article is part of a Special Issue entitled Human Auditory Neuroimaging.
Collapse
Affiliation(s)
- Jyrki Ahveninen
- Harvard Medical School - Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital, Charlestown, MA, USA.
| | | | | |
Collapse
|
61
|
Neural correlates of sound externalization. Neuroimage 2013; 66:22-7. [DOI: 10.1016/j.neuroimage.2012.10.057] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2012] [Revised: 09/26/2012] [Accepted: 10/12/2012] [Indexed: 11/18/2022] Open
|
62
|
Kolarik AJ, Cirstea S, Pardhan S. Evidence for enhanced discrimination of virtual auditory distance among blind listeners using level and direct-to-reverberant cues. Exp Brain Res 2012. [PMID: 23178908 DOI: 10.1007/s00221-012-3340-0] [Citation(s) in RCA: 57] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Abstract
Totally blind listeners often demonstrate better than normal capabilities when performing spatial hearing tasks. Accurate representation of three-dimensional auditory space requires the processing of available distance information between the listener and the sound source; however, auditory distance cues vary greatly depending upon the acoustic properties of the environment, and it is not known which distance cues are important to totally blind listeners. Our data show that totally blind listeners display better performance compared to sighted age-matched controls for distance discrimination tasks in anechoic and reverberant virtual rooms simulated using a room-image procedure. Totally blind listeners use two major auditory distance cues to stationary sound sources, level and direct-to-reverberant ratio, more effectively than sighted controls for many of the virtual distances tested. These results show that significant compensation among totally blind listeners for virtual auditory spatial distance leads to benefits across a range of simulated acoustic environments. No significant differences in performance were observed between listeners with partial non-correctable visual losses and sighted controls, suggesting that sensory compensation for virtual distance does not occur for listeners with partial vision loss.
Collapse
Affiliation(s)
- Andrew J Kolarik
- Vision and Eye Research Unit (VERU), Postgraduate Medical Institute, Anglia Ruskin University, Eastings 204, East Road, Cambridge, CB1 1PT, UK.
| | | | | |
Collapse
|
63
|
Altmann CF, Matsuhashi M, Votinov M, Goto K, Mima T, Fukuyama H. Visual distance cues modulate neuromagnetic auditory N1m responses. Clin Neurophysiol 2012; 123:2273-80. [DOI: 10.1016/j.clinph.2012.04.004] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2011] [Revised: 03/06/2012] [Accepted: 04/02/2012] [Indexed: 10/28/2022]
|
64
|
Calcagno ER, Abregú EL, Eguía MC, Vergara R. The role of vision in auditory distance perception. Perception 2012; 41:175-92. [PMID: 22670346 DOI: 10.1068/p7153] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
Abstract
In humans, multisensory interaction is an important strategy for improving the detection of stimuli of different nature and reducing the variability of response. It is known that the presence of visual information affects the auditory perception in the horizontal plane (azimuth), but there are few researches that study the influence of vision in the auditory distance perception. In general, the data obtained from these studies are contradictory and do not completely define the way in which visual cues affect the apparent distance of a sound source. Here psychophysical experiments on auditory distance perception in humans are performed, including and excluding visual cues. The results show that the apparent distance from the source is affected by the presence of visual information and that subjects can store in their memory a representation of the environment that later improves the perception of distance.
Collapse
Affiliation(s)
- Esteban R Calcagno
- Laboratorio de Acústica y Percepción Sonora (LAPSo), Universidad Nacional de Quilmes, Av Calchaqui 5800, Florencio Varela, Buenos Aires, Argentina
| | | | | | | |
Collapse
|
65
|
Gourévitch B, Brette R. The impact of early reflections on binaural cues. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2012; 132:9-27. [PMID: 22779451 DOI: 10.1121/1.4726052] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
Animals live in cluttered auditory environments, where sounds arrive at the two ears through several paths. Reflections make sound localization difficult, and it is thought that the auditory system deals with this issue by isolating the first wavefront and suppressing later signals. However, in many situations, reflections arrive too early to be suppressed, for example, reflections from the ground in small animals. This paper examines the implications of these early reflections on binaural cues to sound localization, using realistic models of reflecting surfaces and a spherical model of diffraction by the head. The fusion of direct and reflected signals at each ear results in interference patterns in binaural cues as a function of frequency. These cues are maximally modified at frequencies related to the delay between direct and reflected signals, and therefore to the spatial location of the sound source. Thus, natural binaural cues differ from anechoic cues. In particular, the range of interaural time differences is substantially larger than in anechoic environments. Reflections may potentially contribute binaural cues to distance and polar angle when the properties of the reflecting surface are known and stable, for example, for reflections on the ground.
Collapse
Affiliation(s)
- Boris Gourévitch
- Equipe Audition, Département d'Etudes Cognitives, Ecole Normale Supérieure, 29 rue d'Ulm, 75005 Paris, France
| | | |
Collapse
|
66
|
Hioka Y, Niwa K, Sakauchi S, Furuya K, Haneda Y. Estimating Direct-to-Reverberant Energy Ratio Using D/R Spatial Correlation Matrix Model. ACTA ACUST UNITED AC 2011. [DOI: 10.1109/tasl.2011.2134091] [Citation(s) in RCA: 40] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
67
|
Kopčo N, Shinn-Cunningham BG. Effect of stimulus spectrum on distance perception for nearby sources. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2011; 130:1530-1541. [PMID: 21895092 PMCID: PMC3188969 DOI: 10.1121/1.3613705] [Citation(s) in RCA: 28] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/30/2010] [Revised: 06/01/2011] [Accepted: 06/24/2011] [Indexed: 05/31/2023]
Abstract
The effects of stimulus frequency and bandwidth on distance perception were examined for nearby sources in simulated reverberant space. Sources to the side [containing reverberation-related cues and interaural level difference (ILD) cues] and to the front (without ILDs) were simulated. Listeners judged the distance of noise bursts presented at a randomly roving level from simulated distances ranging from 0.15 to 1.7 m. Six stimuli were tested, varying in center frequency (300-5700 Hz) and bandwidth (200-5400 Hz). Performance, measured as the correlation between simulated and response distances, was worse for frontal than for lateral sources. For both simulated directions, performance was inversely proportional to the low-frequency stimulus cutoff, independent of stimulus bandwidth. The dependence of performance on frequency was stronger for frontal sources. These correlation results were well summarized by considering how mean response, as opposed to response variance, changed with stimulus direction and spectrum: (1) little bias was observed for lateral sources, but listeners consistently overestimated distance for frontal nearby sources; (2) for both directions, increasing the low-frequency cut-off reduced the range of responses. These results are consistent with the hypothesis that listeners used a direction-independent but frequency-dependent mapping of a reverberation-related cue, not the ILD cue, to judge source distance.
Collapse
Affiliation(s)
- Norbert Kopčo
- Department of Cybernetics and Artificial Intelligence, Technical University of Košice, Košice, Slovakia.
| | | |
Collapse
|
68
|
Agganis BT, Muday JA, Schirillo JA. Visual Biasing of Auditory Localization in Azimuth and Depth. Percept Mot Skills 2010; 111:872-92. [DOI: 10.2466/22.24.27.pms.111.6.872-892] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Correctly integrating sensory information across different modalities is a vital task, yet there are illusions which cause the incorrect localization of multisensory stimuli. A common example of these phenomena is the “ventriloquism effect.” In this illusion, the localization of auditory signals is biased by the presence of visual stimuli. For instance, when a light and sound are simultaneously presented, observers may erroneously locate the sound closer to the light than its actual position. While this phenomenon has been studied extensively in azimuth at a single depth, little is known about the interactions of stimuli at different depth planes. In the current experiment, virtual acoustics and stereo-image displays were used to test the integration of visual and auditory signals across azimuth and depth. The results suggest that greater variability in the localization of sounds in depth may lead to a greater bias from visual stimuli in depth than in azimuth. These results offer interesting implications for understanding multisensory integration.
Collapse
|
69
|
Jeub M, Schafer M, Esch T, Vary P. Model-Based Dereverberation Preserving Binaural Cues. ACTA ACUST UNITED AC 2010. [DOI: 10.1109/tasl.2010.2052156] [Citation(s) in RCA: 60] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
70
|
Auditorily-induced illusory self-motion: A review. ACTA ACUST UNITED AC 2009; 61:240-55. [DOI: 10.1016/j.brainresrev.2009.07.001] [Citation(s) in RCA: 75] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2008] [Revised: 07/12/2009] [Accepted: 07/14/2009] [Indexed: 11/19/2022]
|
71
|
Abstract
This paper introduces the major phenomena of binaural hearing. The sounds arriving at the two ears are rarely the same: usually one ear will be partially shadowed from the sound source by the head, and the sound will also have to travel further to get to that ear. The resulting differences in interaural level and time can be detected by the auditory system and can be used to determine the direction of the source of sound. They also facilitate improvements in the detectability of a target sound masked by some other sound from some other direction. In many circumstances there is a special emphasis to the onset of a sound, which helps to perceptually suppress the complex patterns of reflections and reverberations that are present in most listening environments; yet, the auditory system is often insensitive to-and cannot take advantage of-fast dynamic changes within a sound.
Collapse
Affiliation(s)
- Michael A Akeroyd
- MRC Institute of Hearing Research (Scottish Section), Glasgow Royal Infirmary, Glasgow, UK.
| |
Collapse
|
72
|
Goerlitz HR, Greif S, Siemers BM. Cues for acoustic detection of prey: insect rustling sounds and the influence of walking substrate. ACTA ACUST UNITED AC 2008; 211:2799-806. [PMID: 18723538 DOI: 10.1242/jeb.019596] [Citation(s) in RCA: 59] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022]
Abstract
When insects walk, they generally produce sounds. These can reveal the walkers' presence and location to potential predators such as owls, bats and nocturnal primates. Additionally, predators might extract information on taxon, palatability, size or profitability from the rustling sounds. In contrast to ear morphology, hearing physiology and psychoacoustics of acoustically oriented predators, little attention has hitherto been paid to the acoustic structure and information content of prey sounds. An important element in the ecology of acoustic prey detection remained virtually unexplored: the influence of the substrate type on rustling sounds. In this study, we analysed amplitude and frequency parameters from insects walking on various natural substrates, in both Germany (Carabus beetles) and Madagascar (various beetles and cockroaches). The data show that rustling sound amplitude and frequency content depend on substrate type. On moist substrates arthropods produced less intense and less broadband rustling sounds than on dry substrates. Sound pressure level was reduced by about 6 dB, halving the detection range for the predator. For a given insect, rustling sound amplitude increased with walking speed. Finally, we found that the previously established correlation of arthropod size and rustling amplitude holds across multiple substrates. Based on these data, we provide for the first time estimates of realistic detection distances in the field. These distances range from below 1 m to over 13 m, depending on the substrate, insect mass, walking speed and background noise level. These estimates are crucial for an understanding of the foraging ecology, foraging efficiency and sensory ecology of acoustic predators.
Collapse
Affiliation(s)
- Holger R Goerlitz
- Ludwig-Maximilians-University, Department Biology II, Neurobiology, Grosshaderner Strasse 2, D-82152 Martinsried, Germany
| | | | | |
Collapse
|
73
|
Sayles M, Winter IM. Reverberation challenges the temporal representation of the pitch of complex sounds. Neuron 2008; 58:789-801. [PMID: 18549789 DOI: 10.1016/j.neuron.2008.03.029] [Citation(s) in RCA: 42] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2007] [Revised: 03/07/2008] [Accepted: 03/28/2008] [Indexed: 10/22/2022]
Abstract
Accurate neural coding of the pitch of complex sounds is an essential part of auditory scene analysis; differences in pitch help segregate concurrent sounds, while similarities in pitch can help group sounds from a common source. In quiet, nonreverberant backgrounds, pitch can be derived from timing information in broadband high-frequency auditory channels and/or from frequency and timing information carried in narrowband low-frequency auditory channels. Recording from single neurons in the cochlear nucleus of anesthetized guinea pigs, we show that the neural representation of pitch based on timing information is severely degraded in the presence of reverberation. This degradation increases with both increasing reverberation strength and channel bandwidth. In a parallel human psychophysical pitch-discrimination task, reverberation impaired the ability to distinguish a high-pass harmonic sound from noise. Together, these findings explain the origin of perceptual difficulties experienced by both normal-hearing and hearing-impaired listeners in reverberant spaces.
Collapse
Affiliation(s)
- Mark Sayles
- Centre for the Neural Basis of Hearing, The Physiological Laboratory, Downing Street, Cambridge, CB2 3EG, UK
| | | |
Collapse
|
74
|
Larsen E, Iyer N, Lansing CR, Feng AS. On the minimum audible difference in direct-to-reverberant energy ratio. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2008; 124:450-61. [PMID: 18646989 PMCID: PMC2677334 DOI: 10.1121/1.2936368] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/15/2023]
Abstract
The goals of this study were to measure sensitivity to the direct-to-reverberant energy ratio (D/R) across a wide range of D/R values and to gain insight into which cues are used in the discrimination process. The main finding is that changes in D/R are discriminated primarily based on spectral cues. Temporal cues may be used but only when spectral cues are diminished or not available, while sensitivity to interaural cross-correlation is too low to be useful in any of the conditions tested. These findings are based on an acoustic analysis of these variables and the results of two psychophysical experiments. The first experiment employs wideband noise with two values for onset and offset times to determine the D/R just-noticeable difference at -10, 0, 10, and 20 dB D/R. This yielded substantially higher sensitivity to D/R at 0 and 10 dB D/R (2-3 dB) than has been reported previously, while sensitivity is much lower at -10 and 20 dB D/R. The second experiment consists of three parts where specific cues to D/R are reduced or removed, which enabled the specified rank ordering of the cues. The acoustic analysis and psychophysical experiments also provide an explanation for the "auditory horizon effect."
Collapse
Affiliation(s)
- Erik Larsen
- Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801, USA.
| | | | | | | |
Collapse
|
75
|
Takahashi TT, Keller CH, Nelson BS, Spitzer MW, Bala ADS, Whitchurch EA. Object localization in cluttered acoustical environments. BIOLOGICAL CYBERNETICS 2008; 98:579-586. [PMID: 18491167 DOI: 10.1007/s00422-008-0232-2] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/01/2007] [Accepted: 03/15/2008] [Indexed: 05/26/2023]
Abstract
In nature, sounds from objects of interest arrive at the ears accompanied by sound waves from other actively emitting objects and by reflections off of nearby surfaces. Despite the fact that all of these waveforms sum at the eardrums, humans with normal hearing effortlessly segregate one sound source from another. Our laboratory is investigating the neural basis of this perceptual feat, often called the "cocktail party effect", using the barn owl as an animal model. The barn owl, renowned for its ability to localize sounds and its spatiotopic representation of auditory space, is an established model for spatial hearing. Here, we briefly review the neural basis of sound-localization of a single sound source in an anechoic environment and then generalize the ideas developed therein to cases in which there are multiple, concomitant sound sources and acoustical reflection.
Collapse
Affiliation(s)
- T T Takahashi
- Institute of Neuroscience, University of Oregon, Eugene, OR 97405, USA.
| | | | | | | | | | | |
Collapse
|
76
|
Kim DO, Moiseff A, Bradley Turner J, Gull J. Acoustic cues underlying auditory distance in barn owls. Acta Otolaryngol 2008; 128:382-7. [PMID: 18368570 DOI: 10.1080/00016480701840114] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Abstract
CONCLUSIONS We conclude that: (1) among several cues examined, the monaural cue of direct-to-reverberant (D/R) ratio in the ipsilateral ear provides the most information about sound-source distance; (2) interaural level difference (ILD) provides less information about sound-source distance; and (3) a comprehensive theory of three-dimensional auditory localization must incorporate the fact that all of the major acoustic cues change with distance. OBJECTIVE Neural mechanisms underlying auditory localization of distance are poorly understood. The present study was an initial step toward filling this gap in knowledge. MATERIALS AND METHODS The binaural room impulse responses of adult barn owls were measured. The sound source was placed at various distances (up to 80 cm) and azimuths (0-90 degrees) relative to the owl's head, with the elevation kept at 0 degrees . RESULTS We determined the value of each cue for a 3-10 kHz band, and found that: (1) D/R ratio of signal amplitudes provided the most information about sound-source distance; (2) the ipsilateral D/R ratio represented distance more clearly than the contralateral or binaural-average D/R ratios; (3) ILD of direct signals increased with decreasing distance under certain conditions; (3) interaural time difference (ITD) of direct signals increased with decreasing distance at 90 degrees azimuth; and (4) the spectral patterns of ILD and the monaural direct signals changed with distance in complex ways.
Collapse
|
77
|
Simon HJ, Levitt H. Effect of dual sensory loss on auditory localization: implications for intervention. Trends Amplif 2007; 11:259-72. [PMID: 18003869 PMCID: PMC4111533 DOI: 10.1177/1084713807308209] [Citation(s) in RCA: 14] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Our sensory systems are remarkable in several respects. They are extremely sensitive, they each perform more than one function, and they interact in a complementary way, thereby providing a high degree of redundancy that is particularly helpful should one or more sensory systems be impaired. In this article, the problem of dual hearing and vision loss is addressed. A brief description is provided on the use of auditory cues in vision loss, the use of visual cues in hearing loss, and the additional difficulties encountered when both sensory systems are impaired. A major focus of this article is the use of sound localization by normal hearing, hearing impaired, and blind individuals and the special problem of sound localization in people with dual sensory loss.
Collapse
Affiliation(s)
- Helen J Simon
- Smith-Kettlewell Eye Research Institute, San Francisco, California, USA.
| | | |
Collapse
|
78
|
Mercado E, Green SR, Schneider JN. Understanding auditory distance estimation by humpback whales: a computational approach. Behav Processes 2007; 77:231-42. [PMID: 18068910 DOI: 10.1016/j.beproc.2007.10.007] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2007] [Revised: 10/07/2007] [Accepted: 10/09/2007] [Indexed: 11/26/2022]
Abstract
Ranging, the ability to judge the distance to a sound source, depends on the presence of predictable patterns of attenuation. We measured long-range sound propagation in coastal waters to assess whether humpback whales might use frequency degradation cues to range singing whales. Two types of neural networks, a multi-layer and a single-layer perceptron, were trained to classify recorded sounds by distance traveled based on their frequency content. The multi-layer network successfully classified received sounds, demonstrating that the distorting effects of underwater propagation on frequency content provide sufficient cues to estimate source distance. Normalizing received sounds with respect to ambient noise levels increased the accuracy of distance estimates by single-layer perceptrons, indicating that familiarity with background noise can potentially improve a listening whale's ability to range. To assess whether frequency patterns predictive of source distance were likely to be perceived by whales, recordings were pre-processed using a computational model of the humpback whale's peripheral auditory system. Although signals processed with this model contained less information than the original recordings, neural networks trained with these physiologically based representations estimated source distance more accurately, suggesting that listening whales should be able to range singers using distance-dependent changes in frequency content.
Collapse
Affiliation(s)
- E Mercado
- Department of Psychology, University at Buffalo, The State University of New York, Buffalo, NY 14260, USA.
| | | | | |
Collapse
|
79
|
Akeroyd MA, Gatehouse S, Blaschke J. The detection of differences in the cues to distance by elderly hearing-impaired listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 121:1077-89. [PMID: 17348530 PMCID: PMC3563070 DOI: 10.1121/1.2404927] [Citation(s) in RCA: 23] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2023]
Abstract
This experiment measured the capability of hearing-impaired individuals to discriminate differences in the cues to the distance of spoken sentences. The stimuli were generated synthetically, using a room-image procedure to calculate the direct sound and first 74 reflections for a source placed in a 7 x 9 m room, and then presenting each of those sounds individually through a circular array of 24 loudspeakers. Seventy-seven listeners participated, aged 22-83 years and with hearing levels from -5 to 59 dB HL. In conditions where a substantial change in overall level due to the inverse-square law was available as a cue, the elderly hearing-impaired listeners did not perform any different from control groups. In other conditions where that cue was unavailable (so leaving the direct-to-reverberant relationship as a cue), either because the reverberant field dominated the direct sound or because the overall level had been artificially equalized, hearing-impaired listeners performed worse than controls. There were significant correlations with listeners' self-reported distance capabilities as measured by the "Speech, Spatial, and Qualities of Hearing" questionnaire [S. Gatehouse and W. Noble, Int. J. Audiol. 43, 85-99 (2004)]. The results demonstrate that hearing-impaired listeners show deficits in the ability to use some of the cues which signal auditory distance.
Collapse
Affiliation(s)
- Michael A Akeroyd
- MRC Institute of Hearing Research (Scottish Section), Glasgow Royal Infirmary, Alexandra Parade, Glasgow, G31 2ER, United Kingdom.
| | | | | |
Collapse
|
80
|
GOERLITZ HR, SIEMERS BM. Sensory ecology of prey rustling sounds: acoustical features and their classification by wild Grey Mouse Lemurs. Funct Ecol 2007. [DOI: 10.1111/j.1365-2435.2006.01212.x] [Citation(s) in RCA: 34] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
81
|
KITAGAWA NORIMICHI, SPENCE CHARLES. Audiotactile multisensory interactions in human information processing. JAPANESE PSYCHOLOGICAL RESEARCH 2006. [DOI: 10.1111/j.1468-5884.2006.00317.x] [Citation(s) in RCA: 48] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
82
|
Russell MK, Schneider AL. Sound Source Perception in a Two-Dimensional Setting: Comparison of Action and Nonaction-Based Response Tasks. ECOLOGICAL PSYCHOLOGY 2006. [DOI: 10.1207/s15326969eco1803_4] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/31/2022]
|
83
|
Abstract
Robust perception requires that information from by our five different senses be combined at some central level to produce a single unified percept of the world. Recent theory and evidence from many laboratories suggests that the combination does not occur in a rigid, hardwired fashion, but follows flexible situation-dependent rules that allow information to be combined with maximal efficiency. In this review we discuss recent evidence from our laboratories investigating how information from auditory and visual modalities is combined. The results support the notion of Bayesian combination. We also examine temporal alignment of auditory and visual signals, and show that perceived simultaneity does not depend solely on neural latencies, but involves active processes that compensate, for example, for the physical delay introduced by the relatively slow speed of sound. Finally, we go on to show that although visual and auditory information is combined to maximize efficiency, attentional resources for the two modalities are largely independent.
Collapse
Affiliation(s)
- David Burr
- Dipartimento di Psicologia, Università degli Studi di Firenze, Via S. Nicolò 89, Firenze, Italy.
| | | |
Collapse
|
84
|
Watkins AJ. Perceptual compensation for effects of reverberation in speech identification. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2005; 118:249-62. [PMID: 16119347 DOI: 10.1121/1.1923369] [Citation(s) in RCA: 27] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
Listeners were asked to identify modified recordings of the words "sir" and "stir," which were spoken by an adult male British-English speaker. Steps along a continuum between the words were obtained by a pointwise interpolation of their temporal-envelopes. These test words were embedded in a longer "context" utterance, and played with different amounts of reverberation. Increasing only the test-word's reverberation shifts the listener's category boundary so that more "sir"-identifications are made. This effect reduces when the context's reverberation is also increased, indicating perceptual compensation that is informed by the context. Experiment 1 finds that compensation is more prominent in rapid speech, that it varies between rooms, that it is more prominent when the test-word's reverberation is high, and that it increases with the context's reverberation. Further experiments show that compensation persists when the room is switched between the context and the test word, when presentation is monaural, and when the context is reversed. However, compensation reduces when the context's reverberation pattern is reversed, as well as when noise-versions of the context are used. "Tails" that reverberation introduces at the ends of sounds and at spectral transitions may inform the compensation mechanism about the amount of reflected sound in the signal.
Collapse
Affiliation(s)
- Anthony J Watkins
- School of Psychology, The University of Reading, Reading RG6 6AL, United Kingdom.
| |
Collapse
|
85
|
Abstract
Recent studies have shown that our ability to match the timing of inputs from objects that can be both seen and heard is highly adaptable and takes into account target depth and the relatively slow speed of sound.
Collapse
|
86
|
Shinn-Cunningham BG, Kopco N, Martin TJ. Localizing nearby sound sources in a classroom: binaural room impulse responses. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2005; 117:3100-15. [PMID: 15957778 DOI: 10.1121/1.1872572] [Citation(s) in RCA: 54] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/03/2023]
Abstract
Binaural room impulse responses (BRIRs) were measured in a classroom for sources at different azimuths and distances (up to 1 m) relative to a manikin located in four positions in a classroom. When the listener is far from all walls, reverberant energy distorts signal magnitude and phase independently at each frequency, altering monaural spectral cues, interaural phase differences, and interaural level differences. For the tested conditions, systematic distortion (comb-filtering) from an early intense reflection is only evident when a listener is very close to a wall, and then only in the ear facing the wall. Especially for a nearby source, interaural cues grow less reliable with increasing source laterality and monaural spectral cues are less reliable in the ear farther from the sound source. Reverberation reduces the magnitude of interaural level differences at all frequencies; however, the direct-sound interaural time difference can still be recovered from the BRIRs measured in these experiments. Results suggest that bias and variability in sound localization behavior may vary systematically with listener location in a room as well as source location relative to the listener, even for nearby sources where there is relatively little reverberant energy.
Collapse
Affiliation(s)
- Barbara G Shinn-Cunningham
- Boston University Hearing Research Center and Departments of Cognitive and Neural Systems and Biomedical Engineering, 677 Beacon Street, Boston, Massachusetts 02215, USA.
| | | | | |
Collapse
|
87
|
Voss P, Lassonde M, Gougoux F, Fortin M, Guillemot JP, Lepore F. Early- and late-onset blind individuals show supra-normal auditory abilities in far-space. Curr Biol 2005; 14:1734-8. [PMID: 15458644 DOI: 10.1016/j.cub.2004.09.051] [Citation(s) in RCA: 199] [Impact Index Per Article: 10.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/18/2004] [Revised: 08/10/2004] [Accepted: 08/10/2004] [Indexed: 10/26/2022]
Abstract
Blind individuals manifest remarkable abilities in navigating through space despite their lack of vision. They have previously been shown to perform normally or even supra-normally in tasks involving spatial hearing in near space, a region that, however, can be calibrated with sensory-motor feedback. Here we show that blind individuals not only properly map auditory space beyond their peri-personal environment but also demonstrate supra-normal performance when subtle acoustic cues for target location and distance must be used to carry out the task. Moreover, it is generally postulated that such abilities rest in part on cross-modal cortical reorganizations, particularly in the immature brain, where important synaptogenesis is still possible. Nonetheless, we show for the first time that even late-onset blind subjects develop above-normal spatial abilities, suggesting that significant compensation can occur in the adult.
Collapse
Affiliation(s)
- Patrice Voss
- Centre de Recherche en Neuropsychologie et Cognition, Université de Montréal, Case Postale 6128, Succursale Centre-Ville, Montréal, Québec, H3C 3J7, Canada
| | | | | | | | | | | |
Collapse
|
88
|
Alais D, Carlile S. Synchronizing to real events: subjective audiovisual alignment scales with perceived auditory depth and speed of sound. Proc Natl Acad Sci U S A 2005; 102:2244-7. [PMID: 15668388 PMCID: PMC548526 DOI: 10.1073/pnas.0407034102] [Citation(s) in RCA: 63] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Because of the slow speed of sound relative to light, acoustic and visual signals from a distant event often will be received asynchronously. Here, using acoustic signals with a robust cue to sound source distance, we show that judgments of perceived temporal alignment with a visual marker depend on the depth simulated in the acoustic signal. For distant sounds, a large delay of sound relative to vision is required for the signals to be perceived as temporally aligned. For nearer sources, the time lag corresponding to audiovisual alignment is smaller and scales at rate approximating the speed of sound. Thus, when robust cues to auditory distance are present, the brain can synchronize disparate audiovisual signals to external events despite considerable differences in time of arrival at the perceiver. This ability is functionally important as it allows auditory and visual signals to be synchronized to the external event that caused them.
Collapse
Affiliation(s)
- David Alais
- Auditory Neuroscience Laboratory, Department of Physiology and Institute for Biomedical Research, School of Medical Science, University of Sydney, Sydney NSW 2006, Australia.
| | | |
Collapse
|
89
|
Oving AB, Veltman JA, Bronkhorst AW. Effectiveness of 3-D Audio for Warnings in the Cockpit. ACTA ACUST UNITED AC 2004. [DOI: 10.1207/s15327108ijap1403_3] [Citation(s) in RCA: 17] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/31/2022]
|
90
|
Blumsack JT. Audiological assessment, rehabilitation, and spatial hearing considerations associated with visual impairment in adults: an overview. Am J Audiol 2004; 12:76-83. [PMID: 14964322 DOI: 10.1044/1059-0889(2003/014)] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022] Open
Abstract
The purpose of this paper is to increase awareness and interest among audiologists regarding the needs of adults who have both hearing loss and visual impairment, particularly people who are blind and travel independently. Case history, audiometric testing, and rehabilitation issues are considered, and extensive discussion of spatial hearing research as it relates to orientation and mobility is provided.
Collapse
Affiliation(s)
- Judith T Blumsack
- Department of Communication Sciences and Disorders, Florida State University, Tallahassee 32306-1200, USA.
| |
Collapse
|
91
|
|
92
|
Dent ML, Dooling RJ. Investigations of the precedence effect in budgerigars: the perceived location of auditory images. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2003; 113:2159-2169. [PMID: 12703726 DOI: 10.1121/1.1560161] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/24/2023]
Abstract
The perceived location of auditory images has been recently studied in budgerigars [Dent and Dooling, J. Acoust. Soc. Am. 113, 2146-2158 (2003)]. Those results suggested that budgerigars (Melopsittacus undulatus) perceive precedence effect stimuli in a manner similar to humans and other animals. Here we extend those experiments to include the effects of intensity on the perceived location of auditory images and the perceived location of paired stimuli from multiple locations in space. We measured the abilities of budgerigars to discriminate between paired stimuli separated in time, intensity, and/or location. Increasing the intensity of a lag stimulus disrupted localization dominance. Budgerigars also perceived simultaneously presented (away from the midline) stimuli as very similar to a single sound presented from the midline, much like the phantom image reported in humans. The perception of paired stimuli from one side of the head versus two sides of the head was also examined and showed that the spatial cues available in these stimuli are important and that echoes are not perceptually inaccessible during localization dominance conditions. The results from these experiments add further data showing the precedence effect in budgerigars is similar to that found in humans and other animals.
Collapse
Affiliation(s)
- Micheal L Dent
- Department of Psychology, University of Maryland, College Park, Maryland 20742, USA
| | | |
Collapse
|
93
|
Schnupp JWH, Booth J, King AJ. Modeling individual differences in ferret external ear transfer functions. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2003; 113:2021-30. [PMID: 12703713 DOI: 10.1121/1.1547460] [Citation(s) in RCA: 41] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/22/2023]
Abstract
Individual variations in head and outer ear size, as well as growth of these structures during development, can markedly alter the values of the binaural and monaural cues which form the basis for auditory localization. This study investigated individual differences in the directional component of the head-related transfer function of both adult and juvenile ferrets. In line with previous studies in humans and cats, intersubject spectral differences were found to be reduced by scaling one of the directional transfer functions on a log-frequency axis. The optimal scale factor correlated most highly with pinna cavity height. Optimal frequency scaling reduced interear spectral difference equally well for adult-juvenile comparisons as for comparisons between pairs of adult ears. This illustrates that the developmental changes in localization cue values should be at least partly predictable on the basis of the expected growth rate of the outer ear structures. Predictions of interaural time differences (ITDs) were also derived from the physical dimensions of the head. ITDs were found to be poorly fitted by the spherical head model, while much better predictions could be derived from a model based on von Mises spherical basis functions. Together, these findings show how more accurate estimates of spatial cue values can be made from knowledge of the dimensions of the head and outer ears, and may facilitate the generation of virtual acoustic space stimuli in the absence of acoustical measurements from individual subjects.
Collapse
Affiliation(s)
- Jan W H Schnupp
- University Laboratory of Physiology, Parks Road, Oxford OX1 3PT, United Kingdom.
| | | | | |
Collapse
|
94
|
Zahorik P. Direct-to-reverberant energy ratio sensitivity. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2002; 112:2110-2117. [PMID: 12430822 DOI: 10.1121/1.1506692] [Citation(s) in RCA: 38] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/24/2023]
Abstract
Although the ratio of direct-to-reverberant sound energy is known to be an important acoustic cue to sound source distance, human sensitivity to changes in this cue is largely unknown. Here, direct-to-reverberant energy discrimination thresholds were measured for six listeners using virtual sound source techniques that allow for convenient and precise control of this stimulus parameter. Four different types of source stimuli were tested: a 50 ms noise burst with abrupt onset/offset, a 300 ms duration noise burst with gradual onset/offset, a speech syllable, and an impulse. Over a range of direct-to-reverberant ratios from 0 to 20 dB, an adaptive 2AFC procedure (3-down, 1-up) was used to measure discrimination thresholds. For all stimuli, these thresholds ranged from 5 to 6 dB. A post hoc fitting procedure confirmed that slopes of the psychometric functions were homogeneous across stimulus conditions and listeners. These threshold results suggest that direct-to-reverberant energy ratio by itself provides only a course coding of sound source distance, because threshold values correspond to greater than 2-fold changes in physical distance for the acoustic environment under examination.
Collapse
Affiliation(s)
- Pavel Zahorik
- Waisman Center, University of Wisconsin-Madison, 53705, USA.
| |
Collapse
|
95
|
Abstract
In the present study we report neuropsychological evidence of the existence of an auditory peripersonal space representation around the head in humans and its characteristics. In a group of right brain-damaged patients with tactile extinction, we found that a sound delivered near the ipsilesional side of the head (20 cm) strongly extinguished a tactile stimulus delivered to the contralesional side of the head (cross-modal auditory-tactile extinction). By contrast, when an auditory stimulus was presented far from the head (70 cm), cross-modal extinction was dramatically reduced. This spatially specific cross-modal extinction was most consistently found (i.e., both in the front and back spaces) when a complex sound was presented, like a white noise burst. Pure tones produced spatially specific cross-modal extinction when presented in the back space, but not in the front space. In addition, the most severe cross-modal extinction emerged when sounds came from behind the head, thus showing that the back space is more sensitive than the front space to the sensory interaction of auditory-tactile inputs. Finally, when cross-modal effects were investigated by reversing the spatial arrangement of cross-modal stimuli (i.e., touch on the right and sound on the left), we found that an ipsilesional tactile stimulus, although inducing a small amount of cross-modal tactile-auditory extinction, did not produce any spatial-specific effect. Therefore, the selective aspects of cross-modal interaction found near the head cannot be explained by a competition between a damaged left spatial representation and an intact right spatial representation. Thus, consistent with neurophysiological evidence from monkeys, our findings strongly support the existence, in humans, of an integrated cross-modal system coding auditory and tactile stimuli near the body, that is, in the peripersonal space.
Collapse
Affiliation(s)
- Alessandro Farnè
- CNC Centro Neuroscienze Cognitive, Dipartimento di Psicologia, Università di Bologna, Italy.
| | | |
Collapse
|
96
|
Philbeck JW, Mershon DH. Knowledge about typical source output influences perceived auditory distance. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2002; 111:1980-1983. [PMID: 12051417 DOI: 10.1121/1.1471899] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/23/2023]
|
97
|
Zahorik P. Assessing auditory distance perception using virtual acoustics. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2002; 111:1832-1846. [PMID: 12002867 DOI: 10.1121/1.1458027] [Citation(s) in RCA: 74] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/23/2023]
Abstract
In most naturally occurring situations, multiple acoustic properties of the sound reaching a listener's ears change as sound source distance changes. Because many of these acoustic properties, or cues, can be confounded with variation in the acoustic properties of the source and the environment, the perceptual processes subserving distance localization likely combine and weight multiple cues in order to produce stable estimates of sound source distance. Here, this cue-weighting process is examined psychophysically, using a method of virtual acoustics that allows precise measurement and control of the acoustic cues thought to be salient for distance perception in a representative large-room environment. Though listeners' judgments of sound source distance are found to consistently and exponentially underestimate true distance, the perceptual weight assigned to two primary distance cues (intensity and direct-to-reverberant energy ratio) varies substantially as a function of both sound source type (noise and speech) and angular position (0 degrees and 90 degrees relative to the median plane). These results suggest that the cue-weighting process is flexible, and able to adapt to individual distance cues that vary as a result of source properties and environmental conditions.
Collapse
Affiliation(s)
- Pavel Zahorik
- Waisman Center, University of Wisconsin-Madison, 53705, USA.
| |
Collapse
|
98
|
|
99
|
Abstract
BACKGROUND Visual capture is an important perceptual phenomenon in which the spatial location of a visual target influences the perceived location of a related auditory target. Little is known about visual capture in a distance dimension. METHODS Two groups of listeners judged the apparent distances of five loudspeaker sound sources, extending from 1 to 5 m. In one group, each listener was allowed to view the loudspeaker array. In the second group, listeners were blindfolded for the duration of the experiment. RESULTS No visual capture effects were observed. Instead, the addition of vision was found to both improve distance judgment accuracy and lower judgment variability compared with the auditory-only stimulus. Auditory-only accuracy was found to substantially improve over the course of the experiment, however. CONCLUSIONS Visual capture in distance is perhaps less general than suggested by past research, a result that has important implications for the display of spatial layout under conditions where vision is either missing or degraded.
Collapse
Affiliation(s)
- P Zahorik
- Department of Psychology, University of California, Santa Barbara, 93106-9660, USA.
| |
Collapse
|
100
|
Abstract
At a listener's ears, sound source power and sound source distance are confounded in measures of acoustic intensity, a physical property long thought to be the primary determinate of loudness. Although the relationship between sound source loudness and power is well known when source distance is fixed, relatively little is known about source loudness under conditions of varying distance. Here we show a robust loudness constancy, similar in many ways to visual size constancy, that results under distance-varying conditions that produce inaccurate estimates of source distance. Our results suggest that the auditory system does not require accurate distance estimates to judge source loudness, even when distance is variable. We offer an alternative explanation of loudness constancy based solely on a reverberant sound energy cue.
Collapse
Affiliation(s)
- P Zahorik
- Department of Psychology and Waisman Center, University of Wisconsin-Madison, Madison, Wisconsin 53705, USA
| | | |
Collapse
|