601
|
Production and Perception of Speech Intonation in Pediatric Cochlear Implant Recipients and Individuals with Normal Hearing. Ear Hear 2008; 29:336-51. [DOI: 10.1097/aud.0b013e318168d94d] [Citation(s) in RCA: 98] [Impact Index Per Article: 6.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
602
|
Bingabr M, Espinoza-Varas B, Loizou PC. Simulating the effect of spread of excitation in cochlear implants. Hear Res 2008; 241:73-9. [PMID: 18556160 DOI: 10.1016/j.heares.2008.04.012] [Citation(s) in RCA: 68] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/06/2007] [Revised: 04/26/2008] [Accepted: 04/28/2008] [Indexed: 10/22/2022]
Abstract
A model was developed to simulate acoustically the effects of excitation spread in cochlear implants (CI). Based on neurophysiologic data, the proposed model simulates the electrical-current decay rate associated with broad and narrow types of excitation, such as those produced by monopolar and bipolar electrode configurations. The effect of excitation spread on speech intelligibility was simulated in normal-hearing subjects by varying the slopes of the synthesis bands in the noise vocoder. Sentences and monosyllabic words processed via 4-16 channels of stimulation with varying degrees of excitation spread were presented to normal-hearing listeners for identification. Results showed significant interaction between spectral resolution (number of channels) and spread of excitation. The effect of narrowing the excitation spread was minimal when the spectral resolution was sufficiently good (>8 channels) but it was significant when the spectral resolution was poor (4 channels). A significant decrement in performance was observed for extremely narrow excitation spread. This outcome is partly consistent with behavioral data obtained with cochlear implant studies in that CI users tend to do as well or better with monopolar stimulation than with bipolar stimulation.
Collapse
Affiliation(s)
- Mohamed Bingabr
- Department of Engineering and Physics, University of Central Oklahoma, 100 North University Drive, Edmond, OK 73034, USA.
| | | | | |
Collapse
|
603
|
Apoux F, Bacon SP. Differential contribution of envelope fluctuations across frequency to consonant identification in quiet. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2008; 123:2792. [PMID: 18529195 PMCID: PMC2811548 DOI: 10.1121/1.2897916] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/03/2006] [Revised: 01/22/2008] [Accepted: 02/25/2008] [Indexed: 05/26/2023]
Abstract
Two experiments investigated the effects of critical bandwidth and frequency region on the use of temporal envelope cues for speech. In both experiments, spectral details were reduced using vocoder processing. In experiment 1, consonant identification scores were measured in a condition for which the cutoff frequency of the envelope extractor was half the critical bandwidth (HCB) of the auditory filters centered on each analysis band. Results showed that performance is similar to those obtained in conditions for which the envelope cutoff was set to 160 Hz or above. Experiment 2 evaluated the impact of setting the cutoff frequency of the envelope extractor to values of 4, 8, and 16 Hz or to HCB in one or two contiguous bands for an eight-band vocoder. The cutoff was set to 16 Hz for all the other bands. Overall, consonant identification was not affected by removing envelope fluctuations above 4 Hz in the low- and high-frequency bands. In contrast, speech intelligibility decreased as the cutoff frequency was decreased in the midfrequency region from 16 to 4 Hz. The behavioral results were fairly consistent with a physical analysis of the stimuli, suggesting that clearly measurable envelope fluctuations cannot be attenuated without affecting speech intelligibility.
Collapse
Affiliation(s)
- Frédéric Apoux
- Psychoacoustics Laboratory, Department of Speech and Hearing Science, Arizona State University, PO Box 870102, Tempe, Arizona 85287-0102, USA.
| | | |
Collapse
|
604
|
Izzo AD, Walsh JT, Ralph H, Webb J, Bendett M, Wells J, Richter CP. Laser stimulation of auditory neurons: effect of shorter pulse duration and penetration depth. Biophys J 2008; 94:3159-66. [PMID: 18192375 PMCID: PMC2275688 DOI: 10.1529/biophysj.107.117150] [Citation(s) in RCA: 100] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2007] [Accepted: 10/31/2007] [Indexed: 12/25/2022] Open
Abstract
We have pioneered what we believe is a novel method of stimulating cochlear neurons, using pulsed infrared radiation, based on the hypothesis that optical radiation can provide more spatially selective stimulation of the cochlea than electric current. Very little of the available optical parameter space has been used for optical stimulation of neurons. Here, we use a pulsed diode laser (1.94 microm) to stimulate auditory neurons of the gerbil. Radiant exposures measured at CAP threshold are similar for pulse durations of 5, 10, 30, and 100 micros, but greater for 300-micros-long pulses. There is evidence that water absorption of optical radiation is a significant factor in optical stimulation. Heat-transfer-based analysis of the data indicates that potential structures involved in optical stimulation of cochlear neurons have a dimension on the order of approximately 10 microm. The implications of these data could direct further research and design of an optical cochlear implant.
Collapse
Affiliation(s)
- Agnella D Izzo
- Department of Otolaryngology, Northwestern University, Chicago, Illinois 60611, USA.
| | | | | | | | | | | | | |
Collapse
|
605
|
Goupell MJ, Laback B, Majdak P, Baumgartner WD. Effects of upper-frequency boundary and spectral warping on speech intelligibility in electrical stimulation. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2008; 123:2295-309. [PMID: 18397034 PMCID: PMC3061454 DOI: 10.1121/1.2831738] [Citation(s) in RCA: 12] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
Speech understanding was tested for seven listeners using 12-electrode Med-El cochlear implants (CIs) and six normal-hearing listeners using a CI simulation. Eighteen different types of processing were evaluated, which varied the frequency-to-tonotopic place mapping and the upper boundary of the frequency and stimulation range. Spectrally unwarped and warped conditions were included. Unlike previous studies on this topic, the lower boundary of the frequency and stimulation range was fixed while the upper boundary was varied. For the unwarped conditions, only eight to ten channels were needed in both quiet and noise to achieve no significant degradation in speech understanding compared to the normal 12-electrode speech processing. The unwarped conditions were often the best conditions for understanding speech; however, small changes in frequency-to-place mapping (<0.77 octaves for the most basal electrode) yielded no significant degradation in performance from the nearest unwarped condition. A second experiment measured the effect of feedback training for both the unwarped and warped conditions. Improvements were found for the unwarped and frequency-expanded conditions, but not for the compressed condition. These results have implications for new CI processing strategies, such as the inclusion of spectral localization cues.
Collapse
Affiliation(s)
- Matthew J Goupell
- Acoustics Research Institute, Austrian Academy of Sciences, Wohllebengasse 12-14, A-1040 Vienna, Austria.
| | | | | | | |
Collapse
|
606
|
Hallum LE, Chen SC, Cloherty SL, Lovell NH. Psychophysics of prosthetic vision: II. stochastic sampling, the phosphene image, and noise. CONFERENCE PROCEEDINGS : ... ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL CONFERENCE 2008; 2006:1634-7. [PMID: 17946471 DOI: 10.1109/iembs.2006.259957] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Stimulation of the diseased retina via an intraocular electrode array is a proposed means of restoring some vision to the profoundly blind. A prosthetic device to this end would involve post-implantation calibration (analogous to cochlear implant fitting), wherein the subject indicates those discrete positions in the visual field where luminous percepts are elicited. This procedure would be a source of noise, because the indicated positions would only approximate the actual positions in the visual field. Put differently, the procedure introduces sampling jitter, and would therefore affect clinical outcomes such as mobility and reading speeds. The nature of this noise is the concern of the present paper; we derive an expression for the noise power spectrum as it relates to the statistical nature of the sampling jitter. We show that, generally, jitter has greater effect on higher spatial-frequencies, that is, those areas of the implantee's visual perception that represent fine detail are more prone to noise. More specifically, the noise spectrum depends on the characteristic function of the random variable describing the sampling jitter. Our results signal the need for experimental work that characterizes sampling jitter in implantees, plus the need for simulations that allow a better understanding of perception and the noisy phosphene image.
Collapse
Affiliation(s)
- Luke E Hallum
- Graduate Sch. of Biomed. Eng., New South Wales Univ., Sydney, NSW
| | | | | | | |
Collapse
|
607
|
Lim HH, Lenarz T, Anderson DJ, Lenarz M. The auditory midbrain implant: effects of electrode location. Hear Res 2008; 242:74-85. [PMID: 18348902 DOI: 10.1016/j.heares.2008.02.003] [Citation(s) in RCA: 44] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/03/2007] [Revised: 02/06/2008] [Accepted: 02/11/2008] [Indexed: 11/19/2022]
Abstract
The auditory midbrain implant (AMI) is a new hearing prosthesis designed for stimulation of the inferior colliculus in patients who do not receive sufficient benefit from cochlear or brainstem prostheses. We have begun clinical trials in which three patients have been implanted with the AMI. Although the intended target was the central nucleus of the inferior colliculus (ICC), the electrode array was implanted into different locations across patients (i.e., ICC, dorsal cortex of inferior colliculus, lateral lemniscus). In this paper, we will summarize the effects of electrical stimulation of these different midbrain regions on various psychophysical properties and speech perception performance. The patient implanted within the intended target, the ICC, exhibited the greatest improvements in hearing performance. However, this patient has not yet achieved open-set speech perception to the performance level typically observed for cochlear implant patients, which we believe is partially due to the location of the array within the ICC. We will present findings from previous AMI studies in guinea pigs demonstrating the existence of spatially distinct functional output regions within the ICC and suggesting that further improvements in performance may be achieved by stimulating within a rostral-ventral region. Remaining questions include if a similar organization exists in the human ICC and if stimulation of its rostral-ventral region with currently available strategies (i.e., those designed for cochlear implants) can restore sufficient speech perception.
Collapse
Affiliation(s)
- Hubert H Lim
- Otorhinolaryngology Department, Hannover Medical University, Carl-Neuberg-Strasse1, Gebaeude K5, Ebene 1, Raum 4010, Hannover 30625, Germany.
| | | | | | | |
Collapse
|
608
|
Stohl JS, Throckmorton CS, Collins LM. Assessing the pitch structure associated with multiple rates and places for cochlear implant users. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2008; 123:1043-1053. [PMID: 18247906 DOI: 10.1121/1.2821980] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
Cochlear implant subjects continue to experience difficulty understanding speech in noise and performing pitch-based musical tasks. Acoustic model studies have suggested that transmitting additional fine structure via multiple stimulation rates is a potential mechanism for addressing these issues [Nie et al., IEEE Trans. Biomed. Eng. 52, 64-73 (2005); Throckmorton et al., Hear. Res. 218, 30-42 (2006)]; however, results from preliminary cochlear implant studies have been less compelling. Multirate speech processing algorithms previously assumed a place-dependent pitch structure in that a basal electrode would always elicit a higher pitch percept than an apical electrode, independent of stimulation rate. Some subjective evidence contradicts this assumption [H. J. McDermott and C. M. McKay, J. Acoust. Soc. Am. 101, 1622-1630 (1997); R. V. Shannon, Hear. Res. 11, 157-189 (1983)]. The purpose of this study is to test the hypothesis that the introduction of multiple rates may invalidate the tonotopic pitch structure resulting from place-pitch alone. The SPEAR3 developmental speech processor was used to collect psychophysical data from five cochlear implant users to assess the tonotopic structure for stimuli presented at two rates on all active electrodes. Pitch ranking data indicated many cases where pitch percepts overlapped across electrodes and rates. Thus, the results from this study suggest that pitch-based tuning across rate and electrode may be necessary to optimize performance of a multirate sound processing strategy in cochlear implant subjects.
Collapse
Affiliation(s)
- Joshua S Stohl
- Department of Electrical and Computer Engineering, Duke University, P.O. Box 90291, Durham, North Carolina 27708-0291, USA.
| | | | | |
Collapse
|
609
|
|
610
|
Hopkins K, Moore BCJ, Stone MA. Effects of moderate cochlear hearing loss on the ability to benefit from temporal fine structure information in speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2008; 123:1140-53. [PMID: 18247914 PMCID: PMC2688774 DOI: 10.1121/1.2824018] [Citation(s) in RCA: 134] [Impact Index Per Article: 8.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2023]
Abstract
Speech reception thresholds (SRTs) were measured with a competing talker background for signals processed to contain variable amounts of temporal fine structure (TFS) information, using nine normal-hearing and nine hearing-impaired subjects. Signals (speech and background talker) were bandpass filtered into channels. Channel signals for channel numbers above a "cut-off channel" (CO) were vocoded to remove TFS information, while channel signals for channel numbers of CO and below were left unprocessed. Signals from all channels were combined. As a group, hearing-impaired subjects benefited less than normal-hearing subjects from the additional TFS information that was available as CO increased. The amount of benefit varied between hearing-impaired individuals, with some showing no improvement in SRT and one showing an improvement similar to that for normal-hearing subjects. The reduced ability to take advantage of TFS information in speech may partially explain why subjects with cochlear hearing loss get less benefit from listening in a fluctuating background than normal-hearing subjects. TFS information may be important in identifying the temporal "dips" in such a background.
Collapse
Affiliation(s)
- Kathryn Hopkins
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, United Kingdom.
| | | | | |
Collapse
|
611
|
Stone MA, Moore BCJ. Effects of spectro-temporal modulation changes produced by multi-channel compression on intelligibility in a competing-speech task. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2008; 123:1063-76. [PMID: 18247908 DOI: 10.1121/1.2821969] [Citation(s) in RCA: 38] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
These experiments are concerned with the intelligibility of target speech in the presence of a background talker. Using a noise vocoder, Stone and Moore [J. Acoust. Soc. Am. 114, 1023-1034 (2003)] showed that single-channel fast-acting compression degraded intelligibility, but slow compression did not. Stone and Moore [J. Acoust. Soc. Am. 116, 2311-2323 (2004)] showed that intelligibility was lower when fast single-channel compression was applied to the target and background after mixing rather than before, and suggested that this was partly due to compression after mixing introducing "comodulation" between the target and background talkers. Experiment 1 here showed a similar effect for multi-channel compression. In experiment 2, intelligibility was measured as a function of the speed of multi-channel compression applied after mixing. For both eight- and 12-channel vocoders with one compressor per channel, intelligibility decreased as compression speed increased. For the eight-channel vocoder, a compressor that only affected modulation depth for rates below 2 Hz still reduced intelligibility. Experiment 3 used 12- or 18-channel vocoders. There were between 1 and 12 compression channels, and four speeds of compression. Intelligibility decreased as the number and speed of compression channels increased. The results are interpreted using several measures of the effects of compression, especially "across-source modulation correlation."
Collapse
Affiliation(s)
- Michael A Stone
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, England.
| | | |
Collapse
|
612
|
Xin Luo, Fu QJ, Galvin JJ. Vocal emotion recognition by normal-hearing listeners and cochlear implant users. Trends Amplif 2008; 11:301-15. [PMID: 18003871 DOI: 10.1177/1084713807305301] [Citation(s) in RCA: 118] [Impact Index Per Article: 7.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
The present study investigated the ability of normal-hearing listeners and cochlear implant users to recognize vocal emotions. Sentences were produced by 1 male and 1 female talker according to 5 target emotions: angry, anxious, happy, sad, and neutral. Overall amplitude differences between the stimuli were either preserved or normalized. In experiment 1, vocal emotion recognition was measured in normal-hearing and cochlear implant listeners; cochlear implant subjects were tested using their clinically assigned processors. When overall amplitude cues were preserved, normal-hearing listeners achieved near-perfect performance, whereas listeners with cochlear implant recognized less than half of the target emotions. Removing the overall amplitude cues significantly worsened mean normal-hearing and cochlear implant performance. In experiment 2, vocal emotion recognition was measured in listeners with cochlear implant as a function of the number of channels (from 1 to 8) and envelope filter cutoff frequency (50 vs 400 Hz) in experimental speech processors. In experiment 3, vocal emotion recognition was measured in normal-hearing listeners as a function of the number of channels (from 1 to 16) and envelope filter cutoff frequency (50 vs 500 Hz) in acoustic cochlear implant simulations. Results from experiments 2 and 3 showed that both cochlear implant and normal-hearing performance significantly improved as the number of channels or the envelope filter cutoff frequency was increased. The results suggest that spectral, temporal, and overall amplitude cues each contribute to vocal emotion recognition. The poorer cochlear implant performance is most likely attributable to the lack of salient pitch cues and the limited functional spectral resolution.
Collapse
Affiliation(s)
- Xin Luo
- Department of Auditory Implants and Perception, House Ear Institute, Los Angeles, California 90057, USA.
| | | | | |
Collapse
|
613
|
Auditory midbrain implant: histomorphologic effects of long-term implantation and electric stimulation of a new deep brain stimulation array. Otol Neurotol 2008; 28:1045-52. [PMID: 18043431 DOI: 10.1097/mao.0b013e318159e74f] [Citation(s) in RCA: 27] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
HYPOTHESIS Chronic implantation and electric stimulation with a human prototype auditory midbrain implant (AMI) array within the inferior colliculus achieves minimal neuronal damage and does not cause any severe complications. BACKGROUND An AMI array has been developed for patients with neural deafness and, based on animal studies, has shown to possess potential as an auditory prosthesis in humans. To investigate the safety of the AMI for clinical use, we characterized the histomorphologic effects of chronic implantation and stimulation within its target structure, the inferior colliculus. METHODS Eight cats were chronically implanted for 3 months, and histologic sections were analyzed to assess long-term tissue effects. Four of the 8 cats were additionally stimulated for 60 days (4 h/d) starting 4 weeks after implantation to assess if clinically relevant stimuli further affected the tissue response. RESULTS In general, both neurons and neuropil surrounding the implant track were apparently unaffected, whereas a fibrillary sheath (approximately 50 microm thick) developed around the array. There was a significant decrease in neuron density 50 to 100 microm away from the track with a significantly elevated number of glial cells out to approximately 250 to 350 microm. Chronic stimulation seemed to improve the tissue response and neuronal survival around the implant, although further studies are needed to confirm this finding. CONCLUSION The histomorphologic effects and extent of neuronal damage observed for our AMI array are similar to those of other neural implants currently and safely used in humans. The minimal tissue damage surrounding the implanted array is encouraging with regard to the safety of the array for human use.
Collapse
|
614
|
Sheldon S, Pichora-Fuller MK, Schneider BA. Effect of age, presentation method, and learning on identification of noise-vocoded words. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2008; 123:476-488. [PMID: 18177175 DOI: 10.1121/1.2805676] [Citation(s) in RCA: 35] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
Noise vocoding was used to investigate the ability of younger and older adults with normal audiometric thresholds in the speech range to use amplitude envelope cues to identify words. In Experiment 1, four 50-word lists were tested, with each word presented initially with one frequency band and the number of bands being incremented until it was correctly identified by the listener. Both age groups required an average of 5.25 bands for 50% correct word identification and performance improved across the four lists. In Experiment 2, the same participants who completed Experiment 1 identified words in four blocked noise-vocoded conditions (16, 8, 4, 2 bands). Compared to Experiment 1, both age groups required more bands to reach the 50% correct word identification threshold in Experiment 2, 6.13, and 8.55 bands, respectively, with younger adults outperforming older adults. Experiment 3 was identical to Experiment 2 except the participants had no prior experience with noise-vocoded speech. Again, younger adults outperformed older adults, with thresholds of 6.67 and 8.97 bands, respectively. The finding of age effects in Experiments 2 and 3, but not in Experiment 1, seems more likely to be related to differences in the presentation methods than to experience with noise vocoding.
Collapse
Affiliation(s)
- Signy Sheldon
- Department of Psychology, University of Toronto, 3359 Mississauga Road North, Mississauga, Ontario L5L 1C6, Canada
| | | | | |
Collapse
|
615
|
Xu L, Pfingst BE. Spectral and temporal cues for speech recognition: implications for auditory prostheses. Hear Res 2007; 242:132-40. [PMID: 18249077 DOI: 10.1016/j.heares.2007.12.010] [Citation(s) in RCA: 69] [Impact Index Per Article: 4.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/30/2007] [Revised: 12/16/2007] [Accepted: 12/19/2007] [Indexed: 11/30/2022]
Abstract
Features of stimulation important for speech recognition in people with normal hearing and in people using implanted auditory prostheses include spectral information represented by place of stimulation along the tonotopic axis and temporal information represented in low-frequency envelopes of the signal. The relative contributions of these features to speech recognition and their interactions have been studied using vocoder-like simulations of cochlear implant speech processors presented to listeners with normal hearing. In these studies, spectral/place information was manipulated by varying the number of channels and the temporal-envelope information was manipulated by varying the lowpass cutoffs of the envelope extractors. Consonant and vowel recognition in quiet reached plateau at 8 and 12 channels and lowpass cutoff frequencies of 16 Hz and 4 Hz, respectively. Phoneme (especially vowel) recognition in noise required larger numbers of channels. Lexical tone recognition required larger numbers of channels and higher lowpass cutoff frequencies. There was a tradeoff between spectral/place and temporal-envelope requirements. Most current auditory prostheses seem to deliver adequate temporal-envelope information, but the number of effective channels is suboptimal, particularly for speech recognition in noise, lexical tone recognition, and music perception.
Collapse
Affiliation(s)
- Li Xu
- School of Hearing, Speech and Language Sciences, Ohio University, Athens, OH 45701, USA.
| | | |
Collapse
|
616
|
Electrical stimulation of the midbrain for hearing restoration: insight into the functional organization of the human central auditory system. J Neurosci 2007; 27:13541-51. [PMID: 18057212 DOI: 10.1523/jneurosci.3123-07.2007] [Citation(s) in RCA: 57] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
The cochlear implant can restore speech perception in patients with sensorineural hearing loss. However, it is ineffective for those without an implantable cochlea or a functional auditory nerve. These patients can be implanted with the auditory brainstem implant (ABI), which stimulates the surface of the cochlear nucleus. Unfortunately, the ABI has achieved limited success in its main patient group [i.e., those with neurofibromatosis type 2 (NF2)] and requires a difficult surgical procedure. These limitations have motivated us to develop a new hearing prosthesis that stimulates the midbrain with a penetrating electrode array. We recently implanted three patients with the auditory midbrain implant (AMI), and it has proven to be safe with minimal movement over time. The AMI provides loudness, pitch, temporal, and directional cues, features that have shown to be important for speech perception and more complex sound processing. Thus far, all three patients obtain enhancements in lip reading capabilities and environmental awareness and some improvements in speech perception comparable with that of NF2 ABI patients. Considering that our midbrain target is more surgically exposable than the cochlear nucleus, this argues for the use of the AMI as an alternative to the ABI. Fortunately, we were able to stimulate different midbrain regions in our patients and investigate the functional organization of the human central auditory system. These findings provide some insight into how we may need to stimulate the midbrain to improve hearing performance with the AMI.
Collapse
|
617
|
Fu QJ, Galvin JJ. Maximizing cochlear implant patients' performance with advanced speech training procedures. Hear Res 2007; 242:198-208. [PMID: 18295992 DOI: 10.1016/j.heares.2007.11.010] [Citation(s) in RCA: 89] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/04/2007] [Revised: 11/21/2007] [Accepted: 11/22/2007] [Indexed: 11/19/2022]
Abstract
Advances in implant technology and speech processing have provided great benefit to many cochlear implant patients. However, some patients receive little benefit from the latest technology, even after many years' experience with the device. Moreover, even the best cochlear implant performers have great difficulty understanding speech in background noise, and music perception and appreciation remain major challenges. Recent studies have shown that targeted auditory training can significantly improve cochlear implant patients' speech recognition performance. Such benefits are not only observed in poorly performing patients, but also in good performers under difficult listening conditions (e.g., speech noise, telephone speech, music, etc.). Targeted auditory training has also been shown to enhance performance gains provided by new implant devices and/or speech processing strategies. These studies suggest that cochlear implantation alone may not fully meet the needs of many patients, and that additional auditory rehabilitation may be needed to maximize the benefits of the implant device. Continuing research will aid in the development of efficient and effective training protocols and materials, thereby minimizing the costs (in terms of time, effort and resources) associated with auditory rehabilitation while maximizing the benefits of cochlear implantation for all recipients.
Collapse
Affiliation(s)
- Qian-Jie Fu
- Department of Auditory Implants and Perception, House Ear Institute, 2100 West Third Street, Los Angeles, CA 90057, USA.
| | | |
Collapse
|
618
|
Kuchta J. Twenty-five years of auditory brainstem implants: perspectives. ACTA NEUROCHIRURGICA. SUPPLEMENT 2007; 97:443-9. [PMID: 17691334 DOI: 10.1007/978-3-211-33081-4_51] [Citation(s) in RCA: 16] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/28/2023]
Abstract
The auditory brainstem implant (ABI) provides auditory sensations, recognition of environmental sounds and aid in spoken communication in more than 300 patients worldwide. It is no more a device under investigation but it is widely accepted for the treatment of patients who have lost hearing due to bilateral tumors of the vestibulocochlear nerve. Most of these patients are completely deaf when the implant is switched off. In contrast to the cochlear implants (CI), only few of the implanted patients achieve open-set speech recognition without the help of visual cues. In the last few years, patients with lesions other than tumors have also been implanted. Auditory perceptual performance in patients who are deaf due to trauma, cochlea aplasia or other non-tumor lesions of the cochlea or the vestibulocochlear nerve turned out to be much better than in NF2 tumor patients. Until recently, the target region for ABI implantation has been the ventral cochlear nucleus (CN). The electrodes are implanted via the translabyrinthine or retrosigmoid approach. Currently, new targets along the central auditory pathways and new, minimally invasive techniques for implantation are under investigation. These techniques may further improve auditory perceptual performance in ABI patients and provide hearing to a variety of types of central deafness.
Collapse
Affiliation(s)
- J Kuchta
- Department of Neurosurgery, Cologne University Hospital, Germany.
| |
Collapse
|
619
|
Turner CW, Reiss LAJ, Gantz BJ. Combined acoustic and electric hearing: preserving residual acoustic hearing. Hear Res 2007; 242:164-71. [PMID: 18164883 DOI: 10.1016/j.heares.2007.11.008] [Citation(s) in RCA: 104] [Impact Index Per Article: 6.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/26/2007] [Revised: 11/21/2007] [Accepted: 11/22/2007] [Indexed: 11/17/2022]
Abstract
The topic of this review is the strategy of preserving residual acoustic hearing in the implanted ear to provide combined electrical stimulation and acoustic hearing as a rehabilitative strategy for sensorineural hearing loss. This chapter will concentrate on research done with the Iowa/Nucleus 10 mm Hybrid device, but we will also attempt to summarize strategies and results from other groups around the world who use slightly different approaches. A number of studies have shown that preserving residual acoustic hearing in the implanted ear is a realistic goal for many patients with severe high-frequency hearing loss. The addition of the electric stimulation to their existing acoustic hearing can provide increased speech recognition for these patients. In addition, the preserved acoustic hearing can offer considerable advantages, as compared to a traditional cochlear implant, for tasks such as speech recognition in backgrounds or appreciation of music and other situations where the poor frequency resolution of electric stimulation has been a disadvantage.
Collapse
Affiliation(s)
- Christopher W Turner
- Department of Speech Pathology and Audiology, 121B SHC Building, University of Iowa, Iowa City, IA 52242, United States.
| | | | | |
Collapse
|
620
|
Shivdasani MN, Mauger SJ, Rathbone GD, Paolini AG. Inferior colliculus responses to multichannel microstimulation of the ventral cochlear nucleus: implications for auditory brain stem implants. J Neurophysiol 2007; 99:1-13. [PMID: 17928560 DOI: 10.1152/jn.00629.2007] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Multichannel techniques were used to assess the frequency specificity of activation in the central nucleus of the inferior colliculus (CIC) produced by electrical stimulation of localized regions within the ventral cochlear nucleus (VCN). Data were recorded in response to pure tones from 141 and 193 multiunit clusters in the rat VCN and the CIC, respectively. Of 141 VCN sites, 126 were individually stimulated while recording responses in the CIC. A variety of CIC response types were seen with an increase in both electrical and acoustic stimulation levels. The majority of sites exhibited monotonic rate-level types acoustically, whereas spike rate saturation was achieved predominantly with electrical stimulation. In 20.6% of the 364 characteristic frequency aligned VCN-CIC pairs, the CIC sites did not respond to stimulation. In 26% of the 193 CIC sites, a high correlation was observed between acoustic tuning and electrical tuning obtained through VCN stimulation. A high degree of frequency specificity was found in 58% of the 118 lowest threshold VCN-CIC pairs. This was dependent on electrode placement within the VCN because a higher degree of frequency specificity was achieved with stimulation of medial, central, and posterolateral VCN regions than more anterolateral regions. Broadness of acoustic tuning in the CIC played a role in frequency-specific activation. Narrowly tuned CIC sites showed the lowest degree of frequency specificity on stimulation of the anterolateral VCN regions. These data provide significant implications for auditory brain stem implant electrode placement, current localization, power requirements, and facilitation of information transfer to higher brain centers.
Collapse
Affiliation(s)
- Mohit N Shivdasani
- The Bionic Ear Institute, East Melbourne Victoria, Melbourne, Victoria, Australia
| | | | | | | |
Collapse
|
621
|
Whitmal NA, Poissant SF, Freyman RL, Helfer KS. Speech intelligibility in cochlear implant simulations: Effects of carrier type, interfering noise, and subject experience. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 122:2376-88. [PMID: 17902872 DOI: 10.1121/1.2773993] [Citation(s) in RCA: 35] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
Channel vocoders using either tone or band-limited noise carriers have been used in experiments to simulate cochlear implant processing in normal-hearing listeners. Previous results from these experiments have suggested that the two vocoder types produce speech of nearly equal intelligibility in quiet conditions. The purpose of this study was to further compare the performance of tone and noise-band vocoders in both quiet and noisy listening conditions. In each of four experiments, normal-hearing subjects were better able to identify tone-vocoded sentences and vowel-consonant-vowel syllables than noise-vocoded sentences and syllables, both in quiet and in the presence of either speech-spectrum noise or two-talker babble. An analysis of consonant confusions for listening in both quiet and speech-spectrum noise revealed significantly different error patterns that were related to each vocoder's ability to produce tone or noise output that accurately reflected the consonant's manner of articulation. Subject experience was also shown to influence intelligibility. Simulations using a computational model of modulation detection suggest that the noise vocoder's disadvantage is in part due to the intrinsic temporal fluctuations of its carriers, which can interfere with temporal fluctuations that convey speech recognition cues.
Collapse
Affiliation(s)
- Nathaniel A Whitmal
- Department of Communication Disorders, University of Massachusetts, Amherst, Massachusetts 01003, USA.
| | | | | | | |
Collapse
|
622
|
Drennan WR, Won JH, Dasika VK, Rubinstein JT. Effects of temporal fine structure on the lateralization of speech and on speech understanding in noise. J Assoc Res Otolaryngol 2007; 8:373-83. [PMID: 17332969 PMCID: PMC2538430 DOI: 10.1007/s10162-007-0074-y] [Citation(s) in RCA: 31] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2006] [Accepted: 01/17/2007] [Indexed: 11/26/2022] Open
Abstract
This study evaluated the role of temporal fine structure in the lateralization and understanding of speech in six normal-hearing listeners. Interaural time differences (ITDs) were introduced to invoke lateralization. Speech reception thresholds (SRTs) were evaluated in backgrounds of two-talker babble and speech-shaped noise. Two-syllable words with ITDs of 0 and 700 micros were used as targets. A vocoder technique, which systematically randomized fine structure, was used to evaluate the effects of fine structure on these tasks. Randomization of temporal fine structure was found to significantly reduce the ability of normal-hearing listeners to lateralize words, although for many listeners, good lateralization performance was achieved with as much as 80% fine-structure randomization. Most listeners demonstrated some rudimentary ability to lateralize with 100% fine-structure randomization. When ITDs were 0 micros, randomization of fine structure had a much greater effect on SRT in two-talker babble than in speech-shaped noise. Binaural advantages were also observed. In steady noise, the difference in SRT between words with 0- vs 700-micros ITDs was, on average, 6 dB with no fine-structure randomization and 2 dB with 100% fine-structure randomization. In two-talker babble this difference was 1.9 dB and, for most listeners, showed little effect of the degree of fine-structure randomization. These results suggest that (1) improved delivery of temporal fine structure would improve speech understanding in noise for implant recipients, (2) bilateral implant recipients might benefit from temporal envelope ITDs, and (3) improved delivery of temporal information could improve binaural benefits.
Collapse
Affiliation(s)
- Ward R Drennan
- VM Bloedel Hearing Research Center, University of Washington, Box 357923, Seattle, WA 98195-7923, USA.
| | | | | | | |
Collapse
|
623
|
Xu L, Zheng Y. Spectral and temporal cues for phoneme recognition in noise. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 122:1758. [PMID: 17927435 DOI: 10.1121/1.2767000] [Citation(s) in RCA: 34] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/24/2023]
Abstract
Cochlear implant users receive limited spectral and temporal information. Their speech recognition deteriorates dramatically in noise. The aim of the present study was to determine the relative contributions of spectral and temporal cues to speech recognition in noise. Spectral information was manipulated by varying the number of channels from 2 to 32 in a noise-excited vocoder. Temporal information was manipulated by varying the low-pass cutoff frequency of the envelope extractor from 1 to 512 Hz. Ten normal-hearing, native speakers of English participated in tests of phoneme recognition using vocoder processed consonants and vowels under three conditions (quiet, and +6 and 0 dB signal-to-noise ratios). The number of channels required for vowel-recognition performance to plateau increased from 12 in quiet to 16-24 in the two noise conditions. However, for consonant recognition, no further improvement in performance was evident when the number of channels was > or =12 in any of the three conditions. The contribution of temporal cues for phoneme recognition showed a similar pattern in both quiet and noise conditions. Similar to the quiet conditions, there was a trade-off between temporal and spectral cues for phoneme recognition in noise.
Collapse
Affiliation(s)
- Li Xu
- School of Hearing, Speech and Language Sciences, Ohio University, Athens, Ohio 45701, USA.
| | | |
Collapse
|
624
|
Carroll J, Zeng FG. Fundamental frequency discrimination and speech perception in noise in cochlear implant simulations. Hear Res 2007; 231:42-53. [PMID: 17604581 PMCID: PMC2034333 DOI: 10.1016/j.heares.2007.05.004] [Citation(s) in RCA: 29] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/08/2007] [Revised: 05/01/2007] [Accepted: 05/10/2007] [Indexed: 10/23/2022]
Abstract
Increasing the number of channels at low frequencies improves discrimination of fundamental frequency (F0) in cochlear implants (Geurts, L., Wouters, J., 2004. Better place-coding of the fundamental frequency in cochlear implants. J. Acoust. Soc. Am. 115 (2), 844-852). We conducted three experiments to test whether improved F0 discrimination can be translated into increased speech intelligibility in noise in a cochlear implant simulation. The first experiment measured F0 discrimination and speech intelligibility in quiet as a function of channel density over different frequency regions. The results from this experiment showed a tradeoff in performance between F0 discrimination and speech intelligibility with a limited number of channels. The second experiment tested whether improved F0 discrimination and optimizing this tradeoff could improve speech performance with a competing talker. However, improved F0 discrimination did not improve speech intelligibility in noise. The third experiment identified the critical number of channels needed at low frequencies to improve speech intelligibility in noise. The result showed that, while 16 channels below 500Hz were needed to observe any improvement in speech intelligibility in noise, even 32 channels did not achieve normal performance. Theoretically, these results suggest that without accurate spectral coding, F0 discrimination and speech perception in noise are two independent processes. Practically, the present results illustrate the need to increase the number of independent channels in cochlear implants.
Collapse
Affiliation(s)
- Jeff Carroll
- Hearing and Speech Research Laboratory, Department of Biomedical Engineering, University of California, Irvine, CA 92697, USA.
| | | |
Collapse
|
625
|
Lim HH, Anderson DJ. Spatially distinct functional output regions within the central nucleus of the inferior colliculus: implications for an auditory midbrain implant. J Neurosci 2007; 27:8733-43. [PMID: 17687050 PMCID: PMC6672938 DOI: 10.1523/jneurosci.5127-06.2007] [Citation(s) in RCA: 37] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
The inferior colliculus central nucleus (ICC) has potential as a new site for an auditory prosthesis [i.e., auditory midbrain implant (AMI)] for deaf patients who cannot benefit from cochlear implants (CIs). We have previously shown that ICC stimulation achieves lower thresholds, greater dynamic ranges, and more localized, frequency-specific primary auditory cortex (A1) activation than CI stimulation. However, we also observed that stimulation location along the caudorostral (isofrequency) dimension of the ICC affects thresholds and frequency specificity in A1, suggesting possible differences in functional (output) organization within the ICC. In this study, we electrically stimulated different regions along the isofrequency laminas of the ICC and recorded the corresponding A1 activity in ketamine-anesthetized guinea pigs using multisite probes to systematically assess ICC stimulation location effects. Our results indicate that stimulation of more rostral and somewhat ventral regions within an ICC lamina achieves lower thresholds, smaller discriminable level steps, and larger evoked potentials in A1. We also observed longer first spike latencies, which correlated with reduced spiking precision, when stimulating in more caudal and dorsal ICC regions. These findings suggest that at least two spatially distinct functional output regions exist along an ICC lamina: a caudal-dorsal region and a rostral-ventral region. The AMI will be implanted along the tonotopic axis of the ICC to achieve frequency-specific activation. However, stimulation location along the ICC laminas affects response properties that have shown to be important for speech perception performance, and needs to be considered when implanting future AMI patients.
Collapse
Affiliation(s)
- Hubert H. Lim
- Kresge Hearing Research Institute, Department of Biomedical Engineering, and Department of Electrical Engineering and Computer Science, University of Michigan, Ann Arbor, Michigan 48109
| | - David J. Anderson
- Kresge Hearing Research Institute, Department of Biomedical Engineering, and Department of Electrical Engineering and Computer Science, University of Michigan, Ann Arbor, Michigan 48109
| |
Collapse
|
626
|
Lenarz T, Lim HH, Reuter G, Patrick JF, Lenarz M. The auditory midbrain implant: a new auditory prosthesis for neural deafness-concept and device description. Otol Neurotol 2007; 27:838-43. [PMID: 16936570 DOI: 10.1097/01.mao.0000232010.01116.e9] [Citation(s) in RCA: 80] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
The auditory midbrain implant (AMI) is a new central auditory prosthesis designed for penetrating stimulation of the human inferior colliculus. The major group of candidates for the AMI consists of neurofibromatosis type 2 (NF2) patients who develop neural deafness because of growth and/or surgical removal of bilateral acoustic neuromas. Because of the absence of a viable auditory nerve, these patients cannot benefit from cochlear implants. An alternative solution has been the auditory brainstem implant (ABI), which stimulates the cochlear nucleus. However, speech perception performance in NF2 ABI patients has been limited. The fact that the ABI is able to produce high levels of speech perception in nontumor patients (with inaccessible cochleae or posttraumatic damage to the cochlear nerve) suggests that limitations in ABI performance in NF2 patients may be associated with cochlear nucleus damage caused by the tumors or the tumor removal process. Thus, stimulation of the auditory midbrain proximal to the damaged cochlear nucleus may be a better alternative for hearing restoration in NF2 patients. We propose the central nucleus of the inferior colliculus (ICC) as the potential site. A penetrating electrode array aligned along the well-defined tonotopic gradient of the ICC should selectively activate different frequency regions, which is an important elementfor supporting good speech understanding. The goal of this article is to present the ICC as an alternative site for an auditory implant for NF2 patients and to describe the design of the first human prototype AMI. Practical considerations for implementation of the AMI will also be discussed.
Collapse
Affiliation(s)
- Thomas Lenarz
- Otorhinolaryngology Department, Medical University of Hannover, Germany.
| | | | | | | | | |
Collapse
|
627
|
Litvak LM, Spahr AJ, Saoji AA, Fridman GY. Relationship between perception of spectral ripple and speech recognition in cochlear implant and vocoder listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 122:982-91. [PMID: 17672646 DOI: 10.1121/1.2749413] [Citation(s) in RCA: 144] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/08/2023]
Abstract
Spectral resolution has been reported to be closely related to vowel and consonant recognition in cochlear implant (CI) listeners. One measure of spectral resolution is spectral modulation threshold (SMT), which is defined as the smallest detectable spectral contrast in the spectral ripple stimulus. SMT may be determined by the activation pattern associated with electrical stimulation. In the present study, broad activation patterns were simulated using a multi-band vocoder to determine if similar impairments in speech understanding scores could be produced in normal-hearing listeners. Tokens were first decomposed into 15 logarithmically spaced bands and then re-synthesized by multiplying the envelope of each band by matched filtered noise. Various amounts of current spread were simulated by adjusting the drop-off of the noise spectrum away from the peak (40-5 dBoctave). The average SMT (0.25 and 0.5 cyclesoctave) increased from 6.3 to 22.5 dB, while average vowel identification scores dropped from 86% to 19% and consonant identification scores dropped from 93% to 59%. In each condition, the impairments in speech understanding were generally similar to those found in CI listeners with similar SMTs, suggesting that variability in spread of neural activation largely accounts for the variability in speech perception of CI listeners.
Collapse
Affiliation(s)
- Leonid M Litvak
- Advanced Bionics Corporation, Sylmar, California 91342, USA.
| | | | | | | |
Collapse
|
628
|
Luo X, Fu QJ. Frequency modulation detection with simultaneous amplitude modulation by cochlear implant users. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 122:1046-54. [PMID: 17672652 DOI: 10.1121/1.2751258] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2023]
Abstract
To better represent fine structure cues in cochlear implants (CIs), recent research has proposed varying the stimulation rate based on slowly varying frequency modulation (FM) information. The present study investigated the abilities of CI users to detect FM with simultaneous amplitude modulation (AM). FM detection thresholds (FMDTs) for 10-Hz sinusoidal FM and upward frequency sweeps were measured as a function of standard frequency (75-1000 Hz). Three AM conditions were tested, including (1) No AM, (2) 20-Hz Sinusoidal AM (SAM) with modulation depths of 10%, 20%, or 30%, and (3) Noise AM (NAM), in which the amplitude was randomly and uniformly varied over a range of 1, 2, or 3 dB, relative to the reference amplitude. Results showed that FMDTs worsened with increasing standard frequencies, and were lower for sinusoidal FM than for upward frequency sweeps. Simultaneous AM significantly interfered with FM detection; FMDTs were significantly poorer with simultaneous NAM than with SAM. Besides, sinusoidal FMDTs significantly worsened when the starting phase of simultaneous SAM was randomized. These results suggest that FM and AM in CI partly share a common loudness-based coding mechanism and the feasibility of "FM+AM" strategies for CI speech processing may be limited.
Collapse
Affiliation(s)
- Xin Luo
- Department of Auditory Implants and Perception, House Ear Institute, Los Angeles, California 90057, USA.
| | | |
Collapse
|
629
|
Abstract
The speech reception performance of a recipient of the Clarion CII implant was evaluated with a comprehensive set of tests. The same tests were administered for a group of six subjects with normal hearing. Scores for the implant subject were not different from the scores for the normal-hearing subjects, for seven of the nine tests, including the most difficult test used in standard clinical practice. These results are both surprising and encouraging, in that the implant provides only a very crude mimicking of only some aspects of the normal physiology.
Collapse
Affiliation(s)
- Blake S Wilson
- RTI International, Research Triangle Park, NC 27709, USA.
| | | |
Collapse
|
630
|
Kidd GR, Watson CS, Gygi B. Individual differences in auditory abilities. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 122:418-35. [PMID: 17614500 DOI: 10.1121/1.2743154] [Citation(s) in RCA: 116] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2023]
Abstract
Performance on 19 auditory discrimination and identification tasks was measured for 340 listeners with normal hearing. Test stimuli included single tones, sequences of tones, amplitude-modulated and rippled noise, temporal gaps, speech, and environmental sounds. Principal components analysis and structural equation modeling of the data support the existence of a general auditory ability and four specific auditory abilities. The specific abilities are (1) loudness and duration (overall energy) discrimination; (2) sensitivity to temporal envelope variation; (3) identification of highly familiar sounds (speech and nonspeech); and (4) discrimination of unfamiliar simple and complex spectral and temporal patterns. Examination of Scholastic Aptitude Test (SAT) scores for a large subset of the population revealed little or no association between general or specific auditory abilities and general intellectual ability. The findings provide a basis for research to further specify the nature of the auditory abilities. Of particular interest are results suggestive of a familiar sound recognition (FSR) ability, apparently specialized for sound recognition on the basis of limited or distorted information. This FSR ability is independent of normal variation in both spectral-temporal acuity and of general intellectual ability.
Collapse
Affiliation(s)
- Gary R Kidd
- Department of Speech and Hearing Sciences, Indiana University, Bloomington, Indiana 47405, USA
| | | | | |
Collapse
|
631
|
Won JH, Drennan WR, Rubinstein JT. Spectral-ripple resolution correlates with speech reception in noise in cochlear implant users. J Assoc Res Otolaryngol 2007; 8:384-92. [PMID: 17587137 PMCID: PMC2538435 DOI: 10.1007/s10162-007-0085-8] [Citation(s) in RCA: 202] [Impact Index Per Article: 11.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2007] [Accepted: 04/27/2007] [Indexed: 10/23/2022] Open
Abstract
Speech perception ability in noise is one of the most practical measures of success with a cochlear implant; however, with experience, this ability can change dramatically over time, making it a less than ideal tool for comparing performance among different processing strategies. This study examined performance on a spectral discrimination task and compared it to speech perception in noise. An adaptive procedure was used to determine the spectral-ripple density that subjects could discriminate. A closed-set, forced-choice adaptive procedure was used to determine speech reception thresholds for words in two-talker babble and in speech-shaped, steady-state noise. Spectral-ripple thresholds (ripples/octave) were significantly correlated with speech reception thresholds (dB SNR) in noise for 29 cochlear implant users (r = -0.55, p = 0.002 in two-talker babble; r = -0.62, p = 0.0004 in steady-state noise), demonstrating that better spectral resolution was associated with better speech perception in noise. A significant correlation was also found between the spectral-ripple discrimination ability and word recognition in quiet (r = 0.50, p = 0.009). In addition, test-retest reliability for spectral-ripple discrimination was good, and no learning was observed. The present study demonstrates that the spectral-ripple discrimination test, which is time efficient and nonlinguistic, would be a useful tool to evaluate cochlear implant performance with different signal processing strategies.
Collapse
Affiliation(s)
- Jong Ho Won
- VM Bloedel Hearing Research Center, University of Washington, P.O. Box 357923, Seattle, WA 98195-7923, USA.
| | | | | |
Collapse
|
632
|
Spriet A, Van Deun L, Eftaxiadis K, Laneau J, Moonen M, van Dijk B, van Wieringen A, Wouters J. Speech understanding in background noise with the two-microphone adaptive beamformer BEAM in the Nucleus Freedom Cochlear Implant System. Ear Hear 2007; 28:62-72. [PMID: 17204899 DOI: 10.1097/01.aud.0000252470.54246.54] [Citation(s) in RCA: 129] [Impact Index Per Article: 7.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVE This paper evaluates the benefit of the two-microphone adaptive beamformer BEAM in the Nucleus Freedom cochlear implant (CI) system for speech understanding in background noise by CI users. DESIGN A double-blind evaluation of the two-microphone adaptive beamformer BEAM and a hardware directional microphone was carried out with five adult Nucleus CI users. The test procedure consisted of a pre- and post-test in the lab and a 2-wk trial period at home. In the pre- and post-test, the speech reception threshold (SRT) with sentences and the percentage correct phoneme scores for CVC words were measured in quiet and background noise at different signal-to-noise ratios. Performance was assessed for two different noise configurations (with a single noise source and with three noise sources) and two different noise materials (stationary speech-weighted noise and multitalker babble). During the 2-wk trial period at home, the CI users evaluated the noise reduction performance in different listening conditions by means of the SSQ questionnaire. In addition to the perceptual evaluation, the noise reduction performance of the beamformer was measured physically as a function of the direction of the noise source. RESULTS Significant improvements of both the SRT in noise (average improvement of 5-16 dB) and the percentage correct phoneme scores (average improvement of 10-41%) were observed with BEAM compared to the standard hardware directional microphone. In addition, the SSQ questionnaire and subjective evaluation in controlled and real-life scenarios suggested a possible preference for the beamformer in noisy environments. CONCLUSIONS The evaluation demonstrates that the adaptive noise reduction algorithm BEAM in the Nucleus Freedom CI-system may significantly increase the speech perception by cochlear implantees in noisy listening conditions. This is the first monolateral (adaptive) noise reduction strategy actually implemented in a mainstream commercial CI.
Collapse
Affiliation(s)
- Ann Spriet
- ExpORL, Dept. Neurosciences, KU Leuven, Leuven, Belgium.
| | | | | | | | | | | | | | | |
Collapse
|
633
|
Kong YY, Carlyon RP. Improved speech recognition in noise in simulated binaurally combined acoustic and electric stimulation. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 121:3717-27. [PMID: 17552722 DOI: 10.1121/1.2717408] [Citation(s) in RCA: 57] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/15/2023]
Abstract
Speech recognition in noise improves with combined acoustic and electric stimulation compared to electric stimulation alone [Kong et al., J. Acoust. Soc. Am. 117, 1351-1361 (2005)]. Here the contribution of fundamental frequency (F0) and low-frequency phonetic cues to speech recognition in combined hearing was investigated. Normal-hearing listeners heard vocoded speech in one ear and low-pass (LP) filtered speech in the other. Three listening conditions (vocode-alone, LP-alone, combined) were investigated. Target speech (average F0=120 Hz) was mixed with a time-reversed masker (average F0=172 Hz) at three signal-to-noise ratios (SNRs). LP speech aided performance at all SNRs. Low-frequency phonetic cues were then removed by replacing the LP speech with a LP equal-amplitude harmonic complex, frequency and amplitude modulated by the F0 and temporal envelope of voiced segments of the target. The combined hearing advantage disappeared at 10 and 15 dB SNR, but persisted at 5 dB SNR. A similar finding occurred when, additionally, F0 contour cues were removed. These results are consistent with a role for low-frequency phonetic cues, but not with a combination of F0 information between the two ears. The enhanced performance at 5 dB SNR with F0 contour cues absent suggests that voicing or glimpsing cues may be responsible for the combined hearing benefit.
Collapse
Affiliation(s)
- Ying-Yee Kong
- MRC-Cognition & Brain Sciences Unit, 15 Chaucer Road, Cambridge CB2 2EF, United Kingdom.
| | | |
Collapse
|
634
|
Middlebrooks JC, Snyder RL. Auditory prosthesis with a penetrating nerve array. J Assoc Res Otolaryngol 2007; 8:258-79. [PMID: 17265124 PMCID: PMC2538356 DOI: 10.1007/s10162-007-0070-2] [Citation(s) in RCA: 109] [Impact Index Per Article: 6.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2006] [Accepted: 12/15/2006] [Indexed: 11/28/2022] Open
Abstract
Contemporary auditory prostheses ("cochlear implants") employ arrays of stimulating electrodes implanted in the scala tympani of the cochlea. Such arrays have been implanted in some 100,000 profoundly or severely deaf people worldwide and arguably are the most successful of present-day neural prostheses. Nevertheless, most implant users show poor understanding of speech in noisy backgrounds, poor pitch recognition, and poor spatial hearing, even when using bilateral implants. Many of these limitations can be attributed to the remote location of stimulating electrodes relative to excitable cochlear neural elements. That is, a scala tympani electrode array lies within a bony compartment filled with electrically conductive fluid. Moreover, scala tympani arrays typically do not extend to the apical turn of the cochlea in which low frequencies are represented. In the present study, we have tested in an animal model an alternative to the conventional cochlear implant: a multielectrode array implanted directly into the auditory nerve. We monitored the specificity of stimulation of the auditory pathway by recording extracellular unit activity at 32 sites along the tonotopic axis of the inferior colliculus. The results demonstrate the activation of specific auditory nerve populations throughout essentially the entire frequency range that is represented by characteristic frequencies in the inferior colliculus. Compared to conventional scala tympani stimulation, thresholds for neural excitation are as much as 50-fold lower and interference between electrodes stimulated simultaneously is markedly reduced. The results suggest that if an intraneural stimulating array were incorporated into an auditory prosthesis system for humans, it could offer substantial improvement in hearing replacement compared to contemporary cochlear implants.
Collapse
Affiliation(s)
- John C. Middlebrooks
- Kresge Hearing Research Institute, Department of Otorhinolaryngology, University of Michigan, Ann Arbor, MI 48109-0506 USA
| | - Russell L. Snyder
- Epstein Laboratory, Department of Otolaryngology, Head and Neck Surgery, University of California, San Francisco, CA 94143-0526 USA
- Department of Psychology, Utah State University, Logan, UT 84322-2810 USA
| |
Collapse
|
635
|
Izzo AD, Walsh JT, Jansen ED, Bendett M, Webb J, Ralph H, Richter CP. Optical parameter variability in laser nerve stimulation: a study of pulse duration, repetition rate, and wavelength. IEEE Trans Biomed Eng 2007; 54:1108-14. [PMID: 17554829 PMCID: PMC3471085 DOI: 10.1109/tbme.2007.892925] [Citation(s) in RCA: 115] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Pulsed lasers can evoke neural activity from motor as well as sensory neurons in vivo. Lasers allow more selective spatial resolution of stimulation than the conventional electrical stimulation. To date, few studies have examined pulsed, mid-infrared laser stimulation of nerves and very little of the available optical parameter space has been studied. In this study, a pulsed diode laser, with wavelength between 1.844-1.873 microm, was used to elicit compound action potentials (CAPs) from the auditory system of the gerbil. We found that pulse durations as short as 35 micros elicit a CAP from the cochlea. In addition, repetition rates up to 13 Hz can continually stimulate cochlear spiral ganglion cells for extended periods of time. Varying the wavelength and, therefore, the optical penetration depth, allowed different populations of neurons to be stimulated. The technology of optical stimulation could significantly improve cochlear implants, which are hampered by a lack of spatial selectivity.
Collapse
Affiliation(s)
- Agnella D. Izzo
- Department of Biomedical Engineering, Northwestern University, Evanston, IL 60208 USA. She is also with the Department of Otolaryngology, Feinberg Medical School, Northwestern University, Chicago, IL 60611 USA
| | - Joseph T. Walsh
- Department of Biomedical Engineering, Northwestern University, Evanston, IL 60208 USA
| | - E. Duco Jansen
- Department of Biomedical Engineering, Vanderbilt University, Nashville, TN 37235 USA
| | | | - Jim Webb
- Aculight Corporation, Bothell, WA 98011 USA
| | | | - Claus-Peter Richter
- Department of Otolaryngology, Feinberg Medical School, Northwestern University, Chicago, IL 60611 USA
| |
Collapse
|
636
|
Green T, Katiri S, Faulkner A, Rosen S. Talker intelligibility differences in cochlear implant listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 121:EL223-9. [PMID: 17552573 DOI: 10.1121/1.2720938] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/15/2023]
Abstract
People vary in the intelligibility of their speech. This study investigated whether across-talker intelligibility differences observed in normally-hearing listeners are also found in cochlear implant (CI) users. Speech perception for male, female, and child pairs of talkers differing in intelligibility was assessed with actual and simulated CI processing and in normal hearing. While overall speech recognition was, as expected, poorer for CI users, differences in intelligibility across talkers were consistent across all listener groups. This suggests that the primary determinants of intelligibility differences are preserved in the CI-processed signal, though no single critical acoustic property could be identified.
Collapse
Affiliation(s)
- Tim Green
- Department of Phonetics and Linguistics, University College London, 4 Stephenson Way, London NW1 2HE, United Kingdom
| | | | | | | |
Collapse
|
637
|
Başkent D, Eiler CL, Edwards B. Using genetic algorithms with subjective input from human subjects: implications for fitting hearing aids and cochlear implants. Ear Hear 2007; 28:370-80. [PMID: 17485986 DOI: 10.1097/aud.0b013e318047935e] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVE To present a comprehensive analysis of the feasibility of genetic algorithms (GA) for finding the best fit of hearing aids or cochlear implants for individual users in clinical or research settings, where the algorithm is solely driven by subjective human input. DESIGN Due to varying pathology, the best settings of an auditory device differ for each user. It is also likely that listening preferences vary at the same time. The settings of a device customized for a particular user can only be evaluated by the user. When optimization algorithms are used for fitting purposes, this situation poses a difficulty for a systematic and quantitative evaluation of the suitability of the fitting parameters produced by the algorithm. In the present study, an artificial listening environment was generated by distorting speech using a noiseband vocoder. The settings produced by the GA for this listening problem could objectively be evaluated by measuring speech recognition and comparing the performance to the best vocoder condition where speech was least distorted. Nine normal-hearing subjects participated in the study. The parameters to be optimized were the number of vocoder channels, the shift between the input frequency range and the synthesis frequency range, and the compression-expansion of the input frequency range over the synthesis frequency range. The subjects listened to pairs of sentences processed with the vocoder, and entered a preference for the sentence with better intelligibility. The GA modified the solutions iteratively according to the subject preferences. The program converged when the user ranked the same set of parameters as the best in three consecutive steps. The results produced by the GA were analyzed for quality by measuring speech intelligibility, for test-retest reliability by running the GA three times with each subject, and for convergence properties. RESULTS Speech recognition scores averaged across subjects were similar for the best vocoder solution and for the solutions produced by the GA. The average number of iterations was 8 and the average convergence time was 25.5 minutes. The settings produced by different GA runs for the same subject were slightly different; however, speech recognition scores measured with these settings were similar. Individual data from subjects showed that in each run, a small number of GA solutions produced poorer speech intelligibility than for the best setting. This was probably a result of the combination of the inherent randomness of the GA, the convergence criterion used in the present study, and possible errors that the users might have made during the paired comparisons. On the other hand, the effect of these errors was probably small compared to the other two factors, as a comparison between subjective preferences and objective measures showed that for many subjects the two were in good agreement. CONCLUSIONS The results showed that the GA was able to produce good solutions by using listener preferences in a relatively short time. For practical applications, the program can be made more robust by running the GA twice or by not using an automatic stopping criterion, and it can be made faster by optimizing the number of the paired comparisons completed in each iteration.
Collapse
Affiliation(s)
- Deniz Başkent
- Starkey Hearing Research Center, Berkeley, California 94704, USA.
| | | | | |
Collapse
|
638
|
Lin YS, Lee FP, Huang IS, Peng SC. Continuous improvement in Mandarin lexical tone perception as the number of channels increased: a simulation study of cochlear implant. Acta Otolaryngol 2007; 127:505-14. [PMID: 17453477 DOI: 10.1080/00016480600951434] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Abstract
CONCLUSION With reference to English phoneme recognition, where performance usually does not improve after six or eight channels in cochlear implants (CIs), increasing total channel numbers continuously improved perception of Mandarin tones. OBJECTIVE To test our hypothesis that current CI strategies might be modified to improve Mandarin lexical tonal perception. MATERIALS AND METHODS Lexical tonal perception tests using 48 monosyllables in Mandarin Chinese were conducted in 32 native Mandarin speakers with normal hearing. The performance of tonal perception was compared among the controlled factors, which were total channel number, number of channels allocated to the F0 spectrum, and whether there were spectral shifts in the electrode configuration. The experimental condition that preserves fine structure was used as a comparison. RESULTS The signal processing strategy using 16 channels--which is technically possible with current CI devices--produced better tonal perception than those using 12 or 8 channels. Increasing the number of fundamental channels did not improve tonal perception, and spectral shifts did not change tonal perception. An experimental condition (FiC12) that preserves the fine structure produced significantly better overall scores for tone perception than other experimental conditions with envelope strategies.
Collapse
Affiliation(s)
- Yung-Song Lin
- Department of Otolaryngology, Taipei Medical University, Chi Mei Medical Center, Tainan city, Taiwan, ROC.
| | | | | | | |
Collapse
|
639
|
Dingemanse JG, Frijns JHM, Briaire JJ. Psychophysical assessment of spatial spread of excitation in electrical hearing with single and dual electrode contact maskers. Ear Hear 2007; 27:645-57. [PMID: 17086076 DOI: 10.1097/01.aud.0000246683.29611.1b] [Citation(s) in RCA: 20] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVE To evaluate psychophysically the spatial spread of excitation in electrical hearing with a new dual contact masker and to investigate under which conditions it is possible to stimulate fibers in the immediate neighborhood of an electrode contact, which were not excited by neighboring electrode contacts. DESIGN In this study a psychophysical forward masking paradigm with a dual contact masker was used to avoid off-site listening, the electrical analogue of off-frequency listening. The masker stimulus (300 msec) is presented nonsimultaneously on two electrode contacts, one on the apical side and another on the basal side of the probe contact, followed by a probe stimulus of 20 msec.Unmasked probe thresholds were compared with masked ones at a number of masker-probe distances, whereas growth of masking curves were measured for a fixed masker contact pair. Standard selectivity measurements (single contact masking) and the recovery of forward masking with one masker contact were included for comparison with existing methods. All experiments were carried out with six participants who use the Clarion CII device with a HiFocus I electrode array. RESULTS For dual contact masking the amount of masking was significantly greater than for single contact masking and the width of the masking patterns was on average 1.1 mm broader than for single contact masking, resulting in a broad region of excitation, with masker-probe overlap for distances greater than 3 mm. Masking widths for dual and single contact masking were highly correlated. Growth of masking curves were highly nonlinear. They showed a strong elevation of the slope that starts for most subjects around the middle of the dynamic range or above. For 4 out of 6 subjects, no probe threshold was found above a masker amplitude of about 400-500 microA. The ratio of the maximum measurable masked probe thresholds and unmasked probe threshold ranged from 1.7 to 2.6 (S4 excluded). Recovery of masking functions follow an exponential decay. Time constants tau for the recovery process ranged from 21.6 msec to 114.9 msec. CONCLUSIONS With a dual contact masker (1) off-site listening can be avoided, leading to larger estimates of the width of excitation patterns than in single contact masking, (2) it can be estimated for which stimulation level there is complete overlap of excitation patterns of adjacent electrode contacts, (3) it can be shown that stimulation of nerve fibers in the immediate neighborhood of an electrode contact which were not excited by neighboring electrode contacts is only possible if the probe stimulation amplitude is sufficiently high in comparison with amplitudes on neighboring contacts.
Collapse
|
640
|
Oxenham AJ, Simonson AM, Turicchia L, Sarpeshkar R. Evaluation of companding-based spectral enhancement using simulated cochlear-implant processing. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 121:1709-16. [PMID: 17407907 DOI: 10.1121/1.2434757] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2023]
Abstract
This study tested a time-domain spectral enhancement algorithm that was recently proposed by Turicchia and Sarpeshkar [IEEE Trans. Speech Audio Proc. 13, 243-253 (2005)]. The algorithm uses a filter bank, with each filter channel comprising broadly tuned amplitude compression, followed by more narrowly tuned expansion (companding). Normal-hearing listeners were tested in their ability to recognize sentences processed through a noise-excited envelope vocoder that simulates aspects of cochlear-implant processing. The sentences were presented in a steady background noise at signal-to-noise ratios of 0, 3, and 6 dB and were either passed directly through an envelope vocoder, or were first processed by the companding algorithm. Using an eight-channel envelope vocoder, companding produced small but significant improvements in speech reception. Parametric variations of the companding algorithm showed that the improvement in intelligibility was robust to changes in filter tuning, whereas decreases in the time constants resulted in a decrease in intelligibility. Companding continued to provide a benefit when the number of vocoder frequency channels was increased to sixteen. When integrated within a sixteen-channel cochlear-implant simulator, companding also led to significant improvements in sentence recognition. Thus, companding may represent a readily implementable way to provide some speech recognition benefits to current cochlear-implant users.
Collapse
Affiliation(s)
- Andrew J Oxenham
- Research Laboratory of Electronics, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA.
| | | | | | | |
Collapse
|
641
|
Izzo AD, Suh E, Pathria J, Walsh JT, Whitlon DS, Richter CP. Selectivity of neural stimulation in the auditory system: a comparison of optic and electric stimuli. JOURNAL OF BIOMEDICAL OPTICS 2007; 12:021008. [PMID: 17477715 DOI: 10.1117/1.2714296] [Citation(s) in RCA: 65] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
Pulsed, mid-infrared lasers were recently investigated as a method to stimulate neural activity. There are significant benefits of optically stimulating nerves over electrically stimulating, in particular the application of more spatially confined neural stimulation. We report results from experiments in which the gerbil auditory system was stimulated by optical radiation, acoustic tones, or electric current. Immunohistochemical staining for the protein c-FOS revealed the spread of excitation. We demonstrate a spatially selective activation of neurons using a laser; only neurons in the direct optical path are stimulated. This pattern of c-FOS labeling is in contrast to that after electrical stimulation. Electrical stimulation leads to a large, more spatially extended population of labeled, activated neurons. In the auditory system, optical stimulation of nerves could have a significant impact on the performance of cochlear implants, which can be limited by the electric current spread.
Collapse
Affiliation(s)
- Agnella D Izzo
- Northwestern University, Biomedical Engineering, 2145 Sheridan Road, Tech E310, Evanston, Illinois 60208, USA
| | | | | | | | | | | |
Collapse
|
642
|
Hallum LE, Dagnelie G, Suaning GJ, Lovell NH. Simulating auditory and visual sensorineural prostheses: a comparative review. J Neural Eng 2007; 4:S58-71. [PMID: 17325417 DOI: 10.1088/1741-2560/4/1/s08] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
Microelectronic vision prosthesis proposes to render luminous spots (so-called phosphenes) in the visual field of the otherwise blind subject by way of an implanted array of stimulating electrodes, and in doing so restore some spatial vision. There are now many research teams worldwide working towards a therapeutic device, analogous to the cochlear implant, for the profoundly blind. Despite the similarities between the cochlear implant and vision prostheses, there are few instances in the literature where the two approaches are compared and contrasted with a mind to informing the science and engineering of the latter. This is the focus of the present review; specifically, our interest is psychophysics and signal processing. Firstly, we examine the cochlear implant, and review a handful of psychophysical work: the acoustic simulation of cochlear implants and the method used. We focus on the use of normally hearing subjects (played coloured noise bands or sine waves) as a means of investigating cochlear-implant efficacy and speech processing algorithms. These results provide guidance to vision researchers, for they address the interpretation of simulation data, and flag key areas, such as 'artificial' perception in the presence of noise, that require experimental work in coming years. Secondly, we provide an up-to-date review of the body of analogous psychophysical work: the visual simulation, involving normal observers, of microelectronic vision prosthesis. These simulations allow predictions as to the likely clinical efficacy of the prosthesis; indeed, results to date suggest that a number on the order of 100 implanted electrodes will afford subjects mobility and recognition of faces (and other complex stimuli), while even fewer electrodes facilitate reading printed text and very simple visuomanual tasks. Further, the simulations allow investigations of image and signal processing strategies, plus they provide researchers in the field, and other interested persons, a perceptual experience that approximates what a prosthesis will likely afford implantees.
Collapse
Affiliation(s)
- L E Hallum
- Graduate School of Biomedical Engineering, University of New South Wales, Sydney 2052, Australia
| | | | | | | |
Collapse
|
643
|
van der Beek FB, Soede W, Frijns JHM. Evaluation of the Benefit for Cochlear Implantees of Two Assistive Directional Microphone Systems in an Artificial Diffuse Noise Situation. Ear Hear 2007; 28:99-110. [PMID: 17204902 DOI: 10.1097/aud.0b013e31802d0a55] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVE People with cochlear implants have severe problems with speech understanding in noisy surroundings. This study evaluates and quantifies the effect of two assistive directional microphone systems compared to the standard headpiece microphone on speech perception in quiet surroundings and in background noise, in a laboratory setting developed to reflect a situation whereby the listener is disturbed by a noise with a mainly diffuse character due to many sources in a reverberant room. DESIGN Thirteen postlingually deafened patients, implanted in the Leiden University Medical Centre with the Clarion CII device, participated in the study. An experimental set-up with 8 uncorrelated steady-state noise sources was used to test speech perception on monosyllabic words. Each subject was tested with a standard headpiece microphone, and the two assistive directional microphones, TX3 Handymic by Phonak and the Linkit array microphone by Etymotic Research. Testing was done in quiet at a level of 65 dB SPL and with decreasing signal-to-noise ratios (SNR) down to -15 dB. RESULTS Using the assistive directional microphones, speech recognition in background noise improved substantially and was not affected in quiet. At an SNR of 0 dB, the average CVC scores improved from 45% for the headpiece microphone to 67% and 62% for the TX3 Handymic and the Linkit respectively. Compared to the headpiece, the Speech Reception Threshold (SRT) improved by 8.2 dB SNR and 5.9 dB SNR for the TX3 Handymic and the Linkit respectively. The gain in SRT for TX3 Handymic and Linkit was neither correlated to the SRT score with headpiece nor the duration of CI-use. CONCLUSION The speech recognition test in background noise showed a clear benefit from the assistive directional microphones for cochlear implantees compared to the standard microphone. In a noisy environment, the significant benefit from these assistive device microphones may allow understanding of speech with greater ease.
Collapse
Affiliation(s)
- F B van der Beek
- ENT Department, Leiden University Medical Centre, Leiden, The Netherlands
| | | | | |
Collapse
|
644
|
Grant KW, Tufts JB, Greenberg S. Integration efficiency for speech perception within and across sensory modalities by normal-hearing and hearing-impaired individuals. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 121:1164-76. [PMID: 17348537 DOI: 10.1121/1.2405859] [Citation(s) in RCA: 15] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2023]
Abstract
In face-to-face speech communication, the listener extracts and integrates information from the acoustic and optic speech signals. Integration occurs within the auditory modality (i.e., across the acoustic frequency spectrum) and across sensory modalities (i.e., across the acoustic and optic signals). The difficulties experienced by some hearing-impaired listeners in understanding speech could be attributed to losses in the extraction of speech information, the integration of speech cues, or both. The present study evaluated the ability of normal-hearing and hearing-impaired listeners to integrate speech information within and across sensory modalities in order to determine the degree to which integration efficiency may be a factor in the performance of hearing-impaired listeners. Auditory-visual nonsense syllables consisting of eighteen medial consonants surrounded by the vowel [a] were processed into four nonoverlapping acoustic filter bands between 300 and 6000 Hz. A variety of one, two, three, and four filter-band combinations were presented for identification in auditory-only and auditory-visual conditions: A visual-only condition was also included. Integration efficiency was evaluated using a model of optimal integration. Results showed that normal-hearing and hearing-impaired listeners integrated information across the auditory and visual sensory modalities with a high degree of efficiency, independent of differences in auditory capabilities. However, across-frequency integration for auditory-only input was less efficient for hearing-impaired listeners. These individuals exhibited particular difficulty extracting information from the highest frequency band (4762-6000 Hz) when speech information was presented concurrently in the next lower-frequency band (1890-2381 Hz). Results suggest that integration of speech information within the auditory modality, but not across auditory and visual modalities, affects speech understanding in hearing-impaired listeners.
Collapse
Affiliation(s)
- Ken W Grant
- Walter Reed Army Medical Center Army Audiology and Speech Center, Washington, DC 20307-5001, USA.
| | | | | |
Collapse
|
645
|
Chang JE, Bai JY, Zeng FG. Unintelligible low-frequency sound enhances simulated cochlear-implant speech recognition in noise. IEEE Trans Biomed Eng 2007; 53:2598-601. [PMID: 17152439 DOI: 10.1109/tbme.2006.883793] [Citation(s) in RCA: 75] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Speech can be recognized by multiple acoustic cues in both frequency and time domains. These acoustic cues are often thought to be redundant. One example is the low-frequency sound component below 300 Hz, which is not even transmitted by the majority of communication devices including telephones. Here, we showed that this low-frequency sound component, although unintelligible when presented alone, could improve the functional signal-to-noise ratio (SNR) by 10-15 dB for speech recognition in noise when presented in combination with a cochlear-implant simulation. A similar low-frequency enhancement effect could be obtained by presenting the low-frequency sound component to one ear and the cochlear-implant simulation to the other ear. However, a high-frequency sound could not produce a similar speech enhancement in noise. We argue that this low-frequency enhancement effect cannot be due to linear addition of intelligibility between low- and high-frequency components or an increase in the physical SNR. We suggest a brain-based mechanism that uses the voice pitch cue in the low-frequency sound to first segregate the target voice from the competing voice and then to group appropriate temporal envelope cues in the target voice for robust speech recognition under realistic listening situations.
Collapse
Affiliation(s)
- Janice E Chang
- Bioengineering Dept., University of California, Berkeley, CA 94720, USA
| | | | | |
Collapse
|
646
|
Sit JJ, Simonson AM, Oxenham AJ, Faltys MA, Sarpeshkar R. A Low-Power Asynchronous Interleaved Sampling Algorithm for Cochlear Implants That Encodes Envelope and Phase Information. IEEE Trans Biomed Eng 2007; 54:138-49. [PMID: 17260865 DOI: 10.1109/tbme.2006.883819] [Citation(s) in RCA: 39] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
Cochlear implants currently fail to convey phase information, which is important for perceiving music, tonal languages, and for hearing in noisy environments. We propose a bio-inspired asynchronous interleaved sampling (AIS) algorithm that encodes both envelope and phase information, in a manner that may be suitable for delivery to cochlear implant users. Like standard continuous interleaved sampling (CIS) strategies, AIS naturally meets the interleaved-firing requirement, which is to stimulate only one electrode at a time, minimizing electrode interactions. The majority of interspike intervals are distributed over 1-4 ms, thus staying within the absolute refractory limit of neurons, and form a more natural, pseudostochastic pattern of firing due to complex channel interactions. Stronger channels are selected to fire more often but the strategy ensures that weaker channels are selected to fire in proportion to their signal strength as well. The resulting stimulation rates are considerably lower than those of most modern implants, saving power yet delivering higher potential performance. Correlations with original sounds were found to be significantly higher in AIS reconstructions than in signal reconstructions using only envelope information. Two perceptual tests on normal-hearing listeners verified that the reconstructed signals enabled better melody and speech recognition in noise than those processed using tone-excited envelope-vocoder simulations of cochlear implant processing. Thus, our strategy could potentially save power and improve hearing performance in cochlear implant users.
Collapse
Affiliation(s)
- Ji-Jon Sit
- Massachusetts Institute of Technology (MIT), Cambridge, MA 02139, USA
| | | | | | | | | |
Collapse
|
647
|
Abstract
Because there are many parameters in the cochlear implant (CI) device that can be optimized for individual patients, it is important to estimate a parameter's effect before patient evaluation. In this paper, Mel-frequency cepstrum coefficients (MFCCs) were used to estimate the acoustic vowel space for vowel stimuli processed by the CI simulations. The acoustic space was then compared to vowel recognition performance by normal-hearing subjects listening to the same processed speech. Five CI speech processor parameters were simulated to produce different degree of spectral resolution, spectral smearing, spectral warping, spectral shifting, and amplitude distortion. The acoustic vowel space was highly correlated with normal hearing subjects' vowel recognition performance for parameters that affected the spectral channels and spectral smearing. However, the acoustic vowel space was not significantly correlated with perceptual performance for parameters that affected the degree of spectral warping, spectral shifting, and amplitude distortion. In particular, while spectral warping and shifting did not significantly reshape the acoustic space, vowel recognition performance was significantly affected by these parameters. The results from the acoustic analysis suggest that the CI device can preserve phonetic distinctions under conditions of spectral warping and shifting. Auditory training may help CI patients better perceive these speech cues transmitted by their speech processors.
Collapse
Affiliation(s)
- Chuping Liu
- Department of Electrical Engineering, University of Southern California, Los Angeles, CA 90007, USA.
| | | |
Collapse
|
648
|
Iverson P, Smith CA, Evans BG. Vowel recognition via cochlear implants and noise vocoders: effects of formant movement and duration. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2006; 120:3998-4006. [PMID: 17225426 DOI: 10.1121/1.2372453] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
Previous work has demonstrated that normal-hearing individuals use fine-grained phonetic variation, such as formant movement and duration, when recognizing English vowels. The present study investigated whether these cues are used by adult postlingually deafened cochlear implant users, and normal-hearing individuals listening to noise-vocoder simulations of cochlear implant processing. In Experiment 1, subjects gave forced-choice identification judgments for recordings of vowels that were signal processed to remove formant movement and/or equate vowel duration. In Experiment 2, a goodness-optimization procedure was used to create perceptual vowel space maps (i.e., best exemplars within a vowel quadrilateral) that included F1, F2, formant movement, and duration. The results demonstrated that both cochlear implant users and normal-hearing individuals use formant movement and duration cues when recognizing English vowels. Moreover, both listener groups used these cues to the same extent, suggesting that postlingually deafened cochlear implant users have category representations for vowels that are similar to those of normal-hearing individuals.
Collapse
Affiliation(s)
- Paul Iverson
- Department of Phonetics and Linguistics, University College London, 4 Stephenson Way, London NW1 2HE, United Kingdom
| | | | | |
Collapse
|
649
|
Chang YP, Fu QJ. Effects of talker variability on vowel recognition in cochlear implants. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2006; 49:1331-41. [PMID: 17197499 DOI: 10.1044/1092-4388(2006/095)] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
PURPOSE To investigate the effects of talker variability on vowel recognition by cochlear implant (CI) users and by normal-hearing (NH) participants listening to 4-channel acoustic CI simulations. METHOD CI users were tested with their clinically assigned speech processors. For NH participants, 3 CI processors were simulated, using different combinations of carrier type and temporal envelope cutoff frequency (noise band/160 Hz, sine wave/160 Hz, and sine wave/20 Hz). Vowel recognition was measured for 4 talkers, presented in either a single-talker context (1 talker per test block) or a multi-talker context (4 talkers per test block). RESULTS CI users' vowel recognition was significantly poorer in the multi-talker context than in the single-talker context. When noise-band carriers were used in the simulations, NH performance was not significantly affected by talker variability. However, when sine-wave carriers were used in the simulations, NH performance was significantly affected by talker variability in both envelope filter conditions. CONCLUSIONS Because fundamental frequency was not preserved by the 20-Hz envelope filter and only partially preserved by the 160-Hz envelope filter, both spectral and temporal cues contributed to the talker variability effects observed with sine-wave carriers. Similarly, spectral and temporal cues may have contributed to the talker variability effects observed with CI participants.
Collapse
Affiliation(s)
- Yi-ping Chang
- University of Southern California, Los Angeles, USA.
| | | |
Collapse
|
650
|
Lenarz M, Lim HH, Patrick JF, Anderson DJ, Lenarz T. Electrophysiological validation of a human prototype auditory midbrain implant in a guinea pig model. J Assoc Res Otolaryngol 2006; 7:383-98. [PMID: 17075701 PMCID: PMC2504634 DOI: 10.1007/s10162-006-0056-5] [Citation(s) in RCA: 30] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2006] [Accepted: 09/04/2006] [Indexed: 12/19/2022] Open
Abstract
The auditory midbrain implant (AMI) is a new treatment for hearing restoration in patients with neural deafness or surgically inaccessible cochleae who cannot benefit from cochlear implants (CI). This includes neurofibromatosis type II (NF2) patients who, due to development and/or removal of vestibular schwannomas, usually experience complete damage of their auditory nerves. Although the auditory brainstem implant (ABI) provides sound awareness and aids lip-reading capabilities for these NF2 patients, it generally only achieves hearing performance levels comparable with a single-channel CI. In collaboration with Cochlear Ltd. (Lane Cove, Australia), we developed a human prototype AMI, which is designed for electrical stimulation along the well-defined tonotopic gradient of the inferior colliculus central nucleus (ICC). Considering that better speech perception and hearing performance has been correlated with a greater number of discriminable frequency channels of information available, the ability of the AMI to effectively activate discrete frequency regions within the ICC may enable better hearing performance than achieved by the ABI. Therefore, the goal of this study was to investigate if our AMI array could achieve low-threshold, frequency-specific activation within the ICC, and whether the levels for ICC activation via AMI stimulation were within safe limits for human application. We electrically stimulated different frequency regions within the ICC via the AMI array and recorded the corresponding neural activity in the primary auditory cortex (A1) using a multisite silicon probe in ketamine-anesthetized guinea pigs. Based on our results, AMI stimulation achieves lower thresholds and more localized, frequency-specific activation than CI stimulation. Furthermore, AMI stimulation achieves cortical activation with current levels that are within safe limits for central nervous system stimulation. This study confirms that our AMI design is sufficient for ensuring safe and effective activation of the ICC, and warrants further studies to translate the AMI into clinical application.
Collapse
Affiliation(s)
- Minoo Lenarz
- Otorhinolaryngology Department, Medical University of Hannover, Hannover, Lower Saxony, 30625 Germany
| | - Hubert H. Lim
- Otorhinolaryngology Department, Medical University of Hannover, Hannover, Lower Saxony, 30625 Germany
- Kresge Hearing Research Institute, University of Michigan, Ann Arbor, MI 48109 USA
- Cochlear Limited, Lane Cove, NSW 2066 Australia
- Medizinische Hochschule Hannover, HNO Klinik, Sekretariat Prof. Lenarz, Gebaeude K5, Ebene 1, Raum 4010, Carl-Neuberg-Str.1, Hannover, D-30625 Germany
| | | | - David J. Anderson
- Kresge Hearing Research Institute, University of Michigan, Ann Arbor, MI 48109 USA
| | - Thomas Lenarz
- Otorhinolaryngology Department, Medical University of Hannover, Hannover, Lower Saxony, 30625 Germany
| |
Collapse
|