1
|
Anand S, Park Y, Shrivastav R, Eddins DA. Evaluating the Effect of Voice Quality Covariance on Auditory-Perceptual Evaluation Using a Novel Two-Dimensional Magnitude Estimation Task. J Speech Lang Hear Res 2023; 66:4849-4859. [PMID: 37902504 PMCID: PMC11001379 DOI: 10.1044/2023_jslhr-23-00226] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/04/2023] [Revised: 08/12/2023] [Accepted: 09/03/2023] [Indexed: 10/31/2023]
Abstract
PURPOSE Most people with dysphonia present with voices that vary along more than one voice quality (VQ) dimension. This study sought to examine the effect of covariance between breathy and rough VQ in natural voices. METHOD A two-dimensional matrix of 16 /a/ vowels was selected such that two VQ dimensions (breathiness and roughness) were sampled on a 4-point severity scale (none, mild, moderate, and severe). Ten listeners evaluated 480 stimuli (16 stimuli × 10 repetitions × 3 blocks) on one-dimensional magnitude estimation (1DME) tasks and a novel two-dimensional magnitude estimation (2DME) task that allowed for simultaneous measurement of breathiness and roughness. RESULTS Data indicated high intra- and interrater reliabilities for both breathiness and roughness in the 2DME and 1DME tasks. Correlation analyses revealed a strong correlation between 2DME and 1DME judgments for breathiness and roughness (r > .95). There was also a minimal correlation between breathy and rough VQ in the 2DME task (r < .10). CONCLUSIONS Covarying roughness or breathiness had less impact on the perception of the other VQ in natural dysphonic voices in 2DME compared to 1DME. An understanding and quantification of the perceptual interactions among the dimensions will aid in the refinement of computational models and in the establishment of the validity of clinical scales for VQ perception.
Collapse
Affiliation(s)
- Supraja Anand
- Department of Communication Sciences & Disorders, University of South Florida, Tampa
| | - Yeonggwang Park
- Department of Communication Sciences and Disorders, University of Central Florida, Orlando
| | - Rahul Shrivastav
- Office of the Provost & Executive Vice President, Indiana University Bloomington
| | - David A. Eddins
- Department of Communication Sciences and Disorders, University of Central Florida, Orlando
| |
Collapse
|
2
|
Lelo de Larrea-Mancera ES, Stavropoulos T, Carrillo AA, Menon KN, Hoover EC, Eddins DA, Gallun FJ, Seitz AR. Validation of the adaptive scan method in the quest for time-efficient methods of testing auditory processes. Atten Percept Psychophys 2023; 85:2797-2810. [PMID: 37349625 PMCID: PMC10600050 DOI: 10.3758/s13414-023-02743-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 06/01/2023] [Indexed: 06/24/2023]
Abstract
A major barrier to the clinical application of psychophysical testing of central auditory processes is the time required to obtain precise estimates of different listening abilities. In this study, we validate a novel adaptive scan (AS) method of threshold estimation that is designed to adapt on a range of values around threshold rather than on a single threshold value. This method has the advantage of providing the listener with greater familiarity with the stimulus characteristics near threshold while maintaining precise measurement and increasing time-efficiency. Additionally, we explore the time-efficiency of AS through comparison with two more conventional adaptive algorithms and the method of constant stimuli in two common psychophysical tasks: the detection of a gap in noise and the detection of a tone in noise. Seventy undergraduates without hearing complaints were tested using all four methods. The AS method provided similar threshold estimates with similar precision to those from the other adaptive methods and, thus, it is a valid adaptive method of psychophysical testing. We also provide an analysis of the AS method based on precision metrics to propose a shortened version of the algorithm that maximizes the time/precision tradeoff and can achieve similar thresholds to the adaptive methods tested in the validation. This work lays the foundation for using AS across a wide variety of psychophysical assessments and experimental situations where different levels of precision and/or time-efficiency may be required.
Collapse
Affiliation(s)
- E S Lelo de Larrea-Mancera
- Northeastern University, College of Science, Psychology; Center for Cognitive and Brain Health, Boston, MA, USA.
- University of California Riverside, UCR Brain Game Center for Mental Fitness and Wellbeing, Riverside, CA, USA.
| | - T Stavropoulos
- Northeastern University, College of Science, Psychology; Center for Cognitive and Brain Health, Boston, MA, USA
| | - A A Carrillo
- University of California Riverside, UCR Brain Game Center for Mental Fitness and Wellbeing, Riverside, CA, USA
| | - K N Menon
- Department of Hearing and Speech Sciences, University of Maryland at College Park, College Park, MD, USA
| | - E C Hoover
- Department of Hearing and Speech Sciences, University of Maryland at College Park, College Park, MD, USA
| | - D A Eddins
- University of South Florida, Communication Sciences and Disorders, Tampa, FL, USA
| | - F J Gallun
- Oregon Health & Science University, Oregon Hearing Research Center, Portland, OR, USA
| | - A R Seitz
- Northeastern University, College of Science, Psychology; Center for Cognitive and Brain Health, Boston, MA, USA
- University of California Riverside, UCR Brain Game Center for Mental Fitness and Wellbeing, Riverside, CA, USA
| |
Collapse
|
3
|
Park Y, Baker Brehm S, Kelchner L, Weinrich B, McElfresh K, Anand S, Shrivastav R, de Alarcon A, Eddins DA. Effects of Vibratory Source on Auditory-Perceptual and Bio-Inspired Computational Measures of Pediatric Voice Quality. J Voice 2023:S0892-1997(23)00254-0. [PMID: 37739862 PMCID: PMC10950844 DOI: 10.1016/j.jvoice.2023.08.016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2023] [Revised: 08/11/2023] [Accepted: 08/14/2023] [Indexed: 09/24/2023]
Abstract
OBJECTIVE The vibratory source for voicing in children with dysphonia is classified into three categories including a glottal vibratory source (GVS) observed in those with vocal lesions or hyperfunction; supraglottal vibratory sources (SGVS) observed secondary to laryngeal airway injuries, malformations, or reconstruction surgeries; and a combination of both glottal and supraglottal vibratory sources called mixed vibratory source (MVS). This study evaluated the effects of vibratory source on three primary dimensions of voice quality (breathiness, roughness, and strain) in children with GVS, SGVS, and MVS using single-variable matching tasks and computational measures obtained from bio-inspired auditory models. METHODS A total of 44 dysphonic voice samples from children aged 4-11 years were selected. Seven listeners rated breathiness, roughness, and strain of 1000-ms /ɑ/ samples using single-variable matching tasks. Computational estimates of pitch strength, amplitude modulation filterbank output, and sharpness were obtained through custom-designed MATLAB algorithms. RESULTS Perceived roughness and strain were significantly higher in children with SGVS and MVS compared to children with GVS. Among the computational measures, only the modulation filterbank output resulted in significant differences among vibratory sources; a posthoc test revealed that children with SGVS had greater amplitude modulation than children with GVS, as expected from their rougher voice quality. CONCLUSIONS The results indicate that the output of an auditory amplitude modulation filterbank model may capture characteristics of SGVS that are strongly related to the rough voice quality.
Collapse
Affiliation(s)
- Yeonggwang Park
- Department of Communication Sciences and Disorders, University of Central Florida, Orlando, Florida.
| | - Susan Baker Brehm
- Department of Speech Pathology and Audiology, Miami University, Oxford, Ohio; Division of Speech-Language Pathology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio
| | - Lisa Kelchner
- Division of Speech-Language Pathology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio; Department of Communication Sciences and Disorders, University of Cincinnati, Cincinnati, Ohio
| | - Barbara Weinrich
- Department of Speech Pathology and Audiology, Miami University, Oxford, Ohio; Division of Speech-Language Pathology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio
| | - Kevin McElfresh
- Division of Speech-Language Pathology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio
| | - Supraja Anand
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | - Rahul Shrivastav
- Office of the Provost & Executive Vice President, Indiana University, Bloomington, Indiana
| | - Alessandro de Alarcon
- Pediatric Otolaryngology Head & Neck Surgery, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio
| | - David A Eddins
- Department of Communication Sciences and Disorders, University of Central Florida, Orlando, Florida
| |
Collapse
|
4
|
Abstract
Head position at any point in time plays a fundamental role in shaping the auditory information that reaches a listener, information that continuously changes as the head moves and reorients to different listening situations. The connection between hearing science and the kinesthetics of head movement has gained interest due to technological advances that have increased the feasibility of providing behavioral and biological feedback to assistive listening devices that can interpret movement patterns that reflect listening intent. Increasing evidence also shows that the negative impact of hearing deficits on mobility, gait, and balance may be mitigated by prosthetic hearing device intervention. Better understanding of the relationships between head movement, full body kinetics, and hearing health, should lead to improved signal processing strategies across a range of assistive and augmented hearing devices. The purpose of this review is to introduce the wider hearing community to the kinesiology of head movement and to place it in the context of hearing and communication with the goal of expanding the field of ecologically-specific listener behavior.
Collapse
Affiliation(s)
- Nathan C. Higgins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
| | - Daniel A. Pupo
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
- School of Aging Studies, University of South Florida, Tampa, FL, United States
| | - Erol J. Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
| | - David A. Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
| |
Collapse
|
5
|
Hutchison P, Maeda H, Formby C, Small BJ, Eddins DA, Eddins AC. Acoustic deprivation modulates central gain in human auditory brainstem and cortex. Hear Res 2023; 428:108683. [PMID: 36599259 PMCID: PMC9872081 DOI: 10.1016/j.heares.2022.108683] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/23/2021] [Revised: 12/16/2022] [Accepted: 12/26/2022] [Indexed: 12/28/2022]
Abstract
Beyond reduced audibility, there is convincing evidence that the auditory system adapts according to the principles of homeostatic plasticity in response to a hearing loss. Such compensatory changes include modulation of central auditory gain mechanisms. Earplugging is a common experimental method that has been used to introduce a temporary, reversible hearing loss that induces changes consistent with central gain modulation. In the present study, young, normal-hearing adult participants wore a unilateral earplug for two weeks, during which we measured changes in the acoustic reflex threshold (ART), loudness perception, and cortically-evoked (40 Hz) auditory steady-state response (ASSR) to assess potential modulation in central gain with reduced peripheral input. The ART decreased on average by 8 to 10 dB during the treatment period, with modest increases in loudness perception after one week but not after two weeks of earplug use. Significant changes in both the magnitude and hemispheric laterality of source-localized cortical ASSR measures revealed asymmetrical changes in stimulus-driven cortical activity over time. The ART results following unilateral earplugging are consistent with the literature and suggest that homeostatic plasticity is evident in the brainstem. The novel findings from the cortical ASSR in the present study indicates that reduced peripheral input induces adaptive homeostatic plasticity reflected as both an increase in central gain in the auditory brainstem and reduced cortical activity ipsilateral to the deprived ear. Both the ART and the novel use of the 40-Hz ASSR provide sensitive measures of central gain modulation in the brainstem and cortex of young, normal hearing listeners, and thus may be useful in future studies with other clinical populations.
Collapse
Affiliation(s)
- Peter Hutchison
- Department of Communication Sciences and Disorders, University of South Florida, 4202 E. Fowler Ave., PCD 1017, Tampa, FL 33620, USA
| | - Hannah Maeda
- Department of Communication Sciences and Disorders, University of South Florida, 4202 E. Fowler Ave., PCD 1017, Tampa, FL 33620, USA
| | - Craig Formby
- Department of Communication Sciences and Disorders, University of South Florida, 4202 E. Fowler Ave., PCD 1017, Tampa, FL 33620, USA
| | - Brent J Small
- School of Aging Studies, University of South Florida, 4202 E. Fowler Ave., PCD 1017, Tampa, FL 33620, USA
| | - David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, 4202 E. Fowler Ave., PCD 1017, Tampa, FL 33620, USA; Department of Chemical and Biomedical Engineering, University of South Florida, 4202 E. Fowler Ave., PCD 1017, Tampa, FL 33620, USA
| | - Ann Clock Eddins
- Department of Communication Sciences and Disorders, University of South Florida, 4202 E. Fowler Ave., PCD 1017, Tampa, FL 33620, USA; School of Communication Sciences and Disorders, University of Central Florida, 4364 Scorpius Street, Orlando, FL 32816, USA.
| |
Collapse
|
6
|
Park Y, Anand S, Gifford SM, Shrivastav R, Eddins DA. Development and Validation of a Single-Variable Comparison Stimulus for Matching Strained Voice Quality Using a Psychoacoustic Framework. J Speech Lang Hear Res 2023; 66:16-29. [PMID: 36516473 PMCID: PMC10023177 DOI: 10.1044/2022_jslhr-22-00280] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/16/2022] [Revised: 08/17/2022] [Accepted: 09/01/2022] [Indexed: 06/17/2023]
Abstract
PURPOSE Acoustic and perceptual quantification of vocal strain has been a vexing problem for years. To increase measurement rigor, a suitable single-variable matching stimulus for strain was developed and validated, based on the matching stimulus used previously for breathy and rough voice qualities. METHOD A set of 21 comparison stimuli for a single-variable matching task (SVMT) was synthesized based on a speech-shaped sawtooth waveform mixed with speech-shaped noise. Variable bandpass filter gain in mid-to-high frequencies achieved a wide range of computed sharpness (in constant sharpness steps) and served as the independent variable for the SVMT. Ten natural /ɑ/ stimuli with a wide range of the primary voice quality of strain and a minimum of breathiness or roughness were selected and assessed using the SVMT. Natural voice samples and synthetic comparison stimuli were also assessed using a perceptual magnitude estimation (ME) task. RESULTS ME data validated the correspondence of the set of comparison stimuli to varying perceived strain. Perceived strain magnitudes of the comparison stimuli increased significantly and linearly with computed sharpness (r 2 = .99). A linear regression revealed that strain matching values were significantly predicted by computed sharpness (r 2 = .96) and perceived strain magnitudes (r 2 = .95) of the natural voice stimuli. CONCLUSION The perception of vocal strain is strongly associated with computed sharpness and is captured accurately and precisely using an SVMT, in which the independent variable is the bandpass filter gain (in steps of equal sharpness) applied to the comparison stimuli.
Collapse
Affiliation(s)
- Yeonggwang Park
- Department of Communication Sciences & Disorders, University of South Florida, Tampa
| | - Supraja Anand
- Department of Communication Sciences & Disorders, University of South Florida, Tampa
| | - Sophia M. Gifford
- Department of Communication Sciences & Disorders, University of South Florida, Tampa
| | - Rahul Shrivastav
- Office of the Provost & Executive Vice President, Indiana University, Bloomington
| | - David A. Eddins
- Department of Communication Sciences & Disorders, University of South Florida, Tampa
| |
Collapse
|
7
|
Eddins AC, Ozmeral EJ, Eddins DA. Aging alters across-hemisphere cortical dynamics during binaural temporal processing. Front Neurosci 2023; 16:1060172. [PMID: 36703999 PMCID: PMC9871896 DOI: 10.3389/fnins.2022.1060172] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2022] [Accepted: 12/16/2022] [Indexed: 01/11/2023] Open
Abstract
Differences in the timing and intensity of sounds arriving at the two ears provide fundamental binaural cues that help us localize and segregate sounds in the environment. Neural encoding of these cues is commonly represented asymmetrically in the cortex with stronger activation in the hemisphere contralateral to the perceived spatial location. Although advancing age is known to degrade the perception of binaural cues, less is known about how the neural representation of such cues is impacted by age. Here, we use electroencephalography (EEG) to investigate age-related changes in the hemispheric distribution of interaural time difference (ITD) encoding based on cortical auditory evoked potentials (CAEPs) and derived binaural interaction component (BIC) measures in ten younger and ten older normal-hearing adults. Sensor-level analyses of the CAEP and BIC showed age-related differences in global field power, where older listeners had significantly larger responses than younger for both binaural metrics. Source-level analyses showed hemispheric differences in auditory cortex activity for left and right lateralized stimuli in younger adults, consistent with a contralateral activation model for processing ITDs. Older adults, however, showed reduced hemispheric asymmetry across ITDs, despite having overall larger responses than younger adults. Further, when averaged across ITD condition to evaluate changes in cortical asymmetry over time, there was a significant shift in laterality corresponding to the peak components (P1, N1, P2) in the source waveform that also was affected by age. These novel results demonstrate across-hemisphere cortical dynamics during binaural temporal processing that are altered with advancing age.
Collapse
Affiliation(s)
- Ann Clock Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
- School of Communication Sciences and Disorders, University of Central Florida, Orlando, FL, United States
| | - Erol J. Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
| | - David A. Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, United States
| |
Collapse
|
8
|
Park Y, Anand S, Kopf LM, Shrivastav R, Eddins DA. Interactions Between Breathy and Rough Voice Qualities and Their Contributions to Overall Dysphonia Severity. J Speech Lang Hear Res 2022; 65:4071-4084. [PMID: 36260821 PMCID: PMC9940885 DOI: 10.1044/2022_jslhr-22-00012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/03/2023]
Abstract
PURPOSE Dysphonic voices typically present multiple voice quality dimensions. This study investigated potential interactions between perceived breathiness and roughness and their contributions to overall dysphonia severity. METHOD Synthetic stimuli based on four talkers were created to systematically map out potential interactions. For each talker, a stimulus matrix composed of 49 stimuli (seven breathiness steps × seven roughness steps) was created by varying aspiration noise and open quotient to manipulate breathiness and superimposing amplitude modulation of varying depths to simulate roughness. One-dimensional matching (1DMA) and magnitude estimation (1DME) tasks were used to measure perceived breathiness, roughness, their potential interactions, and overall dysphonia severity. Additional 1DME tasks were used to assess a set of natural stimuli that varied along both breathiness and roughness. RESULTS For the synthetic stimuli, the 1DMA task indicated little interaction between the two voice qualities. For the 1DME task, breathiness magnitude was influenced by roughness step to a greater extent than roughness magnitude was influenced by breathiness step. The additive contributions of breathiness and roughness to overall severity gradually diminished with increasing breathiness and roughness steps, possibly reflecting a ceiling effect in the 1DME task. For the natural stimuli, little consistent interaction was observed between breathiness and roughness. CONCLUSIONS The matching task revealed minimal interaction between perceived breathiness and roughness, whereas the magnitude estimation task revealed some interaction between the two qualities and their cumulative contributions to overall dysphonia severity. Task differences are discussed in terms of differences in response bias and the role of perceptual anchors. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.21313701.
Collapse
Affiliation(s)
- Yeonggwang Park
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| | - Supraja Anand
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| | - Lisa M. Kopf
- Department of Speech, Language and Hearing Sciences, The George Washington University, Washington, DC
| | - Rahul Shrivastav
- Office of the Provost and Executive Vice President, Indiana University, Bloomington
- Department of Communicative Sciences and Disorders, Michigan State University, East Lansing
| | - David A. Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
- Department of Communicative Sciences and Disorders, Michigan State University, East Lansing
| |
Collapse
|
9
|
Park Y, Anand S, Ozmeral EJ, Shrivastav R, Eddins DA. Predicting Perceived Vocal Roughness Using a Bio-Inspired Computational Model of Auditory Temporal Envelope Processing. J Speech Lang Hear Res 2022; 65:2748-2758. [PMID: 35867607 PMCID: PMC9911094 DOI: 10.1044/2022_jslhr-22-00101] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/15/2022] [Revised: 04/14/2022] [Accepted: 04/25/2022] [Indexed: 06/15/2023]
Abstract
PURPOSE Vocal roughness is often present in many voice disorders but the assessment of roughness mainly depends on the subjective auditory-perceptual evaluation and lacks acoustic correlates. This study aimed to apply the concept of roughness in general sound quality perception to vocal roughness assessment and to characterize the relationship between vocal roughness and temporal envelop fluctuation measures obtained from an auditory model. METHOD Ten /ɑ/ recordings with a wide range of roughness were selected from an existing database. Ten listeners rated the roughness of the recordings in a single-variable matching task. Temporal envelope fluctuations of the recordings were analyzed with an auditory processing model of amplitude modulation that utilizes a modulation filterbank of different modulation frequencies. Pitch strength and the smoothed cepstral peak prominence were also obtained for comparison. RESULTS Individual simple regression models yielded envelope standard deviation from a modulation filter with a low center frequency (64.3 Hz) as a statistically significant predictor of vocal roughness with a strong coefficient of determination (r 2 = .80). Pitch strength and CPPS were not significant predictors of roughness. CONCLUSION This result supports the possible utility of envelope fluctuation measures from an auditory model as objective correlates of vocal roughness.
Collapse
Affiliation(s)
- Yeonggwang Park
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| | - Supraja Anand
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| | - Erol J. Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| | - Rahul Shrivastav
- Office of the Provost & Executive Vice President, Indiana University Bloomington
| | - David A. Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| |
Collapse
|
10
|
Lelo de Larrea-Mancera ES, Stavropoulos T, Carrillo AA, Cheung S, He YJ, Eddins DA, Molis MR, Gallun FJ, Seitz AR. Remote auditory assessment using Portable Automated Rapid Testing (PART) and participant-owned devices. J Acoust Soc Am 2022; 152:807. [PMID: 36050190 PMCID: PMC9355663 DOI: 10.1121/10.0013221] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/14/2022] [Revised: 07/11/2022] [Accepted: 07/13/2022] [Indexed: 05/23/2023]
Abstract
Remote testing of auditory function can be transformative to both basic research and hearing healthcare; however, historically, many obstacles have limited remote collection of reliable and valid auditory psychometric data. Here, we report performance on a battery of auditory processing tests using a remotely administered system, Portable Automatic Rapid Testing. We compare a previously reported dataset collected in a laboratory setting with the same measures using uncalibrated, participant-owned devices in remote settings (experiment 1, n = 40) remote with and without calibrated hardware (experiment 2, n = 36) and laboratory with and without calibrated hardware (experiment 3, n = 58). Results were well-matched across datasets and had similar reliability, but overall performance was slightly worse than published norms. Analyses of potential nuisance factors such as environmental noise, distraction, or lack of calibration failed to provide reliable evidence that these factors contributed to the observed variance in performance. These data indicate feasibility of remote testing of suprathreshold auditory processing using participants' own devices. Although the current investigation was limited to young participants without hearing difficulties, its outcomes demonstrate the potential for large-scale, remote hearing testing of more hearing-diverse populations both to advance basic science and to establish the clinical viability of auditory remote testing.
Collapse
Affiliation(s)
| | - Trevor Stavropoulos
- Brain Game Center, University of California, Riverside, 1201 University Avenue #204, Riverside, California 92507, USA
| | - Audrey Anna Carrillo
- Brain Game Center, University of California, Riverside, 1201 University Avenue #204, Riverside, California 92507, USA
| | - Sierra Cheung
- Brain Game Center, University of California, Riverside, 1201 University Avenue #204, Riverside, California 92507, USA
| | - Yue J He
- Brain Game Center, University of California, Riverside, 1201 University Avenue #204, Riverside, California 92507, USA
| | - David A Eddins
- University of South Florida, 4202 East Fowler Avenue, Tampa, Florida 33620, USA
| | - Michelle R Molis
- Oregon Health and Science University, 3181 Southwest Sam Jackson Park Road, Portland, Oregon 97239-3098, USA
| | - Frederick J Gallun
- Oregon Health and Science University, 3181 Southwest Sam Jackson Park Road, Portland, Oregon 97239-3098, USA
| | - Aaron R Seitz
- Psychology Department, University of California, Riverside, 900 University Avenue, Riverside, California 92507, USA
| |
Collapse
|
11
|
Gallun FJ, Coco L, Koerner TK, de Larrea-Mancera ESL, Molis MR, Eddins DA, Seitz AR. Relating Suprathreshold Auditory Processing Abilities to Speech Understanding in Competition. Brain Sci 2022; 12:brainsci12060695. [PMID: 35741581 PMCID: PMC9221421 DOI: 10.3390/brainsci12060695] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2022] [Revised: 05/17/2022] [Accepted: 05/25/2022] [Indexed: 11/28/2022] Open
Abstract
(1) Background: Difficulty hearing in noise is exacerbated in older adults. Older adults are more likely to have audiometric hearing loss, although some individuals with normal pure-tone audiograms also have difficulty perceiving speech in noise. Additional variables also likely account for speech understanding in noise. It has been suggested that one important class of variables is the ability to process auditory information once it has been detected. Here, we tested a set of these “suprathreshold” auditory processing abilities and related them to performance on a two-part test of speech understanding in competition with and without spatial separation of the target and masking speech. Testing was administered in the Portable Automated Rapid Testing (PART) application developed by our team; PART facilitates psychoacoustic assessments of auditory processing. (2) Methods: Forty-one individuals (average age 51 years), completed assessments of sensitivity to temporal fine structure (TFS) and spectrotemporal modulation (STM) detection via an iPad running the PART application. Statistical models were used to evaluate the strength of associations between performance on the auditory processing tasks and speech understanding in competition. Age and pure-tone-average (PTA) were also included as potential predictors. (3) Results: The model providing the best fit also included age and a measure of diotic frequency modulation (FM) detection but none of the other potential predictors. However, even the best fitting models accounted for 31% or less of the variance, supporting work suggesting that other variables (e.g., cognitive processing abilities) also contribute significantly to speech understanding in noise. (4) Conclusions: The results of the current study do not provide strong support for previous suggestions that suprathreshold processing abilities alone can be used to explain difficulties in speech understanding in competition among older adults. This discrepancy could be due to the speech tests used, the listeners tested, or the suprathreshold tests chosen. Future work with larger numbers of participants is warranted, including a range of cognitive tests and additional assessments of suprathreshold auditory processing abilities.
Collapse
Affiliation(s)
- Frederick J. Gallun
- Oregon Hearing Research Center, Oregon Health & Science University, Portland, OR 97239, USA; (L.C.); (T.K.K.)
- VA RR&D National Center for Rehabilitative Auditory Research, VA Portland Health Care System, Portland, OR 97239, USA;
- Correspondence: ; Tel.: +1-503-494-4331
| | - Laura Coco
- Oregon Hearing Research Center, Oregon Health & Science University, Portland, OR 97239, USA; (L.C.); (T.K.K.)
- VA RR&D National Center for Rehabilitative Auditory Research, VA Portland Health Care System, Portland, OR 97239, USA;
| | - Tess K. Koerner
- Oregon Hearing Research Center, Oregon Health & Science University, Portland, OR 97239, USA; (L.C.); (T.K.K.)
- VA RR&D National Center for Rehabilitative Auditory Research, VA Portland Health Care System, Portland, OR 97239, USA;
| | | | - Michelle R. Molis
- VA RR&D National Center for Rehabilitative Auditory Research, VA Portland Health Care System, Portland, OR 97239, USA;
| | - David A. Eddins
- Department of Communication Science & Disorders, University of South Florida, Tampa, FL 33620, USA;
| | - Aaron R. Seitz
- Department of Psychology, University of California, Riverside, CA 92521, USA; (E.S.L.d.L.-M.); (A.R.S.)
| |
Collapse
|
12
|
Lertpoompunya A, Ozmeral EJ, Higgins NC, Eddins AC, Eddins DA. Large group differences in binaural sensitivity are represented in preattentive responses from auditory cortex. J Neurophysiol 2022; 127:660-672. [PMID: 35108112 PMCID: PMC8896993 DOI: 10.1152/jn.00360.2021] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2021] [Revised: 01/04/2022] [Accepted: 01/25/2022] [Indexed: 11/22/2022] Open
Abstract
Correlated sounds presented to two ears are perceived as compact and centrally lateralized, whereas decorrelation between ears leads to intracranial image widening. Though most listeners have fine resolution for perceptual changes in interaural correlation (IAC), some investigators have reported large variability in IAC thresholds, and some normal-hearing listeners even exhibit seemingly debilitating IAC thresholds. It is unknown whether or not this variability across individuals and outlier manifestations are a product of task difficulty, poor training, or a neural deficit in the binaural auditory system. The purpose of this study was first to identify listeners with normal and abnormal IAC resolution, second to evaluate the neural responses elicited by IAC changes, and third to use a well-established model of binaural processing to determine a potential explanation for observed individual variability. Nineteen subjects were enrolled in the study, eight of whom were identified as poor performers in the IAC-threshold task. Global scalp responses (N1 and P2 amplitudes of an auditory change complex) in the individuals with poor IAC behavioral thresholds were significantly smaller than for listeners with better IAC resolution. Source-localized evoked responses confirmed this group effect in multiple subdivisions of the auditory cortex, including Heschl's gyrus, planum temporale, and the temporal sulcus. In combination with binaural modeling results, this study provides objective electrophysiological evidence of a binaural processing deficit linked to internal noise, that corresponds to very poor IAC thresholds in listeners that otherwise have normal audiometric profiles and lack spatial hearing complaints.NEW & NOTEWORTHY Group differences in the perception of interaural correlation (IAC) were observed in human adults with normal audiometric sensitivity. These differences were reflected in cortical-evoked activity measured via electroencephalography (EEG). For some participants, weak representation of the binaural cue at the cortical level in preattentive N1-P2 cortical responses may be indicative of a potential processing deficit. Such a deficit may be related to a poorly understood condition known as hidden hearing loss.
Collapse
Affiliation(s)
- Angkana Lertpoompunya
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
- Department of Communication Sciences and Disorders, Mahidol University, Bangkok, Thailand
| | - Erol J Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | - Nathan C Higgins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | - Ann C Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
- Department of Communication Sciences and Disorders, Mahidol University, Bangkok, Thailand
| | - David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
- Department of Communication Sciences and Disorders, Mahidol University, Bangkok, Thailand
| |
Collapse
|
13
|
Devaraju DS, Kemp A, Eddins DA, Shrivastav R, Chandrasekaran B, Hampton Wray A. Effects of Task Demands on Neural Correlates of Acoustic and Semantic Processing in Challenging Listening Conditions. J Speech Lang Hear Res 2021; 64:3697-3706. [PMID: 34403278 DOI: 10.1044/2021_jslhr-21-00006] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Purpose Listeners shift their listening strategies between lower level acoustic information and higher level semantic information to prioritize maximum speech intelligibility in challenging listening conditions. Although increasing task demands via acoustic degradation modulates lexical-semantic processing, the neural mechanisms underlying different listening strategies are unclear. The current study examined the extent to which encoding of lower level acoustic cues is modulated by task demand and associations with lexical-semantic processes. Method Electroencephalography was acquired while participants listened to sentences in the presence of four-talker babble that contained either higher or lower probability final words. Task difficulty was modulated by time available to process responses. Cortical tracking of speech-neural correlates of acoustic temporal envelope processing-were estimated using temporal response functions. Results Task difficulty did not affect cortical tracking of temporal envelope of speech under challenging listening conditions. Neural indices of lexical-semantic processing (N400 amplitudes) were larger with increased task difficulty. No correlations were observed between the cortical tracking of temporal envelope of speech and lexical-semantic processes, even after controlling for the effect of individualized signal-to-noise ratios. Conclusions Cortical tracking of the temporal envelope of speech and semantic processing are differentially influenced by task difficulty. While increased task demands modulated higher level semantic processing, cortical tracking of the temporal envelope of speech may be influenced by task difficulty primarily when the demand is manipulated in terms of acoustic properties of the stimulus, consistent with an emerging perspective in speech perception.
Collapse
Affiliation(s)
- Dhatri S Devaraju
- Department of Communication Science and Disorders, University of Pittsburgh, PA
| | - Amy Kemp
- Department of Communication Sciences and Special Education, University of Georgia, Athens
| | - David A Eddins
- Department of Communication Sciences & Disorders, University of South Florida, Tampa
| | | | | | - Amanda Hampton Wray
- Department of Communication Science and Disorders, University of Pittsburgh, PA
| |
Collapse
|
14
|
Palandrani KN, Hoover EC, Stavropoulos T, Seitz AR, Isarangura S, Gallun FJ, Eddins DA. Temporal integration of monaural and dichotic frequency modulation. J Acoust Soc Am 2021; 150:745. [PMID: 34470296 PMCID: PMC8337085 DOI: 10.1121/10.0005729] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/16/2020] [Revised: 06/17/2021] [Accepted: 07/02/2021] [Indexed: 05/06/2023]
Abstract
Frequency modulation (FM) detection at low modulation frequencies is commonly used as an index of temporal fine-structure processing. The present study evaluated the rate of improvement in monaural and dichotic FM across a range of test parameters. In experiment I, dichotic and monaural FM detection was measured as a function of duration and modulator starting phase. Dichotic FM thresholds were lower than monaural FM thresholds and the modulator starting phase had no effect on detection. Experiment II measured monaural FM detection for signals that differed in modulation rate and duration such that the improvement with duration in seconds (carrier) or cycles (modulator) was compared. Monaural FM detection improved monotonically with the number of modulation cycles, suggesting that the modulator is extracted prior to detection. Experiment III measured dichotic FM detection for shorter signal durations to test the hypothesis that dichotic FM relies primarily on the signal onset. The rate of improvement decreased as duration increased, which is consistent with the use of primarily onset cues for the detection of dichotic FM. These results establish that improvement with duration occurs as a function of the modulation cycles at a rate consistent with the independent-samples model for monaural FM, but later cycles contribute less to detection in dichotic FM.
Collapse
Affiliation(s)
- Katherine N Palandrani
- Department of Communication Sciences and Disorders, University of Maryland, College Park, Maryland 20742, USA
| | - Eric C Hoover
- Department of Communication Sciences and Disorders, University of Maryland, College Park, Maryland 20742, USA
| | - Trevor Stavropoulos
- Brain Game Center, University of California Riverside, Riverside, California 92521, USA
| | - Aaron R Seitz
- Department of Psychology, University of California Riverside, Riverside, California 92521, USA
| | - Sittiprapa Isarangura
- Department of Communication Sciences and Disorders, Mahidol University, Phaya Thai, Bangkok 10400, Thailand
| | - Frederick J Gallun
- Oregon Hearing Research Center, Oregon Health and Science University, Portland, Oregon 97239, USA
| | - David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida 33620, USA
| |
Collapse
|
15
|
Ozmeral EJ, Eddins DA, Eddins AC. Selective auditory attention modulates cortical responses to sound location change in younger and older adults. J Neurophysiol 2021; 126:803-815. [PMID: 34288759 DOI: 10.1152/jn.00609.2020] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/01/2023] Open
Abstract
The present study measured scalp potentials in response to low-frequency, narrowband noise bursts changing location in the front, azimuthal plane. At question was whether selective auditory attention has a modulatory effect on the cortical encoding of spatial change and whether older listeners with normal-hearing thresholds would show depressed cortical representation for spatial changes relative to younger listeners. Young and older normal-hearing listeners were instructed to either passively listen to the stimulus presentation or actively attend to a single location (either 30° left or right of midline) and detect when a noise stream moved to the attended location. Prominent peaks of the electroencephalographic scalp waveforms were compared across groups, locations, and attention conditions. In addition, an opponent-channel model of spatial coding was performed to capture the effect of attention on spatial-change tuning. Younger listeners showed not only larger responses overall but a greater dynamic range in their response to location changes. Results suggest that younger listeners were acquiring and encoding key spatial cues at early cortical processing areas. On the other hand, each group exhibited modulatory effects of attention to spatial-change tuning, indicating that both younger and older listeners selectively attend to space in a manner that amplifies the available signal.NEW & NOTEWORTHY In complex acoustic scenes, listeners take advantage of spatial cues to selectively attend to sounds that are deemed immediately relevant. At the neural level, selective attention amplifies electrical responses to spatial changes. We tested whether older and younger listeners have comparable modulatory effects of attention to stimuli moving in the free field. Results indicate that although older listeners do have depressed overall responses, selective attention enhances spatial-change tuning in younger and older listeners alike.
Collapse
Affiliation(s)
- Erol J Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | - David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | - Ann Clock Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| |
Collapse
|
16
|
Eddins DA, Anand S, Lang A, Shrivastav R. Developing Clinically Relevant Scales of Breathy and Rough Voice Quality. J Voice 2021; 35:663.e9-663.e16. [PMID: 31932189 PMCID: PMC10413723 DOI: 10.1016/j.jvoice.2019.12.021] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2019] [Revised: 12/22/2019] [Accepted: 12/23/2019] [Indexed: 01/14/2023]
Abstract
The most common measurement tools used in the perceptual evaluation of voice quality yield ordinal data and thus do not support the establishment of mathematical relationships among different measurement values. This makes their interpretation challenging. Among the many desirable features of any psychophysical measurement tool is the ability to quantify the difference between two or more measurements and the ability to interpret the measurements in a manner that is related to the experience of the observer. The former allows one to compare among measurements using simple mathematics, while the latter allows that comparison to be interpreted in constructive ways. In this paper we describe the development of standard measurement scales for two dimensions of voice quality, following an approach that has been applied successfully to the perception of loudness. The scales follow step-by-step procedures used to develop the sone scale of loudness, which ties physical units to the perceptual estimates of loudness magnitude. Goals of the current work include development of analogous scales for the perception of breathy and rough voice qualities. First, the relationship between perceived voice quality and physical units were established using single-variable matching tasks. Second, the relationship between a change in physical units from the single-variable matching tasks and perceived voice quality magnitude were established using magnitude estimation tasks. Third, single reference points were identified on breathy and rough continuums. Finally, all points on the newly established voice quality continuums were rescaled relative to these arbitrary reference points. The proposed breathiness and roughness scales result in ratio-level data with standard measurement units that support quantitative comparisons of perceptual judgments. Such judgments can be used, for example, to compare magnitude of change pre- and post-treatment.
Collapse
Affiliation(s)
- David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida.
| | - Supraja Anand
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | | | - Rahul Shrivastav
- Office of the Vice President for Instruction, University of Georgia, Athens, Georgia
| |
Collapse
|
17
|
Stavropoulos TA, Isarangura S, Hoover EC, Eddins DA, Seitz AR, Gallun FJ. Exponential spectro-temporal modulation generation. J Acoust Soc Am 2021; 149:1434. [PMID: 33765775 PMCID: PMC8097710 DOI: 10.1121/10.0003604] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/12/2020] [Revised: 01/20/2021] [Accepted: 02/06/2021] [Indexed: 05/23/2023]
Abstract
Traditionally, real-time generation of spectro-temporally modulated noise has been performed on a linear amplitude scale, partially due to computational constraints. Experiments often require modulation that is sinusoidal on a logarithmic amplitude scale as a result of the many perceptual and physiological measures which scale linearly with exponential changes in the signal magnitude. A method is presented for computing exponential spectro-temporal modulation, showing that it can be expressed analytically as a sum over linearly offset sidebands with component amplitudes equal to the values of the modified Bessel function of the first kind. This approach greatly improves the efficiency and precision of stimulus generation over current methods, facilitating real-time generation for a broad range of carrier and envelope signals.
Collapse
Affiliation(s)
- Trevor A Stavropoulos
- Brain Game Center for Mental Fitness and Well-being, University of California, Riverside, California 92521, USA
| | - Sittiprapa Isarangura
- Department of Communication Sciences and Disorders, Mahidol University, Bangkok, Thailand
| | - Eric C Hoover
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| | - David A Eddins
- Auditory and Speech Science Laboratory, University of South Florida, Tampa, Florida 33612, USA
| | - Aaron R Seitz
- Brain Game Center for Mental Fitness and Well-being, University of California, Riverside, California 92521, USA
| | - Frederick J Gallun
- National Center for Rehabilitative Auditory Research, Portland VA Medical Center, Portland, Oregon 97239, USA
| |
Collapse
|
18
|
Eddins DA. Select Papers From the 8th Aging and Speech Communication Conference. J Speech Lang Hear Res 2021; 64:299-301. [PMID: 33561358 DOI: 10.1044/2021_jslhr-21-00031] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Purpose The purpose of this introduction is to briefly describe the nature of the conference, Aging and Speech Communication: An International and Interdisciplinary Research Conference, and to introduce the articles featured in this forum that represent the nature of the biennial conference.
Collapse
Affiliation(s)
- David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL
| |
Collapse
|
19
|
Yellamsetty A, Ozmeral EJ, Budinsky RA, Eddins DA. A Comparison of Environment Classification Among Premium Hearing Instruments. Trends Hear 2021; 25:2331216520980968. [PMID: 33749410 PMCID: PMC7989119 DOI: 10.1177/2331216520980968] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2020] [Revised: 11/20/2020] [Accepted: 11/24/2020] [Indexed: 11/21/2022] Open
Abstract
Hearing aids classify acoustic environments into multiple, generic classes for the purposes of guiding signal processing. Information about environmental classification is made available to the clinician for fitting, counseling, and troubleshooting purposes. The goal of this study was to better inform scientists and clinicians about the nature of that information by comparing the classification schemes among five premium hearing instruments in a wide range of acoustic scenes including those that vary in signal-to-noise ratio and overall level (dB SPL). Twenty-eight acoustic scenes representing various prototypical environments were presented to five premium devices mounted on an acoustic manikin. Classification measures were recorded from the brand-specific fitting software then recategorized to generic labels to conceal the device company, including (a) Speech in Quiet, (b) Speech in Noise, (c) Noise, and (d) Music. Twelve normal-hearing listeners also classified each scene. The results revealed a variety of similarities and differences among the five devices and the human subjects. Where some devices were highly dependent on input overall level, others were influenced markedly by signal-to-noise ratio. Differences between human and hearing aid classification were evident for several speech and music scenes. Environmental classification is the heart of the signal processing strategy for any given device, providing key input to subsequent decision-making. Comprehensive assessment of environmental classification is essential when considering the cost of signal processing errors, the potential impact for typical wearers, and the information that is available for use by clinicians. The magnitude of differences among devices is remarkable and to be noted.
Collapse
Affiliation(s)
- Anusha Yellamsetty
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida, United States
| | - Erol J. Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida, United States
| | - Robert A. Budinsky
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida, United States
| | - David A. Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida, United States
| |
Collapse
|
20
|
Lelo de Larrea-Mancera ES, Stavropoulos T, Hoover EC, Eddins DA, Gallun FJ, Seitz AR. Portable Automated Rapid Testing (PART) for auditory assessment: Validation in a young adult normal-hearing population. J Acoust Soc Am 2020; 148:1831. [PMID: 33138479 PMCID: PMC7541091 DOI: 10.1121/10.0002108] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/26/2019] [Revised: 09/14/2020] [Accepted: 09/16/2020] [Indexed: 05/23/2023]
Abstract
This study aims to determine the degree to which Portable Automated Rapid Testing (PART), a freely available program running on a tablet computer, is capable of reproducing standard laboratory results. Undergraduate students were assigned to one of three within-subject conditions that examined repeatability of performance on a battery of psychoacoustical tests of temporal fine structure processing, spectro-temporal amplitude modulation, and targets in competition. The repeatability condition examined test/retest with the same system, the headphones condition examined the effects of varying headphones (passive and active noise-attenuating), and the noise condition examined repeatability in the presence of recorded cafeteria noise. In general, performance on the test battery showed high repeatability, even across manipulated conditions, and was similar to that reported in the literature. These data serve as validation that suprathreshold psychoacoustical tests can be made accessible to run on consumer-grade hardware and perform in less controlled settings. This dataset also provides a distribution of thresholds that can be used as a normative baseline against which auditory dysfunction can be identified in future work.
Collapse
Affiliation(s)
| | - Trevor Stavropoulos
- Brain Game Center, University of California Riverside, 1201 University Avenue, Riverside California 92521, USA
| | - Eric C Hoover
- University of Maryland, College Park, Maryland 20742, USA
| | | | | | - Aaron R Seitz
- Psychology Department, University of California, Riverside, 900 University Avenue, Riverside, California 92521, USA
| |
Collapse
|
21
|
Ozmeral EJ, Hoover EC, Gabbidon P, Eddins DA. Development of the Continuous Number Identification Test (CNIT): feasibility of dynamic assessment of speech intelligibility. Int J Audiol 2020; 59:434-442. [PMID: 32003257 DOI: 10.1080/14992027.2020.1718782] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Abstract
Objective: The present study was motivated by a need for a speech intelligibility test capable of indexing dynamic changes in the environment and adaptive processing in hearing aids. The Continuous Number Identification Test (CNIT) was developed to meet these aims.Design: From one location in the free field, speech was presented in noise (∼2 words/s) with a 100-ms inter-word interval. On average, every fourth word was a target digit and all other words were monosyllabic words. Non-numeric words had a fixed presentation level such that the dominant signal-to-noise-ratio (SNR) was held at +6 dB SNR relative to background maskers. To prevent ceiling effects, however, targets were presented at a user-specific SNR, determined by an initial adaptive-tracking procedure that estimated the 79.4% speech reception threshold.Study sample: Ten normal-hearing listeners participated.Results: The CNIT showed comparable psychometric qualities of other established speech tests for long time scales (Exp. 1). Target-location changes did not affect performance on the CNIT (Exp. 2), but the test did show high temporal resolution in assessing sudden changes to SNR (Exp. 3).Conclusions: The CNIT is highly customisable, and the initial experiments tested feasibility of its primary features which set it apart from currently available speech-in-noise tests.
Collapse
Affiliation(s)
- Erol J Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, USA
| | - Eric C Hoover
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| | - Patricia Gabbidon
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, USA
| | - David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, FL, USA.,Department of Chemical and Biomedical Engineering, University of South Florida, Tampa, FL, USA
| |
Collapse
|
22
|
Anand S, Kopf LM, Shrivastav R, Eddins DA. Objective Indices of Perceived Vocal Strain. J Voice 2019; 33:838-845. [DOI: 10.1016/j.jvoice.2018.06.005] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2018] [Revised: 06/06/2018] [Accepted: 06/07/2018] [Indexed: 10/28/2022]
|
23
|
Isarangura S, Eddins AC, Ozmeral EJ, Eddins DA. The Effects of Duration and Level on Spectral Modulation Perception. J Speech Lang Hear Res 2019; 62:3876-3886. [PMID: 31638883 PMCID: PMC7838824 DOI: 10.1044/2019_jslhr-h-18-0449] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/09/2018] [Revised: 04/02/2019] [Accepted: 07/23/2019] [Indexed: 05/30/2023]
Abstract
Purpose Spectral modulation detection is an increasingly common assay of suprathreshold auditory perception and has been correlated with speech perception performance. Here, the potential effects of stimulus duration and stimulus presentation level on spectral modulation detection were investigated. Method Spectral modulation detection thresholds were measured as a function of modulation frequency in young, normal-hearing listeners. The standard stimulus was a bandpass noise, and signal stimuli were created by superimposing sinusoidal spectral modulation on the bandpass noise carrier. The modulation was sinusoidal on a log2 frequency axis and a log10 (dB) amplitude scale with a random starting phase (0-2π radians). In 1 experiment, stimulus durations were 50, 100, 200, or 400 ms (at fixed level 81 dB SPL). In a 2nd experiment, stimuli were presented at sensation levels of 10, 20, 30, 40, and 60 dB SL (fixed at a duration of 400 ms). Results Spectral modulation detection thresholds were similarly low for the 400- and 200-ms durations, increased slightly for the 100-ms duration, and increased markedly for the 50-ms duration. Thresholds were lowest for 40 dB SL; increased slightly for 20, 30, and 60 dB SL; and markedly higher for the 10-dB SL condition. Conclusions The increase in thresholds for the shortest durations and lowest sensational levels is consistent with previous investigations of auditory spectral profile analysis. The effects of presentation level and stimulus duration are important considerations in the context of understanding potential relationships between the perception of spectral cues and speech perception, when designing investigations and interpreting data related to spectral envelope perception, and in the context of models of auditory perception. As examples, 2 simple models based on auditory nerve output that have been used to explain spectrotemporal modulation in previous investigations produced an output inconsistent with the present results. Plain language summary Intensity variations across audio frequency lead to spectral shapes that are essential and sometimes signature features of various sounds in the environment, including speech. Here, we show how laboratory measures of spectral shape perception depend on presentation level and stimulus duration.
Collapse
Affiliation(s)
- Sittiprapa Isarangura
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| | - Ann C. Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| | - Erol J. Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| | - David A. Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| |
Collapse
|
24
|
Anand S, Skowronski MD, Shrivastav R, Eddins DA. Perceptual and Quantitative Assessment of Dysphonia Across Vowel Categories. J Voice 2019; 33:473-481. [DOI: 10.1016/j.jvoice.2017.12.018] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2017] [Accepted: 12/21/2017] [Indexed: 10/16/2022]
|
25
|
Ozmeral EJ, Eddins DA, Eddins AC. Electrophysiological responses to lateral shifts are not consistent with opponent-channel processing of interaural level differences. J Neurophysiol 2019; 122:737-748. [PMID: 31242052 DOI: 10.1152/jn.00090.2019] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Cortical encoding of auditory space relies on two major peripheral cues, interaural time difference (ITD) and interaural level difference (ILD) of the sounds arriving at a listener's ears. In much of the precortical auditory pathway, ITD and ILD cues are processed independently, and it is assumed that cue integration is a higher order process. However, there remains debate on how ITDs and ILDs are encoded in the cortex and whether they share a common mechanism. The present study used electroencephalography (EEG) to measure evoked cortical potentials from narrowband noise stimuli with imposed binaural cue changes. Previous studies have similarly tested ITD shifts to demonstrate that neural populations broadly favor one spatial hemifield over the other, which is consistent with an opponent-channel model that computes the relative activity between broadly tuned neural populations. However, it is still a matter of debate whether the same coding scheme applies to ILDs and, if so, whether processing the two binaural cues is distributed across similar regions of the cortex. The results indicate that ITD and ILD cues have similar neural signatures with respect to the monotonic responses to shift magnitude; however, the direction of the shift did not elicit responses equally across cues. Specifically, ITD shifts evoked greater responses for outward than inward shifts, independently of the spatial hemifield of the shift, whereas ILD-shift responses were dependent on the hemifield in which the shift occurred. Active cortical structures showed only minor overlap between responses to cues, suggesting the two are not represented by the same pathway.NEW & NOTEWORTHY Interaural time differences (ITDs) and interaural level differences (ILDs) are critical to locating auditory sources in the horizontal plane. The higher order perceptual feature of auditory space is thought to be encoded together by these binaural differences, yet evidence of their integration in cortex remains elusive. Although present results show some common effects between the two cues, key differences were observed that are not consistent with an ITD-like opponent-channel process for ILD encoding.
Collapse
Affiliation(s)
- Erol J Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | - David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida.,Department of Chemical and Biomedical Engineering, University of South Florida, Tampa, Florida
| | - Ann Clock Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| |
Collapse
|
26
|
Hoover EC, Kinney BN, Bell KL, Gallun FJ, Eddins DA. A Comparison of Behavioral Methods for Indexing the Auditory Processing of Temporal Fine Structure Cues. J Speech Lang Hear Res 2019; 62:2018-2034. [PMID: 31145649 PMCID: PMC6808371 DOI: 10.1044/2019_jslhr-h-18-0217] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/30/2018] [Revised: 12/20/2018] [Accepted: 02/13/2019] [Indexed: 05/12/2023]
Abstract
Purpose Growing evidence supports the inclusion of perceptual tests that quantify the processing of temporal fine structure (TFS) in clinical hearing assessment. Many tasks have been used to evaluate TFS in the laboratory that vary greatly in the stimuli used and whether the judgments require monaural or binaural comparisons of TFS. The purpose of this study was to compare laboratory measures of TFS for inclusion in a battery of suprathreshold auditory tests. A subset of available TFS tasks were selected on the basis of potential clinical utility and were evaluated using metrics that focus on characteristics important for clinical use. Method TFS measures were implemented in replication of studies that demonstrated clinical utility. Monaural, diotic, and dichotic measures were evaluated in 11 young listeners with normal hearing. Measures included frequency modulation (FM) tasks, harmonic frequency shift detection, interaural phase difference (TFS-low frequency), interaural time difference (ITD), monaural gap duration discrimination, and tone detection in noise with and without a difference in interaural phase (N0S0, N0Sπ). Data were compared with published results and evaluated with metrics of consistency and efficiency. Results Thresholds obtained were consistent with published data. There was no evidence of predictive relationships among the measures consistent with a homogenous group. The most stable tasks across repeated testing were TFS-low frequency, diotic and dichotic FM, and N0Sπ. Monaural and diotic FM had the lowest normalized variance and were the most efficient accounting for differences in total test duration, followed by ITD. Conclusions Despite a long stimulus duration, FM tasks dominated comparisons of consistency and efficiency. Small differences separated the dichotic tasks FM, ITD, and N0Sπ. Future comparisons following procedural optimization of the tasks will evaluate clinical efficiency in populations with impairment.
Collapse
Affiliation(s)
- Eric C. Hoover
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| | - Brianna N. Kinney
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| | - Karen L. Bell
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| | - Frederick J. Gallun
- National Center for Rehabilitative Auditory Research, Portland VA Medical Center, Oregon
- Department of Otolaryngology–Head and Neck Surgery, Oregon Health and Science University, Portland
| | - David A. Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa
| |
Collapse
|
27
|
Eddins AC, Ozmeral EJ, Eddins DA. How aging impacts the encoding of binaural cues and the perception of auditory space. Hear Res 2018; 369:79-89. [PMID: 29759684 PMCID: PMC6196106 DOI: 10.1016/j.heares.2018.05.001] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/15/2017] [Revised: 04/23/2018] [Accepted: 05/02/2018] [Indexed: 10/17/2022]
Abstract
Over the years, the effect of aging on auditory function has been investigated in animal models and humans in an effort to characterize age-related changes in both perception and physiology. Here, we review how aging may impact neural encoding and processing of binaural and spatial cues in human listeners with a focus on recent work by the authors as well as others. Age-related declines in monaural temporal processing, as estimated from measures of gap detection and temporal fine structure discrimination, have been associated with poorer performance on binaural tasks that require precise temporal processing. In lateralization and localization tasks, as well as in the detection of signals in noise, marked age-related changes have been demonstrated in both behavioral and electrophysiological measures and have been attributed to declines in neural synchrony and reduced central inhibition with advancing age. Evidence for such mechanisms, however, are influenced by the task (passive vs. attending) and the stimulus paradigm (e.g., static vs. continuous with dynamic change). That is, cortical auditory evoked potentials (CAEP) measured in response to static interaural time differences (ITDs) are larger in older versus younger listeners, consistent with reduced inhibition, while continuous stimuli with dynamic ITD changes lead to smaller responses in older compared to younger adults, suggestive of poorer neural synchrony. Additionally, the distribution of cortical activity is broader and less asymmetric in older than younger adults, consistent with the hemispheric asymmetry reduction in older adults model of cognitive aging. When older listeners attend to selected target locations in the free field, their CAEP components (N1, P2, P3) are again consistently smaller relative to younger listeners, and the reduced asymmetry in the distribution of cortical activity is maintained. As this research matures, proper neural biomarkers for changes in spatial hearing can provide objective evidence of impairment and targets for remediation. Future research should focus on the development and evaluation of effective approaches for remediating these spatial processing deficits associated with aging and hearing loss.
Collapse
Affiliation(s)
- Ann Clock Eddins
- Department of Communication Sciences and Disorders, University of South Florida, USA.
| | - Erol J Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, USA
| | - David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, USA; Department of Chemical and Biomedical Engineering, University of South Florida, USA
| |
Collapse
|
28
|
Ozmeral EJ, Eddins AC, Eddins DA. How Do Age and Hearing Loss Impact Spectral Envelope Perception? J Speech Lang Hear Res 2018; 61:2376-2385. [PMID: 30178062 PMCID: PMC6195040 DOI: 10.1044/2018_jslhr-h-18-0056] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/13/2018] [Revised: 05/09/2018] [Accepted: 05/16/2018] [Indexed: 06/01/2023]
Abstract
Purpose The goal was to evaluate the potential effects of increasing hearing loss and advancing age on spectral envelope perception. Method Spectral modulation detection was measured as a function of spectral modulation frequency from 0.5 to 8.0 cycles/octave. The spectral modulation task involved discrimination of a noise carrier (3 octaves wide from 400 to 3200 Hz) with a flat spectral envelope from a noise having a sinusoidal spectral envelope across a logarithmic audio frequency scale. Spectral modulation transfer functions (SMTFs; modulation threshold vs. modulation frequency) were computed and compared 4 listener groups: young normal hearing, older normal hearing, older with mild hearing loss, and older with moderate hearing loss. Estimates of the internal spectral contrast were obtained by computing excitation patterns. Results SMTFs for young listeners with normal hearing were bandpass with a minimum modulation detection threshold at 2 cycles/octave, and older listeners with normal hearing were remarkably similar to those of the young listeners. SMTFs for older listeners with mild and moderate hearing loss had a low-pass rather than a bandpass shape. Excitation patterns revealed that limited spectral resolution dictated modulation detection thresholds at high but not low spectral modulation frequencies. Even when factoring out (presumed) differences in frequency resolution among groups, the spectral envelope perception was worse for the group with moderate hearing loss than the other 3 groups. Conclusions The spectral envelope perception as measured by spectral modulation detection thresholds is compromised by hearing loss at higher spectral modulation frequencies, consistent with predictions of reduced spectral resolution known to accompany sensorineural hearing loss. Spectral envelope perception is not negatively impacted by advancing age at any spectral modulation frequency between 0.5 and 8.0 cycles/octave.
Collapse
Affiliation(s)
- Erol J. Ozmeral
- Department of Communication Sciences & Disorders, University of South Florida, Tampa
| | - Ann C. Eddins
- Department of Communication Sciences & Disorders, University of South Florida, Tampa
| | - David A. Eddins
- Department of Communication Sciences & Disorders, University of South Florida, Tampa
| |
Collapse
|
29
|
Abstract
Spectral modulation transfer functions (SMTFs) were measured in 49 young (18-35 years of age) normal-hearing listeners. Noise carriers spanned six octaves from 200 to 12 800 Hz. Sinusoidal (on a log-amplitude scale) spectral modulation with random starting phase was superimposed on the carrier at spectral modulation frequencies of 0.25, 0.5, 1.0, 2.0, 4.0, and 8.0 cycles/octave. Modulation detection thresholds (in dB) yielded SMTFs that were bandpass in nature, consistent with previous investigations reporting data for only a few subjects. Thresholds were notably consistent across subjects despite minimal practice. Population statistics are reported that may serve as reference data for future studies.
Collapse
Affiliation(s)
- Eric C Hoover
- Department of Communication Sciences and Disorders, University of South Florida, 4202 East Fowler Avenue, PCD 1017, Tampa, Florida 32620, USA
| | - Ann C Eddins
- Department of Communication Sciences and Disorders, University of South Florida, 4202 East Fowler Avenue, PCD 1017, Tampa, Florida 32620, USA
| | - David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, 4202 East Fowler Avenue, PCD 1017, Tampa, Florida 32620, USA
| |
Collapse
|
30
|
Kopf LM, Skowronski MD, Anand S, Eddins DA, Shrivastav R. The Perception of Breathiness in the Voices of Pediatric Speakers. J Voice 2017; 33:204-213. [PMID: 29162356 DOI: 10.1016/j.jvoice.2017.09.024] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2017] [Revised: 09/27/2017] [Accepted: 09/28/2017] [Indexed: 10/18/2022]
Abstract
BACKGROUND The perception of pediatric voice quality has been investigated using clinical protocols developed for adult voices and acoustic analyses designed to identify important physical parameters associated with normal and dysphonic pediatric voices. Laboratory investigations of adult dysphonia have included sophisticated methods, including a psychoacoustic approach that involves a single-variable matching task (SVMT), characterized by high inter- and intra-listener reliability, and analyses that include bio-inspired models of auditory perception that have provided valuable information regarding adult voice quality. OBJECTIVES To establish the utility of a psychoacoustic approach to the investigation of voice quality perception in the context of pediatric voices? METHODS Six listeners judged the breathiness of 20 synthetic vowel stimuli using an SVMT. To support comparisons with previous data, stimuli were modeled after four pediatric speakers and synthesized using Klatt with five parameter settings that influence the perception of breathiness. The population average breathiness judgments were modeled with acoustic measures of loudness ratio, pitch strength, and cepstral peak. RESULTS Listeners reliably judged the perceived breathiness of pediatric voices, as with previous investigations of breathiness in adult dysphonic voices. Breathiness judgments were accurately modeled by loudness ratio (r2 = 0.93), pitch strength (r2 = 0.91), and cepstral peak (r2 = 0.82). Model accuracy was not affected significantly by including stimulus fundamental frequency and was slightly higher for pediatric than for adult voices. CONCLUSIONS The SVMT proved robust for pediatric voices spanning a wide range of breathiness. The data indicate that this is a promising approach for future investigation of pediatric voice quality.
Collapse
Affiliation(s)
- Lisa M Kopf
- Department of Communication Sciences and Disorders, University of Northern Iowa, Cedar Falls, Iowa
| | | | - Supraja Anand
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | - David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida.
| | - Rahul Shrivastav
- Office of the Vice President for Instruction, University of Georgia, Athens, Georgia
| |
Collapse
|
31
|
Ozmeral EJ, Eddins DA, Eddins AC. Reduced temporal processing in older, normal-hearing listeners evident from electrophysiological responses to shifts in interaural time difference. J Neurophysiol 2016; 116:2720-2729. [PMID: 27683889 PMCID: PMC5133308 DOI: 10.1152/jn.00560.2016] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2016] [Accepted: 09/24/2016] [Indexed: 11/22/2022] Open
Abstract
Previous electrophysiological studies of interaural time difference (ITD) processing have demonstrated that ITDs are represented by a nontopographic population rate code. Rather than narrow tuning to ITDs, neural channels have broad tuning to ITDs in either the left or right auditory hemifield, and the relative activity between the channels determines the perceived lateralization of the sound. With advancing age, spatial perception weakens and poor temporal processing contributes to declining spatial acuity. At present, it is unclear whether age-related temporal processing deficits are due to poor inhibitory controls in the auditory system or degraded neural synchrony at the periphery. Cortical processing of spatial cues based on a hemifield code are susceptible to potential age-related physiological changes. We consider two distinct predictions of age-related changes to ITD sensitivity: declines in inhibitory mechanisms would lead to increased excitation and medial shifts to rate-azimuth functions, whereas a general reduction in neural synchrony would lead to reduced excitation and shallower slopes in the rate-azimuth function. The current study tested these possibilities by measuring an evoked response to ITD shifts in a narrow-band noise. Results were more in line with the latter outcome, both from measured latencies and amplitudes of the global field potentials and source-localized waveforms in the left and right auditory cortices. The measured responses for older listeners also tended to have reduced asymmetric distribution of activity in response to ITD shifts, which is consistent with other sensory and cognitive processing models of aging.
Collapse
Affiliation(s)
- Erol J Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | - David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| | - Ann C Eddins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida
| |
Collapse
|
32
|
Eddins DA, Kopf LM, Shrivastav R. The psychophysics of roughness applied to dysphonic voice. J Acoust Soc Am 2015; 138:3820-5. [PMID: 26723336 PMCID: PMC4691258 DOI: 10.1121/1.4937753] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/01/2015] [Revised: 11/19/2015] [Accepted: 11/25/2015] [Indexed: 05/19/2023]
Abstract
Roughness is a sound quality that has been related to the amplitude modulation characteristics of the acoustic stimulus. Roughness also is considered one of the primary elements of voice quality associated with natural variations across normal voices and is a salient feature of many dysphonic voices. It is known that the roughness of tonal stimuli is dependent on the frequency and depth of amplitude modulation and on the carrier frequency. Here, it is determined if similar dependencies exist for voiced speech stimuli. Knowledge of such dependencies can lead to a better understanding of the acoustic characteristics of vocal roughness along the continuum of normal to dysphonic and may facilitate computational estimates of vocal roughness. Synthetic vowel stimuli were modeled after talkers selected from the Satloff/Heman-Ackah disordered voice database. To parametrically control amplitude modulation frequency and depth, synthesized stimuli had minimal amplitude fluctuations, and amplitude modulation was superimposed with the desired frequency and depth. Perceptual roughness judgments depended on amplitude modulation frequency and depth in a manner that closely matched data from tonal carriers. The dependence of perceived roughness on amplitude modulation frequency and depth closely matched the roughness of sinusoidal carriers as reported by Fastl and Zwicker [(2007) Psychoacoustics: Facts and Models, 3rd ed. (Springer, New York)].
Collapse
Affiliation(s)
- David A Eddins
- Department of Communication Sciences and Disorders, University of South Florida, 4202 East Fowler Avenue, PCD 1017, Tampa, Florida 33620, USA
| | - Lisa M Kopf
- Department of Communicative Sciences and Disorders, Michigan State University, 1026 Red Cedar Road, Room 109, Oyer Speech and Hearing Building, East Lansing Michigan 48824, USA
| | - Rahul Shrivastav
- Office of the Vice President for Instruction, University of Georgia, 0309 New College, Athens, Georgia 30602, USA
| |
Collapse
|
33
|
Kelly MR, Amuso VJ, Eddins DA, Borkholder DA. The focal index as a singular metric for beamforming effectiveness. J Acoust Soc Am 2014; 136:2654-2664. [PMID: 25373966 DOI: 10.1121/1.4896465] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
Hearing-impaired individuals frequently cite intelligibility problems in multi-talker environments. Microphone arrays performing time-delay beamforming address conditions of poor signal-to-noise ratio by spatially filtering incoming sound. Existing beam pattern metrics including peak side lobe level, integrated side lobe level, beamwidth, and planar directivity index fail to quantitatively capture all elements essential for improving speech intelligibility in multi-talker situations. The focal index (FI) was developed to address these deficiencies. Simulations were performed to exemplify the robust nature of the FI and to demonstrate the utility of this metric for driving array parameter selection. Beam patterns were generated and the metrics were calculated and evaluated against the strict unidirectional requirements for the array. Array performance was assessed by human subjects in a speech recognition task that incorporated competing speech from multiple locations. Simulations of array output were presented under conditions differing in array sparsity. The resulting human subject data were used to demonstrate the linear relationship (R(2) > 0.975) between speech-intelligibility-weighted FI (SII-FI) and the signal-to-noise ratio thresholds for 20% and 80% correct responses. Data indicate that the FI and SII-FI are robust singular metrics for determining the effectiveness of the array.
Collapse
Affiliation(s)
- Matthew R Kelly
- Electrical and Microelectronic Engineering, Rochester Institute of Technology, 79 Lomb Memorial Drive, Rochester, New York 14623
| | - Vincent J Amuso
- Electrical and Microelectronic Engineering, Rochester Institute of Technology, 79 Lomb Memorial Drive, Rochester, New York 14623
| | - David A Eddins
- Communication Sciences and Disorders, University of South Florida, 4202 East Fowler Avenue, PCD 1017, Tampa, Florida 33620
| | - David A Borkholder
- Microsystems Engineering, Rochester Institute of Technology, 168 Lomb Memorial Drive, Rochester, New York 14623
| |
Collapse
|
34
|
|
35
|
Eddins DA. Hearing Aids. Second Edition, Harvey Dillon. Ear Hear 2013. [DOI: 10.1097/01.aud.0000436254.15629.5b] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
36
|
Eddins DA, Shrivastav R. Psychometric properties associated with perceived vocal roughness using a matching task. J Acoust Soc Am 2013; 134:EL294-300. [PMID: 24116533 PMCID: PMC3779263 DOI: 10.1121/1.4819183] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/02/2013] [Accepted: 08/09/2013] [Indexed: 05/19/2023]
Abstract
A psychophysical matching paradigm has been used to better quantify voice quality under laboratory conditions. The goals of this study were to establish which of two candidate comparison stimuli would best ensure that the range of perceived vocal roughness could be adequately bracketed using a matching task and to provide a general solution to the problem of estimating vocal roughness. Psychometric functions for roughness matching indicated that a speech-like sawtooth-plus-noise complex (20 dB signal-to-noise ratio) amplitude modulated by a sinusoidal function raised to the 4th power yielded a comparison stimulus with a perceptual dynamic range well suited for roughness matching.
Collapse
Affiliation(s)
- David A Eddins
- Department of Communication Sciences and Disorders, Global Center for Hearing and Speech Research, University of South Florida, 4202 East Fowler Avenue, PCD 1017, Tampa, Florida 33620
| | | |
Collapse
|
37
|
Zecker SG, Hoffman HJ, Frisina R, Dubno JR, Dhar S, Wallhagen M, Kraus N, Griffith JW, Walton JP, Eddins DA, Newman C, Victorson D, Warrier CM, Wilson RH. Audition assessment using the NIH Toolbox. Neurology 2013; 80:S45-8. [PMID: 23479544 DOI: 10.1212/wnl.0b013e3182872dd2] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022] Open
Abstract
The NIH Toolbox project has assembled measurement tools to assess a wide range of human perception and ability across the lifespan. As part of this initiative, a small but comprehensive battery of auditory tests has been assembled. The main tool of this battery, pure-tone thresholds, measures the ability of people to hear at specific frequencies. Pure-tone thresholds have long been considered the "gold standard" of auditory testing, and are normally obtained in a clinical setting by highly trained audiologists. For the purposes of the Toolbox project, an automated procedure (NIH Toolbox Threshold Hearing Test) was developed that allows nonspecialists to administer the test reliably. Three supplemental auditory tests are also included in the Toolbox auditory test battery: assessment of middle-ear function (tympanometry), speech perception in noise (the NIH Toolbox Words-in-Noise Test), and self-assessment of hearing impairment (the NIH Toolbox Hearing Handicap Inventory Ages 18-64 and the NIH Toolbox Hearing Handicap Inventory Ages 64+). Tympanometry can help differentiate conductive from sensorineural pathology. The NIH Toolbox Words-in-Noise Test measures a listener's ability to perceive words in noisy situations. This ability is not necessarily predicted by a person's pure-tone thresholds; some people with normal hearing have difficulty extracting meaning from speech sounds heard in a noisy context. The NIH Toolbox Hearing Handicap Inventory focuses on how a person's perceived hearing status affects daily life. The test was constructed to include emotional and social/situational subscales, with specific questions about how hearing impairment may affect one's emotional state or limit participation in specific activities. The 4 auditory tests included in the Toolbox auditory test battery cover a range of auditory abilities and provide a snapshot of a participant's auditory capacity.
Collapse
|
38
|
Sabin AT, Clark CA, Eddins DA, Wright BA. Different patterns of perceptual learning on spectral modulation detection between older hearing-impaired and younger normal-hearing adults. J Assoc Res Otolaryngol 2012; 14:283-94. [PMID: 23229719 DOI: 10.1007/s10162-012-0363-y] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2012] [Accepted: 11/14/2012] [Indexed: 10/27/2022] Open
Abstract
Young adults with normal hearing (YNH) can improve their sensitivity to basic acoustic features with practice. However, it is not known to what extent the influence of the same training regimen differs between YNH listeners and older listeners with hearing impairment (OHI)--the largest population seeking treatment in audiology clinics. To examine this issue, we trained OHI listeners on a basic auditory task (spectral modulation detection) using a training regimen previously administered to YNH listeners (≈ 1 h/session for seven sessions on a single condition). For the trained conditions on which pretraining performance was not already at asymptote, the YNH listeners who received training learned more than matched controls who received none, but that learning did not generalize to any untrained spectral modulation frequency. In contrast, the OHI-trained listeners and controls learned similar amounts on the trained condition, implying no effect of the training itself. However, surprisingly the OHI-trained listeners improved over the training phase and on an untrained spectral modulation frequency. These population differences suggest that learning consolidated more slowly, and that training modified an aspect of processing that had broader tuning to spectral modulation frequency, in OHI than YNH listeners. More generally, these results demonstrate that conclusions about perceptual learning that come from examination of one population do not necessarily apply to another.
Collapse
Affiliation(s)
- Andrew T Sabin
- Department of Communication Sciences and Disorders, Northwestern University, Evanston, IL 60208, USA.
| | | | | | | |
Collapse
|
39
|
Abstract
PURPOSE Perceptual estimates of voice quality obtained using rating scales are subject to contextual biases that influence how individuals assign numbers to estimate the magnitude of vocal quality. Because rating scales are commonly used in clinical settings, assessments of voice quality are also subject to the limitations of these scales. Instead, a matching task can be used to obtain objective measures of voice quality, thereby facilitating model development and tools for clinical use. METHOD Twenty-seven individuals participated in a rating task or at least 1 of 3 matching tests (named after their modulation functions: SINE, SQUARE, POWER) to quantify the degree of roughness in dysphonic voice stimuli. Participants evaluated the roughness of 34 voice samples using an amplitude-modulated complex carrier. RESULTS The matching thresholds were highly correlated with the ratings estimates. Reliability of thresholds did not significantly differ across tasks, but linear regressions showed that the POWER test resulted in larger perceptual distances. CONCLUSIONS A matching task can be used to obtain reliable estimates of roughness in dysphonic voices. The POWER comparison is recommended because the variability in matching thresholds across the range of roughness was evenly distributed, and the perceptual distances between stimuli were maximized.
Collapse
Affiliation(s)
- Sona Patel
- University of Florida, Gainesville, FL, USA.
| | | | | |
Collapse
|
40
|
Abstract
The purpose of this study was to establish the potential utility of adaptive tracking procedures, relative to the method of constant stimuli, when examining consonant identification in cafeteria noise or four-talker babble. Thresholds for consonant identification were comparable for the two methods, while psychometric function slopes derived from the two methods were statistically different. To obtain accurate slope estimates from adaptive procedures, the target percent correct and change in variability with signal-to-noise ratio should be considered carefully. However, adaptive procedures are accurate and more efficient than the method of constant stimuli when the purpose is to estimate phoneme identification threshold.
Collapse
Affiliation(s)
- Chang Liu
- Department of Communication Sciences and Disorders, 1 University Station A1100, The University of Texas at Austin Austin, Texas 78712, USA.
| | | |
Collapse
|
41
|
Abstract
PURPOSE In this experiment, a single comparison stimulus was developed as a reference in a perceptual matching task for the quantification of breathy voice quality. Perceptual judgments of a set of synthetic voice samples were compared to previous data obtained using multiple comparison stimuli "customized" for different voices (Patel, Shrivastav, & Eddins, 2010). METHOD Five male and 5 female samples of the vowel /a/ were selected from the Kay Elemetrics Disordered Voice Database and resynthesized using a Klatt synthesizer. Eleven samples were created for each base voice by manipulating the aspiration noise level. Five samples from each continuum were evaluated in a perceptual matching task in which a single sawtooth and noise comparison stimulus was used to obtain breathiness judgments. Linear regression was used to compare measurements obtained using the new comparison stimulus against the customized comparison stimuli. RESULTS Results indicated that the noncustomized sawtooth comparison provides reliability and perceptual distances between stimuli similar to those obtained using customized comparison stimuli. CONCLUSION A single-variable matching task using a single comparison stimulus can be used to obtain perceptual estimates of breathiness across voices and experiments in a laboratory setting. This technique will help develop models of voice-quality perception.
Collapse
Affiliation(s)
- Sona Patel
- University of Florida, Gainesville, FL, USA.
| | | | | |
Collapse
|
42
|
Abstract
Two sounds with the same pitch may vary from each other based on saliency of their pitch sensation. This perceptual attribute is called "pitch strength." The study of voice pitch strength may be important in quantifying of normal and pathological qualities. The present study investigated how pitch strength varies across normal and dysphonic voices. A set of voices (vowel /a/) selected from the Kay Elemetrics Disordered Voice Database served as the stimuli. These stimuli demonstrated a wide range of voice quality. Ten listeners judged the pitch strength of these stimuli in an anchored magnitude estimation task. On a given trial, listeners heard three different stimuli. The first stimulus represented very low pitch strength (wide-band noise), the second stimulus consisted of the target voice and the third stimulus represented very high pitch strength (pure tone). Listeners estimated pitch strength of the target voice by positioning a continuous slider labeled with values between 0 and 1, reflecting the two anchor stimuli. Results revealed that listeners can judge pitch strength reliably in dysphonic voices. Moderate to high correlations with perceptual judgments of voice quality suggest that pitch strength may contribute to voice quality judgments.
Collapse
Affiliation(s)
- Rahul Shrivastav
- Malcom Randall VAMC and University of Florida, Dauer Hall, P.O. Box 117420, Gainesville, Florida 32611, USA.
| | | | | |
Collapse
|
43
|
Eddins DA, Liu C. Psychometric properties of the coordinate response measure corpus with various types of background interference. J Acoust Soc Am 2012; 131:EL177-EL183. [PMID: 22352619 PMCID: PMC3277602 DOI: 10.1121/1.3678680] [Citation(s) in RCA: 23] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/16/2011] [Accepted: 12/29/2011] [Indexed: 05/29/2023]
Abstract
The coordinate response measure (CRM) corpus has gained broad acceptance as a research tool for investigating speech intelligibility in background competition and has been widely used in studies of informational masking. The purpose of this study is to establish the psychometric characteristics of CRM target-word identification in various backgrounds with the goal of being able to determine when it is appropriate or not to use adaptive threshold procedures with the CRM corpus. Target-word identification performance based on adaptive tracking mapped directly onto the monotonic psychometric functions obtained for two-talker, four-talker, and cafeteria noise interferers.
Collapse
Affiliation(s)
- David A Eddins
- Department of Communication Sciences & Disorders, Global Center for Hearing & Speech Research, University of South Florida, 4202 East Fowler Avenue, PCD 1017, Tampa, Florida 33620, USA.
| | | |
Collapse
|
44
|
Shrivastav R, Camacho A, Patel S, Eddins DA. A model for the prediction of breathiness in vowels. J Acoust Soc Am 2011; 129:1605-15. [PMID: 21428523 PMCID: PMC3077964 DOI: 10.1121/1.3543993] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/02/2010] [Revised: 10/29/2010] [Accepted: 12/29/2010] [Indexed: 05/19/2023]
Abstract
The perception of breathiness in vowels is cued by multiple acoustic cues, including changes in aspiration noise (AH) and the open quotient (OQ) [Klatt and Klatt, J. Acoust. Soc. Am. 87(2), 820-857 (1990)]. A loudness model can be used to determine the extent to which AH masks the harmonic components in voice. The resulting "partial loudness" (PL) and loudness of AH ["noise loudness" (NL)] have been shown to be good predictors of perceived breathiness [Shrivastav and Sapienza, J. Acoust. Soc. Am. 114(1), 2217-2224 (2003)]. The levels of AH and OQ were systematically manipulated for ten synthetic vowels. Perceptual judgments of breathiness were obtained and regression functions to predict breathiness from the ratio of NL to PL (η) were derived. Results show that breathiness can be modeled as a power function of η. The power parameter of this function appears to be affected by the fundamental frequency of the vowel. A second experiment was conducted to determine if the resulting power function could estimate breathiness in a different set of voices. The breathiness of these stimuli, both natural and synthetic, was determined in a listening test. The model estimates of breathiness were highly correlated with perceptual data but the absolute predicted values showed some discrepancies.
Collapse
Affiliation(s)
- Rahul Shrivastav
- Department of Communication Sciences and Disorders, University of Florida and Malcom Randall VAMC, Dauer Hall, P.O. Box 117420, Gainesville, Florida 32611, USA.
| | | | | | | |
Collapse
|
45
|
Allen PD, Eddins DA. Presbycusis phenotypes form a heterogeneous continuum when ordered by degree and configuration of hearing loss. Hear Res 2010; 264:10-20. [PMID: 20144701 DOI: 10.1016/j.heares.2010.02.001] [Citation(s) in RCA: 33] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/22/2009] [Revised: 02/01/2010] [Accepted: 02/01/2010] [Indexed: 10/19/2022]
Abstract
Many reports have documented age-by-frequency increases in average auditory thresholds in various human populations. Despite this, the prevalence of different patterns of hearing loss in presbycusis remains uncertain. We examined 'presbycusis phenotypes' in a database of 960 subjects (552 female, 408 male, 18-92 years) that each had 30 measures of peripheral hearing sensitivity: pure tone audiograms for left and right ears from 0.25 to 8 kHz and DPOAE for each ear with F(mean)=1-6.4 kHz. Surprisingly, the hearing phenotypes did not naturally separate into discrete classes of presbycusis. Principal component (PC) analysis revealed that two principal components account for 74% of the variance among the 30 measures of hearing. The two components represent the overall degree (PC1) and configuration of loss (Flat vs. Sloping; PC2) and the phenotypes form a continuum when plotted against them. A heuristic partitioning of this continuum produced classes of presbycusis that vary in their degree of Sloping or Flat hearing loss, suggesting that the previously reported sub-types of presbycusis arise from the categorical segregation of a continuous and heterogeneous distribution. Further, most phenotypes lie intermediate to the extremes of either Flat or Sloping loss, indicating that if audiometric configuration does predict presbycusis etiology, then a mixed origin is the most prevalent.
Collapse
Affiliation(s)
- Paul D Allen
- Department of Neurobiology and Anatomy, University of Rochester School of Medicine and Dentistry, 601 Elmwood Avenue, Box 603, Rochester, NY 14642, USA.
| | | |
Collapse
|
46
|
|
47
|
Saoji AA, Litvak L, Spahr AJ, Eddins DA. Spectral modulation detection and vowel and consonant identifications in cochlear implant listeners. J Acoust Soc Am 2009; 126:955-8. [PMID: 19739707 DOI: 10.1121/1.3179670] [Citation(s) in RCA: 53] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/08/2023]
Abstract
Speech understanding by cochlear implant listeners may be limited by their ability to perceive complex spectral envelopes. Here, spectral envelope perception was characterized by spectral modulation transfer functions in which modulation detection thresholds became poorer with increasing spectral modulation frequency (SMF). Thresholds at low SMFs, less likely to be influenced by spectral resolution, were correlated with vowel and consonant identifications [Litvak, L. M. et al. (2008). J. Acoust. Soc. Am. 122, 982-991] for the same listeners; while thresholds at higher SMFs, more likely to be affected by spectral resolution, were not. Results indicate that the perception of broadly spaced spectral features is important for speech perception.
Collapse
Affiliation(s)
- Aniket A Saoji
- Auditory Research and Development, Advanced Bionics Corporation, 12740 San Fernando Road, Sylmar, CA 91342, USA
| | | | | | | |
Collapse
|
48
|
Patel S, Shrivastav R, Eddins DA. Perceptual distances of breathy voice quality: a comparison of psychophysical methods. J Voice 2009; 24:168-77. [PMID: 19185451 DOI: 10.1016/j.jvoice.2008.08.002] [Citation(s) in RCA: 28] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2008] [Accepted: 08/04/2008] [Indexed: 10/21/2022]
Abstract
Experiments to study voice quality have typically used rating scales or direct magnitude estimation to obtain listener judgments. Unfortunately, the data obtained using these tasks are context dependent, which makes it difficult to compare perceptual judgments of voice quality across experiments. The present experiment describes a simple matching task to quantify voice quality. The data obtained through this task were compared to perceptual judgments obtained using rating scale and direct magnitude estimation tasks to determine whether the three tasks provide equivalent perceptual distances across stimuli. Ten synthetic vowel continua that varied in terms of their aspiration noise were evaluated for breathiness using each of the three tasks. Linear and nonlinear regressions were used to compare the perceptual distances between stimuli obtained through each technique. Results show that the perceptual distances estimated from matching and direct magnitude estimation task are similar, but both differ from the rating scale task, suggesting that the matching task provides perceptual distances with ratio-level measurement properties. The matching task is advantageous for measurement of vocal quality because it provides reliable measurement with ratio-level scale properties. It allows the use of a fixed reference signal for all comparisons, thus allowing researchers to directly compare findings across different experiments.
Collapse
Affiliation(s)
- Sona Patel
- Department of Communication Sciences and Disorders, University of Florida, Gainesville, Florida 32611, USA
| | | | | |
Collapse
|
49
|
Abstract
The goal of this study was to measure the effects of global spectral manipulations on vowel identification by progressively high-pass filtering vowel stimuli in the spectral modulation domain. Twelve American-English vowels, naturally spoken by a female talker, were subjected to varied degrees of high-pass filtering in the spectral modulation domain, with cutoff frequencies of 0.0, 0.5, 1.0, 1.5, and 2.0 cycles/octave. Identification performance for vowels presented at 70 dB sound pressure level with and without spectral modulation filtering was measured for five normal-hearing listeners. Results indicated that vowel identification performance was progressively degraded as the spectral modulation cutoff frequency increased. Degradation of vowel identification was greater for back vowels than for front or central vowels. Detailed acoustic analyses indicated that spectral modulation filtering resulted in a more crowded vowel space (F1xF2), reduced spectral contrast, and reduced spectral tilt relative to the original unfiltered vowels. Changes in the global spectral features produced by spectral modulation filtering were associated with substantial reduction in vowel identification. The results indicated that the spectral cues critical for vowel identification were represented by spectral modulation frequencies below 2 cycles/octave. These results are considered in terms of the interactions among spectral shape perception, spectral smearing, and speech perception.
Collapse
Affiliation(s)
- Chang Liu
- Department of Communication Sciences and Disorders, University of Texas at Austin, 1 University Station A1100, Austin, Texas 78712, USA.
| | | |
Collapse
|
50
|
Abstract
The goal of this study was to measure detection thresholds for 12 isolated American English vowels naturally spoken by three male and three female talkers for young normal-hearing listeners in the presence of a long-term speech-shaped (LTSS) noise, which was presented at 70 dB sound pressure level. The vowel duration was equalized to 170 ms and the spectrum of the LTSS noise was identical to the long-term average spectrum of 12-talker babble. Given the same duration, detection thresholds for vowels differed by 19 dB across the 72 vowels. Thresholds for vowel detection showed a roughly U-shaped pattern as a function of the vowel category across talkers with lowest thresholds at /i/ and /ae/ vowels and highest thresholds at /u/ vowel in general. Both vowel category and talker had a significant effect on vowel detectability. Detection thresholds predicted from three excitation pattern metrics by using a simulation model were well matched with thresholds obtained from human listeners, suggesting that listeners could use a constant metric in the excitation pattern of the vowel to detect the signal in noise independent of the vowel category and talker. Application of the simulation model to predict thresholds of vowel detection in noise was also discussed.
Collapse
Affiliation(s)
- Chang Liu
- Department of Communication Sciences and Disorders, University of Texas at Austin, 1 University Station A1100, Austin, Texas 78712, USA.
| | | |
Collapse
|