1
|
Mazumdar B, De la Mora N, Roberts T, Swiderski A, Kapantzoglou M, Fergadiotis G. Response Latencies During Confrontation Picture Naming in Aphasia: Are Proxy Measurements Sufficient? JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:1548-1557. [PMID: 38557214 PMCID: PMC11087083 DOI: 10.1044/2024_jslhr-23-00452] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/31/2023] [Revised: 12/13/2023] [Accepted: 01/31/2024] [Indexed: 04/04/2024]
Abstract
PURPOSE Anomia, or word-finding difficulty, is a prevalent and persistent feature of aphasia, a neurogenic language disorder affecting millions of people in the United States. Anomia assessments are essential for measuring performance and monitoring outcomes in clinical settings. This study aims to evaluate the reliability of response time (RT) annotation based on spectrograms and assess the predictive utility of proxy RTs collected during computerized naming tests. METHOD Archival data from 10 people with aphasia were used. Trained research assistants phonemically transcribed participants' responses, and RTs were generated from the onset of picture stimulus to the initial phoneme of the first complete attempt. RTs were measured in two ways: hand-generated RTs (from spectrograms) and proxy RTs (automatically extracted online). Interrater agreement was evaluated based on interclass correlation coefficients and generalizability theory tools including variance partitioning and the φ-coefficient. The predictive utility of proxy RTs was evaluated within a linear mixed-effects framework. RESULTS RT annotation reliability showed near-perfect agreement across research assistants (φ-coefficient = .93), and the variance accounted for by raters was negligible. Furthermore, proxy RTs significantly and strongly predicted hand-annotated RTs (R2 = ~0.82), suggesting their utility as an alternative measure. CONCLUSIONS The study confirms the reliability of RT annotation and demonstrates the predictive utility of proxy RTs in estimating RTs during computerized naming tests. Incorporating proxy RTs can enhance clinical assessments, providing additional information for cognitive measurement. Further research with larger samples and exploring the impact of using proxy RTs in different psychometric models could optimize clinical protocols and improve communication interventions for individuals with aphasia.
Collapse
Affiliation(s)
- Barnali Mazumdar
- Department of Communication Sciences & Disorders, Louisiana State University, Baton Rouge
| | - Nora De la Mora
- Department of Speech & Hearing Sciences, Portland State University, OR
| | - Teresa Roberts
- Department of Speech & Hearing Sciences, Portland State University, OR
| | - Alexander Swiderski
- Department of Communication Science and Disorders, University of Pittsburgh, PA
| | | | | |
Collapse
|
2
|
Fergadiotis G, Casilio M, Dickey MW, Steel S, Nicholson H, Fleegle M, Swiderski A, Hula WD. Item Response Theory Modeling of the Verb Naming Test. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2023; 66:1718-1739. [PMID: 37000934 PMCID: PMC10457085 DOI: 10.1044/2023_jslhr-22-00458] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/05/2022] [Revised: 12/19/2022] [Accepted: 01/23/2023] [Indexed: 05/11/2023]
Abstract
PURPOSE Item response theory (IRT) is a modern psychometric framework with several advantageous properties as compared with classical test theory. IRT has been successfully used to model performance on anomia tests in individuals with aphasia; however, all efforts to date have focused on noun production accuracy. The purpose of this study is to evaluate whether the Verb Naming Test (VNT), a prominent test of action naming, can be successfully modeled under IRT and evaluate its reliability. METHOD We used responses on the VNT from 107 individuals with chronic aphasia from AphasiaBank. Unidimensionality and local independence, two assumptions prerequisite to IRT modeling, were evaluated using factor analysis and Yen's Q 3 statistic (Yen, 1984), respectively. The assumption of equal discrimination among test items was evaluated statistically via nested model comparisons and practically by using correlations of resulting IRT-derived scores. Finally, internal consistency, marginal and empirical reliability, and conditional reliability were evaluated. RESULTS The VNT was found to be sufficiently unidimensional with the majority of item pairs demonstrating adequate local independence. An IRT model in which item discriminations are constrained to be equal demonstrated fit equivalent to a model in which unique discrimination parameters were estimated for each item. All forms of reliability were strong across the majority of IRT ability estimates. CONCLUSIONS Modeling the VNT using IRT is feasible, yielding ability estimates that are both informative and reliable. Future efforts are needed to quantify the validity of the VNT under IRT and determine the extent to which it measures the same construct as other anomia tests. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.22329235.
Collapse
Affiliation(s)
| | - Marianne Casilio
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN
| | - Michael Walsh Dickey
- Department of Communication Science and Disorders, University of Pittsburgh, PA
- VA Pittsburgh Healthcare System, PA
| | - Stacey Steel
- Department of Speech & Hearing Sciences, Portland State University, OR
| | - Hannele Nicholson
- U.S. Department of Veterans Affairs, VA Minneapolis Healthcare System, MN
| | - Mikala Fleegle
- Department of Speech & Hearing Sciences, Portland State University, OR
| | - Alexander Swiderski
- Department of Communication Science and Disorders, University of Pittsburgh, PA
- VA Pittsburgh Healthcare System, PA
| | - William D Hula
- Department of Communication Science and Disorders, University of Pittsburgh, PA
- VA Pittsburgh Healthcare System, PA
| |
Collapse
|
3
|
Casilio M, Fergadiotis G, Salem AC, Gale RC, McKinney-Bock K, Bedrick S. ParAlg: A Paraphasia Algorithm for Multinomial Classification of Picture Naming Errors. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2023; 66:966-986. [PMID: 36791263 PMCID: PMC10461785 DOI: 10.1044/2022_jslhr-22-00255] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/06/2022] [Revised: 10/05/2022] [Accepted: 11/21/2022] [Indexed: 06/18/2023]
Abstract
PURPOSE A preliminary version of a paraphasia classification algorithm (henceforth called ParAlg) has previously been shown to be a viable method for coding picture naming errors. The purpose of this study is to present an updated version of ParAlg, which uses multinomial classification, and comprehensively evaluate its performance when using two different forms of transcribed input. METHOD A subset of 11,999 archival responses produced on the Philadelphia Naming Test were classified into six cardinal paraphasia types using ParAlg under two transcription configurations: (a) using phonemic transcriptions for responses exclusively (phonemic-only) and (b) using phonemic transcriptions for nonlexical responses and orthographic transcriptions for lexical responses (orthographic-lexical). Agreement was quantified by comparing ParAlg-generated paraphasia codes between configurations and relative to human-annotated codes using four metrics (positive predictive value, sensitivity, specificity, and F1 score). An item-level qualitative analysis of misclassifications under the best performing configuration was also completed to identify the source and nature of coding discrepancies. RESULTS Agreement between ParAlg-generated and human-annotated codes was high, although the orthographic-lexical configuration outperformed phonemic-only (weighted-average F1 scores of .78 and .87, respectively). A qualitative analysis of the orthographic-lexical configuration revealed a mix of human- and ParAlg-related misclassifications, the former of which were related primarily to phonological similarity judgments whereas the latter were due to semantic similarity assignment. CONCLUSIONS ParAlg is an accurate and efficient alternative to manual scoring of paraphasias, particularly when lexical responses are orthographically transcribed. With further development, it has the potential to be a useful software application for anomia assessment. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.22087763.
Collapse
|
4
|
Nickels L, Lampe LF, Mason C, Hameau S. Investigating the influence of semantic factors on word retrieval: Reservations, results and recommendations. Cogn Neuropsychol 2022; 39:113-154. [PMID: 35972430 DOI: 10.1080/02643294.2022.2109958] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/31/2023]
Abstract
There is consensus that word retrieval starts with activation of semantic representations. However, in adults without language impairment, relatively little attention has been paid to the effects of the semantic attributes of to-be-retrieved words. This paper, therefore, addresses the question of which item-inherent semantic factors influence word retrieval. Specifically, it reviews the literature on a selection of these factors: imageability, concreteness, number of semantic features, typicality, intercorrelational density, featural distinctiveness, concept distinctiveness, animacy, semantic neighbourhood density, semantic similarity, operativity, valence, and arousal. It highlights several methodological challenges in this field, and has a focus on the insights from studies with people with aphasia where the effects of these variables are more prevalent. The paper concludes that further research simultaneously examining the effects of different semantic factors that are likely to affect lexical co-activation, and the interaction of these variables, would be fruitful, as would suitably scaled computational modelling of these effects in unimpaired language processing and in language impairment. Such research would enable the refinement of theories of semantic processing and word production, and potentially have implications for diagnosis and treatment of semantic and lexical impairments.
Collapse
Affiliation(s)
- Lyndsey Nickels
- School of Psychological Sciences, Macquarie University, Sydney, Australia.,Macquarie Centre for Reading, Macquarie University, Sydney, Australia
| | - Leonie F Lampe
- School of Psychological Sciences, Macquarie University, Sydney, Australia.,Department of Linguistics, University of Potsdam, Potsdam, Germany
| | - Catherine Mason
- School of Psychological Sciences, Macquarie University, Sydney, Australia.,Macquarie Centre for Reading, Macquarie University, Sydney, Australia
| | - Solène Hameau
- School of Psychological Sciences, Macquarie University, Sydney, Australia.,Macquarie Centre for Reading, Macquarie University, Sydney, Australia
| |
Collapse
|
5
|
Evans WS, Hula WD, Quique Y, Starns JJ. How Much Time Do People With Aphasia Need to Respond During Picture Naming? Estimating Optimal Response Time Cutoffs Using a Multinomial Ex-Gaussian Approach. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:599-614. [PMID: 32073336 DOI: 10.1044/2019_jslhr-19-00255] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Purpose Aphasia is a language disorder caused by acquired brain injury, which generally involves difficulty naming objects. Naming ability is assessed by measuring picture naming, and models of naming performance have mostly focused on accuracy and excluded valuable response time (RT) information. Previous approaches have therefore ignored the issue of processing efficiency, defined here in terms of optimal RT cutoff, that is, the shortest deadline at which individual people with aphasia produce their best possible naming accuracy performance. The goals of this study were therefore to (a) develop a novel model of aphasia picture naming that could accurately account for RT distributions across response types; (b) use this model to estimate the optimal RT cutoff for individual people with aphasia; and (c) explore the relationships between optimal RT cutoff, accuracy, naming ability, and aphasia severity. Method A total of 4,021 naming trials across 10 people with aphasia were scored for accuracy and RT onset. Data were fit using a novel ex-Gaussian multinomial RT model, which was then used to characterize individual optimal RT cutoffs. Results Overall, the model fitted the empirical data well and provided reliable individual estimates of optimal RT cutoff in picture naming. Optimal cutoffs ranged between approximately 5 and 10 s, which has important implications for assessment and treatment. There was no direct relationship between aphasia severity, naming RT, and optimal RT cutoff. Conclusion The multinomial ex-Gaussian modeling approach appears to be a promising and straightforward way to estimate optimal RT cutoffs in picture naming in aphasia. Limitations and future directions are discussed.
Collapse
Affiliation(s)
- William S Evans
- Geriatric Research Education and Clinical Center, VA Healthcare System, Pittsburgh, PA
- Department of Communication Sciences and Disorders, University of Pittsburgh, PA
| | - William D Hula
- Geriatric Research Education and Clinical Center, VA Healthcare System, Pittsburgh, PA
- Department of Communication Sciences and Disorders, University of Pittsburgh, PA
| | - Yina Quique
- Geriatric Research Education and Clinical Center, VA Healthcare System, Pittsburgh, PA
- Department of Communication Sciences and Disorders, University of Pittsburgh, PA
| | | |
Collapse
|
6
|
Hula WD, Fergadiotis G, Swiderski AM, Silkes JP, Kellough S. Empirical Evaluation of Computer-Adaptive Alternate Short Forms for the Assessment of Anomia Severity. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:163-172. [PMID: 31851861 PMCID: PMC7213484 DOI: 10.1044/2019_jslhr-l-19-0213] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/10/2019] [Revised: 08/12/2019] [Accepted: 08/22/2019] [Indexed: 05/09/2023]
Abstract
Purpose The purpose of this study was to verify the equivalence of 2 alternate test forms with nonoverlapping content generated by an item response theory (IRT)-based computer-adaptive test (CAT). The Philadelphia Naming Test (PNT; Roach, Schwartz, Martin, Grewal, & Brecher, 1996)was utilized as an item bank in a prospective, independent sample of persons with aphasia. Method Two alternate CAT short forms of the PNT were administered to a sample of 25 persons with aphasia who were at least 6 months postonset and received no treatment for 2 weeks before or during the study. The 1st session included administration of a 30-item PNT-CAT, and the 2nd session, conducted approximately 2 weeks later, included a variable-length PNT-CAT that excluded items administered in the 1st session and terminated when the modeled precision of the ability estimate was equal to or greater than the value obtained in the 1st session. The ability estimates were analyzed in a Bayesian framework. Results The 2 test versions correlated highly (r = .89) and obtained means and standard deviations that were not credibly different from one another. The correlation and error variance between the 2 test versions were well predicted by the IRT measurement model. Discussion The results suggest that IRT-based CAT alternate forms may be productively used in the assessment of anomia. IRT methods offer advantages for the efficient and sensitive measurement of change over time. Future work should consider the potential impact of differential item functioning due to person factors and intervention-specific effects, as well as expanding the item bank to maximize the clinical utility of the test. Supplemental Material https://doi.org/10.23641/asha.11368040.
Collapse
Affiliation(s)
- William D. Hula
- Geriatric Research Education and Clinical Center, VA Pittsburgh Healthcare System, PA
- Department of Communication Science and Disorders, University of Pittsburgh, PA
| | | | - Alexander M. Swiderski
- Department of Communication Science and Disorders, University of Pittsburgh, PA
- Research and Development Service, VA Pittsburgh Healthcare System, PA
| | - JoAnn P. Silkes
- Department of Speech and Hearing Sciences, University of Washington, Seattle
| | - Stacey Kellough
- Research and Development Service, VA Pittsburgh Healthcare System, PA
| |
Collapse
|
7
|
Fergadiotis G, Hula WD, Swiderski AM, Lei CM, Kellough S. Enhancing the Efficiency of Confrontation Naming Assessment for Aphasia Using Computer Adaptive Testing. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:1724-1738. [PMID: 31158037 PMCID: PMC6808378 DOI: 10.1044/2018_jslhr-l-18-0344] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/24/2018] [Revised: 11/21/2018] [Accepted: 12/16/2018] [Indexed: 05/09/2023]
Abstract
Purpose In this study, we investigated the agreement between the 175-item Philadelphia Naming Test (PNT; Roach, Schwartz, Martin, Grewal, & Brecher, 1996 ) and a 30-item computer adaptive PNT (PNT-CAT; Fergadiotis, Kellough, & Hula, 2015 ; Hula, Kellough, & Fergadiotis, 2015 ) created using item response theory (IRT) methods. Method The full PNT and the PNT-CAT were administered to 47 participants with aphasia in counterbalanced order. Latent trait-naming ability estimates for the 2 PNT versions were analyzed in a Bayesian framework, and the agreement between them was evaluated using correlation and measures of constant, variable, and total error. We also evaluated the extent to which individual pairwise differences were credibly greater than 0 and whether the IRT measurement model provided an adequate indication of the precision of individual score estimates. Results The agreement between the PNT and the PNT-CAT was strong, as indicated by high correlation ( r = .95, 95% CI [.92, .97]), negligible bias, and low variable and total error. The number of statistically robust pairwise score differences did not credibly exceed the Type I error rate, and the precision of individual score estimates was reasonably well predicted by the IRT model. Discussion The strong agreement between the full PNT and the PNT-CAT suggests that the latter is a suitable measurement of anomia in group studies. The relatively robust estimates of score precision also suggest that the PNT-CAT can be useful for the clinical assessment of anomia in individual cases. Finally, the IRT methods used to construct the PNT-CAT provide a framework for additional development to further reduce measurement error. Supplemental Material https://doi.org/10.23641/asha.8202176.
Collapse
Affiliation(s)
| | - William D. Hula
- Geriatric Research, Education and Clinical Center, VA Pittsburgh Healthcare System, PA
- Department of Communication Science & Disorders, University of Pittsburgh, PA
| | - Alexander M. Swiderski
- Department of Communication Science & Disorders, University of Pittsburgh, PA
- VA Pittsburgh Healthcare System, PA
| | - Chia-Ming Lei
- Department of Communication Science & Disorders, University of Pittsburgh, PA
- VA Pittsburgh Healthcare System, PA
| | | |
Collapse
|