1
|
Rachel M, Jia H, Amina A, Perez-Garcia M, Kumar M, Wicherts JM. Psychometric evaluation of the computerized battery for neuropsychological evaluation of children (BENCI) among school aged children in the context of HIV in an urban Kenyan setting. BMC Psychiatry 2023; 23:373. [PMID: 37248481 DOI: 10.1186/s12888-023-04880-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/13/2023] [Accepted: 05/17/2023] [Indexed: 05/31/2023] Open
Abstract
INTRODUCTION Culturally validated neurocognitive measures for children in Low- and Middle-Income Countries are important in the timely and correct identification of neurocognitive impairments. Such measures can inform development of interventions for children exposed to additional vulnerabilities like HIV infection. The Battery for Neuropsychological Evaluation of Children (BENCI) is an openly available, computerized neuropsychological battery specifically developed to evaluate neurocognitive impairment. This study adapted the BENCI and evaluated its reliability and validity in Kenya. METHODOLOGY The BENCI was adapted using translation and back-translation from Spanish to English. The psychometric properties were evaluated in a case-control study of 328 children (aged 6 - 14 years) living with HIV and 260 children not living with HIV in Kenya. We assessed reliability, factor structure, and measurement invariance with respect to HIV. Additionally, we examined convergent validity of the BENCI using tests from the Kilifi Toolkit. RESULTS Internal consistencies (0.49 < α < 0.97) and test-retest reliabilities (-.34 to .81) were sufficient-to-good for most of the subtests. Convergent validity was supported by significant correlations between the BENCI's Verbal memory and Kilifi's Verbal List Learning (r = .41), the BENCI's Visual memory and Kilifi's Verbal List Learning (r = .32) and the BENCI's Planning total time test and Kilifi's Tower Test (r = -.21) and the BENCI's Abstract Reasoning test and Kilifi's Raven's Progressive Matrix (r = .21). The BENCI subtests highlighted meaningful differences between children living with HIV and those not living with HIV. After some minor adaptions, a confirmatory four-factor model consisting of flexibility, fluency, reasoning and working memory fitted well (χ2 = 135.57, DF = 51, N = 604, p < .001, RMSEA = .052, CFI = .944, TLI = .914) and was partially scalar invariant between HIV positive and negative groups. CONCLUSION The English version of the BENCI formally translated for use in Kenya can be further adapted and integrated in clinical and research settings as a valid and reliable cognitive test battery.
Collapse
Affiliation(s)
- Maina Rachel
- Department of Methodology and Statistics, Tilburg University, Tilburg, Netherlands.
- Brain and Mind Institute, Aga Khan University, Nairobi, 10834-00400, Kenya.
| | - He Jia
- Department of Methodology and Statistics, Tilburg University, Tilburg, Netherlands
| | - Abubakar Amina
- Institute for Human Development, Aga Khan University, Nairobi, Kenya
| | - Miguel Perez-Garcia
- Mind, Brain and Behavior Research Center (CIMCYC), University of Granada, Granada, Spain
| | - Manasi Kumar
- Brain and Mind Institute, Aga Khan University, Nairobi, 10834-00400, Kenya
| | - Jelte M Wicherts
- Department of Methodology and Statistics, Tilburg University, Tilburg, Netherlands
| |
Collapse
|
2
|
Thomas ML, Brown GG, Patt VM, Duffy JR. Latent Variable Modeling and Adaptive Testing for Experimental Cognitive Psychopathology Research. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2021; 81:155-181. [PMID: 33456066 PMCID: PMC7797961 DOI: 10.1177/0013164420919898] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The adaptation of experimental cognitive tasks into measures that can be used to quantify neurocognitive outcomes in translational studies and clinical trials has become a key component of the strategy to address psychiatric and neurological disorders. Unfortunately, while most experimental cognitive tests have strong theoretical bases, they can have poor psychometric properties, leaving them vulnerable to measurement challenges that undermine their use in applied settings. Item response theory-based computerized adaptive testing has been proposed as a solution but has been limited in experimental and translational research due to its large sample requirements. We present a generalized latent variable model that, when combined with strong parametric assumptions based on mathematical cognitive models, permits the use of adaptive testing without large samples or the need to precalibrate item parameters. The approach is demonstrated using data from a common measure of working memory-the N-back task-collected across a diverse sample of participants. After evaluating dimensionality and model fit, we conducted a simulation study to compare adaptive versus nonadaptive testing. Computerized adaptive testing either made the task 36% more efficient or score estimates 23% more precise, when compared to nonadaptive testing. This proof-of-concept study demonstrates that latent variable modeling and adaptive testing can be used in experimental cognitive testing even with relatively small samples. Adaptive testing has the potential to improve the impact and replicability of findings from translational studies and clinical trials that use experimental cognitive tasks as outcome measures.
Collapse
|
3
|
Thomas ML. Advances in applications of item response theory to clinical assessment. Psychol Assess 2019; 31:1442-1455. [PMID: 30869966 DOI: 10.1037/pas0000597] [Citation(s) in RCA: 25] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/11/2022]
Abstract
Item response theory (IRT) is moving to the forefront of methodologies used to develop, evaluate, and score clinical measures. Funding agencies and test developers are routinely supporting IRT work, and the theory has become closely tied to technological advances within the field. As a result, familiarity with IRT has grown increasingly relevant to mental health research and practice. But to what end? This article reviews advances in applications of IRT to clinical measurement in an effort to identify tangible improvements that can be attributed to the methodology. Although IRT shares similarities with classical test theory and factor analysis, the approach has certain practical benefits, but also limitations, when applied to measurement challenges. Major opportunities include the use of computerized adaptive tests to prevent conditional measurement error, multidimensional models to prevent misinterpretation of scores, and analyses of differential item functioning to prevent bias. Whereas these methods and technologies were once only discussed as future possibilities, they are now accessible because of recent support of IRT-focused clinical research. Despite this, much work still remains in widely disseminating methods and technologies from IRT into mental health research and practice. Clinicians have been reluctant to fully embrace the approach, especially in terms or prospective test development and adaptive item administration. Widespread use of IRT technologies will require continued cooperation among psychometricians, clinicians, and other stakeholders. There are also many opportunities to expand the methodology, especially with respect to integrating modern measurement theory with models from personality and cognitive psychology as well as neuroscience. (PsycINFO Database Record (c) 2019 APA, all rights reserved).
Collapse
|
4
|
Thomas ML, Brown GG, Gur RC, Moore TM, Patt VM, Risbrough VB, Baker DG. A signal detection-item response theory model for evaluating neuropsychological measures. J Clin Exp Neuropsychol 2018; 40:745-760. [PMID: 29402152 PMCID: PMC6050112 DOI: 10.1080/13803395.2018.1427699] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
Abstract
INTRODUCTION Models from signal detection theory are commonly used to score neuropsychological test data, especially tests of recognition memory. Here we show that certain item response theory models can be formulated as signal detection theory models, thus linking two complementary but distinct methodologies. We then use the approach to evaluate the validity (construct representation) of commonly used research measures, demonstrate the impact of conditional error on neuropsychological outcomes, and evaluate measurement bias. METHOD Signal detection-item response theory (SD-IRT) models were fitted to recognition memory data for words, faces, and objects. The sample consisted of U.S. Infantry Marines and Navy Corpsmen participating in the Marine Resiliency Study. Data comprised item responses to the Penn Face Memory Test (PFMT; N = 1,338), Penn Word Memory Test (PWMT; N = 1,331), and Visual Object Learning Test (VOLT; N = 1,249), and self-report of past head injury with loss of consciousness. RESULTS SD-IRT models adequately fitted recognition memory item data across all modalities. Error varied systematically with ability estimates, and distributions of residuals from the regression of memory discrimination onto self-report of past head injury were positively skewed towards regions of larger measurement error. Analyses of differential item functioning revealed little evidence of systematic bias by level of education. CONCLUSIONS SD-IRT models benefit from the measurement rigor of item response theory-which permits the modeling of item difficulty and examinee ability-and from signal detection theory-which provides an interpretive framework encompassing the experimentally validated constructs of memory discrimination and response bias. We used this approach to validate the construct representation of commonly used research measures and to demonstrate how nonoptimized item parameters can lead to erroneous conclusions when interpreting neuropsychological test data. Future work might include the development of computerized adaptive tests and integration with mixture and random-effects models.
Collapse
Affiliation(s)
- Michael L. Thomas
- Department of Psychiatry, University of California San Diego, La Jolla, CA
- VA Center of Excellence for Stress and Mental Health (CESAMH), San Diego, CA
| | - Gregory G. Brown
- Department of Psychiatry, University of California San Diego, La Jolla, CA
- VISN-22 Mental Illness, Research, Education and Clinical Center (MIRECC), VA San Diego Healthcare System, San Diego, CA
| | - Ruben C. Gur
- Department of Psychiatry, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA
| | - Tyler M. Moore
- Department of Psychiatry, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA
| | - Virginie M. Patt
- Department of Psychiatry, University of California San Diego, La Jolla, CA
- Joint Doctoral Program in Clinical Psychology, San Diego State University/University of California, San Diego, CA
| | - Victoria B. Risbrough
- Department of Psychiatry, University of California San Diego, La Jolla, CA
- VA Center of Excellence for Stress and Mental Health (CESAMH), San Diego, CA
| | - Dewleen G. Baker
- Department of Psychiatry, University of California San Diego, La Jolla, CA
- VA Center of Excellence for Stress and Mental Health (CESAMH), San Diego, CA
| |
Collapse
|
5
|
McDermott TJ, Kirlic N, Aupperle RL. Roadmap for optimizing the clinical utility of emotional stress paradigms in human neuroimaging research. Neurobiol Stress 2018; 8:134-146. [PMID: 29888309 PMCID: PMC5991342 DOI: 10.1016/j.ynstr.2018.05.001] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2017] [Revised: 04/30/2018] [Accepted: 05/03/2018] [Indexed: 01/24/2023] Open
Abstract
The emotional stress response is relevant to a number of psychiatric disorders, including posttraumatic stress disorder (PTSD) in particular. Research using neuroimaging methods such as functional magnetic resonance imaging (fMRI) to probe stress-related neural processing have provided some insights into psychiatric disorders. Treatment providers and individual patients would benefit from clinically useful fMRI paradigms that provide information about patients' current brain state and responses to stress in order to inform the treatment selection process. However, neuroimaging has not yet made a meaningful impact on real-world clinical practice. This lack of clinical utility may be related to a number of basic psychometric properties that are often overlooked during fMRI task development. The goals of the current review are to discuss important methodological considerations for current human fMRI stress-related paradigms and to provide a roadmap for developing methodologically sound and clinically useful paradigms. This would include establishing various aspects of reliability, including internal consistency, test-retest and multi-site, as well as validity, including face, content, construct, and criterion. In addition, the establishment of standardized normative data from a large sample of participants would support our understanding of how any one individual compares to the general population. Addressing these methodological gaps will likely have a powerful effect on improving the replicability of findings and optimize our chances for improving real-world clinical outcomes.
Collapse
Affiliation(s)
- Timothy J. McDermott
- Laureate Institute for Brain Research, Tulsa, OK, United States
- Department of Psychology, University of Tulsa, Tulsa, OK, United States
| | - Namik Kirlic
- Laureate Institute for Brain Research, Tulsa, OK, United States
| | - Robin L. Aupperle
- Laureate Institute for Brain Research, Tulsa, OK, United States
- Department of Community Medicine, University of Tulsa, Tulsa, OK, United States
| |
Collapse
|
6
|
Brown GG, Thomas ML, Patt V. Parametric model measurement: reframing traditional measurement ideas in neuropsychological practice and research. Clin Neuropsychol 2017; 31:1047-1072. [PMID: 28617067 DOI: 10.1080/13854046.2017.1334829] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
OBJECTIVE Neuropsychology is an applied measurement field with its psychometric work primarily built upon classical test theory (CTT). We describe a series of psychometric models to supplement the use of CTT in neuropsychological research and test development. METHOD We introduce increasingly complex psychometric models as measurement algebras, which include model parameters that represent abilities and item properties. Within this framework of parametric model measurement (PMM), neuropsychological assessment involves the estimation of model parameters with ability parameter values assuming the role of test 'scores'. Moreover, the traditional notion of measurement error is replaced by the notion of parameter estimation error, and the definition of reliability becomes linked to notions of item and test information. The more complex PMM approaches incorporate into the assessment of neuropsychological performance formal parametric models of behavior validated in the experimental psychology literature, along with item parameters. These PMM approaches endorse the use of experimental manipulations of model parameters to assess a test's construct representation. Strengths and weaknesses of these models are evaluated by their implications for measurement error conditional upon ability level, sensitivity to sample characteristics, computational challenges to parameter estimation, and construct validity. CONCLUSION A family of parametric psychometric models can be used to assess latent processes of interest to neuropsychologists. By modeling latent abilities at the item level, psychometric studies in neuropsychology can investigate construct validity and measurement precision within a single framework and contribute to a unification of statistical methods within the framework of generalized latent variable modeling.
Collapse
Affiliation(s)
- Gregory G Brown
- a Psychology Service (116B) , VA San Diego Healthcare System , San Diego , CA , USA
| | - Michael L Thomas
- b Department of Psychiatry , University of California , San Diego , CA , USA
| | - Virginie Patt
- c San Diego State University/University of California San Diego Joint Doctoral Program in Clinical Psychology , San Diego , CA , USA
| |
Collapse
|