1
|
Dubuy Y, Hardouin JB, Blanchin M, Sébille V. Identification of sources of DIF using covariates in patient-reported outcome measures: a simulation study comparing two approaches based on Rasch family models. Front Psychol 2023; 14:1191107. [PMID: 37637889 PMCID: PMC10448192 DOI: 10.3389/fpsyg.2023.1191107] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2023] [Accepted: 06/27/2023] [Indexed: 08/29/2023] Open
Abstract
When analyzing patient-reported outcome (PRO) data, sources of differential item functioning (DIF) can be multiple and there may be more than one covariate of interest. Hence, it could be of great interest to disentangle their effects. Yet, in the literature on PRO measures, there are many studies where DIF detection is applied separately and independently for each covariate under examination. With such an approach, the covariates under investigation are not introduced together in the analysis, preventing from simultaneously studying their potential DIF effects on the questionnaire items. One issue, among others, is that it may lead to the detection of false-positive effects when covariates are correlated. To overcome this issue, we developed two new algorithms (namely ROSALI-DIF FORWARD and ROSALI-DIF BACKWARD). Our aim was to obtain an iterative item-by-item DIF detection method based on Rasch family models that enable to adjust group comparisons for DIF in presence of two binary covariates. Both algorithms were evaluated through a simulation study under various conditions aiming to be representative of health research contexts. The performance of the algorithms was assessed using: (i) the rates of false and correct detection of DIF, (ii) the DIF size and form recovery, and (iii) the bias in the latent variable level estimation. We compared the performance of the ROSALI-DIF algorithms to the one of another approach based on likelihood penalization. For both algorithms, the rate of false detection of DIF was close to 5%. The DIF size and form influenced the rates of correct detection of DIF. Rates of correct detection was higher with increasing DIF size. Besides, the algorithm fairly identified homogeneous differences in the item threshold parameters, but had more difficulties identifying non-homogeneous differences. Over all, the ROSALI-DIF algorithms performed better than the penalized likelihood approach. Integrating several covariates during the DIF detection process may allow a better assessment and understanding of DIF. This study provides valuable insights regarding the performance of different approaches that could be undertaken to fulfill this aim.
Collapse
Affiliation(s)
- Yseulys Dubuy
- UMR INSERM 1246, MethodS in Patients-centered outcomes and HEalth ResEarch (SPHERE), Nantes Université, Nantes, France
| | - Jean-Benoit Hardouin
- UMR INSERM 1246, MethodS in Patients-centered outcomes and HEalth ResEarch (SPHERE), Nantes Université, Nantes, France
- Methodology and Biostatistics Unit, CHU Nantes, Nantes Université, Nantes, France
- Public Health Department, CHU Nantes, Nantes Université, Nantes, France
| | - Myriam Blanchin
- UMR INSERM 1246, MethodS in Patients-centered outcomes and HEalth ResEarch (SPHERE), Nantes Université, Nantes, France
| | - Véronique Sébille
- UMR INSERM 1246, MethodS in Patients-centered outcomes and HEalth ResEarch (SPHERE), Nantes Université, Nantes, France
- Methodology and Biostatistics Unit, CHU Nantes, Nantes Université, Nantes, France
| |
Collapse
|
2
|
Vogelsmeier LVDE, Vermunt JK, Bülow A, De Roover K. Evaluating Covariate Effects on ESM Measurement Model Changes with Latent Markov Factor Analysis: A Three-Step Approach. MULTIVARIATE BEHAVIORAL RESEARCH 2023; 58:262-291. [PMID: 34657547 DOI: 10.1080/00273171.2021.1967715] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
Invariance of the measurement model (MM) between subjects and within subjects over time is a prerequisite for drawing valid inferences when studying dynamics of psychological factors in intensive longitudinal data. To conveniently evaluate this invariance, latent Markov factor analysis (LMFA) was proposed. LMFA combines a latent Markov model with mixture factor analysis: The Markov model captures changes in MMs over time by clustering subjects' observations into a few states and state-specific factor analyses reveal what the MMs look like. However, to estimate the model, Vogelsmeier, Vermunt, van Roekel, and De Roover (2019) introduced a one-step (full information maximum likelihood; FIML) approach that is counterintuitive for applied researchers and entails cumbersome model selection procedures in the presence of many covariates. In this paper, we simplify the complex LMFA estimation and facilitate the exploration of covariate effects on state memberships by splitting the estimation in three intuitive steps: (1) obtain states with mixture factor analysis while treating repeated measures as independent, (2) assign observations to the states, and (3) use these states in a discrete- or continuous-time latent Markov model taking into account classification errors. A real data example demonstrates the empirical value.
Collapse
Affiliation(s)
| | | | - Anne Bülow
- Tilburg University
- Erasmus University Rotterdam
| | | |
Collapse
|
3
|
Xu H, Li X, Zhang Z, Grannis S. Score test for assessing the conditional dependence in latent class models and its application to record linkage. J R Stat Soc Ser C Appl Stat 2022. [DOI: 10.1111/rssc.12590] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Affiliation(s)
- Huiping Xu
- Department of Biostatistics and Health Data Science Indiana University Indianapolis Indiana USA
| | - Xiaochun Li
- Department of Biostatistics and Health Data Science Indiana University Indianapolis Indiana USA
| | | | | |
Collapse
|
4
|
Guastadisegni L, Cagnone S, Moustaki I, Vasdekis V. Use of the Lagrange Multiplier Test for Assessing Measurement Invariance Under Model Misspecification. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2022; 82:254-280. [PMID: 35185159 PMCID: PMC8850767 DOI: 10.1177/00131644211020355] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
This article studies the Type I error, false positive rates, and power of four versions of the Lagrange multiplier test to detect measurement noninvariance in item response theory (IRT) models for binary data under model misspecification. The tests considered are the Lagrange multiplier test computed with the Hessian and cross-product approach, the generalized Lagrange multiplier test and the generalized jackknife score test. The two model misspecifications are those of local dependence among items and nonnormal distribution of the latent variable. The power of the tests is computed in two ways, empirically through Monte Carlo simulation methods and asymptotically, using the asymptotic distribution of each test under the alternative hypothesis. The performance of these tests is evaluated by means of a simulation study. The results highlight that, under mild model misspecification, all tests have good performance while, under strong model misspecification, the tests performance deteriorates, especially for false positive rates under local dependence and power for small sample size under misspecification of the latent variable distribution. In general, the Lagrange multiplier test computed with the Hessian approach and the generalized Lagrange multiplier test have better performance in terms of false positive rates while the Lagrange multiplier test computed with the cross-product approach has the highest power for small sample sizes. The asymptotic power turns out to be a good alternative to the classic empirical power because it is less time consuming. The Lagrange tests studied here have been also applied to a real data set.
Collapse
Affiliation(s)
| | | | - Irini Moustaki
- London School of Economics and Political Science, London, UK
| | | |
Collapse
|
5
|
TAID-LCA: Segmentation Algorithm Based on Ternary Trees. MATHEMATICS 2022. [DOI: 10.3390/math10040560] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/01/2023]
Abstract
In this work, a statistical method for the segmentation of samples and/or populations is presented, which is based on a ternary tree structure. This approach overcomes known limitations of other segmentation methods such as CHAID, concerning the multivariate response and the non-symmetric relationship between explanatory and response variables. The multivariate response segmentation problem is handled through latent class models, while the factorial decomposition of the explanatory capability of variables is based on the Non-Symmetrical Correspondence Analysis. Stop criteria based on the CATANOVA index and impurity measures are proposed. A Simulated Annealing based post-pruning strategy is considered to avoid over-fitting relative to the training set and guarantee a better generalization capability for the method.
Collapse
|
6
|
Sinha P, Calfee CS, Delucchi KL. Practitioner's Guide to Latent Class Analysis: Methodological Considerations and Common Pitfalls. Crit Care Med 2021; 49:e63-e79. [PMID: 33165028 PMCID: PMC7746621 DOI: 10.1097/ccm.0000000000004710] [Citation(s) in RCA: 216] [Impact Index Per Article: 72.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/22/2022]
Abstract
Latent class analysis is a probabilistic modeling algorithm that allows clustering of data and statistical inference. There has been a recent upsurge in the application of latent class analysis in the fields of critical care, respiratory medicine, and beyond. In this review, we present a brief overview of the principles behind latent class analysis. Furthermore, in a stepwise manner, we outline the key processes necessary to perform latent class analysis including some of the challenges and pitfalls faced at each of these steps. The review provides a one-stop shop for investigators seeking to apply latent class analysis to their data.
Collapse
Affiliation(s)
- Pratik Sinha
- Department of Medicine, Division of Pulmonary, Critical Care, Allergy and Sleep Medicine; University of California, San Francisco; San Francisco, CA
- Department of Anesthesia; University of California, San Francisco; San Francisco, CA
| | - Carolyn S. Calfee
- Department of Medicine, Division of Pulmonary, Critical Care, Allergy and Sleep Medicine; University of California, San Francisco; San Francisco, CA
- Department of Anesthesia; University of California, San Francisco; San Francisco, CA
| | - Kevin L. Delucchi
- Department of Psychiatry; University of California, San Francisco; San Francisco, CA
| |
Collapse
|
7
|
Lee J, Jung K, Park J. Detecting Conditional Dependence Using Flexible Bayesian Latent Class Analysis. Front Psychol 2020; 11:1987. [PMID: 32903609 PMCID: PMC7438797 DOI: 10.3389/fpsyg.2020.01987] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2019] [Accepted: 07/17/2020] [Indexed: 11/13/2022] Open
Abstract
A fundamental assumption underlying latent class analysis (LCA) is that class indicators are conditionally independent of each other, given latent class membership. Bayesian LCA enables researchers to detect and accommodate violations of this assumption by estimating any number of correlations among indicators with proper prior distributions. However, little is known about how the choice of prior may affect the performance of Bayesian LCA. This article presents a Monte Carlo simulation study that investigates (1) the utility of priors in a range of prior variances (i.e., strongly non-informative to strongly informative priors) in terms of Type I error and power for detecting conditional dependence and (2) the influence of imposing approximate independence on model fit of Bayesian LCA. Simulation results favored the use of a weakly informative prior with large variance-model fit (posterior predictive p-value) was always satisfactory when the class indicators were either independent or dependent. Based on the current findings and the additional literature, this article offers methodological guidelines and suggestions for applied researchers.
Collapse
Affiliation(s)
- Jaehoon Lee
- Department of Educational Psychology and Leadershi, Texas Tech University, Lubbock, TX, United States
| | - Kwanghee Jung
- Department of Educational Psychology and Leadershi, Texas Tech University, Lubbock, TX, United States
| | - Jungkyu Park
- Department of Psychology, Kyungpook National University, Daegu, South Korea
| |
Collapse
|
8
|
Bailey AJ, Farmer EJ, Finn PR. Patterns of polysubstance use and simultaneous co-use in high risk young adults. Drug Alcohol Depend 2019; 205:107656. [PMID: 31706247 PMCID: PMC6901131 DOI: 10.1016/j.drugalcdep.2019.107656] [Citation(s) in RCA: 31] [Impact Index Per Article: 6.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/01/2019] [Revised: 09/12/2019] [Accepted: 09/13/2019] [Indexed: 11/27/2022]
Abstract
BACKGROUND Polysubstance use (PSU) is associated with worse prognosis and poorer physical and mental health compared to single substance use. The current study provides information about PSU patterns by examining a diverse range of alcohol/substance use behaviors ranging from low-level experimentation to indicators of severe abuse. In addition, the current study, for the first time, examines how simultaneous co-use of multiple substances cluster with other more commonly studied PSU behaviors. METHODS Latent Class Analysis was used to identify patterns of substance use, in a sample of young-adults (n = 2098), using 25 items from the Semi-Structured Assessment for the Genetics of Alcoholism (SSAGA-II) including: items assessing severity of problems with alcohol, cannabis, stimulants, opiates, and sedatives; items assessing # of lifetime uses; items assessing simultaneous co-use of all combinations of substances. Then the association of class membership and age, antisocial and impulsive personality, experience seeking, anxiety, and neuroticism was examined using Multinomial Regression. RESULTS Fit indices (i.e. AIC, SSABIC, and entropy) and interpretability of classes supported a five-class solution: "Low Problems" (32% of sample), "Alcohol Primary" (11%), "Alcohol and Cannabis" (25%), "Moderate PSU" (23%), and "Severe PSU" (9%). Simultaneous co-use behaviors discriminated between lower and higher severity groups. Externalizing personality constructs robustly predicted membership in the "Moderate" and "Severe" PSU classes compared to the "Alcohol Primary" class. CONCLUSIONS PSU patterns followed an additive pattern of use with lower severity classes using alcohol/cannabis and more severe classes using other illicit substances in addition. Co-use items provided valuable information about PSU severity.
Collapse
Affiliation(s)
- Allen J Bailey
- Department of Psychological and Brain Sciences, Indiana University-Bloomington, 1101 East 10th Street, Bloomington, IN 47405, USA.
| | - Eli J Farmer
- Department of Psychological and Brain Sciences, Indiana University-Bloomington, 1101 East 10th Street, Bloomington, IN 47405, USA.
| | - Peter R Finn
- Department of Psychological and Brain Sciences, Indiana University-Bloomington, 1101 East 10th Street, Bloomington, IN 47405, USA.
| |
Collapse
|
9
|
Abstract
AbstractDepression in later life is one of the most common mental disorders. Several instruments have been developed to detect the presence or the absence of certain symptoms or emotional disorders, based on cut-off points. However, the use of a cut-off does not allow identification of depression sub-types or distinguish between mild and severe depression. As a result, depression may be under- or over-diagnosed in older people. This paper aims to apply a model-driven approach to classify individuals into distinct sub-groups, based on different combinations of depressive and emotional conditions. This approach is based on two distinct statistical solutions: first, a latent class analysis is applied to the items collected by the depression scale and, according to the final model, the probability of belonging to each class is calculated for every individual. Second, a factor analysis of these classes is performed to obtain a reduced number of clusters for easy interpretation. We use data collected through the EURO-D scale in a large sample of older individuals, participants of the sixth wave of the Survey of Health, Ageing and Retirement in Europe. We show that by using such a model-based approach it is possible to classify individuals in a more accurate way than the simple dichotomisation ‘depressed’ versus ‘non-depressed’.
Collapse
|
10
|
Falk CF, Monroe S. On Lagrange Multiplier Tests in Multidimensional Item Response Theory: Information Matrices and Model Misspecification. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2018; 78:653-678. [PMID: 30147121 PMCID: PMC6096471 DOI: 10.1177/0013164417714506] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Lagrange multiplier (LM) or score tests have seen renewed interest for the purpose of diagnosing misspecification in item response theory (IRT) models. LM tests can also be used to test whether parameters differ from a fixed value. We argue that the utility of LM tests depends on both the method used to compute the test and the degree of misspecification in the initially fitted model. We demonstrate both of these points in the context of a multidimensional IRT framework. Through an extensive Monte Carlo simulation study, we examine the performance of LM tests under varying degrees of model misspecification, model size, and different information matrix approximations. A generalized LM test designed specifically for use under misspecification, which has apparently not been previously studied in an IRT framework, performed the best in our simulations. Finally, we reemphasize caution in using LM tests for model specification searches.
Collapse
Affiliation(s)
- Carl F. Falk
- Michigan State University, East Lansing,
MI, USA
| | | |
Collapse
|
11
|
Oberski DL, Kirchner A, Eckman S, Kreuter F. Evaluating the Quality of Survey and Administrative Data with Generalized Multitrait-Multimethod Models. J Am Stat Assoc 2018. [DOI: 10.1080/01621459.2017.1302338] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
Affiliation(s)
- D. L. Oberski
- Department of Methodology & Statistics, Utrecht University, Utrecht, The Netherlands
| | - A. Kirchner
- Survey Research Division, RTI International, NC
- University of Nebraska, Lincoln, NE
| | - S. Eckman
- Survey Research Division, RTI International, NC
| | - F. Kreuter
- Statistical Methods Group at the Institute for Employment Research, Nürnberg, Germany
- School of Social Science, University of Mannheim, Mannheim, Germany
- Joint Program in Survey Methodology, University of Maryland, College Park, MD
| |
Collapse
|
12
|
van den Bergh M, Schmittmann VD, Vermunt JK. Building Latent Class Trees, With an Application to a Study of Social Capital. METHODOLOGY-EUROPEAN JOURNAL OF RESEARCH METHODS FOR THE BEHAVIORAL AND SOCIAL SCIENCES 2017. [DOI: 10.1027/1614-2241/a000128] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022] Open
Abstract
Abstract. Researchers use latent class (LC) analysis to derive meaningful clusters from sets of categorical variables. However, especially when the number of classes required to obtain a good fit is large, interpretation of the latent classes may not be straightforward. To overcome this problem, we propose an alternative way of performing LC analysis, Latent Class Tree (LCT) modeling. For this purpose, a recursive partitioning procedure similar to divisive hierarchical cluster analysis is used: classes are split until a certain criterion indicates that the fit does not improve. The advantage of the LCT approach compared to the standard LC approach is that it gives a clear insight into how the latent classes are formed and how solutions with different numbers of classes relate. We also propose measures to evaluate the relative importance of the splits. The practical use of the approach is illustrated by the analysis of a data set on social capital.
Collapse
Affiliation(s)
| | | | - Jeroen K. Vermunt
- Department of Methodology and Statistics, Tilburg University, The Netherlands
| |
Collapse
|
13
|
Problems in detecting misfit of latent class models in diagnostic research without a gold standard were shown. J Clin Epidemiol 2015; 74:158-66. [PMID: 26628335 DOI: 10.1016/j.jclinepi.2015.11.012] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2015] [Revised: 11/04/2015] [Accepted: 11/20/2015] [Indexed: 11/21/2022]
Abstract
OBJECTIVES The objective of this study was to evaluate the performance of goodness-of-fit testing to detect relevant violations of the assumptions underlying the criticized "standard" two-class latent class model. Often used to obtain sensitivity and specificity estimates for diagnostic tests in the absence of a gold reference standard, this model relies on assuming that diagnostic test errors are independent. When this assumption is violated, accuracy estimates may be biased: goodness-of-fit testing is often used to evaluate the assumption and prevent bias. STUDY DESIGN AND SETTING We investigate the performance of goodness-of-fit testing by Monte Carlo simulation. The simulation scenarios are based on three empirical examples. RESULTS Goodness-of-fit tests lack power to detect relevant misfit of the standard two-class latent class model at sample sizes that are typically found in empirical diagnostic studies. The goodness-of-fit tests that are based on asymptotic theory are not robust to the sparseness of data. A parametric bootstrap procedure improves the evaluation of goodness of fit in the case of sparse data. CONCLUSION Our simulation study suggests that relevant violation of the local independence assumption underlying the standard two-class latent class model may remain undetected in empirical diagnostic studies, potentially leading to biased estimates of sensitivity and specificity.
Collapse
|
14
|
Beyond the number of classes: separating substantive from non-substantive dependence in latent class analysis. ADV DATA ANAL CLASSI 2015. [DOI: 10.1007/s11634-015-0211-0] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
15
|
van Kollenburg GH, Mulder J, Vermunt JK. Assessing Model Fit in Latent Class Analysis When Asymptotics Do Not Hold. METHODOLOGY-EUROPEAN JOURNAL OF RESEARCH METHODS FOR THE BEHAVIORAL AND SOCIAL SCIENCES 2015. [DOI: 10.1027/1614-2241/a000093] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022] Open
Abstract
The application of latent class (LC) analysis involves evaluating the LC model using goodness-of-fit statistics. To assess the misfit of a specified model, say with the Pearson chi-squared statistic, a p-value can be obtained using an asymptotic reference distribution. However, asymptotic p-values are not valid when the sample size is not large and/or the analyzed contingency table is sparse. Another problem is that for various other conceivable global and local fit measures, asymptotic distributions are not readily available. An alternative way to obtain the p-value for the statistic of interest is by constructing its empirical reference distribution using resampling techniques such as the parametric bootstrap or the posterior predictive check (PPC). In the current paper, we show how to apply the parametric bootstrap and two versions of the PPC to obtain empirical p-values for a number of commonly used global and local fit statistics within the context of LC analysis. The main difference between the PPC using test statistics and the parametric bootstrap is that the former takes into account parameter uncertainty. The PPC using discrepancies has the advantage that it is computationally much less intensive than the other two resampling methods. In a Monte Carlo study we evaluated Type I error rates and power of these resampling methods when used for global and local goodness-of-fit testing in LC analysis. Results show that both the bootstrap and the PPC using test statistics are generally good alternatives to asymptotic p-values and can also be used when (asymptotic) distributions are not known. Nominal Type I error rates were not met when sample size was small and the contingency table has many cells. Overall the PPC using test statistics was somewhat more conservative than the parametric bootstrap. We have also replicated previous research suggesting that the Pearson χ2 statistic should in many cases be preferred over the likelihood-ratio G2 statistic. Power to reject a model for which the number of LCs was one less than in the population was very high, unless sample size was small. When the contingency tables are very sparse, the total bivariate residual (TBVR) statistic, which is based on bivariate relationships, still had very high power, signifying its usefulness in assessing model fit.
Collapse
Affiliation(s)
| | - Joris Mulder
- Department of Methodology and Statistics, Tilburg University, The Netherlands
| | - Jeroen K. Vermunt
- Department of Methodology and Statistics, Tilburg University, The Netherlands
| |
Collapse
|