1
|
Sijtsma K, Ellis JL, Borsboom D. Recognize the Value of the Sum Score, Psychometrics' Greatest Accomplishment. PSYCHOMETRIKA 2024; 89:84-117. [PMID: 38627311 DOI: 10.1007/s11336-024-09964-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/25/2024] [Indexed: 05/02/2024]
Abstract
The sum score on a psychological test is, and should continue to be, a tool central in psychometric practice. This position runs counter to several psychometricians' belief that the sum score represents a pre-scientific conception that must be abandoned from psychometrics in favor of latent variables. First, we reiterate that the sum score stochastically orders the latent variable in a wide variety of much-used item response models. In fact, item response theory provides a mathematically based justification for the ordinal use of the sum score. Second, because discussions about the sum score often involve its reliability and estimation methods as well, we show that, based on very general assumptions, classical test theory provides a family of lower bounds several of which are close to the true reliability under reasonable conditions. Finally, we argue that eventually sum scores derive their value from the degree to which they enable predicting practically relevant events and behaviors. None of our discussion is meant to discredit modern measurement models; they have their own merits unattainable for classical test theory, but the latter model provides impressive contributions to psychometrics based on very few assumptions that seem to have become obscured in the past few decades. Their generality and practical usefulness add to the accomplishments of more recent approaches.
Collapse
Affiliation(s)
- Klaas Sijtsma
- Department of Methodology and Statistics TSB, Tilburg University, PO Box 90153, 5000LE , Tilburg, The Netherlands.
| | - Jules L Ellis
- Open University OF THE NETHERLANDS, Heerlen, The Netherlands
| | | |
Collapse
|
2
|
Spiess M, Jordan P. In models we trust: preregistration, large samples, and replication may not suffice. Front Psychol 2023; 14:1266447. [PMID: 37809287 PMCID: PMC10551181 DOI: 10.3389/fpsyg.2023.1266447] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2023] [Accepted: 09/04/2023] [Indexed: 10/10/2023] Open
Abstract
Despite discussions about the replicability of findings in psychological research, two issues have been largely ignored: selection mechanisms and model assumptions. Both topics address the same fundamental question: Does the chosen statistical analysis tool adequately model the data generation process? In this article, we address both issues and show, in a first step, that in the face of selective samples and contrary to common practice, the validity of inferences, even when based on experimental designs, can be claimed without further justification and adaptation of standard methods only in very specific situations. We then broaden our perspective to discuss consequences of violated assumptions in linear models in the context of psychological research in general and in generalized linear mixed models as used in item response theory. These types of misspecification are oftentimes ignored in the psychological research literature. It is emphasized that the above problems cannot be overcome by strategies such as preregistration, large samples, replications, or a ban on testing null hypotheses. To avoid biased conclusions, we briefly discuss tools such as model diagnostics, statistical methods to compensate for selectivity and semi- or non-parametric estimation. At a more fundamental level, however, a twofold strategy seems indispensable: (1) iterative, cumulative theory development based on statistical methods with theoretically justified assumptions, and (2) empirical research on variables that affect (self-) selection into the observed part of the sample and the use of this information to compensate for selectivity.
Collapse
Affiliation(s)
- Martin Spiess
- Institute of Psychology, Department of Psychology and Human Movement Science, University of Hamburg, Hamburg, Germany
| | | |
Collapse
|
3
|
Ellis JL, Sijtsma K. A Test to Distinguish Monotone Homogeneity from Monotone Multifactor Models. PSYCHOMETRIKA 2023; 88:387-412. [PMID: 36933110 PMCID: PMC10188426 DOI: 10.1007/s11336-023-09905-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/19/2022] [Indexed: 05/17/2023]
Abstract
The goodness-of-fit of the unidimensional monotone latent variable model can be assessed using the empirical conditions of nonnegative correlations (Mokken in A theory and procedure of scale-analysis, Mouton, The Hague, 1971), manifest monotonicity (Junker in Ann Stat 21:1359-1378, 1993), multivariate total positivity of order 2 (Bartolucci and Forcina in Ann Stat 28:1206-1218, 2000), and nonnegative partial correlations (Ellis in Psychometrika 79:303-316, 2014). We show that multidimensional monotone factor models with independent factors also imply these empirical conditions; therefore, the conditions are insensitive to multidimensionality. Conditional association (Rosenbaum in Psychometrika 49(3):425-435, 1984) can detect multidimensionality, but tests of it (De Gooijer and Yuan in Comput Stat Data Anal 55:34-44, 2011) are usually not feasible for realistic numbers of items. The only existing feasible test procedures that can reveal multidimensionality are Rosenbaum's (Psychometrika 49(3):425-435, 1984) Case 2 and Case 5, which test the covariance of two items or two subtests conditionally on the unweighted sum of the other items. We improve this procedure by conditioning on a weighted sum of the other items. The weights are estimated in a training sample from a linear regression analysis. Simulations show that the Type I error rate is under control and that, for large samples, the power is higher if one dimension is more important than the other or if there is a third dimension. In small samples and with two equally important dimensions, using the unweighted sum yields greater power.
Collapse
Affiliation(s)
- Jules L Ellis
- Behavioural Science Institute, Radboud University Nijmegen, P.O.B. 9104, 6500 HE, Nijmegen, The Netherlands.
| | | |
Collapse
|
4
|
Finch WH, French BF, Hazelwood A. A Comparison of Confirmatory Factor Analysis and Network Models for Measurement Invariance Assessment When Indicator Residuals are Correlated. APPLIED PSYCHOLOGICAL MEASUREMENT 2023; 47:106-122. [PMID: 36875291 PMCID: PMC9979199 DOI: 10.1177/01466216231151700] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/18/2023]
Abstract
Social science research is heavily dependent on the use of standardized assessments of a variety of phenomena, such as mood, executive functioning, and cognitive ability. An important assumption when using these instruments is that they perform similarly for all members of the population. When this assumption is violated, the validity evidence of the scores is called into question. The standard approach for assessing the factorial invariance of the measures across subgroups within the population involves multiple groups confirmatory factor analysis (MGCFA). CFA models typically, but not always, assume that once the latent structure of the model is accounted for, the residual terms for the observed indicators are uncorrelated (local independence). Commonly, correlated residuals are introduced after a baseline model shows inadequate fit and inspection of modification indices ensues to remedy fit. An alternative procedure for fitting latent variable models that may be useful when local independence does not hold is based on network models. In particular, the residual network model (RNM) offers promise with respect to fitting latent variable models in the absence of local independence via an alternative search procedure. This simulation study compared the performances of MGCFA and RNM for measurement invariance assessment when local independence is violated, and residual covariances are themselves not invariant. Results revealed that RNM had better Type I error control and higher power compared to MGCFA when local independence was absent. Implications of the results for statistical practice are discussed.
Collapse
|
5
|
Ligtvoet R. Incomplete Tests of Conditional Association for the Assessment of Model Assumptions. PSYCHOMETRIKA 2022; 87:1214-1237. [PMID: 35124767 PMCID: PMC9636116 DOI: 10.1007/s11336-022-09841-1] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/22/2020] [Revised: 09/10/2021] [Indexed: 05/28/2023]
Abstract
Many of the models that have been proposed for response data share the assumptions that define the monotone homogeneity (MH) model. Observable properties that are implied by the MH model allow for these assumptions to be tested. For binary response data, the most restrictive of these properties is called conditional association (CA). All the other properties considered can be considered incomplete tests of CA that alleviate the practical limitations encountered when assessing the MH model assumptions using CA. It is found that the assessment of the MH model assumptions with an incomplete test of CA, rather than CA, is generally associated with a substantial loss of information. We also look at the sensitivity of the observable properties to model violation and discuss the implications of the results. It is argued that more research is required about the extent to which the assumptions and the model specifications influence the inferences made from response data.
Collapse
Affiliation(s)
- Rudy Ligtvoet
- Department Erziehungs- und Sozialwissenschaften, University of Cologne, Germany, Gronewaldstr. 2a, 50931, Cologne, Deutschland.
| |
Collapse
|
6
|
Tutz G. Item Response Thresholds Models: A General Class of Models for Varying Types of Items. PSYCHOMETRIKA 2022; 87:1238-1269. [PMID: 35476176 PMCID: PMC9636304 DOI: 10.1007/s11336-022-09865-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/24/2021] [Revised: 03/24/2022] [Indexed: 06/14/2023]
Abstract
A comprehensive class of models is proposed that can be used for continuous, binary, ordered categorical and count type responses. The difficulty of items is described by difficulty functions, which replace the item difficulty parameters that are typically used in item response models. They crucially determine the response distribution and make the models very flexible with regard to the range of distributions that are covered. The model class contains several widely used models as the binary Rasch model and the graded response model as special cases, allows for simplifications, and offers a distribution free alternative to count type items. A major strength of the models is that they can be used for mixed item formats, when different types of items are combined to measure abilities or attitudes. It is an immediate consequence of the comprehensive modeling approach that allows that difficulty functions automatically adapt to the response distribution. Basic properties of the model class are shown. Several real data sets are used to illustrate the flexibility of the models.
Collapse
Affiliation(s)
- Gerhard Tutz
- Ludwig-Maximilians-Universität München, Akademiestraße 1, 80799, Munich, Germany.
| |
Collapse
|
7
|
Marsman M, Huth K. Idiographic Ising and Divide and Color Models: Encompassing Networks for Heterogeneous Binary Data. MULTIVARIATE BEHAVIORAL RESEARCH 2022:1-28. [PMID: 36434773 DOI: 10.1080/00273171.2022.2135089] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
The Ising model is a graphical model that has played an essential role in network psychometrics. It has been used as a theoretical model to conceptualize psychological concepts and as a statistical model to analyze psychological data. Using graphical models such as the Ising model to analyze psychological data has been heavily critiqued since these data often come from cross-sectional applications. An often voiced concern is the inability of the Ising model to express heterogeneity in the population. The idiographic approach has been posed as an alternative and aims to infer individual network structures. While idiographic networks overcome population heterogeneity, it is unclear how they aggregate into established cross-sectional phenomena. This paper establishes a formal bridge between idiographic and cross-sectional network approaches of the Ising model. We ascertain unique topological structures that characterize individuals and aggregate into an Ising model cross-sectionally. This new formulation supports population heterogeneity while being consistent with cross-sectional phenomena. The proposed theory also establishes a new statistical framework for analyzing populations of idiographic networks for binary variables. The Ising model and the divide and color model are special cases of this new framework. We introduce a Gibbs sampling algorithm to estimate models from this new framework.
Collapse
Affiliation(s)
- M Marsman
- Department of Psychology, University of Amsterdam
| | - K Huth
- Department of Psychology, University of Amsterdam
- Department of Psychiatry, Amsterdam University Medical Center
- Centre for Urban Mental Health, University of Amsterdam
| |
Collapse
|
8
|
Waldorp L, Marsman M. Relations between Networks, Regression, Partial Correlation, and the Latent Variable Model. MULTIVARIATE BEHAVIORAL RESEARCH 2022; 57:994-1006. [PMID: 34397314 DOI: 10.1080/00273171.2021.1938959] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
The Gaussian graphical model (GGM) has become a popular tool for analyzing networks of psychological variables. In a recent article in this journal, Forbes, Wright, Markon, and Krueger (FWMK) voiced the concern that GGMs that are estimated from partial correlations wrongfully remove the variance that is shared by its constituents. If true, this concern has grave consequences for the application of GGMs. Indeed, if partial correlations only capture the unique covariances, then the data that come from a unidimensional latent variable model ULVM should be associated with an empty network (no edges), as there are no unique covariances in a ULVM. We know that this cannot be true, which suggests that FWMK are missing something with their claim. We introduce a connection between the ULVM and the GGM and use that connection to prove that we find a fully-connected and not an empty network associated with a ULVM. We then use the relation between GGMs and linear regression to show that the partial correlation indeed does not remove the common variance.
Collapse
Affiliation(s)
- Lourens Waldorp
- Faculty of Social and Behavioral Sciences, Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands
| | - Maarten Marsman
- Faculty of Social and Behavioral Sciences, Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands
| |
Collapse
|
9
|
de Ron J, Robinaugh DJ, Fried EI, Pedrelli P, Jain FA, Mischoulon D, Epskamp S. Quantifying and addressing the impact of measurement error in network models. Behav Res Ther 2022; 157:104163. [PMID: 36030733 PMCID: PMC10786122 DOI: 10.1016/j.brat.2022.104163] [Citation(s) in RCA: 14] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2021] [Revised: 06/30/2022] [Accepted: 07/12/2022] [Indexed: 11/02/2022]
Abstract
Network psychometric models are often estimated using a single indicator for each node in the network, thus failing to consider potential measurement error. In this study, we investigate the impact of measurement error on cross-sectional network models. First, we conduct a simulation study to evaluate the performance of models based on single indicators as well as models that utilize information from multiple indicators per node, including average scores, factor scores, and latent variables. Our results demonstrate that measurement error impairs the reliability and performance of network models, especially when using single indicators. The reliability and performance of network models improves substantially with increasing sample size and when using methods that combine information from multiple indicators per node. Second, we use empirical data from the STAR*D trial (n = 3,731) to further evaluate the impact of measurement error. In the STAR*D trial, depression symptoms were assessed via three questionnaires, providing multiple indicators per symptom. Consistent with our simulation results, we find that when using sub-samples of this dataset, the discrepancy between the three single-indicator networks (one network per questionnaire) diminishes with increasing sample size. Together, our simulated and empirical findings provide evidence that measurement error can hinder network estimation when working with smaller samples and offers guidance on methods to mitigate measurement error.
Collapse
Affiliation(s)
- Jill de Ron
- Department of Psychological Methods, University of Amsterdam, the Netherlands.
| | - Donald J Robinaugh
- Department of Psychiatry, Massachusetts General Hospital & Harvard Medical School, USA; Department of Applied Psychology, Northeastern University, USA
| | - Eiko I Fried
- Department of Clinical Psychology, Leiden University, the Netherlands
| | - Paola Pedrelli
- Department of Psychiatry, Massachusetts General Hospital & Harvard Medical School, USA
| | - Felipe A Jain
- Department of Psychiatry, Massachusetts General Hospital & Harvard Medical School, USA
| | - David Mischoulon
- Department of Psychiatry, Massachusetts General Hospital & Harvard Medical School, USA
| | - Sacha Epskamp
- Department of Psychological Methods, University of Amsterdam, the Netherlands; Centre for Urban Mental Health, University of Amsterdam, the Netherlands
| |
Collapse
|
10
|
Williams MN, Marques MD, Hill SR, Kerr JR, Ling M. Why are beliefs in different conspiracy theories positively correlated across individuals? Testing monological network versus unidimensional factor model explanations. BRITISH JOURNAL OF SOCIAL PSYCHOLOGY 2022; 61:1011-1031. [PMID: 35083755 DOI: 10.1111/bjso.12518] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2020] [Revised: 12/20/2021] [Accepted: 12/20/2021] [Indexed: 01/21/2023]
Abstract
A substantial minority of the public express belief in conspiracy theories. A robust phenomenon in this area is that people who believe one conspiracy theory are more likely to believe in others. But the reason for this "positive manifold" of belief in conspiracy theories is unclear. One possibility is that a single underlying latent factor (e.g. "conspiracism") causes variation in belief in specific conspiracy theories. Another possibility is that beliefs in various conspiracy theories support one another in a mutually reinforcing network of beliefs (the "monological belief system" theory). While the monological theory has been influential in the literature, the fact that it can be operationalised as a statistical network model has not previously been recognised. In this study, we therefore tested both the unidimensional factor model and a network model. Participants were 1553 American adults recruited via Prolific. Belief in conspiracies was measured using an adapted version of the Belief in Conspiracy Theories Inventory. The fit of the two competing models was evaluated both by using van Bork et al.'s (Psychometrika, 83, 2018, 443, Multivariate Behavioral Research, 56, 2019, 175) method for testing network versus unidimensional factor models, as well as by evaluating goodness of fit to the sample covariance matrix. In both cases, evaluation of fit according to our pre-registered inferential criteria favoured the network model.
Collapse
Affiliation(s)
| | | | | | | | - Mathew Ling
- Deakin University, Melbourne, Victoria, Australia
| |
Collapse
|
11
|
Cooke DJ, Hart SD, Logan C, Michie C. Evaluating the Test Validity of the Comprehensive Assessment of Psychopathic Personality Symptom Rating Sale (CAPP SRS). J Pers Assess 2021; 104:711-722. [PMID: 34739345 DOI: 10.1080/00223891.2021.1998082] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
The Comprehensive Assessment of Psychopathic Personality Symptom Rating Scale (CAPP SRS) is a relatively new measure of psychopathic personality disorder (PPD) based on the CAPP concept map of psychopathy. To investigate the CAPP SRS, we identified the most plausible formal test structure for the test using the framework proposed by Slaney and Maraun, identified an appropriate quantitative characterization of that test structure, and then statistically evaluated it based on analysis of CAPP SRS data collected from a multisite sample of 314 adult male correctional offenders and secure hospital patients in Scotland and England. Overall, the CAPP SRS survived falsification when observed test data were compared to expectations based on the unidimensional monotone latent variable or UMLV model of Holland and Rosenbaum. CAPP SRS composite scores calculated consistent with the UMLV model had good measurement precision and good external validity with respect to scores on an established test of PPD. The findings provide provisional support for the test validity of the CAPP SRS and highlight the importance of theory-driven evaluations of test validity.
Collapse
Affiliation(s)
- David J Cooke
- Glasgow Caledonian University, Glasgow, UK.,Faculty of Psychology, University of Bergen, Bergen, Norway
| | - Stephen D Hart
- Faculty of Psychology, University of Bergen, Bergen, Norway.,Department of Psychology, Simon Fraser University, Burnaby, Canada
| | - Caroline Logan
- Division of Psychology & Mental Health, School of Health Sciences, University of Manchester, Manchester, UK.,Edenfield Centre, Prestwich Hospital, Greater Manchester Mental Health, NHS Foundation Trust, Manchester, UK
| | | |
Collapse
|
12
|
van Bork R, Rhemtulla M, Waldorp LJ, Kruis J, Rezvanifar S, Borsboom D. Latent Variable Models and Networks: Statistical Equivalence and Testability. MULTIVARIATE BEHAVIORAL RESEARCH 2021; 56:175-198. [PMID: 31617420 DOI: 10.1080/00273171.2019.1672515] [Citation(s) in RCA: 49] [Impact Index Per Article: 16.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Networks are gaining popularity as an alternative to latent variable models for representing psychological constructs. Whereas latent variable approaches introduce unobserved common causes to explain the relations among observed variables, network approaches posit direct causal relations between observed variables. While these approaches lead to radically different understandings of the psychological constructs of interest, recent articles have established mathematical equivalences that hold between network models and latent variable models. We argue that the fact that for any model from one class there is an equivalent model from the other class does not mean that both models are equally plausible accounts of the data-generating mechanism. In many cases the constraints that are meaningful in one framework translate to constraints in the equivalent model that lack a clear interpretation in the other framework. Finally, we discuss three diverging predictions for the relation between zero-order correlations and partial correlations implied by sparse network models and unidimensional factor models. We propose a test procedure that compares the likelihoods of these models in light of these diverging implications. We use an empirical example to illustrate our argument.
Collapse
|
13
|
Studying Enhanced Recovery After Surgery (ERAS®) Core Items in Colorectal Surgery: A Causal Model with Latent Variables. World J Surg 2021; 45:928-939. [PMID: 33575826 PMCID: PMC7921056 DOI: 10.1007/s00268-020-05940-1] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/09/2020] [Indexed: 01/30/2023]
Abstract
Background Previous Enhanced Recovery After Surgery (ERAS®) studies have not always taken into account that ERAS interventions depend on baseline covariates and that several confounding variables affect the composite outcomes. Method A causal latent variable model is proposed to analyze data obtained prospectively concerning 1261 patients undergoing elective colorectal surgery within the ERAS protocol. Primary outcomes (composite of any complication, surgical site infection, medical complications, early ready for discharge (TRD), early actual discharge) and secondary outcomes (composite of late bowel function recovery, IV fluid resumption, nasogastric tube replacement, postoperative nausea and vomiting, re-intervention, re-admission, death) are considered along with their multiple dimensions. Results Concerning the primary outcomes, our results evidence three subpopulations of patients: one with probable good outcome, one with possibly prolonged TRD and discharge without complications, and the other one with probable complications and prolonged TRD and discharge. Epidural anesthesia, waiving surgical drainage, and early ambulation, IV fluid stop and urinary catheter removal act favorably, while preoperative hospital stay and blood transfusion act negatively. Concerning the secondary outcomes our results evidence two subpopulations of patients: one with high probability of good outcome and one with high probability of complications. Epidural anesthesia, waiving surgical drainage, early ambulation and IV fluid stop act favorably, while blood transfusion acts negatively also with respect to these secondary outcomes. Conclusion The multivariate causal latent class two-parameter logistic model, a modern statistical method overcoming drawbacks of traditional models to estimate the average causal effects on the treated, allows us to disentangle subpopulations of patients and to evaluate ERAS interventions.
Collapse
|
14
|
Blanchard MA, Heeren A. Why we should move from reductionism and embrace a network approach to parental burnout. New Dir Child Adolesc Dev 2020; 2020:159-168. [PMID: 33084239 DOI: 10.1002/cad.20377] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/28/2022]
Abstract
Network science has allowed varied scientific fields to investigate and visualize complex relations between many variables, and psychology research has begun to adopt a network perspective. In this paper, we consider how leaving behind reductionist approaches and instead embracing a network perspective can advance the field of parental burnout. Although research into parental burnout is in its early stages, we argue that a network approach to parental burnout could set the scene for radically new vistas in parental burnout research. We claim that such an approach can allow simultaneous investigations (and clear visualizations) of many variables related to parental burnout and their interactions, integrates smoothly with prior family systems theories, and prioritizes dynamic research questions. We likewise discuss potential future clinical applications, such as interventions targeting central nodes and treatment personalized to a specific family's network system. We also review practical considerations, limitations, and future directions for researchers interested in applying a network approach to parental burnout research.
Collapse
Affiliation(s)
| | - Alexandre Heeren
- Psychological Sciences Research Institute, UCLouvain, Louvain-la-Neuve, Belgium.,Institute of Neuroscience, UCLouvain, Louvain-la-Neuve, Belgium
| |
Collapse
|
15
|
Jain S, Rosenbaum PR, Reiter JG, Hoffman G, Small DS, Ha J, Hill AS, Wolk DA, Gaulton T, Neuman MD, Eckenhoff RG, Fleisher LA, Silber JH. Using Medicare claims in identifying Alzheimer's disease and related dementias. Alzheimers Dement 2020; 17:10.1002/alz.12199. [PMID: 33090695 PMCID: PMC8296851 DOI: 10.1002/alz.12199] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2020] [Revised: 08/25/2020] [Accepted: 08/29/2020] [Indexed: 12/11/2022]
Abstract
INTRODUCTION This study develops a measure of Alzheimer's disease and related dementias (ADRD) using Medicare claims. METHODS Validation resembles the approach of the American Psychological Association, including (1) content validity, (2) construct validity, and (3) predictive validity. RESULTS We found that four items-a Medicare claim recording ADRD 1 year ago, 2 years ago, 3 years ago, and a total stay of 6 months in a nursing home-exhibit a pattern of association consistent with a single underlying ADRD construct, and presence of any two of these four items predict a direct measure of cognitive function and also future claims for ADRD. DISCUSSION Our four items are internally consistent with the measurement of a single quantity. The presence of any two items do a better job than a single claim when predicting both a direct measure of cognitive function and future ADRD claims.
Collapse
Affiliation(s)
- Siddharth Jain
- Center for Outcomes Research, Children’s Hospital of Philadelphia, Philadelphia, PA
- The Leonard Davis Institute of Health Economics, The University of Pennsylvania, Philadelphia, PA
| | - Paul R. Rosenbaum
- The Leonard Davis Institute of Health Economics, The University of Pennsylvania, Philadelphia, PA
- Department of Statistics, The Wharton School, The University of Pennsylvania, Philadelphia, PA
| | - Joseph G. Reiter
- Center for Outcomes Research, Children’s Hospital of Philadelphia, Philadelphia, PA
| | - Geoffrey Hoffman
- Department of Systems, Populations and Leadership, University of Michigan School of Nursing, Ann Arbor, MI, USA
- University of Michigan’s Institute for Healthcare Policy and Innovation, Ann Arbor, MI, USA
| | - Dylan S. Small
- The Leonard Davis Institute of Health Economics, The University of Pennsylvania, Philadelphia, PA
- Department of Statistics, The Wharton School, The University of Pennsylvania, Philadelphia, PA
| | - JinKyung Ha
- Division of Geriatrics/Institute of Gerontology, University of Michigan, Ann Arbor, MI, USA
| | - Alexander S. Hill
- Center for Outcomes Research, Children’s Hospital of Philadelphia, Philadelphia, PA
| | - David A. Wolk
- Department of Neurology, The Perelman School of Medicine, The University of Pennsylvania
| | - Timothy Gaulton
- The Leonard Davis Institute of Health Economics, The University of Pennsylvania, Philadelphia, PA
- Center for Perioperative Outcomes Research and Transformation, The University of Pennsylvania, Philadelphia, PA
- Department of Anesthesiology and Critical Care, The University of Pennsylvania Perelman School of Medicine, Philadelphia, PA
| | - Mark D. Neuman
- The Leonard Davis Institute of Health Economics, The University of Pennsylvania, Philadelphia, PA
- Center for Perioperative Outcomes Research and Transformation, The University of Pennsylvania, Philadelphia, PA
- Department of Anesthesiology and Critical Care, The University of Pennsylvania Perelman School of Medicine, Philadelphia, PA
| | - Roderic G. Eckenhoff
- Department of Anesthesiology and Critical Care, The University of Pennsylvania Perelman School of Medicine, Philadelphia, PA
| | - Lee A. Fleisher
- The Leonard Davis Institute of Health Economics, The University of Pennsylvania, Philadelphia, PA
- Center for Perioperative Outcomes Research and Transformation, The University of Pennsylvania, Philadelphia, PA
- Department of Anesthesiology and Critical Care, The University of Pennsylvania Perelman School of Medicine, Philadelphia, PA
| | - Jeffrey H. Silber
- Center for Outcomes Research, Children’s Hospital of Philadelphia, Philadelphia, PA
- The Leonard Davis Institute of Health Economics, The University of Pennsylvania, Philadelphia, PA
- Department of Anesthesiology and Critical Care, The University of Pennsylvania Perelman School of Medicine, Philadelphia, PA
- The Departments of Pediatrics, The University of Pennsylvania Perelman School of Medicine, Philadelphia, PA
- Department of Health Care Management, The Wharton School, The University of Pennsylvania, Philadelphia, PA
| |
Collapse
|
16
|
Ellis JL, Pecanka J, Goeman JJ. Gaining power in multiple testing of interval hypotheses via conditionalization. Biostatistics 2020; 21:e65-e79. [PMID: 30247521 DOI: 10.1093/biostatistics/kxy042] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2017] [Revised: 07/12/2018] [Accepted: 08/04/2018] [Indexed: 11/14/2022] Open
Abstract
In this article, we introduce a novel procedure for improving power of multiple testing procedures (MTPs) of interval hypotheses. When testing interval hypotheses the null hypothesis $P$-values tend to be stochastically larger than standard uniform if the true parameter is in the interior of the null hypothesis. The new procedure starts with a set of $P$-values and discards those with values above a certain pre-selected threshold, while the rest are corrected (scaled-up) by the value of the threshold. Subsequently, a chosen family-wise error rate (FWER) or false discovery rate MTP is applied to the set of corrected $P$-values only. We prove the general validity of this procedure under independence of $P$-values, and for the special case of the Bonferroni method, we formulate several sufficient conditions for the control of the FWER. It is demonstrated that this "filtering" of $P$-values can yield considerable gains of power.
Collapse
Affiliation(s)
- Jules L Ellis
- Behavioral Science Institute, Radboud University Nijmegen, Postbus 9104, 6500 HE, Nijmegen, The Netherlands
| | - Jakub Pecanka
- Biomedical Data Sciences, Leiden University Medical Center, Postbus 9600, 2300 RC, Leiden, The Netherlands
| | - Jelle J Goeman
- Biomedical Data Sciences, Leiden University Medical Center, Postbus 9600, 2300 RC, Leiden, The Netherlands
| |
Collapse
|
17
|
Lange J, Dalege J, Borsboom D, van Kleef GA, Fischer AH. Toward an Integrative Psychometric Model of Emotions. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2020; 15:444-468. [PMID: 32040935 PMCID: PMC7059206 DOI: 10.1177/1745691619895057] [Citation(s) in RCA: 34] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/20/2023]
Abstract
Emotions are part and parcel of the human condition, but their nature is debated. Three broad classes of theories about the nature of emotions can be distinguished: affect-program theories, constructionist theories, and appraisal theories. Integrating these broad classes of theories into a unifying theory is challenging. An integrative psychometric model of emotions can inform such a theory because psychometric models are intertwined with theoretical perspectives about constructs. To identify an integrative psychometric model, we delineate properties of emotions stated by emotion theories and investigate whether psychometric models account for these properties. Specifically, an integrative psychometric model of emotions should allow (a) identifying distinct emotions (central in affect-program theories), (b) between- and within-person variations of emotions (central in constructionist theories), and (c) causal relationships between emotion components (central in appraisal theories). Evidence suggests that the popular reflective and formative latent variable models-in which emotions are conceptualized as unobservable causes or consequences of emotion components-cannot account for all properties. Conversely, a psychometric network model-in which emotions are conceptualized as systems of causally interacting emotion components-accounts for all properties. The psychometric network model thus constitutes an integrative psychometric model of emotions, facilitating progress toward a unifying theory.
Collapse
Affiliation(s)
- Jens Lange
- Psychology Research Institute, University of
Amsterdam
| | - Jonas Dalege
- Psychology Research Institute, University of
Amsterdam
| | | | | | | |
Collapse
|
18
|
Chen X. False discovery rate control for multiple testing based on discrete p-values. Biom J 2020; 62:1060-1079. [PMID: 31958180 DOI: 10.1002/bimj.201900163] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2019] [Revised: 10/07/2019] [Accepted: 10/15/2019] [Indexed: 11/08/2022]
Abstract
For multiple testing based on discrete p-values, we propose a false discovery rate (FDR) procedure "BH+" with proven conservativeness. BH+ is at least as powerful as the BH (i.e., Benjamini-Hochberg) procedure when they are applied to superuniform p-values. Further, when applied to mid-p-values, BH+ can be more powerful than it is applied to conventional p-values. An easily verifiable necessary and sufficient condition for this is provided. BH+ is perhaps the first conservative FDR procedure applicable to mid-p-values and to p-values with general distributions. It is applied to multiple testing based on discrete p-values in a methylation study, an HIV study and a clinical safety study, where it makes considerably more discoveries than the BH procedure. In addition, we propose an adaptive version of the BH+ procedure, prove its conservativeness under certain conditions, and provide evidence on its excellent performance via simulation studies.
Collapse
Affiliation(s)
- Xiongzhi Chen
- Department of Mathematics and Statistics, Washington State University, Pullman, WA, USA
| |
Collapse
|
19
|
Abstract
The positive manifold of intelligence has fascinated generations of scholars in human ability. In the past century, various formal explanations have been proposed, including the dominant g factor, the revived sampling theory, and the recent multiplier effect model and mutualism model. In this article, we propose a novel idiographic explanation. We formally conceptualize intelligence as evolving networks in which new facts and procedures are wired together during development. The static model, an extension of the Fortuin-Kasteleyn model, provides a parsimonious explanation of the positive manifold and intelligence's hierarchical factor structure. We show how it can explain the Matthew effect across developmental stages. Finally, we introduce a method for studying growth dynamics. Our truly idiographic approach offers a new view on a century-old construct and ultimately allows the fields of human ability and human learning to coalesce.
Collapse
Affiliation(s)
| | | | | | - Gunter K. J. Maris
- Department of Psychology, University of Amsterdam
- ACTNext by ACT, Inc., Iowa City, Iowa
| |
Collapse
|
20
|
Jewsbury PA. Diagnostic Test Score Validation With a Fallible Criterion. APPLIED PSYCHOLOGICAL MEASUREMENT 2019; 43:579-596. [PMID: 31551637 PMCID: PMC6745629 DOI: 10.1177/0146621618817785] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Criterion-related validation of diagnostic test scores for a construct of interest is complicated by the unavailability of the construct directly. The standard method, Known Group Validation, assumes an infallible reference test in place of the construct, but infallible reference tests are rare. In contrast, Mixed Group Validation allows for a fallible reference test, but has been found to make strong assumptions not appropriate for the majority of diagnostic test validation studies. The Neighborhood model is adapted for the purpose of diagnostic test validation, which makes alternate, but also strong, assumptions. The statistical properties of the Neighborhood model are evaluated and the assumptions are reviewed in the context of diagnostic test validation. Alternatively, strong assumptions may be avoided by estimating only intervals for the validity estimates, instead of point estimates. The Method of Bounds is also adapted for the purpose of diagnostic test validation, and an extension, Method of Bounds-Test Validation, is introduced here for the first time. All three point-estimate methods were found to make strong assumptions concerning the conditional relationships between the tests and the construct of interest, and all three lack robustness to assumption violation. The Method of Bounds-Test Validation was found to perform well across a range of plausible simulated datasets where the point-estimate methods failed. The point-estimate methods are recommended in special cases where the assumptions can be justified, while the interval methods are appropriate more generally.
Collapse
|
21
|
von Davier M, Cho Y, Pan T. Effects of Discontinue Rules on Psychometric Properties of Test Scores. PSYCHOMETRIKA 2019; 84:147-163. [PMID: 30607661 DOI: 10.1007/s11336-018-09652-3] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/16/2017] [Indexed: 06/09/2023]
Abstract
This paper provides results on a form of adaptive testing that is used frequently in intelligence testing. In these tests, items are presented in order of increasing difficulty. The presentation of items is adaptive in the sense that a session is discontinued once a test taker produces a certain number of incorrect responses in sequence, with subsequent (not observed) responses commonly scored as wrong. The Stanford-Binet Intelligence Scales (SB5; Riverside Publishing Company, 2003) and the Kaufman Assessment Battery for Children (KABC-II; Kaufman and Kaufman, 2004), the Kaufman Adolescent and Adult Intelligence Test (Kaufman and Kaufman 2014) and the Universal Nonverbal Intelligence Test (2nd ed.) (Bracken and McCallum 2015) are some of the many examples using this rule. He and Wolfe (Educ Psychol Meas 72(5):808-826, 2012. https://doi.org/10.1177/0013164412441937 ) compared different ability estimation methods in a simulation study for this discontinue rule adaptation of test length. However, there has been no study, to our knowledge, of the underlying distributional properties based on analytic arguments drawing on probability theory, of what these authors call stochastic censoring of responses. The study results obtained by He and Wolfe (Educ Psychol Meas 72(5):808-826, 2012. https://doi.org/10.1177/0013164412441937 ) agree with results presented by DeAyala et al. (J Educ Meas 38:213-234, 2001) as well as Rose et al. (Modeling non-ignorable missing data with item response theory (IRT; ETS RR-10-11), Educational Testing Service, Princeton, 2010) and Rose et al. (Psychometrika 82:795-819, 2017. https://doi.org/10.1007/s11336-016-9544-7 ) in that ability estimates are biased most when scoring the not observed responses as wrong. This scoring is used operationally, so more research is needed in order to improve practice in this field. The paper extends existing research on adaptivity by discontinue rules in intelligence tests in multiple ways: First, an analytical study of the distributional properties of discontinue rule scored items is presented. Second, a simulation is presented that includes additional scoring rules and uses ability estimators that may be suitable to reduce bias for discontinue rule scored intelligence tests.
Collapse
Affiliation(s)
- Matthias von Davier
- National Board of Medical Examiners, 3750 Market Street, Philadelphia, PA, 19104-3102, USA.
| | - Youngmi Cho
- American Institutes for Research, 1000 Thomas Jefferson Street, NW, Washington D.C., 20007, USA
| | - Tianshu Pan
- Pearson, 19500 Bulverde Rd, San Antonio, TX, 78259, USA
| |
Collapse
|
22
|
Maraun MD, Metta A, Hart SD, Fraser-Maraun J, Heene M. The dimensionality of the hare psychopathy checklist-revised, revisited: Its purported multidimensionality might well be artifactual. PERSONALITY AND INDIVIDUAL DIFFERENCES 2019. [DOI: 10.1016/j.paid.2018.09.004] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
|
23
|
Langer JK, Tonge NA, Piccirillo M, Rodebaugh TL, Thompson RJ, Gotlib IH. Symptoms of social anxiety disorder and major depressive disorder: A network perspective. J Affect Disord 2019; 243:531-538. [PMID: 30292147 PMCID: PMC6202058 DOI: 10.1016/j.jad.2018.09.078] [Citation(s) in RCA: 24] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/20/2018] [Revised: 09/05/2018] [Accepted: 09/21/2018] [Indexed: 12/26/2022]
Abstract
BACKGROUND We used network analyses to examine symptoms that may play a role in the co-occurrence of social anxiety disorder (SAD) and major depressive disorder (MDD). Whereas latent variable models examine relations among latent constructs, network analyses have the advantage of characterizing direct relations among the symptoms themselves. METHOD We conducted network modeling on symptoms of social anxiety and depression in a clinical sample of 130 women who met criteria for SAD, MDD, both disorders, or had no lifetime history of mental illness. RESULTS In the resulting network, the core symptoms of social fear and depressed mood appeared at opposite ends of the network and were weakly related; so-called "bridges" between these symptoms appeared to occur via intervening variables. In particular, the worthless variable appeared to play a central role in the network. LIMITATIONS Because our data were cross-sectional, we are unable to draw conclusions about the direction of these effects or whether these variables are related to each other prospectively. CONCLUSIONS Continued testing of these pathways using longitudinal data will help facilitate the development of more effective clinical interventions for these disorders.
Collapse
Affiliation(s)
- Julia K. Langer
- Minneapolis Veterans Affairs Health Care System, Washington University in St. Louis
| | - Natasha A. Tonge
- Department of Psychological and Brain Sciences, Washington University in St. Louis
| | - Marilyn Piccirillo
- Department of Psychological and Brain Sciences, Washington University in St. Louis
| | - Thomas L. Rodebaugh
- Department of Psychological and Brain Sciences, Washington University in St. Louis
| | - Renee J. Thompson
- Department of Psychological and Brain Sciences, Washington University in St. Louis
| | | |
Collapse
|
24
|
Halpin PF, Bergner Y. Psychometric Models of Small Group Collaborations. PSYCHOMETRIKA 2018; 83:941-962. [PMID: 30094746 DOI: 10.1007/s11336-018-9631-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/13/2017] [Revised: 06/20/2018] [Indexed: 06/08/2023]
Abstract
The social combination theory of group problem solving is used to extend existing psychometric models to collaborative settings. A model for pairwise group work is proposed, the implications of the model for assessment design are considered, and its estimation is addressed. The results are illustrated with an empirical example in which dyads work together on a twelfth-grade level mathematics assessment. In conclusion, attention is given to avenues of research that seem most fruitful for advancing current initiatives concerning the assessment of collaboration, teamwork, and related constructs.
Collapse
Affiliation(s)
- Peter F Halpin
- New York University, 246 Greene Street, Office 204, New York, NY, 10002, USA.
| | - Yoav Bergner
- New York University, 246 Greene Street, Office 204, New York, NY, 10002, USA
| |
Collapse
|
25
|
Epskamp S, Waldorp LJ, Mõttus R, Borsboom D. The Gaussian Graphical Model in Cross-Sectional and Time-Series Data. MULTIVARIATE BEHAVIORAL RESEARCH 2018; 53:453-480. [PMID: 29658809 DOI: 10.1080/00273171.2018.1454823] [Citation(s) in RCA: 377] [Impact Index Per Article: 62.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/02/2023]
Abstract
We discuss the Gaussian graphical model (GGM; an undirected network of partial correlation coefficients) and detail its utility as an exploratory data analysis tool. The GGM shows which variables predict one-another, allows for sparse modeling of covariance structures, and may highlight potential causal relationships between observed variables. We describe the utility in three kinds of psychological data sets: data sets in which consecutive cases are assumed independent (e.g., cross-sectional data), temporally ordered data sets (e.g., n = 1 time series), and a mixture of the 2 (e.g., n > 1 time series). In time-series analysis, the GGM can be used to model the residual structure of a vector-autoregression analysis (VAR), also termed graphical VAR. Two network models can then be obtained: a temporal network and a contemporaneous network. When analyzing data from multiple subjects, a GGM can also be formed on the covariance structure of stationary means-the between-subjects network. We discuss the interpretation of these models and propose estimation methods to obtain these networks, which we implement in the R packages graphicalVAR and mlVAR. The methods are showcased in two empirical examples, and simulation studies on these methods are included in the supplementary materials.
Collapse
Affiliation(s)
- Sacha Epskamp
- a Department of Psychological Methods , University of Amsterdam
| | | | - René Mõttus
- b Department of Psychology , University of Edinburgh
| | - Denny Borsboom
- a Department of Psychological Methods , University of Amsterdam
| |
Collapse
|
26
|
Affiliation(s)
- Jingshu Wang
- Department of Statistics, University of Pennsylvania, Philadelphia, PA
| | - Art B. Owen
- Department of Statistics, Stanford University, Stanford, CA
| |
Collapse
|
27
|
van Bork R, Grasman RPPP, Waldorp LJ. Unidimensional factor models imply weaker partial correlations than zero-order correlations. PSYCHOMETRIKA 2018; 83:443-452. [PMID: 29488148 DOI: 10.1007/s11336-018-9607-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/22/2016] [Revised: 02/07/2018] [Indexed: 06/08/2023]
Abstract
In this paper we present a new implication of the unidimensional factor model. We prove that the partial correlation between two observed variables that load on one factor given any subset of other observed variables that load on this factor lies between zero and the zero-order correlation between these two observed variables. We implement this result in an empirical bootstrap test that rejects the unidimensional factor model when partial correlations are identified that are either stronger than the zero-order correlation or have a different sign than the zero-order correlation. We demonstrate the use of the test in an empirical data example with data consisting of fourteen items that measure extraversion.
Collapse
Affiliation(s)
- Riet van Bork
- Department of Psychological Methods, University of Amsterdam, Nieuwe Achtergracht 129-B, 1018 WS, Amsterdam, The Netherlands.
| | - Raoul P P P Grasman
- Department of Psychological Methods, University of Amsterdam, Nieuwe Achtergracht 129-B, 1018 WS, Amsterdam, The Netherlands
| | - Lourens J Waldorp
- Department of Psychological Methods, University of Amsterdam, Nieuwe Achtergracht 129-B, 1018 WS, Amsterdam, The Netherlands
| |
Collapse
|
28
|
Heeren A, Bernstein EE, McNally RJ. Deconstructing trait anxiety: a network perspective. ANXIETY STRESS AND COPING 2018; 31:262-276. [PMID: 29433339 DOI: 10.1080/10615806.2018.1439263] [Citation(s) in RCA: 41] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/05/2023]
Abstract
BACKGROUND AND OBJECTIVES For decades, the dominant paradigm in trait anxiety research has regarded the construct as signifying the underlying cause of the thoughts, feelings, and behaviors that supposedly reflect its presence. Recently, a network theory of personality has appeared. According to this perspective, trait anxiety is a formative construct emerging from interactions among its constitutive features (e.g., thought, feelings, behaviors); it is not a latent cause of these features. DESIGN In this study, we characterized trait anxiety as a network system of interacting elements. METHODS To do so, we estimated a graphical gaussian model via the computation of a regularized partial correlation network in an unselected sample (N = 611). We also implemented modularity-based community detection analysis to test whether the features of trait anxiety cohere as a single network system. RESULTS We find that trait anxiety can indeed be conceptualized as a single, coherent network system of interacting elements. CONCLUSIONS This radically new approach to visualizing trait anxiety may offer an especially informative view of the interplay between its constitutive features. As prior research has implicated trait anxiety as a risk factor for the development of anxiety-related psychopathology, our findings also set the scene for novel research directions.
Collapse
Affiliation(s)
- Alexandre Heeren
- a Department of Psychology , Harvard University , Cambridge , MA , USA.,b Psychological Science Research Institute , Université Catholique de Louvain , Louvain-la-Neuve , Belgium.,c Institute of Neuroscience , Université Catholique de Louvain , Brussels , Belgium
| | - Emily E Bernstein
- a Department of Psychology , Harvard University , Cambridge , MA , USA
| | - Richard J McNally
- a Department of Psychology , Harvard University , Cambridge , MA , USA
| |
Collapse
|
29
|
Epskamp S, Rhemtulla M, Borsboom D. Generalized Network Psychometrics: Combining Network and Latent Variable Models. PSYCHOMETRIKA 2017; 82:904-927. [PMID: 28290111 DOI: 10.1007/s11336-017-9557-x] [Citation(s) in RCA: 245] [Impact Index Per Article: 35.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/24/2015] [Revised: 11/25/2016] [Indexed: 05/21/2023]
Abstract
We introduce the network model as a formal psychometric model, conceptualizing the covariance between psychometric indicators as resulting from pairwise interactions between observable variables in a network structure. This contrasts with standard psychometric models, in which the covariance between test items arises from the influence of one or more common latent variables. Here, we present two generalizations of the network model that encompass latent variable structures, establishing network modeling as parts of the more general framework of structural equation modeling (SEM). In the first generalization, we model the covariance structure of latent variables as a network. We term this framework latent network modeling (LNM) and show that, with LNM, a unique structure of conditional independence relationships between latent variables can be obtained in an explorative manner. In the second generalization, the residual variance-covariance structure of indicators is modeled as a network. We term this generalization residual network modeling (RNM) and show that, within this framework, identifiable models can be obtained in which local independence is structurally violated. These generalizations allow for a general modeling framework that can be used to fit, and compare, SEM models, network models, and the RNM and LNM generalizations. This methodology has been implemented in the free-to-use software package lvnet, which contains confirmatory model testing as well as two exploratory search algorithms: stepwise search algorithms for low-dimensional datasets and penalized maximum likelihood estimation for larger datasets. We show in simulation studies that these search algorithms perform adequately in identifying the structure of the relevant residual or latent networks. We further demonstrate the utility of these generalizations in an empirical example on a personality inventory dataset.
Collapse
Affiliation(s)
- Sacha Epskamp
- University of Amsterdam, PO Box 15906, 1001 NK, Amsterdam, The Netherlands.
| | - Mijke Rhemtulla
- University of Amsterdam, PO Box 15906, 1001 NK, Amsterdam, The Netherlands
| | - Denny Borsboom
- University of Amsterdam, PO Box 15906, 1001 NK, Amsterdam, The Netherlands
| |
Collapse
|
30
|
Lee CP, Chou YH, Liu CY, Hung CI. Dimensionality of the Chinese hospital anxiety depression scale in psychiatric outpatients: Mokken scale and factor analyses. Int J Psychiatry Clin Pract 2017; 21:283-291. [PMID: 28417655 DOI: 10.1080/13651501.2017.1311350] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
OBJECTIVE The hospital anxiety and depression scale (HADS) is a widely used scale of anxiety and depression. However, recent studies have challenged the bi-dimensional scoring of the HADS. The present study was to examine the dimensionality of the Chinese HADS. METHODS We recruited a convenience sample of 214 adult psychiatric outpatients at a medical centre in Taiwan, and they completed the Chinese HADS. We used Mokken scale analysis (MSA), exploratory factor analysis (EFA), exploratory bifactor analysis (EBA) and confirmatory factor analysis (CFA) to examine the dimensionality of the Chinese HADS. RESULTS The Chinese HADS was a moderate Mokken scale (Hs = 0.44), and had a two-factor structure. EBA showed that one general factor, emotional distress, explained 68% of the common variance of the Chinese HADS. CFA confirmed that the bifactor model had the best fit statistics. The items 5 and 7 of the Chinese HADS contributed to structural ambiguity in the Chinese HADS subscales. CONCLUSIONS The sum scores of the Chinese HADS were a reliable and valid unidimensional measure of emotional distress. The Chinese HADS subscales were incapable of differentiating between anxiety and depression. Clinicians and researchers should choose other scales that are specifically designed for measuring anxiety and depression.
Collapse
Affiliation(s)
- Chin-Pang Lee
- a Department of Psychiatry , Chang Gung Memorial Hospital , Linkou , Taiwan.,b Department of Psychiatry, School of Medicine , Chang Gung University , Taoyuan , Taiwan
| | - Ya-Hsin Chou
- a Department of Psychiatry , Chang Gung Memorial Hospital , Linkou , Taiwan.,b Department of Psychiatry, School of Medicine , Chang Gung University , Taoyuan , Taiwan
| | - Chia-Yih Liu
- a Department of Psychiatry , Chang Gung Memorial Hospital , Linkou , Taiwan.,b Department of Psychiatry, School of Medicine , Chang Gung University , Taoyuan , Taiwan
| | - Ching-I Hung
- a Department of Psychiatry , Chang Gung Memorial Hospital , Linkou , Taiwan.,b Department of Psychiatry, School of Medicine , Chang Gung University , Taoyuan , Taiwan
| |
Collapse
|
31
|
Jordan P, Shedden-Mora MC, Löwe B. Psychometric analysis of the Generalized Anxiety Disorder scale (GAD-7) in primary care using modern item response theory. PLoS One 2017; 12:e0182162. [PMID: 28771530 PMCID: PMC5542568 DOI: 10.1371/journal.pone.0182162] [Citation(s) in RCA: 90] [Impact Index Per Article: 12.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2017] [Accepted: 07/13/2017] [Indexed: 11/18/2022] Open
Abstract
OBJECTIVE The Generalized Anxiety Disorder scale (GAD-7) is one of the most frequently used diagnostic self-report scales for screening, diagnosis and severity assessment of anxiety disorder. Its psychometric properties from the view of the Item Response Theory paradigm have rarely been investigated. We aimed to close this gap by analyzing the GAD-7 within a large sample of primary care patients with respect to its psychometric properties and its implications for scoring using Item Response Theory. METHODS Robust, nonparametric statistics were used to check unidimensionality of the GAD-7. A graded response model was fitted using a Bayesian approach. The model fit was evaluated using posterior predictive p-values, item information functions were derived and optimal predictions of anxiety were calculated. RESULTS The sample included N = 3404 primary care patients (60% female; mean age, 52,2; standard deviation 19.2) The analysis indicated no deviations of the GAD-7 scale from unidimensionality and a decent fit of a graded response model. The commonly suggested ultra-brief measure consisting of the first two items, the GAD-2, was supported by item information analysis. The first four items discriminated better than the last three items with respect to latent anxiety. CONCLUSION The information provided by the first four items should be weighted more heavily. Moreover, estimates corresponding to low to moderate levels of anxiety show greater variability. The psychometric validity of the GAD-2 was supported by our analysis.
Collapse
Affiliation(s)
- Pascal Jordan
- Department of Psychosomatic Medicine and Psychotherapy, University Medical Center Hamburg-Eppendorf and Schön Klinik Hamburg Eilbek, Hamburg, Germany
- Department of Psychological Methods, Faculty of Psychology and Movement Sciences. University of Hamburg, Hamburg, Germany
- * E-mail:
| | - Meike C. Shedden-Mora
- Department of Psychosomatic Medicine and Psychotherapy, University Medical Center Hamburg-Eppendorf and Schön Klinik Hamburg Eilbek, Hamburg, Germany
| | - Bernd Löwe
- Department of Psychosomatic Medicine and Psychotherapy, University Medical Center Hamburg-Eppendorf and Schön Klinik Hamburg Eilbek, Hamburg, Germany
| |
Collapse
|
32
|
Fallat S, Lauritzen S, Sadeghi K, Uhler C, Wermuth N, Zwiernik P. Total positivity in Markov structures. Ann Stat 2017. [DOI: 10.1214/16-aos1478] [Citation(s) in RCA: 25] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
33
|
Sijtsma K, van der Ark LA. A tutorial on how to do a Mokken scale analysis on your test and questionnaire data. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2017; 70:137-158. [PMID: 27958642 DOI: 10.1111/bmsp.12078] [Citation(s) in RCA: 102] [Impact Index Per Article: 14.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/29/2016] [Revised: 08/29/2016] [Indexed: 05/28/2023]
Abstract
Over the past decade, Mokken scale analysis (MSA) has rapidly grown in popularity among researchers from many different research areas. This tutorial provides researchers with a set of techniques and a procedure for their application, such that the construction of scales that have superior measurement properties is further optimized, taking full advantage of the properties of MSA. First, we define the conceptual context of MSA, discuss the two item response theory (IRT) models that constitute the basis of MSA, and discuss how these models differ from other IRT models. Second, we discuss dos and don'ts for MSA; the don'ts include misunderstandings we have frequently encountered with researchers in our three decades of experience with real-data MSA. Third, we discuss a methodology for MSA on real data that consist of a sample of persons who have provided scores on a set of items that, depending on the composition of the item set, constitute the basis for one or more scales, and we use the methodology to analyse an example real-data set.
Collapse
|
34
|
Nandakumar R, Stout W. Refinements of Stout’s Procedure for Assessing Latent Trait Unidimensionality. ACTA ACUST UNITED AC 2016. [DOI: 10.3102/10769986018001041] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
This article provides a detailed investigation of Stout’s statistical procedure (the computer program DIMTEST) for testing the hypothesis that an essentially unidimensional latent trait model fits observed binary item response data from a psychological test. One finding was that DIMTEST may fail to perform as desired in the presence of guessing when coupled with many high-discriminating items. A revision of DIMTEST is proposed to overcome this limitation. Also, an automatic approach is devised to determine the size of the assessment subtests. Further, an adjustment is made on the estimated standard error of the statistic on which DIMTEST depends. These three refinements have led to an improved procedure that is shown in simulation studies to adhere closely to the nominal level of signficance while achieving considerably greater power. Finally, DIMTEST is validated on a selection of real data sets.
Collapse
|
35
|
Straat JH, van der Ark LA, Sijtsma K. Using Conditional Association to Identify Locally Independent Item Sets. METHODOLOGY-EUROPEAN JOURNAL OF RESEARCH METHODS FOR THE BEHAVIORAL AND SOCIAL SCIENCES 2016. [DOI: 10.1027/1614-2241/a000115] [Citation(s) in RCA: 24] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022] Open
Abstract
Abstract. The ordinal, unidimensional monotone latent variable model assumes unidimensionality, local independence, and monotonicity, and implies the observable property of conditional association. We investigated three special cases of conditional association and implemented them in a new procedure that aims at identifying locally dependent items, removing these items from the initial item set, and producing an item subset that is locally independent. A simulation study showed that the new procedure correctly identified 89.5% of the model-consistent items and up to 90% of the model-inconsistent items. We recommend using this procedure for selecting locally independent item sets. The procedure may be used in combination with Mokken scale analysis.
Collapse
|
36
|
Abstract
Conditional stochastic ordering is concerned with the stochastic ordering of a pair of probability measures conditional on certain subsets or sub-σ -algebras. Some basic results of conditional stochastic ordering were proved by Whitt. We extend some of Whitt's results and prove a basic relation between stochastic ordering conditional on subsets and stochastic ordering conditional on σ -algebras. In the second part of the paper we consider the ordering of conditional expectations. There are several different formulations of this problem motivated by different types of applications.
Collapse
|
37
|
Abstract
Conditional stochastic ordering is concerned with the stochastic ordering of a pair of probability measures conditional on certain subsets or sub-σ-algebras. Some basic results of conditional stochastic ordering were proved by Whitt. We extend some of Whitt's results and prove a basic relation between stochastic ordering conditional on subsets and stochastic ordering conditional onσ-algebras. In the second part of the paper we consider the ordering of conditional expectations. There are several different formulations of this problem motivated by different types of applications.
Collapse
|
38
|
Verweij AC, Sijtsma K, Koops W. A Mokken Scale for Transitive Reasoning Suited for Longitudinal Research. INTERNATIONAL JOURNAL OF BEHAVIORAL DEVELOPMENT 2016. [DOI: 10.1177/016502549601900115] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
A set of transitivity tasks was solved on two occasions by 634 2nd through 6th grade primary school children. Transitivity scales were constructed for both occasions using Mokken scale analysis. Combining the results, one final scale was constructed for use in future research. The order of the scale tasks according to their difficulty level was the same on both occasions. Analysis of the scale scores of one group retested after 16 weeks and two control groups revealed an absence of effects due to memorisation or test experience. It was concluded that the final scale is useful as a transitivity measure in longitudinal research.
Collapse
|
39
|
Ligtvoet R. Remarks and a Correction of Ligtvoet's Treatment of the Isotonic Partial Credit Model. PSYCHOMETRIKA 2015; 80:514-515. [PMID: 24337959 DOI: 10.1007/s11336-013-9397-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/17/2013] [Indexed: 06/03/2023]
Abstract
This note contains some remarks on Ligtvoet's (Psychometrika, 77:479-494, 2012) treatment of the isotonic partial credit model. The Proposition relating to the observable property MWI is shown to be false.
Collapse
Affiliation(s)
- Rudy Ligtvoet
- Department of Pedagogical and Educational Science, University of Amsterdam, Nieuwe Prinsengracht 130, 1018 VZ, Amsterdam, The Netherlands,
| |
Collapse
|
40
|
Fried EI. Problematic assumptions have slowed down depression research: why symptoms, not syndromes are the way forward. Front Psychol 2015; 6:309. [PMID: 25852621 PMCID: PMC4369644 DOI: 10.3389/fpsyg.2015.00309] [Citation(s) in RCA: 186] [Impact Index Per Article: 20.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2014] [Accepted: 03/04/2015] [Indexed: 12/16/2022] Open
Abstract
Major depression (MD) is a highly heterogeneous diagnostic category. Diverse symptoms such as sad mood, anhedonia, and fatigue are routinely added to an unweighted sum-score, and cutoffs are used to distinguish between depressed participants and healthy controls. Researchers then investigate outcome variables like MD risk factors, biomarkers, and treatment response in such samples. These practices presuppose that (1) depression is a discrete condition, and that (2) symptoms are interchangeable indicators of this latent disorder. Here I review these two assumptions, elucidate their historical roots, show how deeply engrained they are in psychological and psychiatric research, and document that they contrast with evidence. Depression is not a consistent syndrome with clearly demarcated boundaries, and depression symptoms are not interchangeable indicators of an underlying disorder. Current research practices lump individuals with very different problems into one category, which has contributed to the remarkably slow progress in key research domains such as the development of efficacious antidepressants or the identification of biomarkers for depression. The recently proposed network framework offers an alternative to the problematic assumptions. MD is not understood as a distinct condition, but as heterogeneous symptom cluster that substantially overlaps with other syndromes such as anxiety disorders. MD is not framed as an underlying disease with a number of equivalent indicators, but as a network of symptoms that have direct causal influence on each other: insomnia can cause fatigue which then triggers concentration and psychomotor problems. This approach offers new opportunities for constructing an empirically based classification system and has broad implications for future research.
Collapse
Affiliation(s)
- Eiko I. Fried
- Research Group of Quantitative Psychology and Individual Differences, Faculty of Psychology and Educational Sciences, University of LeuvenLeuven, Belgium
| |
Collapse
|
41
|
van Rijn P, Rijmen F. On the explaining-away phenomenon in multivariate latent variable models. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2015; 68:1-22. [PMID: 25469472 DOI: 10.1111/bmsp.12046] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/23/2013] [Revised: 10/24/2014] [Indexed: 06/04/2023]
Abstract
Many probabilistic models for psychological and educational measurements contain latent variables. Well-known examples are factor analysis, item response theory, and latent class model families. We discuss what is referred to as the 'explaining-away' phenomenon in the context of such latent variable models. This phenomenon can occur when multiple latent variables are related to the same observed variable, and can elicit seemingly counterintuitive conditional dependencies between latent variables given observed variables. We illustrate the implications of explaining away for a number of well-known latent variable models by using both theoretical and real data examples.
Collapse
|
42
|
Ligtvoet R. A test for using the sum score to obtain a stochastic ordering of subjects. J MULTIVARIATE ANAL 2015. [DOI: 10.1016/j.jmva.2014.09.003] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
43
|
Manrique-Vallier D, Reiter JP. Bayesian Estimation of Discrete Multivariate Latent Structure Models With Structural Zeros. J Comput Graph Stat 2014. [DOI: 10.1080/10618600.2013.844700] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
44
|
Affiliation(s)
- Paul W. Holland
- Research Statistics Group, Educational Testing Service; Princeton NJ 08541-0001
| | - Dorothy T. Thayer
- Research Statistics Group, Educational Testing Service; Princeton NJ 08541-0001
| |
Collapse
|
45
|
|
46
|
Mislevy RJ, Wu PK. MISSING RESPONSES AND IRT ABILITY ESTIMATION: OMITS, CHOICE, TIME LIMITS, AND ADAPTIVE TESTING. ACTA ACUST UNITED AC 2014. [DOI: 10.1002/j.2333-8504.1996.tb01708.x] [Citation(s) in RCA: 33] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
47
|
Ellis JL. An inequality for correlations in unidimensional monotone latent variable models for binary variables. PSYCHOMETRIKA 2014; 79:303-316. [PMID: 24659373 DOI: 10.1007/s11336-013-9341-5] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/01/2012] [Revised: 11/16/2012] [Indexed: 05/28/2023]
Abstract
It is shown that a unidimensional monotone latent variable model for binary items implies a restriction on the relative sizes of item correlations: The negative logarithm of the correlations satisfies the triangle inequality. This inequality is not implied by the condition that the correlations are nonnegative, the criterion that coefficient H exceeds 0.30, or manifest monotonicity. The inequality implies both a lower bound and an upper bound for each correlation between two items, based on the correlations of those two items with every possible third item. It is discussed how this can be used in Mokken's (A theory and procedure of scale-analysis, Mouton, The Hague, 1971) scale analysis.
Collapse
Affiliation(s)
- Jules L Ellis
- School of Psychology and Artificial Intelligence, Radboud University Nijmegen, P.O. Box 9104, 6500 HE, Nijmegen, The Netherlands,
| |
Collapse
|
48
|
Bazán JL, Branco MD, Bolfarine H. Extensions of the skew-normal ogive item response model. BRAZ J PROBAB STAT 2014. [DOI: 10.1214/12-bjps191] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
49
|
Wermuth N, Marchetti GM. Star graphs induce tetrad correlations: for Gaussian as well as for binary variables. Electron J Stat 2014. [DOI: 10.1214/14-ejs884] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
50
|
Haberman SJ, Sinharay S. Generalized Residuals for General Models for Contingency Tables With Application to Item Response Theory. J Am Stat Assoc 2013. [DOI: 10.1080/01621459.2013.835660] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|