1
|
Ulitzsch E, Zhang S, Pohl S. A Model-Based Approach to the Disentanglement and Differential Treatment of Engaged and Disengaged Item Omissions. MULTIVARIATE BEHAVIORAL RESEARCH 2024; 59:599-619. [PMID: 38594939 DOI: 10.1080/00273171.2024.2307518] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/11/2024]
Abstract
Item omissions in large-scale assessments may occur for various reasons, ranging from disengagement to not being capable of solving the item and giving up. Current response-time-based classification approaches allow researchers to implement different treatments of item omissions presumably going back to different mechanisms. These approaches, however, are limited in that they require a clear-cut decision on the underlying missingness mechanism and do not allow to take the uncertainty in classification into account. We present a response-time-based model-based mixture modeling approach that overcomes this limitation. The approach (a) facilitates disentangling item omissions stemming from disengagement from those going back to solution behavior, (b) considers the uncertainty in omission classification, (c) allows for omission mechanisms to vary on the item-by-examinee level, (d) supports investigating person and item characteristics associated with different types of omission behavior, and (e) gives researchers flexibility in deciding on how to handle different types of omissions. The approach exhibits good parameter recovery under realistic research conditions. We illustrate the approach on data from the Programme for the International Assessment of Adult Competencies 2012 and compare it against previous classification approaches for item omissions.
Collapse
Affiliation(s)
- Esther Ulitzsch
- IPN - Leibniz Institute for Science and Mathematics Education
| | - Susu Zhang
- University of Illinois at Urbana-Champaign
| | | |
Collapse
|
2
|
Couto J, Lebreton M, van Maanen L. Specificity and sensitivity of the fixed-point test for binary mixture distributions. Behav Res Methods 2024; 56:2977-2991. [PMID: 37957433 PMCID: PMC11133060 DOI: 10.3758/s13428-023-02244-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/17/2023] [Indexed: 11/15/2023]
Abstract
When two cognitive processes contribute to a behavioral output-each process producing a specific distribution of the behavioral variable of interest-and when the mixture proportion of these two processes varies as a function of an experimental condition, a common density point should be present in the observed distributions of the data across said conditions. In principle, one can statistically test for the presence (or absence) of a fixed point in experimental data to provide evidence in favor of (or against) the presence of a mixture of processes, whose proportions are affected by an experimental manipulation. In this paper, we provide an empirical diagnostic of this test to detect a mixture of processes. We do so using resampling of real experimental data under different scenarios, which mimic variations in the experimental design suspected to affect the sensitivity and specificity of the fixed-point test (i.e., mixture proportion, time on task, and sample size). Resampling such scenarios with real data allows us to preserve important features of data which are typically observed in real experiments while maintaining tight control over the properties of the resampled scenarios. This is of particular relevance considering such stringent assumptions underlying the fixed-point test. With this paper, we ultimately aim at validating the fixed-point property of binary mixture data and at providing some performance metrics to researchers aiming at testing the fixed-point property on their experimental data.
Collapse
Affiliation(s)
- Joaquina Couto
- Department of Psychology, University of Amsterdam, Amsterdam, Netherlands.
- Department of Experimental Psychology, Utrecht University, Utrecht, Netherlands.
| | - Maël Lebreton
- Swiss Center for Affective Sciences, University of Geneva, Geneva, Switzerland
- Paris School of Economics, Paris, France
| | - Leendert van Maanen
- Department of Experimental Psychology, Utrecht University, Utrecht, Netherlands
| |
Collapse
|
3
|
Ulitzsch E, Shin HJ, Lüdtke O. Accounting for careless and insufficient effort responding in large-scale survey data-development, evaluation, and application of a screen-time-based weighting procedure. Behav Res Methods 2024; 56:804-825. [PMID: 36867339 PMCID: PMC10830617 DOI: 10.3758/s13428-022-02053-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/09/2022] [Indexed: 03/04/2023]
Abstract
Careless and insufficient effort responding (C/IER) poses a major threat to the quality of large-scale survey data. Traditional indicator-based procedures for its detection are limited in that they are only sensitive to specific types of C/IER behavior, such as straight lining or rapid responding, rely on arbitrary threshold settings, and do not allow taking the uncertainty of C/IER classification into account. Overcoming these limitations, we develop a two-step screen-time-based weighting procedure for computer-administered surveys. The procedure allows considering the uncertainty in C/IER identification, is agnostic towards the specific types of C/IE response patterns, and can feasibly be integrated with common analysis workflows for large-scale survey data. In Step 1, we draw on mixture modeling to identify subcomponents of log screen time distributions presumably stemming from C/IER. In Step 2, the analysis model of choice is applied to item response data, with respondents' posterior class probabilities being employed to downweigh response patterns according to their probability of stemming from C/IER. We illustrate the approach on a sample of more than 400,000 respondents being administered 48 scales of the PISA 2018 background questionnaire. We gather supporting validity evidence by investigating relationships between C/IER proportions and screen characteristics that entail higher cognitive burden, such as screen position and text length, relating identified C/IER proportions to other indicators of C/IER as well as by investigating rank-order consistency in C/IER behavior across screens. Finally, in a re-analysis of the PISA 2018 background questionnaire data, we investigate the impact of the C/IER adjustments on country-level comparisons.
Collapse
Affiliation(s)
- Esther Ulitzsch
- IPN-Leibniz Institute for Science and Mathematics Education, Educational Measurement, Olshausenstraße 62, 24118, Kiel, Germany.
- Centre for International Student Assessment, Munich, Germany.
| | | | - Oliver Lüdtke
- IPN-Leibniz Institute for Science and Mathematics Education, Educational Measurement, Olshausenstraße 62, 24118, Kiel, Germany
- Centre for International Student Assessment, Munich, Germany
| |
Collapse
|
4
|
Archambeau K, Couto J, Van Maanen L. Non-parametric mixture modeling of cognitive psychological data: A new method to disentangle hidden strategies. Behav Res Methods 2023; 55:2232-2248. [PMID: 36219308 PMCID: PMC10439044 DOI: 10.3758/s13428-022-01837-0] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 03/14/2022] [Indexed: 11/08/2022]
Abstract
In a wide variety of cognitive domains, participants have access to several alternative strategies to perform a particular task and, on each trial, one specific strategy is selected and executed. Determining how many strategies are used by a participant as well as their identification at a trial level is a challenging problem for researchers. In the current paper, we propose a new method - the non-parametric mixture model - to efficiently disentangle hidden strategies in cognitive psychological data, based on observed response times. The developed method derived from standard hidden Markov modeling. Importantly, we used a model-free approach where a particular shape of a response time distribution does not need to be assumed. This has the considerable advantage of avoiding potentially unreliable results when an inappropriate response time distribution is assumed. Through three simulation studies and two applications to real data, we repeatedly demonstrated that the non-parametric mixture model is able to reliably recover hidden strategies present in the data as well as to accurately estimate the number of concurrent strategies. The results also showed that this new method is more efficient than a standard parametric approach. The non-parametric mixture model is therefore a useful statistical tool for strategy identification that can be applied in many areas of cognitive psychology. To this end, practical guidelines are provided for researchers wishing to apply the non-parametric mixture models on their own data set.
Collapse
Affiliation(s)
- Kim Archambeau
- Department of Psychology, University of Amsterdam, Postbus 15906, 1001 Amsterdam, NK Netherlands
- Center for Research in Cognition and Neurosciences, Université Libre de Bruxelles, Brussels, Belgium
| | - Joaquina Couto
- Department of Psychology, University of Amsterdam, Postbus 15906, 1001 Amsterdam, NK Netherlands
| | - Leendert Van Maanen
- Department of Psychology, University of Amsterdam, Postbus 15906, 1001 Amsterdam, NK Netherlands
- Department of Experimental Psychology, Utrecht University, Heidelberglaan 1 – 3584, Utrecht, CS Netherlands
| |
Collapse
|
5
|
Liu Y, Wang W. Semiparametric Factor Analysis for Item-Level Response Time Data. PSYCHOMETRIKA 2022; 87:666-692. [PMID: 35098450 DOI: 10.1007/s11336-021-09832-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/05/2021] [Revised: 09/27/2021] [Indexed: 06/14/2023]
Abstract
Item-level response time (RT) data can be conveniently collected from computer-based test/survey delivery platforms and have been demonstrated to bear a close relation to a miscellany of cognitive processes and test-taking behaviors. Individual differences in general processing speed can be inferred from item-level RT data using factor analysis. Conventional linear normal factor models make strong parametric assumptions, which sacrifices modeling flexibility for interpretability, and thus are not ideal for describing complex associations between observed RT and the latent speed. In this paper, we propose a semiparametric factor model with minimal parametric assumptions. Specifically, we adopt a functional analysis of variance representation for the log conditional densities of the manifest variables, in which the main effect and interaction functions are approximated by cubic splines. Penalized maximum likelihood estimation of the spline coefficients can be performed by an Expectation-Maximization algorithm, and the penalty weight can be empirically determined by cross-validation. In a simulation study, we compare the semiparametric model with incorrectly and correctly specified parametric factor models with regard to the recovery of data generating mechanism. A real data example is also presented to demonstrate the advantages of the proposed method.
Collapse
Affiliation(s)
- Yang Liu
- Department of Human Development and Quantitative Methodology, University of Maryland, College Park, USA.
| | - Weimeng Wang
- Department of Human Development and Quantitative Methodology, University of Maryland, College Park, USA
| |
Collapse
|
6
|
Ulitzsch E, Pohl S, Khorramdel L, Kroehne U, von Davier M. A Response-Time-Based Latent Response Mixture Model for Identifying and Modeling Careless and Insufficient Effort Responding in Survey Data. PSYCHOMETRIKA 2022; 87:593-619. [PMID: 34855118 PMCID: PMC9166878 DOI: 10.1007/s11336-021-09817-7] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/25/2020] [Revised: 10/11/2021] [Indexed: 06/13/2023]
Abstract
Careless and insufficient effort responding (C/IER) can pose a major threat to data quality and, as such, to validity of inferences drawn from questionnaire data. A rich body of methods aiming at its detection has been developed. Most of these methods can detect only specific types of C/IER patterns. However, typically different types of C/IER patterns occur within one data set and need to be accounted for. We present a model-based approach for detecting manifold manifestations of C/IER at once. This is achieved by leveraging response time (RT) information available from computer-administered questionnaires and integrating theoretical considerations on C/IER with recent psychometric modeling approaches. The approach a) takes the specifics of attentive response behavior on questionnaires into account by incorporating the distance-difficulty hypothesis, b) allows for attentiveness to vary on the screen-by-respondent level, c) allows for respondents with different trait and speed levels to differ in their attentiveness, and d) at once deals with various response patterns arising from C/IER. The approach makes use of item-level RTs. An adapted version for aggregated RTs is presented that supports screening for C/IER behavior on the respondent level. Parameter recovery is investigated in a simulation study. The approach is illustrated in an empirical example, comparing different RT measures and contrasting the proposed model-based procedure against indicator-based multiple-hurdle approaches.
Collapse
Affiliation(s)
- Esther Ulitzsch
- IPN-Leibniz Institute for Science and Mathematics Education, Olshausenstraße 62, 24118, Kiel, Germany.
| | | | | | - Ulf Kroehne
- DIPF-Leibniz Institute for Research and Information in Education, Frankfurt, Germany
| | | |
Collapse
|
7
|
DeCarlo LT. On Joining a Signal Detection Choice Model with Response Time Models. JOURNAL OF EDUCATIONAL MEASUREMENT 2021. [DOI: 10.1111/jedm.12300] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
Affiliation(s)
- Lawrence T. DeCarlo
- Professor of Psychology and Education, Teachers College, Columbia University
| |
Collapse
|
8
|
Domingue BW, Kanopka K, Stenhaug B, Soland J, Kuhfeld M, Wise S, Piech C. Variation in Respondent Speed and its Implications: Evidence from an Adaptive Testing Scenario. JOURNAL OF EDUCATIONAL MEASUREMENT 2021. [DOI: 10.1111/jedm.12291] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Affiliation(s)
| | | | | | | | | | | | - Chris Piech
- Department of Computer Science Stanford University
| |
Collapse
|
9
|
A Practical Cross-Sectional Framework to Contextual Reactivity in Personality: Response Times as Indicators of Reactivity to Contextual Cues. PSYCH 2020. [DOI: 10.3390/psych2040019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
Contextual reactivity refers to the degree in which personality states are affected by contextual cues. Research into contextual reactivity has mainly focused on repeated measurement designs. In this paper, we propose a cross-sectional approach to study contextual reactivity. We argue that contextual reactivity can be operationalized as different response processes which are characterized by different mean response times and different measurement properties. We propose a within-person mixture modeling approach that adopts this idea and which enables studying contextual reactivity in cross-sectional data. We applied the model to data from the Revised Temperament and Character Inventory. Results indicate that we can distinguish between two response specific latent states. We interpret these states as a high contextual reactive state and a low contextual reactive state. From the results it appears that the low contextual reactive state is generally associated with smaller response times and larger discrimination parameters, as compared to the high contextual reactivity state. The utility of this approach in personality research is discussed.
Collapse
|
10
|
Ulitzsch E, von Davier M, Pohl S. A hierarchical latent response model for inferences about examinee engagement in terms of guessing and item-level non-response. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2020; 73 Suppl 1:83-112. [PMID: 31709521 DOI: 10.1111/bmsp.12188] [Citation(s) in RCA: 24] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/29/2019] [Revised: 06/21/2019] [Indexed: 06/10/2023]
Abstract
In low-stakes assessments, test performance has few or no consequences for examinees themselves, so that examinees may not be fully engaged when answering the items. Instead of engaging in solution behaviour, disengaged examinees might randomly guess or generate no response at all. When ignored, examinee disengagement poses a severe threat to the validity of results obtained from low-stakes assessments. Statistical modelling approaches in educational measurement have been proposed that account for non-response or for guessing, but do not consider both types of disengaged behaviour simultaneously. We bring together research on modelling examinee engagement and research on missing values and present a hierarchical latent response model for identifying and modelling the processes associated with examinee disengagement jointly with the processes associated with engaged responses. To that end, we employ a mixture model that identifies disengagement at the item-by-examinee level by assuming different data-generating processes underlying item responses and omissions, respectively, as well as response times associated with engaged and disengaged behaviour. By modelling examinee engagement with a latent response framework, the model allows assessing how examinee engagement relates to ability and speed as well as to identify items that are likely to evoke disengaged test-taking behaviour. An illustration of the model by means of an application to real data is presented.
Collapse
Affiliation(s)
- Esther Ulitzsch
- Methods and Evaluation/Quality Assurance, Freie Universität Berlin, Germany
| | | | - Steffi Pohl
- Methods and Evaluation/Quality Assurance, Freie Universität Berlin, Germany
| |
Collapse
|
11
|
Liu Y, Cheng Y, Liu H. Identifying Effortful Individuals With Mixture Modeling Response Accuracy and Response Time Simultaneously to Improve Item Parameter Estimation. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2020; 80:775-807. [PMID: 32616958 PMCID: PMC7307491 DOI: 10.1177/0013164419895068] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
The responses of non-effortful test-takers may have serious consequences as non-effortful responses can impair model calibration and latent trait inferences. This article introduces a mixture model, using both response accuracy and response time information, to help differentiating non-effortful and effortful individuals, and to improve item parameter estimation based on the effortful group. Two mixture approaches are compared with the traditional response time mixture model (TMM) method and the normative threshold 10 (NT10) method with response behavior effort criteria in four simulation scenarios with regard to item parameter recovery and classification accuracy. The results demonstrate that the mixture methods and the TMM method can reduce the bias of item parameter estimates caused by non-effortful individuals, with the mixture methods showing more advantages when the non-effort severity is high or the response times are not lognormally distributed. An illustrative example is also provided.
Collapse
Affiliation(s)
- Yue Liu
- Beijing Normal University, Beijing,
China
| | - Ying Cheng
- University of Notre Dame, Notre Dame,
IN, USA
| | - Hongyun Liu
- Beijing Normal University, Beijing,
China
- Collaborative Innovation Center of
Assessment toward Basic Education Quality, Beijing Normal University, Beijing,
China
| |
Collapse
|
12
|
Ranger J, Wolgast A. Using Response Times as Collateral Information About Latent Traits in Psychological Tests. METHODOLOGY-EUROPEAN JOURNAL OF RESEARCH METHODS FOR THE BEHAVIORAL AND SOCIAL SCIENCES 2019. [DOI: 10.1027/1614-2241/a000181] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022] Open
Abstract
Abstract. In psychological tests, the time needed to respond to the items provides collateral information about the latent traits of the test takers. This, however, requires a measurement model that incorporates the response times in addition to the responses. Such a measurement model is usually based on a full specification of the response time distribution. In the present article, we suggest a novel modeling approach that requires fewer assumptions. In the approach, the responses are modeled with a unidimensional two-parameter logistic model. The single response times are summed to the scale-specific total testing time which is then related to the latent trait of the two-parameter logistic model via a smooth adaptive Gaussian mixture (SAGM) model. The approach can be justified against the background of the bivariate generalized linear item response theory modeling framework ( Molenaar, Tuerlinckx, & van der Maas, 2015a ). Its utility is investigated in two simulation studies and an empirical example.
Collapse
Affiliation(s)
- Jochen Ranger
- Institut für Psychologie, University of Halle-Wittenberg, Halle, Germany
| | - Anett Wolgast
- Institut für Pädagogik, University of Halle-Wittenberg, Halle, Germany
| |
Collapse
|
13
|
Abstract
Various mixture modeling approaches have been proposed to identify within-subjects differences in the psychological processes underlying responses to psychometric tests. Although valuable, the existing mixture models are associated with at least one of the following three challenges: (1) A parametric distribution is assumed for the response times that—if violated—may bias the results; (2) the response processes are assumed to result in equal variances (homoscedasticity) in the response times, whereas some processes may produce more variability than others (heteroscedasticity); and (3) the different response processes are modeled as independent latent variables, whereas they may be related. Although each of these challenges has been addressed separately, in practice they may occur simultaneously. Therefore, we propose a heteroscedastic hidden Markov mixture model for responses and categorized response times that addresses all the challenges above in a single model. In a simulation study, we demonstrated that the model is associated with acceptable parameter recovery and acceptable resolution to distinguish between various special cases. In addition, the model was applied to the responses and response times of the WAIS-IV block design subtest, to demonstrate its use in practice.
Collapse
|
14
|
Lee YH, Hao J, Man K, Ou L. How Do Test Takers Interact With Simulation-Based Tasks? A Response-Time Perspective. Front Psychol 2019; 10:906. [PMID: 31068876 PMCID: PMC6491860 DOI: 10.3389/fpsyg.2019.00906] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2018] [Accepted: 04/04/2019] [Indexed: 11/13/2022] Open
Abstract
Many traditional educational assessments use multiple-choice items and constructed-response items to measure fundamental skills. Virtual performance assessments, such as game- or simulation-based assessments, are designed recently in the field of educational measurement to measure more integrated skills through the test takers' interactive behaviors within an assessment in a virtual environment. This paper presents a systematic timing study based on data collected from a simulation-based task designed recently at Educational Testing Service. The study is intended to understand the response times in complex simulation-based tasks so as to shed light on possible ways of leveraging response time information in designing, assembling, and scoring of simulation-based tasks. To achieve this objective, a series of five analyses were conducted to first understand the statistical properties of the timing data, and then investigate the relationship between the timing patterns and the test takers' performance on the items/task, demographics, motivation level, personality, and test-taking behaviors through use of different statistical approaches. We found that the five analyses complemented each other and revealed different useful timing aspects of this test-taker sample's behavioral features in the simulation-based task. The findings were also compared with notable existing results in the literature related to timing data.
Collapse
Affiliation(s)
- Yi-Hsuan Lee
- Educational Testing Service, Princeton, NJ, United States
| | - Jiangang Hao
- Educational Testing Service, Princeton, NJ, United States
| | - Kaiwen Man
- Department of Human Development and Quantitative Methodology, Measurement, Statistics and Evaluation Program, University of Maryland at College Park, College Park, MD, United States
| | - Lu Ou
- ACT Inc., Iowa City, IA, United States
| |
Collapse
|
15
|
Wang C, Weiss DJ, Su S. Modeling Response Time and Responses in Multidimensional Health Measurement. Front Psychol 2019; 10:51. [PMID: 30761036 PMCID: PMC6361798 DOI: 10.3389/fpsyg.2019.00051] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2018] [Accepted: 01/09/2019] [Indexed: 12/31/2022] Open
Abstract
This study explored calibrating a large item bank for use in multidimensional health measurement with computerized adaptive testing, using both item responses and response time (RT) information. The Activity Measure for Post-Acute Care is a patient-reported outcomes measure comprised of three correlated scales (Applied Cognition, Daily Activities, and Mobility). All items from each scale are Likert type, so that a respondent chooses a response from an ordered set of four response options. The most appropriate item response theory model for analyzing and scoring these items is the multidimensional graded response model (MGRM). During the field testing of the items, an interviewer read each item to a patient and recorded, on a tablet computer, the patient's responses and the software recorded RTs. Due to the large item bank with over 300 items, data collection was conducted in four batches with a common set of anchor items to link the scale. van der Linden's (2007) hierarchical modeling framework was adopted. Several models, with or without interviewer as a covariate and with or without interaction between interviewer and items, were compared for each batch of data. It was found that the model with the interaction between interviewer and item, when the interaction effect was constrained to be proportional, fit the data best. Therefore, the final hierarchical model with a lognormal model for RT and the MGRM for response data was fitted to all batches of data via a concurrent calibration. Evaluation of parameter estimates revealed that (1) adding response time information did not affect the item parameter estimates and their standard errors significantly; (2) adding response time information helped reduce the standard error of patients' multidimensional latent trait estimates, but adding interviewer as a covariate did not result in further improvement. Implications of the findings for follow up adaptive test delivery design are discussed.
Collapse
Affiliation(s)
- Chun Wang
- College of Education, University of Washington, Seattle, WA, United States
| | - David J Weiss
- Department of Psychology, University of Minnesota, St. Paul, MN, United States
| | - Shiyang Su
- Department of Psychology, University of Central Florida, Orlando, FL, United States
| |
Collapse
|
16
|
De Boeck P, Jeon M. An Overview of Models for Response Times and Processes in Cognitive Tests. Front Psychol 2019; 10:102. [PMID: 30787891 PMCID: PMC6372526 DOI: 10.3389/fpsyg.2019.00102] [Citation(s) in RCA: 75] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2018] [Accepted: 01/14/2019] [Indexed: 11/13/2022] Open
Abstract
Response times (RTs) are a natural kind of data to investigate cognitive processes underlying cognitive test performance. We give an overview of modeling approaches and of findings obtained with these approaches. Four types of models are discussed: response time models (RT as the sole dependent variable), joint models (RT together with other variables as dependent variable), local dependency models (with remaining dependencies between RT and accuracy), and response time as covariate models (RT as independent variable). The evidence from these approaches is often not very informative about the specific kind of processes (other than problem solving, information accumulation, and rapid guessing), but the findings do suggest dual processing: automated processing (e.g., knowledge retrieval) vs. controlled processing (e.g., sequential reasoning steps), and alternative explanations for the same results exist. While it seems well-possible to differentiate rapid guessing from normal problem solving (which can be based on automated or controlled processing), further decompositions of response times are rarely made, although possible based on some of model approaches.
Collapse
Affiliation(s)
- Paul De Boeck
- Department of Psychology, Ohio State University, Columbus, OH, United States
- KU Leuven, Leuven, Belgium
| | - Minjeong Jeon
- Graduate School of Education and Information Studies, University of California, Los Angeles, Los Angeles, CA, United States
| |
Collapse
|
17
|
Ranger J, Wolgast A, Kuhn JT. Robust estimation of the hierarchical model for responses and response times. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:83-107. [PMID: 30051905 DOI: 10.1111/bmsp.12143] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/17/2017] [Revised: 06/04/2018] [Indexed: 06/08/2023]
Abstract
Van der Linden's (2007, Psychometrika, 72, 287) hierarchical model for responses and response times in tests has numerous applications in psychological assessment. The success of these applications requires the parameters of the model to have been estimated without bias. The data used for model fitting, however, are often contaminated, for example, by rapid guesses or lapses of attention. This distorts the parameter estimates. In the present paper, a novel estimation approach is proposed that is robust against contamination. The approach consists of two steps. In the first step, the response time model is fitted on the basis of a robust estimate of the covariance matrix. In the second step, the item response model is extended to a mixture model, which allows for a proportion of irregular responses in the data. The parameters of the mixture model are then estimated with a modified marginal maximum likelihood estimator. The modified marginal maximum likelihood estimator downweights responses of test-takers with unusual response time patterns. As a result, the estimator is resistant to several forms of data contamination. The robustness of the approach is investigated in a simulation study. An application of the estimator is demonstrated with real data.
Collapse
Affiliation(s)
- Jochen Ranger
- Department of Psychology, Martin-Luther-University Halle-Wittenberg, Germany
| | - Anett Wolgast
- Department of Educational Science, Martin-Luther-University Halle-Wittenberg, Germany
| | - Jörg-Tobias Kuhn
- Department of Rehabilitation Science, University of Dortmund, Germany
| |
Collapse
|