1
|
Lyu W, Bolt D. A Psychometric Perspective on the Associations between Response Accuracy and Response Time Residuals. J Intell 2024; 12:74. [PMID: 39195121 DOI: 10.3390/jintelligence12080074] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2024] [Revised: 07/27/2024] [Accepted: 07/29/2024] [Indexed: 08/29/2024] Open
Abstract
We provide an alternative psychometric perspective on the empirical statistical dependencies observed between response accuracy residuals (RARs) and response time residuals (RTRs) in the context of the van der Linden model. This perspective emphasizes the RAR (or parts of the RAR) as being exogenous and having a directional influence on response time. Our simple and theoretically justifiable perspective adds to previous joint response time/accuracy models and comports with recent generalizations of the D-diffusion IRT model incorporating person-by-item interactions, and thus similarly reproduces many of the recently highlighted empirical findings concerning the associations between RARs and RTRs. Using both empirical and simulation-based results, we show how our psychometric perspective has both applied and interpretational implications. Specifically, it would suggest that (1) studies of item parameter estimate heterogeneity in relation to response times may reflect more of a psychometric artifact (due to the exogenous effects of the RARs) as opposed to providing insights about the response process (e.g., the application of different response strategies) and that (2) efforts to use RTRs as indicators of latent proficiency should attend to the anticipated interactions between the latent proficiency and RAR on response times. The validity of our psychometric perspective against alternatives likely relies on appeals to theory; the best perspective to take may vary depending on the test setting.
Collapse
Affiliation(s)
- Weicong Lyu
- College of Education, University of Washington, Seattle, WA 98105, USA
| | - Daniel Bolt
- Department of Educational Psychology, University of Wisconsin, Madison, WI 53706, USA
| |
Collapse
|
2
|
Rodán A, Romero M, Casadevante C, Santacreu J, Montoro PR, Contreras MJ. Getting it right takes time: response time and performance in secondary school students. THE JOURNAL OF GENERAL PSYCHOLOGY 2024; 151:357-373. [PMID: 37906102 DOI: 10.1080/00221309.2023.2275304] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Accepted: 10/20/2023] [Indexed: 11/02/2023]
Abstract
The relation between response time and performance in cognitive tasks is increasingly evident. In the present study, we analyzed the effect of participants' spontaneous speed when responding to a mental rotation task. We carried out a data reanalysis from a previous study where a training of 3 practice sessions of 100 trials each was applied. The procedure was applied to a sample of 21 high school students (11 boys, 10 girls). The relation between response time and performance (hits) across the training trials was analyzed. In addition, we carried out a regression analysis of performance on the learning task as a function of response time on that same task, as well as with the score on two previously applied tests of spatial intelligence and fluid intelligence. Results showed, (a) a significant relationship (r = 0.624) between response time and hits, (b) that the group of participants with longer response times performed better; (c) that participants' response time explained most of the variance of their score on the training task in the regression analysis, although spatial and fluid intelligence scores improved the prediction of performance. Our results suggest that the reflective style achieves greater performance in solving spatial tasks, which could have important practical implications to promote a slower and more reflective style when solving school tasks with spatial components.
Collapse
|
3
|
Kang I, Jeon M. A Recent Development of a Network Approach to Assessment Data: Latent Space Item Response Modeling for Intelligence Studies. J Intell 2024; 12:38. [PMID: 38667705 PMCID: PMC11050824 DOI: 10.3390/jintelligence12040038] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2024] [Revised: 03/17/2024] [Accepted: 03/25/2024] [Indexed: 04/28/2024] Open
Abstract
This article aims to provide an overview of the potential advantages and utilities of the recently proposed Latent Space Item Response Model (LSIRM) in the context of intelligence studies. The LSIRM integrates the traditional Rasch IRT model for psychometric data with the latent space model for network data. The model has person-wise latent abilities and item difficulty parameters, capturing the main person and item effects, akin to the Rasch model. However, it additionally assumes that persons and items can be mapped onto the same metric space called a latent space and distances between persons and items represent further decreases in response accuracy uncaptured by the main model parameters. In this way, the model can account for conditional dependence or interactions between persons and items unexplained by the Rasch model. With two empirical datasets, we illustrate that (1) the latent space can provide information on respondents and items that cannot be captured by the Rasch model, (2) the LSIRM can quantify and visualize potential between-person variations in item difficulty, (3) latent dimensions/clusters of persons and items can be detected or extracted based on their latent positions on the map, and (4) personalized feedback can be generated from person-item distances. We conclude with discussions related to the latent space modeling integrated with other psychometric models and potential future directions.
Collapse
Affiliation(s)
- Inhan Kang
- Department of Psychology, College of Liberal Arts, Yonsei University, Seoul 03722, Republic of Korea;
| | - Minjeong Jeon
- Social Research Methodology, Department of Education, School of Education and Information Studies, University of California, Los Angeles, CA 90095, USA
| |
Collapse
|
4
|
Kang I, Jeon M, Partchev I. A Latent Space Diffusion Item Response Theory Model to Explore Conditional Dependence between Responses and Response Times. PSYCHOMETRIKA 2023; 88:830-864. [PMID: 37316615 DOI: 10.1007/s11336-023-09920-x] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/26/2022] [Indexed: 06/16/2023]
Abstract
Traditional measurement models assume that all item responses correlate with each other only through their underlying latent variables. This conditional independence assumption has been extended in joint models of responses and response times (RTs), implying that an item has the same item characteristics fors all respondents regardless of levels of latent ability/trait and speed. However, previous studies have shown that this assumption is violated in various types of tests and questionnaires and there are substantial interactions between respondents and items that cannot be captured by person- and item-effect parameters in psychometric models with the conditional independence assumption. To study the existence and potential cognitive sources of conditional dependence and utilize it to extract diagnostic information for respondents and items, we propose a diffusion item response theory model integrated with the latent space of variations in information processing rate of within-individual measurement processes. Respondents and items are mapped onto the latent space, and their distances represent conditional dependence and unexplained interactions. We provide three empirical applications to illustrate (1) how to use an estimated latent space to inform conditional dependence and its relation to person and item measures, (2) how to derive diagnostic feedback personalized for respondents, and (3) how to validate estimated results with an external measure. We also provide a simulation study to support that the proposed approach can accurately recover its parameters and detect conditional dependence underlying data.
Collapse
Affiliation(s)
- Inhan Kang
- Yonsei University, 403 Widang Hall, 50 Yonsei-ro, Seodaemun-gu, Seoul, 03722, Republic of Korea.
| | - Minjeong Jeon
- UNIVERSITY OF CALIFORNIA, LOS ANGELES, Los Angeles, USA
| | | |
Collapse
|
5
|
Krämer RJ, Koch M, Levacher J, Schmitz F. Testing Replicability and Generalizability of the Time on Task Effect. J Intell 2023; 11:jintelligence11050082. [PMID: 37233332 DOI: 10.3390/jintelligence11050082] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2022] [Revised: 04/24/2023] [Accepted: 04/25/2023] [Indexed: 05/27/2023] Open
Abstract
The time on task (ToT) effect describes the relationship of the time spent on a cognitive task and the probability of successful task completion. The effect has been shown to vary in size and direction across tests and even within tests, depending on the test taker and item characteristics. Specifically, investing more time has a positive effect on response accuracy for difficult items and low ability test-takers, but a negative effect for easy items and high ability test-takers. The present study sought to test the replicability of this result pattern of the ToT effect across samples independently drawn from the same populations of persons and items. Furthermore, its generalizability was tested in terms of differential correlations across ability tests. To this end, ToT effects were estimated for three different reasoning tests and one test measuring natural sciences knowledge in 10 comparable subsamples with a total N = 2640. Results for the subsamples were highly similar, demonstrating that ToT effects are estimated with sufficient reliability. Generally, faster answers tended to be more accurate, suggesting a relatively effortless processing style. However, with increasing item difficulty and decreasing person ability, the effect flipped to the opposite direction, i.e., higher accuracy with longer processing times. The within-task moderation of the ToT effect can be reconciled with an account on effortful processing or cognitive load. By contrast, the generalizability of the ToT effect across different tests was only moderate. Cross-test relations were stronger in relative terms if performance in the respective tasks was more strongly related. This suggests that individual differences in the ToT effect depend on test characteristics such as their reliabilities but also similarities and differences of their processing requirements.
Collapse
Affiliation(s)
- Raimund J Krämer
- Department of Psychology, University of Duisburg-Essen, Universitätsstraße 2, 45141 Essen, Germany
| | - Marco Koch
- Individual Differences & Psychodiagnostics, Saarland University, Campus A1.3, 66123 Saarbrücken, Germany
| | - Julie Levacher
- Individual Differences & Psychodiagnostics, Saarland University, Campus A1.3, 66123 Saarbrücken, Germany
| | - Florian Schmitz
- Department of Psychology, University of Duisburg-Essen, Universitätsstraße 2, 45141 Essen, Germany
| |
Collapse
|
6
|
Man K, Harring JR, Zhan P. Bridging Models of Biometric and Psychometric Assessment: A Three-Way Joint Modeling Approach of Item Responses, Response Times, and Gaze Fixation Counts. APPLIED PSYCHOLOGICAL MEASUREMENT 2022; 46:361-381. [PMID: 35812811 PMCID: PMC9265489 DOI: 10.1177/01466216221089344] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
Recently, joint models of item response data and response times have been proposed to better assess and understand test takers' learning processes. This article demonstrates how biometric information such as gaze fixation counts obtained from an eye-tracking machine can be integrated into the measurement model. The proposed joint modeling framework accommodates the relations among a test taker's latent ability, working speed and test engagement level via a person-side variance-covariance structure, while simultaneously permitting the modeling of item difficulty, time-intensity, and the engagement intensity through an item-side variance-covariance structure. A Bayesian estimation scheme is used to fit the proposed model to data. Posterior predictive model checking based on three discrepancy measures corresponding to various model components are introduced to assess model-data fit. Findings from a Monte Carlo simulation and results from analyzing experimental data demonstrate the utility of the model.
Collapse
Affiliation(s)
- Kaiwen Man
- University of Alabama, Tuscaloosa, AL, USA
- Kaiwen Man, Educational Research Program, Educational Studies in Psychology, Research Methodology, and Counseling, 313 Carmichael Box 870231, University of Alabama, Tuscaloosa, AL 35487, USA.
| | | | - Peida Zhan
- Zhejiang Normal University, Jinhua, China
| |
Collapse
|
7
|
Kang I, De Boeck P, Ratcliff R. Modeling Conditional Dependence of Response Accuracy and Response Time with the Diffusion Item Response Theory Model. PSYCHOMETRIKA 2022; 87:725-748. [PMID: 34988775 PMCID: PMC9677523 DOI: 10.1007/s11336-021-09819-5] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/30/2021] [Revised: 09/05/2021] [Indexed: 05/26/2023]
Abstract
In this paper, we propose a model-based method to study conditional dependence between response accuracy and response time (RT) with the diffusion IRT model (Tuerlinckx and De Boeck in Psychometrika 70(4):629-650, 2005, https://doi.org/10.1007/s11336-000-0810-3 ; van der Maas et al. in Psychol Rev 118(2):339-356, 2011, https://doi.org/10.1080/20445911.2011.454498 ). We extend the earlier diffusion IRT model by introducing variability across persons and items in cognitive capacity (drift rate in the evidence accumulation process) and variability in the starting point of the decision processes. We show that the extended model can explain the behavioral patterns of conditional dependency found in the previous studies in psychometrics. Variability in cognitive capacity can predict positive and negative conditional dependency and their interaction with the item difficulty. Variability in starting point can account for the early changes in the response accuracy as a function of RT given the person and item effects. By the combination of the two variability components, the extended model can produce the curvilinear conditional accuracy functions that have been observed in psychometric data. We also provide a simulation study to validate the parameter recovery of the proposed model and present two empirical applications to show how to implement the model to study conditional dependency underlying data response accuracy and RTs.
Collapse
Affiliation(s)
- Inhan Kang
- The Ohio State University, 291 Psychology Building, 1835 Neil Avenue, Columbus, OH, 43210, USA.
| | - Paul De Boeck
- The Ohio State University, 291 Psychology Building, 1835 Neil Avenue, Columbus, OH, 43210, USA
| | - Roger Ratcliff
- The Ohio State University, 291 Psychology Building, 1835 Neil Avenue, Columbus, OH, 43210, USA
| |
Collapse
|
8
|
Kang I, De Boeck P, Partchev I. A randomness perspective on intelligence processes. INTELLIGENCE 2022. [DOI: 10.1016/j.intell.2022.101632] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
|
9
|
Jeon M, De Boeck P, Luo J, Li X, Lu ZL. Modeling Within-Item Dependencies in Parallel Data on Test Responses and Brain Activation. PSYCHOMETRIKA 2021; 86:239-271. [PMID: 33486707 DOI: 10.1007/s11336-020-09741-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/10/2018] [Revised: 11/26/2020] [Accepted: 12/04/2020] [Indexed: 06/12/2023]
Abstract
In this paper, we propose a joint modeling approach to analyze dependency in parallel response data. We define two types of dependency: higher-level dependency and within-item conditional dependency. While higher-level dependency can be estimated with common latent variable modeling approaches, within-item conditional dependency is a unique kind of information that is often not captured with extant methods, despite its potential to shed new insights into the relationship between the two types of response data. We differentiate three ways of modeling within-item conditional dependency by conditioning on raw values, expected values, or residual values of the response data, which have different implications in terms of response processes. The proposed approach is illustrated with the example of analyzing parallel data on response accuracy and brain activations from a Theory of Mind assessment. The consequence of ignoring within-item conditional dependency is investigated with empirical and simulation studies in comparison to conventional dependency analysis that focuses exclusively on relationships between latent variables.
Collapse
Affiliation(s)
- Minjeong Jeon
- Department of Education, University of California, Los Angeles, 3141 Moore Hall, 457 Portola Avenue, Los Angeles, CA, 90024, USA.
| | - Paul De Boeck
- Ohio State University, 225 Psychology Building 1835 Neil Avenue, Columbus, OH, 43210, USA
| | - Jevan Luo
- Department of Education, University of California, Los Angeles, 3141 Moore Hall, 457 Portola Avenue, Los Angeles, CA, 90024, USA
| | - Xiangrui Li
- Ohio State University, 225 Psychology Building 1835 Neil Avenue, Columbus, OH, 43210, USA
| | - Zhong-Lin Lu
- Department of Psychology, New York University, 6 Washington Pl, New York, NY, 10003, USA
| |
Collapse
|
10
|
Wang S, Chen Y. Using Response Times and Response Accuracy to Measure Fluency Within Cognitive Diagnosis Models. PSYCHOMETRIKA 2020; 85:600-629. [PMID: 32816238 DOI: 10.1007/s11336-020-09717-2] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/09/2019] [Indexed: 06/11/2023]
Abstract
The recent "Every Student Succeed Act" encourages schools to use an innovative assessment to provide feedback about students' mastery level of grade-level content standards. Mastery of a skill requires the ability to complete the task with not only accuracy but also fluency. This paper offers a new sight on using both response times and response accuracy to measure fluency with cognitive diagnosis model framework. Defining fluency as the highest level of a categorical latent attribute, a polytomous response accuracy model and two forms of response time models are proposed to infer fluency jointly. A Bayesian estimation approach is developed to calibrate the newly proposed models. These models were applied to analyze data collected from a spatial rotation test. Results demonstrate that compared with the traditional CDM that using response accuracy only, the proposed joint models were able to reveal more information regarding test takers' spatial skills. A set of simulation studies were conducted to evaluate the accuracy of model estimation algorithm and illustrate the various degrees of model complexities.
Collapse
|
11
|
Does planning help for execution? The complex relationship between planning and execution. PLoS One 2020; 15:e0237568. [PMID: 32797063 PMCID: PMC7428192 DOI: 10.1371/journal.pone.0237568] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2019] [Accepted: 07/29/2020] [Indexed: 11/21/2022] Open
Abstract
Planning and execution are two important parts of the problem-solving process. Based on related research, it is expected that planning speed and execution speed are positively correlated because of underlying individual differences in general mental speed. While there could also be a direct negative dependency of execution time on planning time, given the hypothesis that an investment in planning contributes to more efficient execution. The positive correlation and negative dependency are not contradictory since the former is a relationship across individuals (at the latent variable level) and the latter is a relationship within individuals (at the manifest variable level) after controlling for across-individual relationships. With two linear mixed model analyses and a factor model analysis, these two different kinds of relationships were examined using dependency analysis. The results supported the above hypotheses. The correlation between the latent variables of planning and execution was found to be positive and the dependency of execution time on planning time was found to be negative in all analyses. Moreover, the negative dependency varied among items and to some extent among persons as well. In summary, this study provides a clearer picture of the relationship between planning and execution and suggests that analyses at different levels may reveal different relationships.
Collapse
|
12
|
Jeon M, De Boeck P, Li X, Lu ZL. Trivariate Theory of Mind Data Analysis with a Conditional Joint Modeling Approach. PSYCHOMETRIKA 2020; 85:398-436. [PMID: 32623558 DOI: 10.1007/s11336-020-09710-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/15/2018] [Revised: 05/24/2020] [Indexed: 06/11/2023]
Abstract
Theory of mind (ToM) is an essential social-cognitive ability to understand one's own and other people's mental states. Neural data as well as behavior data have been utilized in ToM research, but the two types of data have rarely been analyzed together, creating a large gap in the literature. In this paper, we propose and apply a novel joint modeling approach to analyze brain activations with two types of behavioral data, response times and response accuracy, obtained from a multi-item ToM assessment, with the intention to shed new light on the nature of the underlying process of ToM reasoning. Our trivariate data analysis suggested that different levels or kinds of processes might be involved during the ToM assessment, which seem to differ in terms of cognitive efficiency and sensitivity to ToM items and the correctness of item responses. Additional details on the trivariate data analysis results are provided with discussions on their implications for ToM research.
Collapse
Affiliation(s)
- Minjeong Jeon
- Department of Education, University of California, Los Angeles, 3141 Moore Hall, 457 Portola Avenue, Los Angeles, CA, 90024, USA.
| | | | | | | |
Collapse
|
13
|
Man K, Harring JR, Jiao H, Zhan P. Joint Modeling of Compensatory Multidimensional Item Responses and Response Times. APPLIED PSYCHOLOGICAL MEASUREMENT 2019; 43:639-654. [PMID: 31551641 PMCID: PMC6745633 DOI: 10.1177/0146621618824853] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/23/2023]
Abstract
Computer-based testing (CBT) is becoming increasingly popular in assessing test-takers' latent abilities and making inferences regarding their cognitive processes. In addition to collecting item responses, an important benefit of using CBT is that response times (RTs) can also be recorded and used in subsequent analyses. To better understand the structural relations between multidimensional cognitive attributes and the working speed of test-takers, this research proposes a joint-modeling approach that integrates compensatory multidimensional latent traits and response speediness using item responses and RTs. The joint model is cast as a multilevel model in which the structural relation between working speed and accuracy are connected through their variance-covariance structures. The feasibility of this modeling approach is investigated via a Monte Carlo simulation study using a Bayesian estimation scheme. The results indicate that integrating RTs increased model parameter recovery and precision. In addition, Program of International Student Assessment (PISA) 2015 mathematics standard unit items are analyzed to further evaluate the feasibility of the approach to recover model parameters.
Collapse
Affiliation(s)
- Kaiwen Man
- University of Maryland, College Park,
USA
- Authors share the first authorship
| | - Jeffrey R. Harring
- University of Maryland, College Park,
USA
- Authors share the first authorship
| | - Hong Jiao
- University of Maryland, College Park,
USA
- Authors share the first authorship
| | - Peida Zhan
- Zhejiang Normal University, Jinhua,
China
| |
Collapse
|
14
|
Engelhardt L, Goldhammer F. Validating Test Score Interpretations Using Time Information. Front Psychol 2019; 10:1131. [PMID: 31205462 PMCID: PMC6552849 DOI: 10.3389/fpsyg.2019.01131] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2018] [Accepted: 04/29/2019] [Indexed: 11/14/2022] Open
Abstract
A validity approach is proposed that uses processing times to collect validity evidence for the construct interpretation of test scores. The rationale of the approach is based on current research of processing times and on classical validity approaches, providing validity evidence based on relationships with other variables. Within the new approach, convergent validity evidence is obtained if a component skill, that is expected to underlie the task solution process in the target construct, positively moderates the relationship between effective speed and effective ability in the corresponding target construct. Discriminant validity evidence is provided if a component skill, that is not expected to underlie the task solution process in the target construct, does indeed not moderate the speed-ability relation in this target construct. Using data from a study that follows up the German PIAAC sample, this approach was applied to reading competence, assessed with PIAAC literacy items, and to quantitative reasoning, assessed with Number Series. As expected from theory, the effect of speed on ability in the target construct was only moderated by the respective underlying component skill, that is, word meaning activation skill as an underlying component skill of reading competence, and perceptual speed as an underlying component skill of reasoning. Accordingly, no positive interactions were found for the component skill that should not underlie the task solution process, that is, word meaning activation for reasoning and perceptual speed for reading. Furthermore, the study shows the suitability of the proposed validation approach. The use of time information in association with task results brings construct validation closer to the actual response process than widely used correlations of test scores.
Collapse
Affiliation(s)
- Lena Engelhardt
- DIPF – Leibniz Institute for Research and Information in Education, Frankfurt, Germany
| | - Frank Goldhammer
- DIPF – Leibniz Institute for Research and Information in Education, Frankfurt, Germany
- Centre for International Student Assessment (ZIB), Frankfurt, Germany
| |
Collapse
|
15
|
Park JY, Cornillie F, van der Maas HLJ, Van Den Noortgate W. A Multidimensional IRT Approach for Dynamically Monitoring Ability Growth in Computerized Practice Environments. Front Psychol 2019; 10:620. [PMID: 30984068 PMCID: PMC6450197 DOI: 10.3389/fpsyg.2019.00620] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2018] [Accepted: 03/06/2019] [Indexed: 11/13/2022] Open
Abstract
Adaptive learning systems have received an increasing attention as they enable to provide personalized instructions tailored to the behaviors and needs of individual learners. In order to reach this goal, it is desired to have an assessment system, monitoring each learner's ability change in real time. The Elo Rating System (ERS), a popular scoring algorithm for paired competitions, has recently been considered as a fast and flexible method that can assess learning progress in online learning environments. However, it has been argued that a standard ERS may be problematic due to the multidimensional nature of the abilities embedded in learning materials. In order to handle this issue, we propose a system that incorporates a multidimensional item response theory model (MIRT) in the ERS. The basic idea is that instead of updating a single ability parameter from the Rasch model, our method allows a simultaneous update of multiple ability parameters based on a compensatory MIRT model, resulting in a multidimensional extension of the ERS ("M-ERS"). To evaluate the approach, three simulation studies were conducted. Results suggest that the ERS that incorrectly assumes unidimensionality has a seriously lower prediction accuracy compared to the M-ERS. Accounting for both speed and accuracy in M-ERS is shown to perform better than using accuracy data only. An application further illustrates the method using real-life data from a popular educational platform for exercising math skills.
Collapse
Affiliation(s)
- Jung Yeon Park
- Faculty of Psychology and Educational Sciences, imec–ITEC, KU Leuven, Leuven, Belgium
| | - Frederik Cornillie
- Faculty of Psychology and Educational Sciences, imec–ITEC, KU Leuven, Leuven, Belgium
| | | | - Wim Van Den Noortgate
- Faculty of Psychology and Educational Sciences, imec–ITEC, KU Leuven, Leuven, Belgium
| |
Collapse
|
16
|
De Boeck P, Jeon M. An Overview of Models for Response Times and Processes in Cognitive Tests. Front Psychol 2019; 10:102. [PMID: 30787891 PMCID: PMC6372526 DOI: 10.3389/fpsyg.2019.00102] [Citation(s) in RCA: 75] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2018] [Accepted: 01/14/2019] [Indexed: 11/13/2022] Open
Abstract
Response times (RTs) are a natural kind of data to investigate cognitive processes underlying cognitive test performance. We give an overview of modeling approaches and of findings obtained with these approaches. Four types of models are discussed: response time models (RT as the sole dependent variable), joint models (RT together with other variables as dependent variable), local dependency models (with remaining dependencies between RT and accuracy), and response time as covariate models (RT as independent variable). The evidence from these approaches is often not very informative about the specific kind of processes (other than problem solving, information accumulation, and rapid guessing), but the findings do suggest dual processing: automated processing (e.g., knowledge retrieval) vs. controlled processing (e.g., sequential reasoning steps), and alternative explanations for the same results exist. While it seems well-possible to differentiate rapid guessing from normal problem solving (which can be based on automated or controlled processing), further decompositions of response times are rarely made, although possible based on some of model approaches.
Collapse
Affiliation(s)
- Paul De Boeck
- Department of Psychology, Ohio State University, Columbus, OH, United States
- KU Leuven, Leuven, Belgium
| | - Minjeong Jeon
- Graduate School of Education and Information Studies, University of California, Los Angeles, Los Angeles, CA, United States
| |
Collapse
|
17
|
Chen H, De Boeck P, Grady M, Yang CL, Waldschmidt D. Curvilinear dependency of response accuracy on response time in cognitive tests. INTELLIGENCE 2018. [DOI: 10.1016/j.intell.2018.04.001] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
18
|
Molenaar D, de Boeck P. Response Mixture Modeling: Accounting for Heterogeneity in Item Characteristics across Response Times. PSYCHOMETRIKA 2018; 83:279-297. [PMID: 29392567 DOI: 10.1007/s11336-017-9602-9] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/16/2015] [Indexed: 05/26/2023]
Abstract
In item response theory modeling of responses and response times, it is commonly assumed that the item responses have the same characteristics across the response times. However, heterogeneity might arise in the data if subjects resort to different response processes when solving the test items. These differences may be within-subject effects, that is, a subject might use a certain process on some of the items and a different process with different item characteristics on the other items. If the probability of using one process over the other process depends on the subject's response time, within-subject heterogeneity of the item characteristics across the response times arises. In this paper, the method of response mixture modeling is presented to account for such heterogeneity. Contrary to traditional mixture modeling where the full response vectors are classified, response mixture modeling involves classification of the individual elements in the response vector. In a simulation study, the response mixture model is shown to be viable in terms of parameter recovery. In addition, the response mixture model is applied to a real dataset to illustrate its use in investigating within-subject heterogeneity in the item characteristics across response times.
Collapse
Affiliation(s)
- Dylan Molenaar
- Psychological Methods, Department of Psychology, University of Amsterdam, PO box 15906, 1001 NK , Amsterdam, The Netherlands.
| | | |
Collapse
|
19
|
|
20
|
van Rijn PW, Ali US. A Generalized Speed-Accuracy Response Model for Dichotomous Items. PSYCHOMETRIKA 2018; 83:109-131. [PMID: 29164449 DOI: 10.1007/s11336-017-9590-9] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/15/2015] [Indexed: 06/07/2023]
Abstract
We propose a generalization of the speed-accuracy response model (SARM) introduced by Maris and van der Maas (Psychometrika 77:615-633, 2012). In these models, the scores that result from a scoring rule that incorporates both the speed and accuracy of item responses are modeled. Our generalization is similar to that of the one-parameter logistic (or Rasch) model to the two-parameter logistic (or Birnbaum) model in item response theory. An expectation-maximization (EM) algorithm for estimating model parameters and standard errors was developed. Furthermore, methods to assess model fit are provided in the form of generalized residuals for item score functions and saddlepoint approximations to the density of the sum score. The presented methods were evaluated in a small simulation study, the results of which indicated good parameter recovery and reasonable type I error rates for the residuals. Finally, the methods were applied to two real data sets. It was found that the two-parameter SARM showed improved fit compared to the one-parameter SARM in both data sets.
Collapse
Affiliation(s)
| | - Usama S Ali
- Educational Testing Service, Princeton, NJ, USA
- South Valley University, Qena, Egypt
| |
Collapse
|
21
|
van Rijn PW, Ali US. A comparison of item response models for accuracy and speed of item responses with applications to adaptive testing. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2017; 70:317-345. [PMID: 28474769 DOI: 10.1111/bmsp.12101] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/18/2016] [Revised: 03/28/2017] [Indexed: 05/07/2023]
Abstract
We compare three modelling frameworks for accuracy and speed of item responses in the context of adaptive testing. The first framework is based on modelling scores that result from a scoring rule that incorporates both accuracy and speed. The second framework is the hierarchical modelling approach developed by van der Linden (2007, Psychometrika, 72, 287) in which a regular item response model is specified for accuracy and a log-normal model for speed. The third framework is the diffusion framework in which the response is assumed to be the result of a Wiener process. Although the three frameworks differ in the relation between accuracy and speed, one commonality is that the marginal model for accuracy can be simplified to the two-parameter logistic model. We discuss both conditional and marginal estimation of model parameters. Models from all three frameworks were fitted to data from a mathematics and spelling test. Furthermore, we applied a linear and adaptive testing mode to the data off-line in order to determine differences between modelling frameworks. It was found that a model from the scoring rule framework outperformed a hierarchical model in terms of model-based reliability, but the results were mixed with respect to correlations with external measures.
Collapse
Affiliation(s)
| | - Usama S Ali
- Educational Testing Service, Princeton, New Jersey, USA
| |
Collapse
|