1
|
Merhof V, Meiser T. Co-occurring dominance and ideal point processes: A general IRTree framework for multidimensional item responding. Behav Res Methods 2024; 56:7005-7025. [PMID: 38627325 PMCID: PMC11362233 DOI: 10.3758/s13428-024-02405-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 03/13/2024] [Indexed: 08/30/2024]
Abstract
Responding to rating scale items is a multidimensional process, since not only the substantive trait being measured but also additional personal characteristics can affect the respondents' category choices. A flexible model class for analyzing such multidimensional responses are IRTree models, in which rating responses are decomposed into a sequence of sub-decisions. Different response processes can be involved in item responding both sequentially across those sub-decisions and as co-occurring processes within sub-decisions. In the previous literature, modeling co-occurring processes has been exclusively limited to dominance models, where higher trait levels are associated with higher expected scores. However, some response processes may rather follow an ideal point rationale, where the expected score depends on the proximity of a person's trait level to the item's location. Therefore, we propose a new multidimensional IRT model of co-occurring dominance and ideal point processes (DI-MIRT model) as a flexible framework for parameterizing IRTree sub-decisions with multiple dominance processes, multiple ideal point processes, and combinations of both. The DI-MIRT parameterization opens up new application areas for the IRTree model class and allows the specification of a wide range of theoretical assumptions regarding the cognitive processing of item responding. A simulation study shows that IRTree models with DI-MIRT parameterization provide excellent parameter recovery and accurately reflect co-occurring dominance and ideal point processes. In addition, a clear advantage over traditional IRTree models with purely sequential processes is demonstrated. Two application examples from the field of response style analysis highlight the benefits of the general IRTree framework under real-world conditions.
Collapse
Affiliation(s)
- Viola Merhof
- Department of Psychology, University of Mannheim, L 13 15, D-68161, Mannheim, Germany.
| | - Thorsten Meiser
- Department of Psychology, University of Mannheim, L 13 15, D-68161, Mannheim, Germany
| |
Collapse
|
2
|
Chevignard M, Câmara-Costa H, Dellatolas G. Predicting and improving outcome in severe pediatric traumatic brain injury. Expert Rev Neurother 2024; 24:963-983. [PMID: 39140714 DOI: 10.1080/14737175.2024.2389921] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2024] [Accepted: 08/05/2024] [Indexed: 08/15/2024]
Abstract
INTRODUCTION Severe pediatric traumatic brain injury (spTBI), including abusive head trauma (AHT) in young children, is a major public health problem. Long-term consequences of spTBI include a large variety of physical, neurological, biological, cognitive, behavioral and social deficits and impairments. AREAS COVERED The present narrative review summarizes studies and reviews published from January 2019 to February 2024 on spTBI. Significant papers published before 2019 were also included. The article gives coverage to the causes of spTBI, its epidemiology and fatality rates; disparities, inequalities, and socioeconomic factors; critical care; outcomes; and interventions. EXPERT OPINION There are disparities between countries and according to socio-economic factors regarding causes, treatments and outcomes of spTBI. AHT has an overall poor outcome. Adherence to critical care guidelines is imperfect and the evidence-base of guidelines needs further investigations. Neuroimaging and biomarker predictors of outcomes is a rapidly evolving domain. Long-term cognitive, behavioral and psychosocial difficulties are the most prevalent and disabling. Their investigation should make a clear distinction between objective (clinical examination, cognitive tests, facts) and subjective measures (estimations using patient- and proxy-reported questionnaires), considering possible common source bias in reported difficulties. Family/caregiver-focused interventions, ecological approaches, and use of technology in delivery of interventions are recommended to improve long-term difficulties after spTBI.
Collapse
Affiliation(s)
- Mathilde Chevignard
- Rehabilitation Department for Children with Acquired Neurological Injury, Saint Maurice Hospitals, Saint Maurice, France
- Sorbonne Université, CNRS, INSERM, Laboratoire d'Imagerie Biomédicale (LIB), Paris, France
- Sorbonne Université, GRC 24 Handicap Moteur Cognitif et Réadaptation (HaMCRe), AP-HP, Paris, France
| | - Hugo Câmara-Costa
- Rehabilitation Department for Children with Acquired Neurological Injury, Saint Maurice Hospitals, Saint Maurice, France
- Sorbonne Université, CNRS, INSERM, Laboratoire d'Imagerie Biomédicale (LIB), Paris, France
- Sorbonne Université, GRC 24 Handicap Moteur Cognitif et Réadaptation (HaMCRe), AP-HP, Paris, France
| | - Georges Dellatolas
- Sorbonne Université, GRC 24 Handicap Moteur Cognitif et Réadaptation (HaMCRe), AP-HP, Paris, France
| |
Collapse
|
3
|
Merhof V, Böhm CM, Meiser T. Separation of Traits and Extreme Response Style in IRTree Models: The Role of Mimicry Effects for the Meaningful Interpretation of Estimates. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2024; 84:927-956. [PMID: 39318484 PMCID: PMC11418598 DOI: 10.1177/00131644231213319] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/26/2024]
Abstract
Item response tree (IRTree) models are a flexible framework to control self-reported trait measurements for response styles. To this end, IRTree models decompose the responses to rating items into sub-decisions, which are assumed to be made on the basis of either the trait being measured or a response style, whereby the effects of such person parameters can be separated from each other. Here we investigate conditions under which the substantive meanings of estimated extreme response style parameters are potentially invalid and do not correspond to the meanings attributed to them, that is, content-unrelated category preferences. Rather, the response style factor may mimic the trait and capture part of the trait-induced variance in item responding, thus impairing the meaningful separation of the person parameters. Such a mimicry effect is manifested in a biased estimation of the covariance of response style and trait, as well as in an overestimation of the response style variance. Both can lead to severely misleading conclusions drawn from IRTree analyses. A series of simulation studies reveals that mimicry effects depend on the distribution of observed responses and that the estimation biases are stronger the more asymmetrically the responses are distributed across the rating scale. It is further demonstrated that extending the commonly used IRTree model with unidimensional sub-decisions by multidimensional parameterizations counteracts mimicry effects and facilitates the meaningful separation of parameters. An empirical example of the Program for International Student Assessment (PISA) background questionnaire illustrates the threat of mimicry effects in real data. The implications of applying IRTree models for empirical research questions are discussed.
Collapse
Affiliation(s)
| | - Caroline M. Böhm
- Rhineland-Palatinate Technical University of Kaiserslautern-Landau, Germany
| | | |
Collapse
|
4
|
Alagöz ÖEC, Meiser T. Investigating Heterogeneity in Response Strategies: A Mixture Multidimensional IRTree Approach. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2024; 84:957-993. [PMID: 39318480 PMCID: PMC11418595 DOI: 10.1177/00131644231206765] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/26/2024]
Abstract
To improve the validity of self-report measures, researchers should control for response style (RS) effects, which can be achieved with IRTree models. A traditional IRTree model considers a response as a combination of distinct decision-making processes, where the substantive trait affects the decision on response direction, while decisions about choosing the middle category or extreme categories are largely determined by midpoint RS (MRS) and extreme RS (ERS). One limitation of traditional IRTree models is the assumption that all respondents utilize the same set of RS in their response strategies, whereas it can be assumed that the nature and the strength of RS effects can differ between individuals. To address this limitation, we propose a mixture multidimensional IRTree (MM-IRTree) model that detects heterogeneity in response strategies. The MM-IRTree model comprises four latent classes of respondents, each associated with a different set of RS traits in addition to the substantive trait. More specifically, the class-specific response strategies involve (1) only ERS in the "ERS only" class, (2) only MRS in the "MRS only" class, (3) both ERS and MRS in the "2RS" class, and (4) neither ERS nor MRS in the "0RS" class. In a simulation study, we showed that the MM-IRTree model performed well in recovering model parameters and class memberships, whereas the traditional IRTree approach showed poor performance if the population includes a mixture of response strategies. In an application to empirical data, the MM-IRTree model revealed distinct classes with noticeable class sizes, suggesting that respondents indeed utilize different response strategies.
Collapse
|
5
|
Hasselhorn K, Ottenstein C, Meiser T, Lischetzke T. The Effects of Questionnaire Length on the Relative Impact of Response Styles in Ambulatory Assessment. MULTIVARIATE BEHAVIORAL RESEARCH 2024; 59:1043-1057. [PMID: 38779850 DOI: 10.1080/00273171.2024.2354233] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2024]
Abstract
Ambulatory assessment (AA) is becoming an increasingly popular research method in the fields of psychology and life science. Nevertheless, knowledge about the effects that design choices, such as questionnaire length (i.e., number of items per questionnaire), have on AA data quality is still surprisingly restricted. Additionally, response styles (RS), which threaten data quality, have hardly been analyzed in the context of AA. The aim of the current research was to experimentally manipulate questionnaire length and investigate the association between questionnaire length and RS in an AA study. We expected that the group with the longer (82-item) questionnaire would show greater reliance on RS relative to the substantive traits than the group with the shorter (33-item) questionnaire. Students (n = 284) received questionnaires three times a day for 14 days. We used a multigroup two-dimensional item response tree model in a multilevel structural equation modeling framework to estimate midpoint and extreme RS in our AA study. We found that the long questionnaire group showed a greater reliance on RS relative to trait-based processes than the short questionnaire group. Although further validation of our findings is necessary, we hope that researchers consider our findings when planning an AA study in the future.
Collapse
Affiliation(s)
- Kilian Hasselhorn
- Department of Psychology, RPTU Kaiserslautern-Landau, Landau, Germany
| | | | | | - Tanja Lischetzke
- Department of Psychology, RPTU Kaiserslautern-Landau, Landau, Germany
| |
Collapse
|
6
|
Câmara-Costa H, Dellatolas G, Jourdan C, Ruet A, Bayen E, Vallat-Azouvi C, Allain P, Chevignard M, Azouvi P. The 20-item dysexecutive questionnaire after severe traumatic brain injury: Distribution of the total score and its significance. Neuropsychol Rehabil 2024:1-22. [PMID: 39106184 DOI: 10.1080/09602011.2024.2387065] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2024] [Accepted: 07/26/2024] [Indexed: 08/09/2024]
Abstract
TRIAL REGISTRATION ClinicalTrials.gov identifier: NCT01437683..
Collapse
Affiliation(s)
- Hugo Câmara-Costa
- Rehabilitation Department for Children with Acquired Brain Injury; Saint Maurice Hospitals, Saint Maurice, France
- Sorbonne Université, CNRS, INSERM, Laboratoire d'Imagerie Biomédicale (LIB), F-75006, Paris, France
- Sorbonne Université, GRC 24 Handicap Moteur Cognitif et Réadaptation (HaMCRe), AP-HP, F-75013, Paris, France
| | - Georges Dellatolas
- Sorbonne Université, GRC 24 Handicap Moteur Cognitif et Réadaptation (HaMCRe), AP-HP, F-75013, Paris, France
| | - Claire Jourdan
- Physical Medicine and Rehabilitation Department, Lapeyronie Hospital, CHRU, Montpellier, France
| | - Alexis Ruet
- Physical Medicine and Rehabilitation Department, CHU Caen, INSERM U1077, Caen, France
| | - Eléonore Bayen
- Sorbonne Université, CNRS, INSERM, Laboratoire d'Imagerie Biomédicale (LIB), F-75006, Paris, France
- Sorbonne Université, GRC 24 Handicap Moteur Cognitif et Réadaptation (HaMCRe), AP-HP, F-75013, Paris, France
- Assistance Publique-Hôpitaux de Paris (APHP), Groupe Hospitalier Pitié-Salpêtrière, Service De Médecine Physique et Réadaptation, Paris, France
| | | | - Philippe Allain
- Laboratoire de Psychologie des Pays de la Loire, LPPL EA 4638, SFR Confluences, UNIV Angers, Nantes Université, Maison de la recherche Germaine Tillion, Angers, France
- Département de Neurologie, Centre Hospitalier Universitaire d'Angers, Angers, France
| | - Mathilde Chevignard
- Rehabilitation Department for Children with Acquired Brain Injury; Saint Maurice Hospitals, Saint Maurice, France
- Sorbonne Université, CNRS, INSERM, Laboratoire d'Imagerie Biomédicale (LIB), F-75006, Paris, France
- Sorbonne Université, GRC 24 Handicap Moteur Cognitif et Réadaptation (HaMCRe), AP-HP, F-75013, Paris, France
| | - Philippe Azouvi
- AP-HP, GH Paris Saclay, Hôpital Raymond Poincaré, Service de Médecine Physique et de Réadaptation, Garches, France
- Université Paris-Saclay, UVSQ, Inserm, CESP, 94807, Villejuif, France
| |
Collapse
|
7
|
He S, Kern JL. Investigating Directional Invariance in an Item Response Tree Model for Extreme Response Style and Trait-Based Unfolding Responses. APPLIED PSYCHOLOGICAL MEASUREMENT 2024; 48:187-207. [PMID: 39055537 PMCID: PMC11268250 DOI: 10.1177/01466216241261705] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/27/2024]
Abstract
Item response tree (IRTree) approaches have received increasing attention in the response style literature due to their capability to partial out response style latent traits from content-related latent traits by considering separate decisions for agreement and level of agreement. Additionally, it has shown that the functioning of the intensity of agreement decision may depend upon the agreement decision with an item, so that the item parameters and person parameters may differ by direction of agreement; when the parameters across direction are the same, this is called directional invariance. Furthermore, for non-cognitive psychological constructs, it has been argued that the response process may be best described as following an unfolding process. In this study, a family of IRTree models to handle unfolding responses with the agreement decision following the hyperbolic cosine model and the intensity of agreement decision following a graded response model is investigated. This model family also allows for investigation of item- and person-level directional invariance. A simulation study is conducted to evaluate parameter recovery; model parameters are estimated with a fully Bayesian approach using JAGS (Just Another Gibbs Sampler). The proposed modeling scheme is demonstrated with two data examples with multiple model comparisons allowing for varying levels of directional invariance and unfolding versus dominance processes. An approach to visualizing the final model item response functioning is also developed. The article closes with a short discussion about the results.
Collapse
Affiliation(s)
- Siqi He
- University of Illinois at Urbana-Champaign, IL, USA
| | | |
Collapse
|
8
|
Li M, Zhang B, Mou Y. Though Forced, Still Valid: Examining the Psychometric Performance of Forced-Choice Measurement of Personality in Children and Adolescents. Assessment 2024:10731911241255841. [PMID: 38867477 DOI: 10.1177/10731911241255841] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/14/2024]
Abstract
Unveiling the roles personality plays during childhood and adolescence necessitates its accurate measurement, commonly using traditional Likert-type (LK) scales. However, this format is susceptible to various response biases, which can be particularly prevalent in children and adolescents, thus likely undermining measurement accuracy. Forced-choice (FC) scales appear to be a promising alternative because they are largely free from these biases by design. However, some argue that the FC format may not perform satisfactorily in children and adolescents due to its complexity. Little empirical evidence exists regarding the suitability of the FC format for children and adolescents. As such, the current study examined the psychometric performance of an FC measure of the Big Five personality factors in three children and adolescent samples: 5th to 6th graders (N = 428), 7th to 8th graders (N = 449), and 10th to 11th graders (N = 555). Across the three age groups, the FC scale demonstrated a better fit to the Big Five model and better discriminant validity in comparison to the LK counterpart. Personality scores from the FC scale also converged well with those from the LK scale and demonstrated high reliability as well as sizable criterion-related validity. Furthermore, the FC scale had more invariant statements than its LK counterpart across age groups. Overall, we found good evidence showing that FC measurement of personality is suitable for children and adolescents.
Collapse
Affiliation(s)
- Mengtong Li
- University of Illinois Urbana-Champaign, IL, USA
| | - Bo Zhang
- University of Illinois Urbana-Champaign, IL, USA
| | - Yi Mou
- Sun Yat-Sen University, Guangzhou, China
| |
Collapse
|
9
|
Ulitzsch E, Henninger M, Meiser T. Differences in response-scale usage are ubiquitous in cross-country comparisons and a potential driver of elusive relationships. Sci Rep 2024; 14:10890. [PMID: 38740767 PMCID: PMC11091185 DOI: 10.1038/s41598-024-60465-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2023] [Accepted: 04/22/2024] [Indexed: 05/16/2024] Open
Affiliation(s)
- Esther Ulitzsch
- IPN-Leibniz Institute for Science and Mathematics Education, Educational Measurement, Olshausenstraße 62, 24118, Kiel, Germany.
- University of Mannheim, Mannheim, Germany.
| | | | | |
Collapse
|
10
|
Schoenmakers M, Tijmstra J, Vermunt J, Bolsinova M. Correcting for Extreme Response Style: Model Choice Matters. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2024; 84:145-170. [PMID: 38250509 PMCID: PMC10795569 DOI: 10.1177/00131644231155838] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 01/23/2024]
Abstract
Extreme response style (ERS), the tendency of participants to select extreme item categories regardless of the item content, has frequently been found to decrease the validity of Likert-type questionnaire results. For this reason, various item response theory (IRT) models have been proposed to model ERS and correct for it. Comparisons of these models are however rare in the literature, especially in the context of cross-cultural comparisons, where ERS is even more relevant due to cultural differences between groups. To remedy this issue, the current article examines two frequently used IRT models that can be estimated using standard software: a multidimensional nominal response model (MNRM) and a IRTree model. Studying conceptual differences between these models reveals that they differ substantially in their conceptualization of ERS. These differences result in different category probabilities between the models. To evaluate the impact of these differences in a multigroup context, a simulation study is conducted. Our results show that when the groups differ in their average ERS, the IRTree model and MNRM can drastically differ in their conclusions about the size and presence of differences in the substantive trait between these groups. An empirical example is given and implications for the future use of both models and the conceptualization of ERS are discussed.
Collapse
|
11
|
Merhof V, Meiser T. Dynamic Response Strategies: Accounting for Response Process Heterogeneity in IRTree Decision Nodes. PSYCHOMETRIKA 2023; 88:1354-1380. [PMID: 36746887 PMCID: PMC10656330 DOI: 10.1007/s11336-023-09901-0] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/15/2022] [Indexed: 06/18/2023]
Abstract
It is essential to control self-reported trait measurements for response style effects to ensure a valid interpretation of estimates. Traditional psychometric models facilitating such control consider item responses as the result of two kinds of response processes-based on the substantive trait, or based on response styles-and they assume that both of these processes have a constant influence across the items of a questionnaire. However, this homogeneity over items is not always given, for instance, if the respondents' motivation declines throughout the questionnaire so that heuristic responding driven by response styles may gradually take over from cognitively effortful trait-based responding. The present study proposes two dynamic IRTree models, which account for systematic continuous changes and additional random fluctuations of response strategies, by defining item position-dependent trait and response style effects. Simulation analyses demonstrate that the proposed models accurately capture dynamic trajectories of response processes, as well as reliably detect the absence of dynamics, that is, identify constant response strategies. The continuous version of the dynamic model formalizes the underlying response strategies in a parsimonious way and is highly suitable as a cognitive model for investigating response strategy changes over items. The extended model with random fluctuations of strategies can adapt more closely to the item-specific effects of different response processes and thus is a well-fitting model with high flexibility. By using an empirical data set, the benefits of the proposed dynamic approaches over traditional IRTree models are illustrated under realistic conditions.
Collapse
Affiliation(s)
- Viola Merhof
- Department of Psychology, University of Mannheim, L 13 15, 68161, Mannheim, Germany.
| | - Thorsten Meiser
- Department of Psychology, University of Mannheim, L 13 15, 68161, Mannheim, Germany
| |
Collapse
|
12
|
Quirk VL, Kern JL. Using IRTree Models to Promote Selection Validity in the Presence of Extreme Response Styles. J Intell 2023; 11:216. [PMID: 37998715 PMCID: PMC10672242 DOI: 10.3390/jintelligence11110216] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2023] [Revised: 10/16/2023] [Accepted: 11/14/2023] [Indexed: 11/25/2023] Open
Abstract
The measurement of psychological constructs is frequently based on self-report tests, which often have Likert-type items rated from "Strongly Disagree" to "Strongly Agree". Recently, a family of item response theory (IRT) models called IRTree models have emerged that can parse out content traits (e.g., personality traits) from noise traits (e.g., response styles). In this study, we compare the selection validity and adverse impact consequences of noise traits on selection when scores are estimated using a generalized partial credit model (GPCM) or an IRTree model. First, we present a simulation which demonstrates that when noise traits do exist, the selection decisions made based on the IRTree model estimated scores have higher accuracy rates and have less instances of adverse impact based on extreme response style group membership when compared to the GPCM. Both models performed similarly when there was no influence of noise traits on the responses. Second, we present an application using data collected from the Open-Source Psychometrics Project Fisher Temperament Inventory dataset. We found that the IRTree model had a better fit, but a high agreement rate between the model decisions resulted in virtually identical impact ratios between the models. We offer considerations for applications of the IRTree model and future directions for research.
Collapse
|
13
|
Jeon M. Commentary: Explore Conditional Dependencies in Item Response Tree Data. PSYCHOMETRIKA 2023; 88:803-808. [PMID: 37106310 DOI: 10.1007/s11336-023-09915-8] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/19/2023] [Indexed: 06/19/2023]
Abstract
Item response tree (IRTree) models are widely used in various applications for their ability to differentiate sets of sub-responses from polytomous item response data based on a pre-specified tree structure. Lyu et al. (Psychometrika) article highlighted that item slopes are often lower for later nodes than earlier nodes in IRTree applications. Lyu et al. argued that this phenomenon might signal the presence of item-specific factors across nodes. In this commentary, I present a different perspective that conditional dependencies in IRTree data could explain the phenomenon more generally. I illustrate my point with an empirical example, utilizing the latent space item response model that visualizes conditional dependencies in IRTree data. I conclude the commentary with a discussion on the potential of exploring conditional dependencies in IRTree data that goes beyond identifying the sources of conditional dependencies.
Collapse
Affiliation(s)
- Minjeong Jeon
- University of California, Los Angeles, Los Angeles, USA.
| |
Collapse
|
14
|
Lyu W, Bolt DM, Westby S. Exploring the Effects of Item-Specific Factors in Sequential and IRTree Models. PSYCHOMETRIKA 2023; 88:745-775. [PMID: 37326911 DOI: 10.1007/s11336-023-09912-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/06/2022] [Indexed: 06/17/2023]
Abstract
Test items for which the item score reflects a sequential or IRTree modeling outcome are considered. For such items, we argue that item-specific factors, although not empirically measurable, will often be present across stages of the same item. In this paper, we present a conceptual model that incorporates such factors. We use the model to demonstrate how the varying conditional distributions of item-specific factors across stages become absorbed into the stage-specific item discrimination and difficulty parameters, creating ambiguity in the interpretations of item and person parameters beyond the first stage. We discuss implications in relation to various applications considered in the literature, including methodological studies of (1) repeated attempt items; (2) answer change/review, (3) on-demand item hints; (4) item skipping behavior; and (5) Likert scale items. Our own empirical applications, as well as several examples published in the literature, show patterns of violations of item parameter invariance across stages that are highly suggestive of item-specific factors. For applications using sequential or IRTree models as analytical models, or for which the resulting item score might be viewed as outcomes of such a process, we recommend (1) regular inspection of data or analytic results for empirical evidence (or theoretical expectations) of item-specific factors; and (2) sensitivity analyses to evaluate the implications of item-specific factors for the intended inferences or applications.
Collapse
Affiliation(s)
- Weicong Lyu
- University of Wisconsin-Madison, 880 Educational Sciences, 1025 West Johnson Street, Madison, WI, 53706, USA.
| | - Daniel M Bolt
- University of Wisconsin-Madison, 1082 A Educational Sciencesz, 1025 West Johnson Street, Madison, WI, 53706, USA
| | - Samuel Westby
- Northeastern University, 177 Huntington Ave Desk 232O, Boston, MA, 02115, USA
| |
Collapse
|
15
|
Meiser T, Reiber F. Item-Specific Factors in IRTree Models: When They Matter and When They Don't. PSYCHOMETRIKA 2023; 88:739-744. [PMID: 37326912 PMCID: PMC10444655 DOI: 10.1007/s11336-023-09916-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/21/2023] [Indexed: 06/17/2023]
Abstract
Lyu et al. (Psychometrika, 2023) demonstrated that item-specific factors can cause spurious effects on the structural parameters of IRTree models for multiple nested response processes per item. Here, we discuss some boundary conditions and argue that person selection effects on item parameters are not unique to item-specific factors and that the effects presented by Lyu et al. (Psychometrika, 2023) may not generalize to the family of IRTree models as a whole. We conclude with the recommendation that IRTree model specification should be guided by theoretical considerations, rather than driven by data, in order to avoid misinterpretations of parameter differences.
Collapse
Affiliation(s)
- Thorsten Meiser
- Department of Psychology, School of Social Sciences, University of Mannheim, 68131, Mannheim, Germany.
| | - Fabiola Reiber
- Department of Psychology, School of Social Sciences, University of Mannheim, 68131, Mannheim, Germany
| |
Collapse
|
16
|
Somma A, Krueger RF, Markon KE, Gialdi G, Frau C, Fossati A. The joint hierarchical structure of psychopathology and dysfunctional personality domain indicators among community-dwelling adults. Personal Ment Health 2023; 17:3-19. [PMID: 35770737 DOI: 10.1002/pmh.1556] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/04/2021] [Revised: 05/31/2022] [Accepted: 06/10/2022] [Indexed: 11/10/2022]
Abstract
To examine the hierarchical structure of psychopathology and dysfunctional personality domains, 2416 Italian community-dwelling adult volunteers were administered a set of psychometrically sound psychopathology measures and the Personality Inventory for DSM-5 Brief Form+ (PID-5-BF+). Parallel analysis, minimum average partial, and very simple structure results suggested that 1-6 principal components (PCs) should be retained. Goldberg's bass-ackwards model of the joint psychopathology measure and PID-5-BF+ ipsatized domain scale correlation matrix evidenced a hierarchical structure that was consistent with the working model proposed by the Hierarchical Taxonomy of Psychopathology (HiTOP) consortium. Hierarchical agglomerative cluster analysis around latent variables of the psychopathology indicators and PID-5-BF+ domain scales recovered four latent dimensions, which were akin to the corresponding bass-ackwards components and nicely reproduced the HiTOP Internalizing, Externalizing, Thought Disorder, and Eating Pathology dimensions.
Collapse
Affiliation(s)
- Antonella Somma
- School of Psychology, Vita-Salute San Raffaele, Milan, Italy
| | - Robert F Krueger
- Department of Psychology, University of Minnesota, Minneapolis, MN, USA
| | - Kristian E Markon
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City, IA, USA
| | - Giulia Gialdi
- School of Psychology, Vita-Salute San Raffaele, Milan, Italy
| | | | - Andrea Fossati
- School of Psychology, Vita-Salute San Raffaele, Milan, Italy
| |
Collapse
|
17
|
Alarcon GM, Lee MA, Johnson D. A Monte Carlo study of IRTree models' ability to recover item parameters. Front Psychol 2023; 14:1003756. [PMID: 36949921 PMCID: PMC10025501 DOI: 10.3389/fpsyg.2023.1003756] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2022] [Accepted: 02/01/2023] [Indexed: 03/08/2023] Open
Abstract
Item response tree (IRTree) models are theorized to extract response styles from self-report data by utilizing multidimensional item response theory (IRT) models based on theoretical decision processes. Despite the growing popularity of the IRTree framework, there has been little research that has systematically examined the ability of its most popular models to recover item parameters across sample size and test length. This Monte Carlo simulation study explored the ability of IRTree models to recover item parameters based on data created from the midpoint primary process model. Results indicate the IRTree model can adequately recover item parameters early in the decision process model, specifically the midpoint node. However, as the model progresses through the decision hierarchy, item parameters have increased associated error variance. The authors ultimately recommend caution when employing the IRTree framework.
Collapse
Affiliation(s)
- Gene M. Alarcon
- Air Force Research Laboratory, Wright-Patterson Air Force Base, Dayton, Oh, United States
- *Correspondence: Gene M. Alarcon,
| | - Michael A. Lee
- General Dynamics Information Technology, Inc., Atlanta, GA, United States
| | - Dexter Johnson
- Air Force Research Laboratory, Wright-Patterson Air Force Base, Dayton, Oh, United States
| |
Collapse
|
18
|
Lee P, Joo S, Jia Z. Cross‐cultural differences in the use of the “
?
” Response category of the Job Descriptive Index: An application of the item response tree model. INTERNATIONAL JOURNAL OF SELECTION AND ASSESSMENT 2022. [DOI: 10.1111/ijsa.12414] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/31/2022]
Affiliation(s)
- Philseok Lee
- Department of Psychology George Mason University Fairfax Virginia USA
| | - Sean Joo
- Department of Educational Psychology University of Kansas Lawrence Kansas USA
| | - Zihao Jia
- Department of Psychology George Mason University Fairfax Virginia USA
| |
Collapse
|
19
|
Scharl A, Gnambs T. The Impact of Different Methods to Correct for Response Styles on the External Validity of Self-Reports. EUROPEAN JOURNAL OF PSYCHOLOGICAL ASSESSMENT 2022. [DOI: 10.1027/1015-5759/a000731] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
Abstract
Abstract. Response styles (RSs) such as acquiescence represent systematic respondent behaviors in self-report questionnaires beyond the actual item content. They distort trait estimates and contribute to measurement bias in questionnaire-based research. Although various approaches were proposed to correct the influence of RSs, little is known about their relative performance. Because different correction methods formalize the latent traits differently, it is unclear how model choice affects the external validity of the corrected measures. Therefore, the present study on N = 1,000 Dutch respondents investigated the impact of correcting responses to measures of self-esteem and the need for cognition using structural equation models with structured residuals, multidimensional generalized partial credit models, and multinomial processing trees. The study considered three RSs: extreme, midpoint, and acquiescence RS. The results showed homogeneous correlation patterns among the modeled latent and external variables, especially if they were not themselves subject to RSs. In that case, the IRT-based models, including an uncorrected model, still yielded consistent results. Nevertheless, the strength of the effect sizes showed variation.
Collapse
Affiliation(s)
- Anna Scharl
- Leibniz Institute for Educational Trajectories, Bamberg, Germany
| | - Timo Gnambs
- Leibniz Institute for Educational Trajectories, Bamberg, Germany
| |
Collapse
|
20
|
Thompson C, Byrne R, Adams J, Vidgen HA. Development, validation and item reduction of a food literacy questionnaire (IFLQ-19) with Australian adults. Int J Behav Nutr Phys Act 2022; 19:113. [PMID: 36050778 PMCID: PMC9438317 DOI: 10.1186/s12966-022-01351-8] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2022] [Accepted: 08/18/2022] [Indexed: 11/24/2022] Open
Abstract
Background Food literacy is theorised to improve diet quality, nutrition behaviours, social connectedness and food security. The definition and conceptualisation by Vidgen & Gallegos, consisting of 11 theoretical components within the four domains of planning and managing, selecting, preparing and eating, is currently the most highly cited framework. However, a valid and reliable questionnaire is needed to comprehensively measure this conceptualisation. Therefore, this study draws on existing item pools to develop a comprehensive food literacy questionnaire using item response theory. Methods Five hundred Australian adults were recruited in Study 1 to refine a food literacy item pool using principal component analysis (PCA) and item response theory (IRT) which involved detailed item analysis on targeting, responsiveness, validity and reliability. Another 500 participants were recruited in Study 2 to replicate item analysis on validity and reliability on the refined item pool, and 250 of these participants re-completed the food literacy questionnaire to determine its test–retest reliability. Results The PCA saw the 171-item pool reduced to 100-items across 19 statistical components of food literacy. After the thresholds of 26 items were combined, responses to the food literacy questionnaire had ordered thresholds (targeting), acceptable item locations (< -0.01 to + 1.53) and appropriateness of the measurement model (n = 92% expected responses) (responsiveness), met outfit mean-squares MSQ (0.48—1.42) (validity) and had high person, item separation (> 0.99) and test–retest (ICC 2,1 0.55–0.88) scores (reliability). Conclusions We developed a 100-item food literacy questionnaire, the IFLQ-19 to comprehensively address the Vidgen & Gallegos theoretical domains and components with good targeting, responsiveness, reliability and validity in a diverse sample of Australian adults. Supplementary Information The online version contains supplementary material available at 10.1186/s12966-022-01351-8.
Collapse
Affiliation(s)
- Courtney Thompson
- Queensland University of Technology (QUT), Faculty of Health, School of Exercise and Nutrition Sciences, Victoria Park Road, Kelvin Grove, QLD, 4059, Australia.
| | - Rebecca Byrne
- Queensland University of Technology (QUT), Faculty of Health, School of Exercise and Nutrition Sciences, Victoria Park Road, Kelvin Grove, QLD, 4059, Australia.,Queensland University of Technology (QUT), Faculty of Health, School of Exercise and Nutrition Sciences, Centre for Children's Health Research (CCHR), Graham Street, South Brisbane, QLD, 4101, Australia
| | - Jean Adams
- Centre for Diet and Activity Research (CEDAR), MRC Epidemiology Unit, University of Cambridge School of Clinical Medicine, Cambridge, UK
| | - Helen Anna Vidgen
- Queensland University of Technology (QUT), Faculty of Health, School of Exercise and Nutrition Sciences, Victoria Park Road, Kelvin Grove, QLD, 4059, Australia
| |
Collapse
|
21
|
Purpura A, Giorgianni D, Orrù G, Melis G, Sartori G. Identifying single-item faked responses in personality tests: A new TF-IDF-based method. PLoS One 2022; 17:e0272970. [PMID: 36007085 PMCID: PMC9410542 DOI: 10.1371/journal.pone.0272970] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2021] [Accepted: 07/28/2022] [Indexed: 11/19/2022] Open
Abstract
Faking in a psychological test is often observed whenever an examinee may gain an advantage from it. Although techniques are available to identify a faker, they cannot identify the specific questions distorted by faking. This work evaluates the effectiveness of term frequency-inverse document frequency (TF-IDF)—an information retrieval mathematical tool used in search engines and language representations—in identifying single-item faked responses. We validated the technique on three datasets containing responses to the 10-item Big Five questionnaire (total of 694 participants, respectively 221, 243, and 230) in three faking situations. Each participant responded twice, once faking to achieve an objective in one of three contexts (one to obtain child custody and two to land a job) and once honestly. The proposed TF-IDF model has proven very effective in separating honest from dishonest responses—with the honest ones having low TF-IDF values and the dishonest ones having higher values—and in identifying which of the 10 responses to the questionnaire were distorted in the dishonest condition. We also provide examples of the technique in a single-case evaluation.
Collapse
Affiliation(s)
| | - Dora Giorgianni
- Department of General Psychology, University of Padua, Padua, Italy
| | - Graziella Orrù
- Department of Surgical, Medical, Molecular and Critical Area Pathology, University of Pisa, Pisa, Italy
| | - Giulia Melis
- Department of General Psychology, University of Padua, Padua, Italy
| | - Giuseppe Sartori
- Department of General Psychology, University of Padua, Padua, Italy
- * E-mail:
| |
Collapse
|
22
|
Frick S. Modeling Faking in the Multidimensional Forced-Choice Format: The Faking Mixture Model. PSYCHOMETRIKA 2022; 87:773-794. [PMID: 34927219 PMCID: PMC9166892 DOI: 10.1007/s11336-021-09818-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/23/2020] [Revised: 08/31/2021] [Accepted: 10/15/2021] [Indexed: 06/14/2023]
Abstract
The multidimensional forced-choice (MFC) format has been proposed to reduce faking because items within blocks can be matched on desirability. However, the desirability of individual items might not transfer to the item blocks. The aim of this paper is to propose a mixture item response theory model for faking in the MFC format that allows to estimate the fakability of MFC blocks, termed the Faking Mixture model. Given current computing capabilities, within-subject data from both high- and low-stakes contexts are needed to estimate the model. A simulation showed good parameter recovery under various conditions. An empirical validation showed that matching was necessary but not sufficient to create an MFC questionnaire that can reduce faking. The Faking Mixture model can be used to reduce fakability during test construction.
Collapse
Affiliation(s)
- Susanne Frick
- Department of Psychology, School of Social Sciences, Mannheim, Germany.
| |
Collapse
|
23
|
Ulitzsch E, Pohl S, Khorramdel L, Kroehne U, von Davier M. A Response-Time-Based Latent Response Mixture Model for Identifying and Modeling Careless and Insufficient Effort Responding in Survey Data. PSYCHOMETRIKA 2022; 87:593-619. [PMID: 34855118 PMCID: PMC9166878 DOI: 10.1007/s11336-021-09817-7] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/25/2020] [Revised: 10/11/2021] [Indexed: 06/13/2023]
Abstract
Careless and insufficient effort responding (C/IER) can pose a major threat to data quality and, as such, to validity of inferences drawn from questionnaire data. A rich body of methods aiming at its detection has been developed. Most of these methods can detect only specific types of C/IER patterns. However, typically different types of C/IER patterns occur within one data set and need to be accounted for. We present a model-based approach for detecting manifold manifestations of C/IER at once. This is achieved by leveraging response time (RT) information available from computer-administered questionnaires and integrating theoretical considerations on C/IER with recent psychometric modeling approaches. The approach a) takes the specifics of attentive response behavior on questionnaires into account by incorporating the distance-difficulty hypothesis, b) allows for attentiveness to vary on the screen-by-respondent level, c) allows for respondents with different trait and speed levels to differ in their attentiveness, and d) at once deals with various response patterns arising from C/IER. The approach makes use of item-level RTs. An adapted version for aggregated RTs is presented that supports screening for C/IER behavior on the respondent level. Parameter recovery is investigated in a simulation study. The approach is illustrated in an empirical example, comparing different RT measures and contrasting the proposed model-based procedure against indicator-based multiple-hurdle approaches.
Collapse
Affiliation(s)
- Esther Ulitzsch
- IPN-Leibniz Institute for Science and Mathematics Education, Olshausenstraße 62, 24118, Kiel, Germany.
| | | | | | - Ulf Kroehne
- DIPF-Leibniz Institute for Research and Information in Education, Frankfurt, Germany
| | | |
Collapse
|
24
|
Bialo JA, Li H. Fairness and Comparability in Achievement Motivation Items: A Differential Item Functioning Analysis. JOURNAL OF PSYCHOEDUCATIONAL ASSESSMENT 2022. [DOI: 10.1177/07342829221090113] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Achievement motivation is a well-documented predictor of a variety of positive student outcomes. However, given observed group differences in motivation and related outcomes, motivation instruments should be checked for comparable item and scale functioning. Therefore, the purpose of this study was to evaluate measurement scale comparability and differential item functioning (DIF) in PISA 2015 achievement motivation items across gender and ethnicity using pairwise and multiple-group comparisons. In addition, DIF was investigated in relation to a common base group that reflected the sample average. Results indicated DIF between gender groups and between the base group and female students. For ethnicity, DIF was consistently flagged in pairwise comparisons with Black/African American students and Asian students as well as in base group comparisons. However, the identified DIF had little practical implications. Implications from these findings are discussed, and recommendations for future research are made.
Collapse
|
25
|
Self-report response style bias and borderline personality features. CURRENT PSYCHOLOGY 2022. [DOI: 10.1007/s12144-022-03122-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
26
|
Ames AJ. Measuring Response Style Stability Across Constructs With Item Response Trees. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2022; 82:281-306. [PMID: 35185160 PMCID: PMC8850762 DOI: 10.1177/00131644211020103] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Individual response style behaviors, unrelated to the latent trait of interest, may influence responses to ordinal survey items. Response style can introduce bias in the total score with respect to the trait of interest, threatening valid interpretation of scores. Despite claims of response style stability across scales, there has been little research into stability across multiple scales from the beneficial perspective of item response trees. This study examines an extension of the IRTree methodology to include mixed item formats, providing an empirical example of responses to three scales measuring perceptions of social media, climate change, and medical marijuana use. Results show extreme and midpoint response styles were not stable across scales within a single administration and 5-point Likert-type items elicited higher levels of extreme response style than the 4-point items. Latent trait of interest estimation varied, particularly at the lower end of the score distribution, across response style models, demonstrating as appropriate response style model is important for adequate trait estimation using Bayesian Markov chain Monte Carlo estimation.
Collapse
|
27
|
Chen HF, Jin KY. The Impact of Item Feature and Response Preference in a Mixed-Format Design. MULTIVARIATE BEHAVIORAL RESEARCH 2022; 57:208-222. [PMID: 33001710 DOI: 10.1080/00273171.2020.1820308] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
A combination of positively and negatively worded items (termed a mixed-format design) has been widely adopted in personality and attitude assessments. While advocates claim that the inclusion of positively and negatively worded items will encourage respondents to process the items more carefully and avoid response preference, others have reported that negatively worded (NW) items may induce a nuisance factor and contaminate scale scores. The present study examined the extent of the impact of the NW-item feature and further investigated whether a mixed-format design could effectively control acquiescence and the preference for extreme response options using two datasets (Attitude toward Peace Walls, and International Personality Item Pool). A proposed multidimensional item response model was implemented to simultaneously estimate the impact of item feature and response preference. The results suggested that NW items induced an impact on item responses and that affirmative preference was negligible, regardless of the proportion of NW items in a scale. However, participants' extremity preference was large in both balanced and imbalanced mixed-format designs. It concludes that the impact of the NW-item feature is not negligible in a mixed-format scale, which exhibits good control of acquiescence but not extremity preference.
Collapse
Affiliation(s)
- Hui-Fang Chen
- Department of Social and Behavioural Sciences, City University of Hong Kong
| | - Kuan-Yu Jin
- Faculty of Education, University of Hong Kong
| |
Collapse
|
28
|
Clarifying personality measurement in industrial-organizational psychology: The utility of item response tree models. PERSONALITY AND INDIVIDUAL DIFFERENCES 2022. [DOI: 10.1016/j.paid.2021.111410] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
29
|
Lyu W, Bolt DM. A psychometric model for respondent-level anchoring on self-report rating scale instruments. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2022; 75:116-135. [PMID: 34350978 DOI: 10.1111/bmsp.12251] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/11/2020] [Revised: 07/14/2021] [Indexed: 06/13/2023]
Abstract
Among the various forms of response bias that can emerge with self-report rating scale assessments are those related to anchoring, the tendency for respondents to select categories in close proximity to the rating category used for the immediately preceding item. In this study we propose a psychometric model based on a multidimensional nominal model for response style that also simultaneously accommodates a respondent-level anchoring tendency. The model is estimated using a fully Bayesian estimation procedure. By applying this model to a real test data set measuring extraversion, we explore a theory that both response styles and anchoring might be viewed as evidence of a lack of effortful responding. Empirical results show that there is a positive correlation between the strength of midpoint response style and the anchoring effect; further, responses indicative of either anchoring or response style both negatively correlate with response time, consistent with a theory that both phenomena reflect reduced respondent effort. The results support attending to both anchoring and midpoint response style as ways of assessing respondent engagement.
Collapse
Affiliation(s)
- Weicong Lyu
- Department of Educational Psychology, University of Wisconsin-Madison, Wisconsin, USA
| | - Daniel M Bolt
- Department of Educational Psychology, University of Wisconsin-Madison, Wisconsin, USA
| |
Collapse
|
30
|
Alarcon GM, Lee MA. The Relationship of Insufficient Effort Responding and Response Styles: An Online Experiment. Front Psychol 2022; 12:784375. [PMID: 35095672 PMCID: PMC8789874 DOI: 10.3389/fpsyg.2021.784375] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2021] [Accepted: 12/20/2021] [Indexed: 11/13/2022] Open
Abstract
While self-report data is a staple of modern psychological studies, they rely on participants accurately self-reporting. Two constructs that impede accurate results are insufficient effort responding (IER) and response styles. These constructs share conceptual underpinnings and both utilized to reduce cognitive effort when responding to self-report scales. Little research has extensively explored the relationship of the two constructs. The current study explored the relationship of the two constructs across even-point and odd-point scales, as well as before and after data cleaning procedures. We utilized IRTrees, a statistical method for modeling response styles, to examine the relationship between IER and response styles. To capture the wide range of IER metrics available, we employed several forms of IER assessment in our analyses and generated IER factors based on the type of IER being detected. Our results indicated an overall modest relationship between IER and response styles, which varied depending on the type of IER metric being considered or type of scale being evaluated. As expected, data cleaning also changed the relationships of some of the variables. We posit the difference between the constructs may be the degree of cognitive effort participants are willing to expend. Future research and applications are discussed.
Collapse
Affiliation(s)
| | - Michael A. Lee
- General Dynamics Information Technology, Inc., Atlanta, GA, United States
| |
Collapse
|
31
|
Ames AJ, Myers AJ. Explaining Variability in Response Style Traits: A Covariate-Adjusted IRTree. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2021; 81:756-780. [PMID: 34267399 PMCID: PMC8243201 DOI: 10.1177/0013164420969780] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Contamination of responses due to extreme and midpoint response style can confound the interpretation of scores, threatening the validity of inferences made from survey responses. This study incorporated person-level covariates in the multidimensional item response tree model to explain heterogeneity in response style. We include an empirical example and two simulation studies to support the use and interpretation of the model: parameter recovery using Markov chain Monte Carlo (MCMC) estimation and performance of the model under conditions with and without response styles present. Item intercepts mean bias and root mean square error were small at all sample sizes. Item discrimination mean bias and root mean square error were also small but tended to be smaller when covariates were unrelated to, or had a weak relationship with, the latent traits. Item and regression parameters are estimated with sufficient accuracy when sample sizes are greater than approximately 1,000 and MCMC estimation with the Gibbs sampler is used. The empirical example uses the National Longitudinal Study of Adolescent to Adult Health's sexual knowledge scale. Meaningful predictors associated with high levels of extreme response latent trait included being non-White, being male, and having high levels of parental support and relationships. Meaningful predictors associated with high levels of the midpoint response latent trait included having low levels of parental support and relationships. Item-level covariates indicate the response style pseudo-items were less easy to endorse for self-oriented items, whereas the trait of interest pseudo-items were easier to endorse for self-oriented items.
Collapse
|
32
|
Kim N, Bolt DM. A Mixture IRTree Model for Extreme Response Style: Accounting for Response Process Uncertainty. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2021; 81:131-154. [PMID: 33456065 PMCID: PMC7797955 DOI: 10.1177/0013164420913915] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
This paper presents a mixture item response tree (IRTree) model for extreme response style. Unlike traditional applications of single IRTree models, a mixture approach provides a way of representing the mixture of respondents following different underlying response processes (between individuals), as well as the uncertainty present at the individual level (within an individual). Simulation analyses reveal the potential of the mixture approach in identifying subgroups of respondents exhibiting response behavior reflective of different underlying response processes. Application to real data from the Students Like Learning Mathematics (SLM) scale of Trends in International Mathematics and Science Study (TIMSS) 2015 demonstrates the superior comparative fit of the mixture representation, as well as the consequences of applying the mixture on the estimation of content and response style traits. We argue that methodology applied to investigate response styles should attend to the inherent uncertainty of response style influence due to the likely influence of both response styles and the content trait on the selection of extreme response categories.
Collapse
Affiliation(s)
- Nana Kim
- University of Wisconsin–Madison, Madison, WI, USA
- Nana Kim, Department of Educational Psychology, University of Wisconsin–Madison, 1025 West Johnson Street, Madison, WI 53706, USA.
| | | |
Collapse
|
33
|
Lang JW, Tay L. The Science and Practice of Item Response Theory in Organizations. ANNUAL REVIEW OF ORGANIZATIONAL PSYCHOLOGY AND ORGANIZATIONAL BEHAVIOR 2021. [DOI: 10.1146/annurev-orgpsych-012420-061705] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Item response theory (IRT) is a modeling approach that links responses to test items with underlying latent constructs through formalized statistical models. This article focuses on how IRT can be used to advance science and practice in organizations. We describe established applications of IRT as a scale development tool and new applications of IRT as a research and theory testing tool that enables organizational researchers to improve their understanding of workers and organizations. We focus on IRT models and their application in four key research and practice areas: testing, questionnaire responding, construct validation, and measurement equivalence of scores. In so doing, we highlight how novel developments in IRT such as explanatory IRT, multidimensional IRT, random item models, and more complex models of response processes such as ideal point models and tree models can potentially advance existing science and practice in these areas. As a starting point for readers interested in learning IRT and applying recent developments in IRT in their research, we provide concrete examples with data and R code.
Collapse
Affiliation(s)
- Jonas W.B. Lang
- Department of Human Resource Management and Organizational Psychology, Ghent University, B-9000 Gent, Belgium
- Business School, University of Exeter, EX4 4PU Exeter, United Kingdom
| | - Louis Tay
- Department of Psychological Sciences, Purdue University, West Lafayette, Indiana 47907, USA
| |
Collapse
|
34
|
Tutz G, Schauberger G. Uncertainty in Latent Trait Models. APPLIED PSYCHOLOGICAL MEASUREMENT 2020; 44:447-464. [PMID: 32788816 PMCID: PMC7383692 DOI: 10.1177/0146621620920932] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
A model that extends the Rasch model and the Partial Credit Model to account for subject-specific uncertainty when responding to items is proposed. It is demonstrated that ignoring the subject-specific uncertainty may yield biased estimates of model parameters. In the extended version of the model, uncertainty and the underlying trait are linked to explanatory variables. The parameterization allows to identify subgroups that differ in uncertainty and the underlying trait. The modeling approach is illustrated using data on the confidence of citizens in public institutions.
Collapse
|
35
|
Tutz G. Hierarchical Models for the Analysis of Likert Scales in Regression and Item Response Analysis. Int Stat Rev 2020. [DOI: 10.1111/insr.12396] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Affiliation(s)
- Gerhard Tutz
- Department of Statistics Ludwig‐Maximilians‐Universität München Akademiestraße 1 München 80799 Germany
| |
Collapse
|
36
|
Plieninger H. Developing and Applying IR-Tree Models: Guidelines, Caveats, and an Extension to Multiple Groups. ORGANIZATIONAL RESEARCH METHODS 2020. [DOI: 10.1177/1094428120911096] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
IR-tree models assume that categorical item responses can best be explained by multiple response processes. In the present article, guidelines are provided for the development and interpretation of IR-tree models. In more detail, the relationship between a tree diagram, the model equations, and the analysis on the basis of pseudo-items is described. Moreover, it is shown that IR-tree models do not allow conclusions about the sequential order of the processes, and that mistakes in the model specification can have serious consequences. Furthermore, multiple-group IR-tree models are presented as a novel extension of IR-tree models to data from heterogeneous units. This allows, for example, to investigate differences across countries or organizations with respect to core parameters of the IR-tree model. Finally, an empirical example on organizational commitment and response styles is presented.
Collapse
Affiliation(s)
- Hansjörg Plieninger
- School of Social Sciences, Department of Psychology, University of Mannheim, Mannheim, Germany
| |
Collapse
|
37
|
A Novel Partial Credit Extension Using Varying Thresholds to Account for Response Tendencies. JOURNAL OF EDUCATIONAL MEASUREMENT 2020. [DOI: 10.1111/jedm.12268] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
38
|
Falk CF, Ju U. Estimation of Response Styles Using the Multidimensional Nominal Response Model: A Tutorial and Comparison With Sum Scores. Front Psychol 2020; 11:72. [PMID: 32116902 PMCID: PMC7017717 DOI: 10.3389/fpsyg.2020.00072] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2019] [Accepted: 01/10/2020] [Indexed: 11/16/2022] Open
Abstract
Recent years have seen a dramatic increase in item response models for measuring response styles on Likert-type items. These model-based approaches stand in contrast to traditional sum-score-based methods where researchers count the number of times that participants selected certain response options. The multidimensional nominal response model (MNRM) offers a flexible model-based approach that may be intuitive to those familiar with sum score approaches. This paper presents a tutorial on the model along with code for estimating it using three different software packages: flexMIRT®, mirt, and Mplus. We focus on specification and interpretation of response functions. In addition, we provide analytical details on how sum score to scale score conversion can be done with the MNRM. In the context of a real data example, three different scoring approaches are then compared. This example illustrates how sum-score-based approaches can sometimes yield scores that are confounded with substantive content. We expect that the current paper will facilitate further investigations as to whether different substantive conclusions are reached under alternative approaches to measuring response styles.
Collapse
Affiliation(s)
- Carl F Falk
- Department of Psychology, McGill University, Montreal, QC, Canada
| | - Unhee Ju
- Riverside Insights, Itasca, IL, United States
| |
Collapse
|
39
|
Zhang Y, Wang Y. Validity of Three IRT Models for Measuring and Controlling Extreme and Midpoint Response Styles. Front Psychol 2020; 11:271. [PMID: 32153477 PMCID: PMC7049783 DOI: 10.3389/fpsyg.2020.00271] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2019] [Accepted: 02/04/2020] [Indexed: 11/30/2022] Open
Abstract
Response styles, the general tendency to use certain categories of rating scales over others, are a threat to the reliability and validity of self-report measures. The mixed partial credit model, the multidimensional nominal response model, and the item response tree model are three widely used models for measuring extreme and midpoint response styles and correcting their effects. This research aimed to examine and compare their validity by fitting them to empirical data and correlating the content-related factors and the response style-related factors in these models to extraneous criteria. The results showed that the content factors yielded by these models were moderately related to the content criterion and not related to the response style criteria. The response style factors were moderately related to the response style criteria and weakly related to the content criterion. Simultaneous analysis of more than one scale could improve their validity for measuring response styles. These findings indicate that the three models could control and measure extreme and midpoint response styles, though the validity of the mPCM for measuring response styles was not good in some cases. Overall, the multidimensional nominal response model performed slightly better than the other two models.
Collapse
Affiliation(s)
- Yingbin Zhang
- Collaborative Innovation Center of Assessment for Basic Education Quality, Beijing Normal University, Beijing, China.,The Department of Curriculum and Instruction, College of Education, University of Illinois at Urbana-Champaign, Champaign, IL, United States
| | - Yehui Wang
- Collaborative Innovation Center of Assessment for Basic Education Quality, Beijing Normal University, Beijing, China
| |
Collapse
|
40
|
Khorramdel L, Jeon M, Leigh Wang L. Advances in modelling response styles and related phenomena. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:393-400. [PMID: 31721155 DOI: 10.1111/bmsp.12190] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Affiliation(s)
- Lale Khorramdel
- National Board of Medical Examiners (NBME), Center for Advanced Assessments (CAA), Philadelphia, Pennsylvania, USA
- Educational Testing Service (ETS), Psychometrics, Statistics and Data Science (PSDS), Princeton, New Jersey, USA
| | - Minjeong Jeon
- Social Research Methodology, Graduate School of Education & Information Studies, University of California, Los Angeles, USA
| | - Lihshing Leigh Wang
- Quantitative and Mixed-Methods Research Methodologies, University of Cincinnati, Ohio, USA
| |
Collapse
|
41
|
Takagishi M, van de Velden M, Yadohisa H. Clustering preference data in the presence of response-style bias. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:401-425. [PMID: 31049942 DOI: 10.1111/bmsp.12170] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/04/2018] [Revised: 02/22/2019] [Indexed: 06/09/2023]
Abstract
Preference data, such as Likert scale data, are often obtained in questionnaire-based surveys. Clustering respondents based on survey items is useful for discovering latent structures. However, cluster analysis of preference data may be affected by response styles, that is, a respondent's systematic response tendencies irrespective of the item content. For example, some respondents may tend to select ratings at the ends of the scale, which is called an 'extreme response style'. A cluster of respondents with an extreme response style can be mistakenly identified as a content-based cluster. To address this problem, we propose a novel method of clustering respondents based on their indicated preferences for a set of items while correcting for response-style bias. We first introduce a new framework to detect, and correct for, response styles by generalizing the definition of response styles used in constrained dual scaling. We then simultaneously correct for response styles and perform a cluster analysis based on the corrected preference data. A simulation study shows that the proposed method yields better clustering accuracy than the existing methods do. We apply the method to empirical data from four different countries concerning social values.
Collapse
Affiliation(s)
| | | | - Hiroshi Yadohisa
- Facluty of Culture and Information Science, Doshisha University, Japan
| |
Collapse
|
42
|
Khorramdel L, von Davier M, Pokropek A. Combining mixture distribution and multidimensional IRTree models for the measurement of extreme response styles. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:538-559. [PMID: 31385610 DOI: 10.1111/bmsp.12179] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/29/2018] [Revised: 04/21/2019] [Indexed: 05/10/2023]
Abstract
Personality constructs, attitudes and other non-cognitive variables are often measured using rating or Likert-type scales, which does not come without problems. Especially in low-stakes assessments, respondents may produce biased responses due to response styles (RS) that reduce the validity and comparability of the measurement. Detecting and correcting RS is not always straightforward because not all respondents show RS and the ones who do may not do so to the same extent or in the same direction. The present study proposes the combination of a multidimensional IRTree model with a mixture distribution item response theory model and illustrates the application of the approach using data from the Programme for the International Assessment of Adult Competencies (PIAAC). This joint approach allows for the differentiation between different latent classes of respondents who show different RS behaviours and respondents who show RS versus respondents who give (largely) unbiased responses. We illustrate the application of the approach by examining extreme RS and show how the resulting latent classes can be further examined using external variables and process data from computer-based assessments to develop a better understanding of response behaviour and RS.
Collapse
|
43
|
Meiser T, Plieninger H, Henninger M. IRTree models with ordinal and multidimensional decision nodes for response styles and trait-based rating responses. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:501-516. [PMID: 30756379 DOI: 10.1111/bmsp.12158] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/12/2018] [Revised: 11/14/2018] [Indexed: 05/10/2023]
Abstract
IRTree models decompose observed rating responses into sequences of theory-based decision nodes, and they provide a flexible framework for analysing trait-related judgements and response styles. However, most previous applications of IRTree models have been limited to binary decision nodes that reflect qualitatively distinct and unidimensional judgement processes. The present research extends the family of IRTree models for the analysis of response styles to ordinal judgement processes for polytomous decisions and to multidimensional parametrizations of decision nodes. The integration of ordinal judgement processes overcomes the limitation to binary nodes, and it allows researchers to test whether decisions reflect qualitatively distinct response processes or gradual steps on a joint latent continuum. The extension to multidimensional node models enables researchers to specify multiple judgement processes that simultaneously affect the decision between competing response options. Empirical applications highlight the roles of extreme and midpoint response style in rating judgements and show that judgement processes are moderated by different response formats. Model applications with multidimensional decision nodes reveal that decisions among rating categories are jointly informed by trait-related processes and response styles.
Collapse
|
44
|
Jeon M, De Boeck P. Evaluation on types of invariance in studying extreme response bias with an IRTree approach. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:517-537. [PMID: 31292952 DOI: 10.1111/bmsp.12182] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/28/2018] [Revised: 05/17/2019] [Indexed: 05/10/2023]
Abstract
In recent years, item response tree (IRTree) approaches have received increasing attention in the response style literature for their ability to partial out response style latent variables as well as associated item parameters. When an IRTree approach is adopted to measure extreme response styles, directional and content invariance could be assumed at the latent variable and item parameter levels. In this study, we propose to evaluate the empirical validity of these invariance assumptions by employing a general IRTree model with relaxed invariance assumptions. This would allow us to examine extreme response biases, beyond extreme response styles. With three empirical applications of the proposed evaluation, we find that relaxing some of the invariance assumptions improves the model fit, which suggests that not all assumed invariances are empirically supported. Specifically, at the latent variable level, we find reasonable evidence for directional invariance but mixed evidence for content invariance, although we also find that estimated correlations between content-specific extreme response latent variables are high, hinting at the potential presence of a general extreme response tendency. At the item parameter level, we find no directional or content invariance for thresholds and no content invariance for slopes. We discuss how the variant item parameter estimates obtained from a general IRTree model can offer useful insight to help us understand response bias related to extreme responding measured within the IRTree framework.
Collapse
Affiliation(s)
- Minjeong Jeon
- Department of Education, University of California, Los Angeles, California, USA
| | - Paul De Boeck
- Department of Psychology, Ohio State University, Columbus, Ohio, USA
| |
Collapse
|
45
|
Adams DJ, Bolt DM, Deng S, Smith SS, Baker TB. Using multidimensional item response theory to evaluate how response styles impact measurement. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:466-485. [PMID: 30919943 PMCID: PMC6765459 DOI: 10.1111/bmsp.12169] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/10/2018] [Revised: 02/21/2019] [Indexed: 05/25/2023]
Abstract
Multidimensional item response theory (MIRT) models for response style (e.g., Bolt, Lu, & Kim, 2014, Psychological Methods, 19, 528; Falk & Cai, 2016, Psychological Methods, 21, 328) provide flexibility in accommodating various response styles, but often present difficulty in isolating the effects of response style(s) from the intended substantive trait(s). In the presence of such measurement limitations, we consider several ways in which MIRT models are nevertheless useful in lending insight into how response styles may interfere with measurement for a given test instrument. Such a study can also inform whether alternative design considerations (e.g., anchoring vignettes, self-report items of heterogeneous content) that seek to control for response style effects may be helpful. We illustrate several aspects of an MIRT approach using real and simulated analyses.
Collapse
Affiliation(s)
- Daniel J Adams
- Department of Educational Psychology, University of Wisconsin, Madison, Wisconsin, USA
| | - Daniel M Bolt
- Department of Educational Psychology, University of Wisconsin, Madison, Wisconsin, USA
| | | | - Stevens S Smith
- Department of Medicine, University of Wisconsin, Madison, Wisconsin, USA
| | - Timothy B Baker
- Department of Medicine, University of Wisconsin, Madison, Wisconsin, USA
| |
Collapse
|
46
|
Böckenholt U. Assessing item-feature effects with item response tree models. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:486-500. [PMID: 30912584 DOI: 10.1111/bmsp.12163] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/28/2018] [Revised: 01/14/2019] [Indexed: 06/09/2023]
Abstract
Recent applications of item response tree models demonstrate that this model class is well suited to detect midpoint and extremity response style effects in both attitudinal and personality measurements. This paper proposes an extension of this approach that goes beyond measuring response styles and allows us to examine item-feature effects. In a reanalysis of three published data sets, it is shown that the proposed extension captures item-feature effects across affirmative and reverse-worded items in a psychological test. These effects are found to affect directional responses but not midpoint and extremity preferences. Moreover, accounting for item-feature effects substantially improves model fit and interpretation of the construct measurement. The proposed extension can be implemented readily with current software programs that facilitate maximum likelihood estimation of item response models with missing data.
Collapse
Affiliation(s)
- Ulf Böckenholt
- Kellogg School of Management, Northwestern University, Evanston, Illinois, USA
| |
Collapse
|
47
|
Park M, Wu AD. Item Response Tree Models to Investigate Acquiescence and Extreme Response Styles in Likert-Type Rating Scales. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2019; 79:911-930. [PMID: 31488919 PMCID: PMC6713983 DOI: 10.1177/0013164419829855] [Citation(s) in RCA: 17] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
Item response tree (IRTree) models are recently introduced as an approach to modeling response data from Likert-type rating scales. IRTree models are particularly useful to capture a variety of individuals' behaviors involving in item responding. This study employed IRTree models to investigate response styles, which are individuals' tendencies to prefer or avoid certain response categories in a rating scale. Specifically, we introduced two types of IRTree models, descriptive and explanatory models, perceived under a larger modeling framework, called explanatory item response models, proposed by De Boeck and Wilson. This extends the typical application of IRTree models for studying response styles. As a demonstration, we applied the descriptive and explanatory IRTree models to examine acquiescence and extreme response styles in Rosenberg's Self-Esteem Scale. Our findings suggested the presence of two distinct extreme response styles and acquiescence response style in the scale.
Collapse
Affiliation(s)
- Minjeong Park
- University of British Columbia, Vancouver, British Columbia, Canada
- Minjeong Park, Department of Education and Counselling Psychology, and Special Education, University of British Columbia, 2125 Main Mall, Vancouver, British Columbia, V6T 1Z4, Canada.
| | - Amery D. Wu
- University of British Columbia, Vancouver, British Columbia, Canada
| |
Collapse
|
48
|
Leventhal BC. Extreme Response Style: A Simulation Study Comparison of Three Multidimensional Item Response Models. APPLIED PSYCHOLOGICAL MEASUREMENT 2019; 43:322-335. [PMID: 31156283 PMCID: PMC6512164 DOI: 10.1177/0146621618789392] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
Several multidimensional item response models have been proposed for survey responses affected by response styles. Through simulation, this study compares three models designed to account for extreme response tendencies: the IRTree Model, the multidimensional nominal response model, and the modified generalized partial credit model. The modified generalized partial credit model results in the lowest item mean squared error (MSE) across simulation conditions of sample size (500, 1,000), survey length (10, 20), and number of response options (4, 6). The multidimensional nominal response model is equally suitable for surveys measuring one substantive trait using responses to 10 four-option, forced-choice Likert-type items. Based on data validation, comparison of item MSE, and posterior predictive model checking, the IRTree Model is hypothesized to account for additional sources of construct-irrelevant variance.
Collapse
|
49
|
Ma W. A diagnostic tree model for polytomous responses with multiple strategies. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:61-82. [PMID: 29687453 DOI: 10.1111/bmsp.12137] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/18/2017] [Revised: 01/05/2018] [Indexed: 06/08/2023]
Abstract
Constructed-response items have been shown to be appropriate for cognitively diagnostic assessments because students' problem-solving procedures can be observed, providing direct evidence for making inferences about their proficiency. However, multiple strategies used by students make item scoring and psychometric analyses challenging. This study introduces the so-called two-digit scoring scheme into diagnostic assessments to record both students' partial credits and their strategies. This study also proposes a diagnostic tree model (DTM) by integrating the cognitive diagnosis models with the tree model to analyse the items scored using the two-digit rubrics. Both convergent and divergent tree structures are considered to accommodate various scoring rules. The MMLE/EM algorithm is used for item parameter estimation of the DTM, and has been shown to provide good parameter recovery under varied conditions in a simulation study. A set of data from TIMSS 2007 mathematics assessment is analysed to illustrate the use of the two-digit scoring scheme and the DTM.
Collapse
Affiliation(s)
- Wenchao Ma
- The University of Alabama, Tuscaloosa, AL, USA
| |
Collapse
|
50
|
Colombi R, Giordano S, Gottard A, Iannario M. Hierarchical marginal models with latent uncertainty. Scand Stat Theory Appl 2018. [DOI: 10.1111/sjos.12366] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Affiliation(s)
- Roberto Colombi
- Department of ManagementInformation and Production Engineering, University of Bergamo Bergamo Italy
| | - Sabrina Giordano
- Department of Economics, Statistics and Finance “Giovanni Anania”University of Calabria Arcavacata di Rende Italy
| | - Anna Gottard
- Department of Statistics, Computer Science, Applications “Giuseppe Parenti”University of Florence Florence Italy
| | - Maria Iannario
- Department of Political SciencesUniversity of Naples Federico II Naples Italy
| |
Collapse
|