1
|
Manolov R. Does the choice of a linear trend-assessment technique matter in the context of single-case data? Behav Res Methods 2023; 55:4200-4221. [PMID: 36622560 DOI: 10.3758/s13428-022-02013-0] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 10/22/2022] [Indexed: 01/10/2023]
Abstract
Trend is one of the data aspects that is an object of assessment in the context of single-case experimental designs. This assessment can be performed both visually and quantitatively. Given that trend, just like other relevant data features such as level, immediacy, or overlap does not have a single operative definition, a comparison among the existing alternatives is necessary. Previous studies have included illustrations of differences between trend-line fitting techniques using real data. In the current study, I carry out a simulation to study the degree to which different trend-line fitting techniques lead to different degrees of bias, mean square error, and statistical power for a variety of quantifications that entail trend lines. The simulation involves generating both continuous and count data, for several phase lengths, degrees of autocorrelation, and effect sizes (change in level and change in slope). The results suggest that, in general, ordinary least squares estimation performs well in terms of relative bias and mean square error. Especially, a quantification of slope change is associated with better statistical results than quantifying an average difference between conditions on the basis of a projected baseline trend. In contrast, the performance of the split-middle (bisplit) technique is less than optimal.
Collapse
Affiliation(s)
- Rumen Manolov
- Department of Social Psychology and Quantitative Psychology, Faculty of Psychology, University of Barcelona, Passeig de la Vall d'Hebron 171, 08035, Barcelona, Spain.
| |
Collapse
|
2
|
Abstract
The continuation of a baseline pattern of responding into a treatment phase, sometimes referred to as a "transition state," can obscure interpretation of data depicted in single-case experimental designs (SCEDs). For example, when using visual analysis, transition states may lead to the conclusion that the treatment is ineffective. Likewise, the inclusion of overlapping data points in some statistical analyses may lead to conclusions that the treatment had a small effect size and give rise to publication bias. This study reviewed 20 volumes in a journal that publishes primarily SCEDs studies. We defined a transition state as a situation wherein at least the first three consecutive data points of a treatment phase or condition are within the range of the baseline phase or condition. Results indicate that transitions states (a) were present for 7.4% of graphs that met inclusion criteria and (b) occurred for a mean of 4.9 data points before leading to behavior change. We discuss some implications and directions for future research on transition states.
Collapse
|
3
|
Kratochwill TR, Horner RH, Levin JR, Machalicek W, Ferron J, Johnson A. Single-case intervention research design standards: Additional proposed upgrades and future directions. J Sch Psychol 2023; 97:192-216. [PMID: 36914365 DOI: 10.1016/j.jsp.2022.12.002] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2021] [Revised: 05/28/2022] [Accepted: 12/23/2022] [Indexed: 02/25/2023]
Abstract
Single-case intervention research design standards have evolved considerably over the past decade. These standards serve the dual role of assisting in single-case design (SCD) intervention research methodology and as guidelines for literature syntheses within a particular research domain. In a recent article (Kratochwill et al., 2021), we argued for a need to clarify key features of these standards. In this article we offer additional recommendations for SCD research and synthesis standards that have been either underdeveloped or missing in the conduct of research and in literature syntheses. Our recommendations are organized into three categories: expanding design standards, expanding evidence standards, and expanding the applications and consistency of SCDs. The recommendations we advance are for consideration for future standards, research design training, and they are especially important to guide the reporting of SCD intervention investigations as they enter the literature-synthesis phase of evidence-based practice initiatives.
Collapse
Affiliation(s)
| | | | | | | | - John Ferron
- University of South Florida,United States of America
| | - Austin Johnson
- University of California, Riverside, United States of America
| |
Collapse
|
4
|
Lanovaz MJ, Primiani R. Waiting for baseline stability in single-case designs: Is it worth the time and effort? Behav Res Methods 2023; 55:843-854. [PMID: 35469087 PMCID: PMC10027773 DOI: 10.3758/s13428-022-01858-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/05/2022] [Indexed: 11/08/2022]
Abstract
Researchers and practitioners often use single-case designs (SCDs), or n-of-1 trials, to develop and validate novel treatments. Standards and guidelines have been published to provide guidance as to how to implement SCDs, but many of their recommendations are not derived from the research literature. For example, one of these recommendations suggests that researchers and practitioners should wait for baseline stability prior to introducing an independent variable. However, this recommendation is not strongly supported by empirical evidence. To address this issue, we used Monte Carlo simulations to generate graphs with fixed, response-guided, and random baseline lengths while manipulating trend and variability. Then, our analyses compared the type I error rate and power produced by two methods of analysis: the conservative dual-criteria method (a structured visual aid) and a support vector classifier (a model derived from machine learning). The conservative dual-criteria method produced fewer errors when using response-guided decision-making (i.e., waiting for stability) and random baseline lengths. In contrast, waiting for stability did not reduce decision-making errors with the support vector classifier. Our findings question the necessity of waiting for baseline stability when using SCDs with machine learning, but the study must be replicated with other designs and graph parameters that change over time to support our results.
Collapse
Affiliation(s)
- Marc J Lanovaz
- École de psychoéducation, Université de Montréal, C.P. 6128, succursale Centre-Ville, Montreal, QC, H3C 3J7, Canada.
- Centre de recherche de l'Institut universitaire en santé mentale de Montréal, Montreal, Canada.
| | - Rachel Primiani
- École de psychoéducation, Université de Montréal, C.P. 6128, succursale Centre-Ville, Montreal, QC, H3C 3J7, Canada
- Centre de recherche de l'Institut universitaire en santé mentale de Montréal, Montreal, Canada
| |
Collapse
|
5
|
Taylor T, Lanovaz MJ. Agreement between visual inspection and objective analysis methods: A replication and extension. J Appl Behav Anal 2022; 55:986-996. [PMID: 35478098 PMCID: PMC9323513 DOI: 10.1002/jaba.921] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2021] [Revised: 03/14/2022] [Accepted: 03/18/2022] [Indexed: 11/30/2022]
Abstract
Behavior analysts typically rely on visual inspection of single‐case experimental designs to make treatment decisions. However, visual inspection is subjective, which has led to the development of supplemental objective methods such as the conservative dual‐criteria method. To replicate and extend a study conducted by Wolfe et al. (2018) on the topic, we examined agreement between the visual inspection of five raters, the conservative dual‐criteria method, and a machine‐learning algorithm (i.e., the support vector classifier) on 198 AB graphs extracted from clinical data. The results indicated that average agreement between the 3 methods was generally consistent. Mean interrater agreement was 84%, whereas raters agreed with the conservative dual‐criteria method and the support vector classifier on 84% and 85% of graphs, respectively. Our results indicate that both objective methods produce results consistent with visual inspection, which may support their future use.
Collapse
Affiliation(s)
- Tessa Taylor
- University of Canterbury/Te Whare Wānanga o Waitaha.,Paediatric Feeding International
| | | |
Collapse
|
6
|
Ruckle MM, Bednar MK, Suen K, Falligant JM. Brief assessment and treatment of pica using differential reinforcement, response interruption and redirection, and competing stimuli. BEHAVIORAL INTERVENTIONS 2022. [DOI: 10.1002/bin.1881] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Affiliation(s)
- Mary M. Ruckle
- Neurobehavioral Unit Kennedy Krieger Institute Baltimore MD USA
| | - Molly K. Bednar
- Neurobehavioral Unit Kennedy Krieger Institute Baltimore MD USA
- Little Leaves Behavioral Services Silver Spring MD USA
| | - Kevin Suen
- Neurobehavioral Unit Kennedy Krieger Institute Baltimore MD USA
| | - John Michael Falligant
- Neurobehavioral Unit Kennedy Krieger Institute Baltimore MD USA
- Johns Hopkins University School of Medicine Baltimore MD USA
| |
Collapse
|
7
|
Falligant JM, Kranak MP, Hagopian LP. Further Analysis of Advanced Quantitative Methods and Supplemental Interpretative Aids with Single-Case Experimental Designs. Perspect Behav Sci 2022; 45:77-99. [PMID: 35342866 PMCID: PMC8894533 DOI: 10.1007/s40614-021-00313-y] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 08/11/2021] [Indexed: 02/04/2023] Open
Abstract
Reliable and accurate visual analysis of graphically depicted behavioral data acquired using single-case experimental designs (SCEDs) is integral to behavior-analytic research and practice. Researchers have developed a range of techniques to increase reliable and objective visual inspection of SCED data including visual interpretive guides, statistical techniques, and nonstatistical quantitative methods to objectify the visual-analytic interpretation of data to guide clinicians, and ensure a replicable data interpretation process in research. These structured data analytic practices are now more frequently used by behavior analysts and the subject of considerable research within the field of quantitative methods and behavior analysis. First, there are contemporaneous analytic methods that have preliminary support with simulated datasets, but have not been thoroughly examined with nonsimulated clinical datasets. There are a number of relatively new techniques that have preliminary support (e.g., fail-safe k), but require additional research. Other analytic methods (e.g., dual-criteria and conservative dual criteria) have more extensive support, but have infrequently been compared against other analytic methods. Across three studies, we examine how these methods corresponded to clinical outcomes (and one another) for the purpose of replicating and extending extant literature in this area. Implications and recommendations for practitioners and researchers are discussed.
Collapse
Affiliation(s)
- John Michael Falligant
- Kennedy Krieger Institute, Baltimore, MD USA
- Johns Hopkins University School of Medicine, Baltimore, MD USA
| | - Michael P. Kranak
- Oakland University, Rochester, MI USA
- Oakland University Center for Autism, Rochester, MI USA
| | - Louis P. Hagopian
- Kennedy Krieger Institute, Baltimore, MD USA
- Johns Hopkins University School of Medicine, Baltimore, MD USA
| |
Collapse
|
8
|
Dowdy A, Jessel J, Saini V, Peltier C. Structured visual analysis of single-case experimental design data: Developments and technological advancements. J Appl Behav Anal 2021; 55:451-462. [PMID: 34962646 DOI: 10.1002/jaba.899] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2020] [Revised: 11/22/2021] [Accepted: 11/22/2021] [Indexed: 11/05/2022]
Abstract
Visual analysis is the primary method used to interpret single-case experimental design (SCED) data in applied behavior analysis. Research shows that agreement between visual analysts can be suboptimal at times. To address the inconsistent interpretations of SCED data, recent structured visual-analysis technological advancements have been developed. To assess the extent to which structured visual analysis is used to guide or supplement applied behavior analysts' interpretation of SCED graphs, a systematic review between the years of 2015 to 2020 in the Journal of Applied Behavior Analysis was conducted. Findings showed that despite recent efforts to develop structured visual-analysis tools and criteria, these methods are rarely used to analyze SCED data. An overview of structured visual-analysis tools is shared, their utility is delineated, common characteristics are brought to light, and future directions for both research and their clinical use are highlighted.
Collapse
Affiliation(s)
- Art Dowdy
- Department of Teaching and Learning, Temple University
| | | | - Valdeep Saini
- Department of Applied Disability Studies, Brock University
| | - Corey Peltier
- Department of Educational Psychology, University of Oklahoma
| |
Collapse
|
9
|
Lanovaz MJ, Hranchuk K. Machine learning to analyze single-case graphs: A comparison to visual inspection. J Appl Behav Anal 2021; 54:1541-1552. [PMID: 34263923 PMCID: PMC8596748 DOI: 10.1002/jaba.863] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2021] [Revised: 06/04/2021] [Accepted: 06/04/2021] [Indexed: 01/12/2023]
Abstract
Behavior analysts commonly use visual inspection to analyze single‐case graphs, but studies on its reliability have produced mixed results. To examine this issue, we compared the Type I error rate and power of visual inspection with a novel approach—machine learning. Five expert visual raters analyzed 1,024 simulated AB graphs, which differed on number of points per phase, autocorrelation, trend, variability, and effect size. The ratings were compared to those obtained by the conservative dual‐criteria method and two models derived from machine learning. On average, visual raters agreed with each other on only 75% of graphs. In contrast, both models derived from machine learning showed the best balance between Type I error rate and power while producing more consistent results across different graph characteristics. The results suggest that machine learning may support researchers and practitioners in making fewer errors when analyzing single‐case graphs, but replications remain necessary.
Collapse
Affiliation(s)
- Marc J Lanovaz
- École de psychoéducation, Université de Montréal.,Centre de recherche de l'Institut universitaire en santé mentale de Montréal
| | - Kieva Hranchuk
- Behavioural Science and Behavioural Psychology, St. Lawrence College
| |
Collapse
|
10
|
Trudel L, Lanovaz MJ, Préfontaine I. Brief Report: Mobile Technology to Support Parents in Reducing Stereotypy. J Autism Dev Disord 2021; 51:2550-2558. [PMID: 33000395 DOI: 10.1007/s10803-020-04735-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Abstract
Although behavioral interventions have been known to effectively reduce stereotypy in children with ASD, these types of interventions are not accessible to all families. In response to this issue, we evaluated the effects of the iSTIM, an iOS application designed to support parents in the reduction of stereotypy in their child with ASD. We used a series of AB designs to determine the effectiveness of the iSTIM on stereotypy using parents as behavior change agents. The use of iSTIM by the parents led to a reduction in stereotypy for six of seven participants. Our results suggest that the use of technology may be a cost effective and easily accessible method for parents to reduce stereotypy in their child with ASD.
Collapse
Affiliation(s)
- Lydia Trudel
- École de Psychoéducation, Université de Montréal, Succursale Centre-Ville, Montreal, QC, C.P. 6128, H3C 3J7, Canada
| | - Marc J Lanovaz
- École de Psychoéducation, Université de Montréal, Succursale Centre-Ville, Montreal, QC, C.P. 6128, H3C 3J7, Canada. .,Centre de Recherche de l'Institut Universitaire en Santé Mentale de Montréal, Montreal, Canada.
| | - Isabelle Préfontaine
- École de Psychoéducation, Université de Montréal, Succursale Centre-Ville, Montreal, QC, C.P. 6128, H3C 3J7, Canada
| |
Collapse
|
11
|
Implementing Automated Nonparametric Statistical Analysis on Functional Analysis Data: A Guide for Practitioners and Researchers. Perspect Behav Sci 2021; 45:53-75. [DOI: 10.1007/s40614-021-00290-2] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/16/2021] [Indexed: 11/25/2022] Open
|
12
|
Cox AD, Zonneveld KLM, Tardi LD. Further evaluating interobserver reliability and accuracy with and without structured visual‐inspection criteria. BEHAVIORAL INTERVENTIONS 2021. [DOI: 10.1002/bin.1793] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022]
Affiliation(s)
- Alison D. Cox
- Department of Applied Disability Studies Brock University St. Catharines Ontario Canada
| | | | - Laura D. Tardi
- Department of Applied Disability Studies Brock University St. Catharines Ontario Canada
| |
Collapse
|
13
|
Manolov R, Tanious R, Onghena P. Quantitative Techniques and Graphical Representations for Interpreting Results from Alternating Treatment Design. Perspect Behav Sci 2021; 45:259-294. [DOI: 10.1007/s40614-021-00289-9] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/13/2021] [Indexed: 01/11/2023] Open
|
14
|
Kranak MP, Falligant JM, Hausman NL. Application of automated nonparametric statistical analysis in clinical contexts. J Appl Behav Anal 2020; 54:824-833. [PMID: 33084039 DOI: 10.1002/jaba.789] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2020] [Revised: 09/08/2020] [Accepted: 09/08/2020] [Indexed: 01/10/2023]
Abstract
Functional analyses (FAs) provide clinicians with results upon which they design behavioral treatments. Unfortunately, interrater reliability of visual analysis of FA results can be inconsistent. Accordingly, researchers have designed quantitative metrics and visual aids to supplement visual analysis. Recently, Hall et al. (2020) provided a proof of concept for using automated nonparametric statistical analysis (ANSA) to interpret FA data. Their results show promise for ANSA as a supplemental tool. However, they evaluated ANSA with only published FA datasets, which may not be representative of FAs commonly encountered in clinical care. Therefore, the purpose of this replication was to compare ANSA to another validated supplemental aid (i.e., the structured criteria method) and investigate its utility with unpublished clinical FA data. Our results were consistent with Hall et al.'s, indicating ANSA may augment clinical interpretation of FA data. Recommendations for clinical applications of ANSA and future directions for researchers are discussed.
Collapse
Affiliation(s)
| | | | - Nicole L Hausman
- Kennedy Krieger Institute.,Johns Hopkins University School of Medicine
| |
Collapse
|
15
|
Lanovaz MJ, Turgeon S. How Many Tiers Do We Need? Type I Errors and Power in Multiple Baseline Designs. Perspect Behav Sci 2020; 43:605-616. [PMID: 33024931 PMCID: PMC7490309 DOI: 10.1007/s40614-020-00263-x] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022] Open
Abstract
Design quality guidelines typically recommend that multiple baseline designs include at least three demonstrations of effects. Despite its widespread adoption, this recommendation does not appear grounded in empirical evidence. The main purpose of our study was to address this issue by assessing Type I error rate and power in multiple baseline designs. First, we generated 10,000 multiple baseline graphs, applied the dual-criteria method to each tier, and computed Type I error rate and power for different number of tiers showing a clear change. Second, two raters categorized the tiers for 300 multiple baseline graphs to replicate our analyses using visual inspection. When multiple baseline designs had at least three tiers and two or more of these tiers showed a clear change, the Type I error rate remained adequate (< .05) while power also reached acceptable levels (> .80). In contrast, requiring all tiers to show a clear change resulted in overly stringent conclusions (i.e., unacceptably low power). Therefore, our results suggest that researchers and practitioners should carefully consider limitations in power when requiring all tiers of a multiple baseline design to show a clear change in their analyses.
Collapse
Affiliation(s)
- Marc J Lanovaz
- École de psychoéducation, Université de Montréal, C.P. 6128, succursale Centre-Ville, Montreal, QC H3C 3J7 Canada.,Centre de recherche de l'Institut universitaire en santé mentale de Montréal, Montreal, QC Canada
| | - Stéphanie Turgeon
- École de psychoéducation, Université de Montréal, C.P. 6128, succursale Centre-Ville, Montreal, QC H3C 3J7 Canada
| |
Collapse
|
16
|
Falligant JM, Kranak MP, Schmidt JD, Rooker GW. Correspondence between Fail-Safe k and Dual-Criteria Methods: Analysis of Data Series Stability. Perspect Behav Sci 2020; 43:303-319. [PMID: 32647784 DOI: 10.1007/s40614-020-00255-x] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022] Open
Abstract
Barnard-Brak, Richman, Little, and Yang (Behaviour Research and Therapy, 102, 8-15, 2018) developed a structured-criteria metric, fail-safe k, which quantifies the stability of data series within single-case experimental designs (SCEDs) using published baseline and treatment data. Fail-safe k suggests the optimal point in time to change phases (e.g., move from Phase B to Phase C, reverse back to Phase A). However, this tool has not been tested with clinical data obtained in the course of care. Thus, the purpose of the current study was to replicate the procedures described by Barnard-Brak et al. with clinical data. We also evaluated the correspondence between the fail-safe k metric with outcomes obtained via dual-criteria and conservative-dual criteria methods, which are empirically supported methods for evaluating data-series trends within SCEDs. Our results provide some degree of support for use of this approach as a research tool with clinical data, in particular when evaluating small or medium treatment effect sizes, but further research is needed before this can be used widely by practitioners.
Collapse
Affiliation(s)
- John Michael Falligant
- Kennedy Krieger Institute & Johns Hopkins University School of Medicine, 707 North Broadway, Baltimore, MD 21205 USA
| | - Michael P Kranak
- Kennedy Krieger Institute & Johns Hopkins University School of Medicine, 707 North Broadway, Baltimore, MD 21205 USA
| | - Jonathan D Schmidt
- Kennedy Krieger Institute & Johns Hopkins University School of Medicine, 707 North Broadway, Baltimore, MD 21205 USA
| | - Griffin W Rooker
- Kennedy Krieger Institute & Johns Hopkins University School of Medicine, 707 North Broadway, Baltimore, MD 21205 USA
| |
Collapse
|
17
|
Falligant JM, McNulty MK, Kranak MP, Hausman NL, Rooker GW. Evaluating sources of baseline data using dual-criteria and conservative dual-criteria methods: A quantitative analysis. J Appl Behav Anal 2020; 53:2330-2338. [PMID: 32337720 DOI: 10.1002/jaba.710] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2019] [Revised: 02/04/2020] [Accepted: 02/19/2020] [Indexed: 11/11/2022]
Abstract
Scheithauer et al. (2020) recently demonstrated that differences in the source of baseline data extracted from a functional analysis (FA) may not affect subsequent clinical decision-making in comparison to a standard baseline. These outcomes warrant additional quantitative examination, as correspondence of visual analysis has sometimes been reported to be unreliable. In the current study, we quantified the occurrence of false positives within a dataset of FA and baseline data using the dual-criteria (DC) and conservative dual-criteria (CDC) methods. Results of this quantitative analysis suggest that false positives were more likely when using FA data (rather than original baseline data) as the initial treatment baseline. However, both sources of baseline data may have acceptably low levels of false positives for practical use. Overall, the findings provide preliminary quantitative support for the conclusion that determinations of effective treatment may be easily obtained using different sources of baseline data.
Collapse
Affiliation(s)
- John Michael Falligant
- Department of Behavioral Psychology, Kennedy Krieger Institute.,Department of Pediatrics, Johns Hopkins University School of Medicine
| | - Molly K McNulty
- Department of Behavioral Psychology, Kennedy Krieger Institute
| | - Michael P Kranak
- Department of Behavioral Psychology, Kennedy Krieger Institute.,Department of Pediatrics, Johns Hopkins University School of Medicine
| | - Nicole L Hausman
- Department of Behavioral Psychology, Kennedy Krieger Institute.,Department of Psychiatry and Behavioral Sciences, Johns Hopkins University School of Medicine
| | - Griffin W Rooker
- Department of Behavioral Psychology, Kennedy Krieger Institute.,Department of Psychiatry and Behavioral Sciences, Johns Hopkins University School of Medicine
| |
Collapse
|
18
|
Falligant JM, Vetter JA. Quantifying false positives in simulated events using partial interval recording and momentary time sampling with dual‐criteria methods. BEHAVIORAL INTERVENTIONS 2020. [DOI: 10.1002/bin.1707] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Affiliation(s)
- John Michael Falligant
- Department of Behavioral PsychologyKennedy Krieger Institute, Johns Hopkins University School of Medicine Baltimore Maryland
| | - Jennifer A. Vetter
- Department of Behavioral PsychologyKennedy Krieger Institute, Johns Hopkins University School of Medicine Baltimore Maryland
| |
Collapse
|
19
|
Lanovaz MJ, Giannakakos AR, Destras O. Machine Learning to Analyze Single-Case Data: A Proof of Concept. Perspect Behav Sci 2020; 43:21-38. [PMID: 32440643 PMCID: PMC7198678 DOI: 10.1007/s40614-020-00244-0] [Citation(s) in RCA: 28] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Visual analysis is the most commonly used method for interpreting data from single-case designs, but levels of interrater agreement remain a concern. Although structured aids to visual analysis such as the dual-criteria (DC) method may increase interrater agreement, the accuracy of the analyses may still benefit from improvements. Thus, the purpose of our study was to (a) examine correspondence between visual analysis and models derived from different machine learning algorithms, and (b) compare the accuracy, Type I error rate and power of each of our models with those produced by the DC method. We trained our models on a previously published dataset and then conducted analyses on both nonsimulated and simulated graphs. All our models derived from machine learning algorithms matched the interpretation of the visual analysts more frequently than the DC method. Furthermore, the machine learning algorithms outperformed the DC method on accuracy, Type I error rate, and power. Our results support the somewhat unorthodox proposition that behavior analysts may use machine learning algorithms to supplement their visual analysis of single-case data, but more research is needed to examine the potential benefits and drawbacks of such an approach.
Collapse
Affiliation(s)
- Marc J Lanovaz
- 1École de Psychoéducation, Université de Montréal, C.P. 6128, succursale Centre-Ville, Montreal, QC H3C 3J7 Canada
| | | | | |
Collapse
|
20
|
Falligant JM, McNulty MK, Hausman NL, Rooker GW. Using dual‐criteria methods to supplement visual inspection: Replication and extension. J Appl Behav Anal 2019; 53:1789-1798. [DOI: 10.1002/jaba.665] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/07/2019] [Accepted: 09/04/2019] [Indexed: 11/07/2022]
Affiliation(s)
| | | | - Nicole L. Hausman
- Kennedy Krieger Institute & Johns Hopkins University School of Medicine
| | - Griffin W. Rooker
- Kennedy Krieger Institute & Johns Hopkins University School of Medicine
| |
Collapse
|
21
|
Lanovaz MJ, Turgeon S, Cardinal P, Wheatley TL. Using Single-Case Designs in Practical Settings: Is Within-Subject Replication Always Necessary? Perspect Behav Sci 2019; 42:153-162. [PMID: 31976426 PMCID: PMC6701506 DOI: 10.1007/s40614-018-0138-9] [Citation(s) in RCA: 29] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022] Open
Abstract
Behavior analysts have widely adopted and embraced within-subject replication through the use of reversal and multielement designs. However, the withdrawal of treatment, which is central to these designs, may not be desirable, feasible, or even ethical in practical settings. To examine this issue, we extracted 501 ABAB graphs from theses and dissertations to examine to what extent we would have reached correct or incorrect conclusions if we had based our analysis on the initial AB component only. In our first experiment, we examined the proportion of datasets for which the results of the first AB component matched the results of the subsequent phase reversals. In our second experiment, we calculated three effect size estimates for the same datasets to examine whether these measures could predict the relevance of conducting a within-subject replication. Our analyses indicated that the initial effects were successfully replicated at least once in approximately 85% of the cases and that effect size may predict the probability of within-subject replication. Overall, our results support the rather controversial proposition that it may be possible to set threshold values of effect size above which conducting a replication could be considered unnecessary. That said, more research is needed to confirm and examine the generalizability of these results prior to recommending changes in practice.
Collapse
Affiliation(s)
- Marc J. Lanovaz
- Université de Montréal and Centre de Recherche du CHU Sainte-Justine, Université de Montréal, C.P. 6128, succursale Centre-Ville, Montréal, QC H3C 3J7 Canada
| | - Stéphanie Turgeon
- Université de Montréal, C.P. 6128, succ. Centre-Ville, Montréal, QC H3C 3J7 Canada
| | - Patrick Cardinal
- École de Technologie Supérieure, 1100 Notre-Dame St W, Montreal, QC H3C 1K3 Canada
| | - Tara L. Wheatley
- Halton Catholic District School Board, 802 Drury Lane, Burlington, ON L7R 2Y2 Canada
| |
Collapse
|
22
|
Laraway S, Snycerski S, Pradhan S, Huitema BE. An Overview of Scientific Reproducibility: Consideration of Relevant Issues for Behavior Science/Analysis. Perspect Behav Sci 2019; 42:33-57. [PMID: 31976420 PMCID: PMC6701706 DOI: 10.1007/s40614-019-00193-3] [Citation(s) in RCA: 37] [Impact Index Per Article: 7.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022] Open
Abstract
For over a decade, the failure to reproduce findings in several disciplines, including the biomedical, behavioral, and social sciences, have led some authors to claim that there is a so-called "replication (or reproducibility) crisis" in those disciplines. The current article examines: (a) various aspects of the reproducibility of scientific studies, including definitions of reproducibility; (b) published concerns about reproducibility in the scientific literature and public press; (c) variables involved in assessing the success of attempts to reproduce a study; (d) suggested factors responsible for reproducibility failures; (e) types of validity of experimental studies and threats to validity as they relate to reproducibility; and (f) evidence for threats to reproducibility in the behavior science/analysis literature. Suggestions for improving the reproducibility of studies in behavior science and analysis are described throughout.
Collapse
Affiliation(s)
- Sean Laraway
- Department of Psychology, San José State University, San José, CA 95192-0120 USA
| | - Susan Snycerski
- Department of Psychology, San José State University, San José, CA 95192-0120 USA
| | | | | |
Collapse
|
23
|
McDougale CB, Coon JC, Richling SM, O'Rourke S, Rapp JT, Thompson KR, Burkhart BR. Group Procedures for Decreasing Problem Behavior Displayed by Detained Adolescents. Behav Modif 2018; 43:615-638. [PMID: 29902929 DOI: 10.1177/0145445518781314] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
As one component of providing treatment in a residential facility, Brogan, Falligant, and Rapp decreased problem behavior by two groups of detained adolescents using group contingency procedures. The current series of studies evaluated the extent to which group procedures could be extended to other contexts within a residential facility. In Study 1, fixed-time delivery of attention by dormitory staff decreased problem behavior displayed by a group of five to 11 detained adolescents during free periods. In Study 2, rules from a therapist plus contingencies for following those rules increased appropriate line walking during specific transition periods. Subsequently, rules alone maintained appropriate line walking, however, direct training was required to produce appropriate line walking during other transitions. Measures of social validity indicated the procedures and outcomes in both studies were acceptable to facility personnel.
Collapse
|
24
|
A Simulation Study on Two Analytical Techniques for Alternating Treatments Designs. Behav Modif 2018; 43:544-563. [DOI: 10.1177/0145445518777875] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
|
25
|
Krasny-Pacini A, Evans J. Single-case experimental designs to assess intervention effectiveness in rehabilitation: A practical guide. Ann Phys Rehabil Med 2017; 61:164-179. [PMID: 29253607 DOI: 10.1016/j.rehab.2017.12.002] [Citation(s) in RCA: 146] [Impact Index Per Article: 20.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2017] [Revised: 11/29/2017] [Accepted: 12/10/2017] [Indexed: 11/15/2022]
Abstract
Single-case experimental designs (SCED) are experimental designs aiming at testing the effect of an intervention using a small number of patients (typically one to three), using repeated measurements, sequential (±randomized) introduction of an intervention and method-specific data analysis, including visual analysis and specific statistics. The aim of this paper is to familiarise professionals working in different fields of rehabilitation with SCEDs and provide practical advice on how to design and implement a SCED in clinical rehabilitation practice. Research questions suitable for SCEDs and the different types of SCEDs (e.g., alternating treatment designs, introduction/withdrawal designs and multiple baseline designs) are reviewed. Practical steps in preparing a SCED design are outlined. Examples from different rehabilitation domains are provided throughout the paper. Challenging issues such as the choice of the repeated measure, assessment of generalisation, randomization, procedural fidelity, replication and generalizability of findings are discussed. Simple rules and resources for data analysis are presented. The utility of SCEDs in physical and rehabilitation medicine (PRM) are discussed.
Collapse
Affiliation(s)
- Agata Krasny-Pacini
- Institut universitaire de réadaptation Clemenceau-Strasbourg, 45, boulevard Clemenceau, 67082 Strasbourg, France; Service de chirurgie orthopédique infantile, hôpital de Hautepierre, CHU de Strasbourg, avenue Molière, 67098 Strasbourg, France; GRC handicap cognitif et réadaptation (HanCRe), hôpitaux universitaires Pitié-Salpêtière, 75013 Paris, France.
| | - Jonathan Evans
- Institute of Health and Wellbeing, University of Glasgow, The Academic Centre, Gartnavel Royal Hospital, 1055 Great Western Road, Glasgow G12 0XH, United Kingdom
| |
Collapse
|
26
|
Coon JC, Rapp JT. Application of multiple baseline designs in behavior analytic research: Evidence for the influence of new guidelines. BEHAVIORAL INTERVENTIONS 2017. [DOI: 10.1002/bin.1510] [Citation(s) in RCA: 36] [Impact Index Per Article: 5.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
27
|
Lanovaz MJ, Cardinal P, Francis M. Using a Visual Structured Criterion for the Analysis of Alternating-Treatment Designs. Behav Modif 2017; 43:115-131. [PMID: 29094610 DOI: 10.1177/0145445517739278] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Although visual inspection remains common in the analysis of single-case designs, the lack of agreement between raters is an issue that may seriously compromise its validity. Thus, the purpose of our study was to develop and examine the properties of a simple structured criterion to supplement the visual analysis of alternating-treatment designs. To this end, we generated simulated data sets with varying number of points, number of conditions, effect sizes, and autocorrelations, and then measured Type I error rates and power produced by the visual structured criterion (VSC) and permutation analyses. We also validated the results for Type I error rates using nonsimulated data. Overall, our results indicate that using the VSC as a supplement for the analysis of systematically alternating-treatment designs with at least five points per condition generally provides adequate control over Type I error rates and sufficient power to detect most behavior changes.
Collapse
Affiliation(s)
- Marc J Lanovaz
- 1 Université de Montréal, Québec, Canada.,2 Centre de recherche du CHU Sainte-Justine, Montreal, Québec, Canada
| | | | | |
Collapse
|
28
|
Abstract
The frequently used visual analysis of single-case data focuses on data aspects such as level, trend, variability, overlap, immediacy of effect, and consistency of data patterns; most of these aspects are also commonly quantified besides inspecting them visually. The present text focuses on trend, because even linear trend can be operatively defined in several different ways, while there are also different approaches for controlling for baseline trend. We recommend using a quantitative criterion for choosing a trend line fitting technique and comparing baseline and intervention slopes, instead of detrending. We implement our proposal in a free web-based application created specifically for following the What Works Clearinghouse Standards recommendations for visual analysis. This application is especially destined to applied researchers and provides graphical representation of the data, visual aids, and quantifications of the difference between phases in terms of level, trend, and overlap, as well as two quantifications of the immediate effect. An evaluation of the consistency of effects across replications of the AB sequence is also provided. For methodologists and statisticians, we include formulas and examples of the different straight line fitting and detrending techniques to improve the reproducibility of results and simulations.
Collapse
|