1
|
Xie CX, De Simoni A, Eldridge S, Pinnock H, Relton C. Development of a conceptual framework for defining trial efficiency. PLoS One 2024; 19:e0304187. [PMID: 38781167 PMCID: PMC11115328 DOI: 10.1371/journal.pone.0304187] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2023] [Accepted: 05/07/2024] [Indexed: 05/25/2024] Open
Abstract
BACKGROUND Globally, there is a growing focus on efficient trials, yet numerous interpretations have emerged, suggesting a significant heterogeneity in understanding "efficiency" within the trial context. Therefore in this study, we aimed to dissect the multifaceted nature of trial efficiency by establishing a comprehensive conceptual framework for its definition. OBJECTIVES To collate diverse perspectives regarding trial efficiency and to achieve consensus on a conceptual framework for defining trial efficiency. METHODS From July 2022 to July 2023, we undertook a literature review to identify various terms that have been used to define trial efficiency. We then conducted a modified e-Delphi study, comprising an exploratory open round and a subsequent scoring round to refine and validate the identified items. We recruited a wide range of experts in the global trial community including trialists, funders, sponsors, journal editors and members of the public. Consensus was defined as items rated "without disagreement", measured by the inter-percentile range adjusted for symmetry through the UCLA/RAND approach. RESULTS Seventy-eight studies were identified from a literature review, from which we extracted nine terms related to trial efficiency. We then used review findings as exemplars in the Delphi open round. Forty-nine international experts were recruited to the e-Delphi panel. Open round responses resulted in the refinement of the initial nine terms, which were consequently included in the scoring round. We obtained consensus on all nine items: 1) four constructs that collectively define trial efficiency containing scientific efficiency, operational efficiency, statistical efficiency and economic efficiency; and 2) five essential building blocks for efficient trial comprising trial design, trial process, infrastructure, superstructure, and stakeholders. CONCLUSIONS This is the first attempt to dissect the concept of trial efficiency into theoretical constructs. Having an agreed definition will allow better trial implementation and facilitate effective communication and decision-making across stakeholders. We also identified essential building blocks that are the cornerstones of an efficient trial. In this pursuit of understanding, we are not only unravelling the complexities of trial efficiency but also laying the groundwork for evaluating the efficiency of an individual trial or a trial system in the future.
Collapse
Affiliation(s)
- Charis Xuan Xie
- Wolfson Institute of Population Health, Queen Mary University of London, London, England, United Kingdom
| | - Anna De Simoni
- Wolfson Institute of Population Health, Queen Mary University of London, London, England, United Kingdom
| | - Sandra Eldridge
- Wolfson Institute of Population Health, Queen Mary University of London, London, England, United Kingdom
| | - Hilary Pinnock
- Usher Institute, Asthma UK Centre for Applied Research, The University of Edinburgh, Edinburgh, Scotland, United Kingdom
| | - Clare Relton
- Wolfson Institute of Population Health, Queen Mary University of London, London, England, United Kingdom
| |
Collapse
|
2
|
Placzek M, Friede T. Blinded sample size recalculation in adaptive enrichment designs. Biom J 2023; 65:e2000345. [PMID: 35983952 DOI: 10.1002/bimj.202000345] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2020] [Revised: 09/24/2021] [Accepted: 11/07/2021] [Indexed: 12/17/2022]
Abstract
In the precision medicine era, (prespecified) subgroup analyses are an integral part of clinical trials. Incorporating multiple populations and hypotheses in the design and analysis plan, adaptive designs promise flexibility and efficiency in such trials. Adaptations include (unblinded) interim analyses (IAs) or blinded sample size reviews. An IA offers the possibility to select promising subgroups and reallocate sample size in further stages. Trials with these features are known as adaptive enrichment designs. Such complex designs comprise many nuisance parameters, such as prevalences of the subgroups and variances of the outcomes in the subgroups. Additionally, a number of design options including the timepoint of the sample size review and timepoint of the IA have to be selected. Here, for normally distributed endpoints, we propose a strategy combining blinded sample size recalculation and adaptive enrichment at an IA, that is, at an early timepoint nuisance parameters are reestimated and the sample size is adjusted while subgroup selection and enrichment is performed later. We discuss implications of different scenarios concerning the variances as well as the timepoints of blinded review and IA and investigate the design characteristics in simulations. The proposed method maintains the desired power if planning assumptions were inaccurate and reduces the sample size and variability of the final sample size when an enrichment is performed. Having two separate timepoints for blinded sample size review and IA improves the timing of the latter and increases the probability to correctly enrich a subgroup.
Collapse
Affiliation(s)
- Marius Placzek
- Department of Medical Statistics, University Medical Center Göttingen, Göttingen, Germany
| | - Tim Friede
- Department of Medical Statistics, University Medical Center Göttingen, Göttingen, Germany.,DZHK (German Center for Cardiovascular Research), partner site Göttingen, Göttingen, Germany
| |
Collapse
|
3
|
Herrmann C, Kieser M, Rauch G, Pilz M. Optimization of adaptive designs with respect to a performance score. Biom J 2022; 64:989-1006. [PMID: 35426460 DOI: 10.1002/bimj.202100166] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2021] [Revised: 02/09/2022] [Accepted: 02/12/2022] [Indexed: 11/08/2022]
Abstract
Adaptive designs are an increasingly popular method for the adaptation of design aspects in clinical trials, such as the sample size. Scoring different adaptive designs helps to make an appropriate choice among the numerous existing adaptive design methods. Several scores have been proposed to evaluate adaptive designs. Moreover, it is possible to determine optimal two-stage adaptive designs with respect to a customized objective score by solving a constrained optimization problem. In this paper, we use the conditional performance score by Herrmann et al. (2020) as the optimization criterion to derive optimal adaptive two-stage designs. We investigate variations of the original performance score, for example, by assigning different weights to the score components and by incorporating prior assumptions on the effect size. We further investigate a setting where the optimization framework is extended by a global power constraint, and additional optimization of the critical value function next to the stage-two sample size is performed. Those evaluations with respect to the sample size curves and the resulting design's performance can contribute to facilitate the score's usage in practice.
Collapse
Affiliation(s)
- Carolin Herrmann
- Institute of Biometry and Clinical Epidemiology, Charité - Universitätsmedizin Berlin, Berlin, Germany
| | - Meinhard Kieser
- Institute of Medical Biometry, University Hospital Heidelberg, Heidelberg, Germany
| | - Geraldine Rauch
- Institute of Biometry and Clinical Epidemiology, Charité - Universitätsmedizin Berlin, Berlin, Germany
| | - Maximilian Pilz
- Institute of Medical Biometry, University Hospital Heidelberg, Heidelberg, Germany
| |
Collapse
|
4
|
Herrmann C, Kluge C, Pilz M, Kieser M, Rauch G. Improving sample size recalculation in adaptive clinical trials by resampling. Pharm Stat 2021; 20:1035-1050. [PMID: 33792167 DOI: 10.1002/pst.2122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2020] [Revised: 12/16/2020] [Accepted: 03/08/2021] [Indexed: 11/11/2022]
Abstract
Sample size calculations in clinical trials need to be based on profound parameter assumptions. Wrong parameter choices may lead to too small or too high sample sizes and can have severe ethical and economical consequences. Adaptive group sequential study designs are one solution to deal with planning uncertainties. Here, the sample size can be updated during an ongoing trial based on the observed interim effect. However, the observed interim effect is a random variable and thus does not necessarily correspond to the true effect. One way of dealing with the uncertainty related to this random variable is to include resampling elements in the recalculation strategy. In this paper, we focus on clinical trials with a normally distributed endpoint. We consider resampling of the observed interim test statistic and apply this principle to several established sample size recalculation approaches. The resulting recalculation rules are smoother than the original ones and thus the variability in sample size is lower. In particular, we found that some resampling approaches mimic a group sequential design. In general, incorporating resampling of the interim test statistic in existing sample size recalculation rules results in a substantial performance improvement with respect to a recently published conditional performance score.
Collapse
Affiliation(s)
- Carolin Herrmann
- Charité - Universitätsmedizin Berlin, corporate member of Freie Universität Berlin and Humboldt-Universität zu Berlin, Institute of Biometry and Clinical Epidemiology, Charitéplatz 1, 10117 Berlin, Germany
| | - Corinna Kluge
- Charité - Universitätsmedizin Berlin, corporate member of Freie Universität Berlin and Humboldt-Universität zu Berlin, Institute of Biometry and Clinical Epidemiology, Charitéplatz 1, 10117 Berlin, Germany
| | - Maximilian Pilz
- Institute of Medical Biometry and Informatics, University Medical Center Ruprechts-Karls University Heidelberg, Heidelberg, Germany
| | - Meinhard Kieser
- Institute of Medical Biometry and Informatics, University Medical Center Ruprechts-Karls University Heidelberg, Heidelberg, Germany
| | - Geraldine Rauch
- Charité - Universitätsmedizin Berlin, corporate member of Freie Universität Berlin and Humboldt-Universität zu Berlin, Institute of Biometry and Clinical Epidemiology, Charitéplatz 1, 10117 Berlin, Germany
| |
Collapse
|
5
|
Herrmann C, Rauch G. Smoothing Corrections for Improving Sample Size Recalculation Rules in Adaptive Group Sequential Study Designs. Methods Inf Med 2021; 60:1-8. [PMID: 33648007 PMCID: PMC8432271 DOI: 10.1055/s-0040-1721727] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Background
An adequate sample size calculation is essential for designing a successful clinical trial. One way to tackle planning difficulties regarding parameter assumptions required for sample size calculation is to adapt the sample size during the ongoing trial.
This can be attained by adaptive group sequential study designs. At a predefined timepoint, the interim effect is tested for significance. Based on the interim test result, the trial is either stopped or continued with the possibility of a sample size recalculation. Objectives
Sample size recalculation rules have different limitations in application like a high variability of the recalculated sample size. Hence, the goal is to provide a tool to counteract this performance limitation.
Methods
Sample size recalculation rules can be interpreted as functions of the observed interim effect. Often, a “jump” from the first stage's sample size to the maximal sample size at a rather arbitrarily chosen interim effect size is implemented and the curve decreases monotonically afterwards. This jump is one reason for a high variability of the sample size. In this work, we investigate how the shape of the recalculation function can be improved by implementing a smoother increase of the sample size. The design options are evaluated by means of Monte Carlo simulations. Evaluation criteria are univariate performance measures such as the conditional power and sample size as well as a conditional performance score which combines these components.
Results
We demonstrate that smoothing corrections can reduce variability in conditional power and sample size as well as they increase the performance with respect to a recently published conditional performance score for medium and large standardized effect sizes.
Conclusion
Based on the simulation study, we present a tool that is easily implemented to improve sample size recalculation rules. The approach can be combined with existing sample size recalculation rules described in the literature.
Collapse
Affiliation(s)
- Carolin Herrmann
- Charité-Universitätsmedizin Berlin, Corporate member of Freie Universität Berlin, Humboldt-Universität zu Berlin, and Berlin Institute of Health, Institute of Biometry and Clinical Epidemiology, Berlin, Germany
| | - Geraldine Rauch
- Charité-Universitätsmedizin Berlin, Corporate member of Freie Universität Berlin, Humboldt-Universität zu Berlin, and Berlin Institute of Health, Institute of Biometry and Clinical Epidemiology, Berlin, Germany
| |
Collapse
|
6
|
Edwards JM, Walters SJ, Kunz C, Julious SA. A systematic review of the "promising zone" design. Trials 2020; 21:1000. [PMID: 33276810 PMCID: PMC7718653 DOI: 10.1186/s13063-020-04931-w] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2020] [Accepted: 11/25/2020] [Indexed: 12/01/2022] Open
Abstract
Introduction Sample size calculations require assumptions regarding treatment response and variability. Incorrect assumptions can result in under- or overpowered trials, posing ethical concerns. Sample size re-estimation (SSR) methods investigate the validity of these assumptions and increase the sample size if necessary. The “promising zone” (Mehta and Pocock, Stat Med 30:3267–3284, 2011) concept is appealing to researchers for its design simplicity. However, it is still relatively new in the application and has been a source of controversy. Objectives This research aims to synthesise current approaches and practical implementation of the promising zone design. Methods This systematic review comprehensively identifies the reporting of methodological research and of clinical trials using promising zone. Databases were searched according to a pre-specified search strategy, and pearl growing techniques implemented. Results The combined search methods resulted in 270 unique records identified; 171 were included in the review, of which 30 were trials. The median time to the interim analysis was 60% of the original target sample size (IQR 41–73%). Of the 15 completed trials, 7 increased their sample size. Only 21 studies reported the maximum sample size that would be considered, for which the median increase was 50% (IQR 35–100%). Conclusions Promising zone is being implemented in a range of trials worldwide, albeit in low numbers. Identifying trials using promising zone was difficult due to the lack of reporting of SSR methodology. Even when SSR methodology was reported, some had key interim analysis details missing, and only eight papers provided promising zone ranges.
Collapse
Affiliation(s)
- Julia M Edwards
- School of Health and Related Research, The University of Sheffield, Sheffield, UK.
| | - Stephen J Walters
- School of Health and Related Research, The University of Sheffield, Sheffield, UK
| | - Cornelia Kunz
- Boehringer Ingelheim, Biberach an der Riss, Biberach, Germany
| | - Steven A Julious
- School of Health and Related Research, The University of Sheffield, Sheffield, UK
| |
Collapse
|
7
|
Selvaratnam S, Kong L, Wiens DP. Model-robust designs for nonlinear quantile regression. Stat Methods Med Res 2020; 30:221-232. [PMID: 32812499 DOI: 10.1177/0962280220948159] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
We construct robust designs for nonlinear quantile regression, in the presence of both a possibly misspecified nonlinear quantile function and heteroscedasticity of an unknown form. The asymptotic mean-squared error of the quantile estimate is evaluated and maximized over a neighbourhood of the fitted quantile regression model. This maximum depends on the scale function and on the design. We entertain two methods to find designs that minimize the maximum loss. The first is local - we minimize for given values of the parameters and the scale function, using a sequential approach, whereby each new design point minimizes the subsequent loss, given the current design. The second is adaptive - at each stage, the maximized loss is evaluated at quantile estimates of the parameters, and a kernel estimate of scale, and then the next design point is obtained as in the sequential method. In the context of a Michaelis-Menten response model for an estrogen/hormone study, and a variety of scale functions, we demonstrate that the adaptive approach performs as well, in large study sizes, as if the parameter values and scale function were known beforehand and the sequential method applied. When the sequential method uses an incorrectly specified scale function, the adaptive method yields an, often substantial, improvement. The performance of the adaptive designs for smaller study sizes is assessed and seen to still be very favourable, especially so since the prior information required to design sequentially is rarely available.
Collapse
Affiliation(s)
| | - Linglong Kong
- Department of Mathematical and Statistical Sciences, University of Alberta, Alberta, Canada
| | - Douglas P Wiens
- Department of Mathematical and Statistical Sciences, University of Alberta, Alberta, Canada
| |
Collapse
|
8
|
Herrmann C, Pilz M, Kieser M, Rauch G. A new conditional performance score for the evaluation of adaptive group sequential designs with sample size recalculation. Stat Med 2020; 39:2067-2100. [DOI: 10.1002/sim.8534] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2018] [Revised: 12/20/2019] [Accepted: 03/04/2020] [Indexed: 11/06/2022]
Affiliation(s)
- Carolin Herrmann
- Institute of Biometry and Clinical Epidemiology Charité ‐ Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin, Humboldt‐Universität zu Berlin, Berlin Institute of Health Berlin Germany
- Berlin Institute of Health (BIH) Berlin Germany
| | - Maximilian Pilz
- Institute of Medical Biometry and Informatics University Medical Center Ruprechts‐Karls University Heidelberg Heidelberg Germany
| | - Meinhard Kieser
- Institute of Medical Biometry and Informatics University Medical Center Ruprechts‐Karls University Heidelberg Heidelberg Germany
| | - Geraldine Rauch
- Institute of Biometry and Clinical Epidemiology Charité ‐ Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin, Humboldt‐Universität zu Berlin, Berlin Institute of Health Berlin Germany
- Berlin Institute of Health (BIH) Berlin Germany
| |
Collapse
|
9
|
Huang Z, Samuelson F, Tcheuko L, Chen W. Adaptive designs in multi-reader multi-case clinical trials of imaging devices. Stat Methods Med Res 2019; 29:1592-1611. [PMID: 31456480 DOI: 10.1177/0962280219869370] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Evaluation of medical imaging devices often involves clinical studies where multiple readers (MR) read images of multiple cases (MC) for a clinical task, which are often called MRMC studies. In addition to sizing patient cases as is required in most clinical trials, MRMC studies also require sizing readers, since both readers and cases contribute to the uncertainty of the estimated diagnostic performance, which is often measured by the area under the ROC curve (AUC). Due to limited prior information, sizing of such a study is often unreliable. It is desired to adaptively resize the study toward a target power after an interim analysis. Although adaptive methods are available in clinical trials where only the patient sample is sized, such methodologies have not been established for MRMC studies. The challenge lies in the fact that there is a correlation structure in MRMC data and the sizing involves both readers and cases. We develop adaptive MRMC design methodologies to enable study resizing. In particular, we resize the study and adjust the critical value for hypothesis testing simultaneously after an interim analysis to achieve a target power and control the type I error rate in comparing AUCs of two modalities. Analytical results have been derived. Simulations show that the type I error rate is controlled close to the nominal level and the power is adjusted toward the target value under a variety of simulation conditions. We demonstrate the use of our methods in a real-world application comparing two imaging modalities for breast cancer detection.
Collapse
Affiliation(s)
- Zhipeng Huang
- Division of Imaging, Diagnostics, and Software Reliability, OSEL/CDRH, Food and Drug Administration, Silver Spring, MD, USA
| | - Frank Samuelson
- Division of Imaging, Diagnostics, and Software Reliability, OSEL/CDRH, Food and Drug Administration, Silver Spring, MD, USA
| | - Lucas Tcheuko
- Division of Population Health and Science, OS/CTP, Food and Drug Administration, Silver Spring, MD, USA
| | - Weijie Chen
- Division of Imaging, Diagnostics, and Software Reliability, OSEL/CDRH, Food and Drug Administration, Silver Spring, MD, USA
| |
Collapse
|
10
|
Jakobsen JC, Ovesen C, Winkel P, Hilden J, Gluud C, Wetterslev J. Power estimations for non-primary outcomes in randomised clinical trials. BMJ Open 2019; 9:e027092. [PMID: 31175197 PMCID: PMC6588976 DOI: 10.1136/bmjopen-2018-027092] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 12/13/2022] Open
Abstract
OBJECTIVE AND METHODS It is rare that trialists report power estimations of non-primary outcomes. In the present article, we will describe how to define a valid hierarchy of outcomes in a randomised clinical trial, to limit problems with Type I and Type II errors, using considerations on the clinical relevance of the outcomes and power estimations. CONCLUSION Power estimations of non-primary outcomes may guide trialists in classifying non-primary outcomes as secondary or exploratory. The power estimations are simple and if they are used systematically, more appropriate outcome hierarchies can be defined, and trial results will become more interpretable.
Collapse
Affiliation(s)
- Janus Christian Jakobsen
- Copenhagen Trial Unit, Centre for Clinical Intervention Research, Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
- Department of Cardiology, Holbæk Hospital, Holbæk, Denmark
| | - Christian Ovesen
- Department of Neurology, Bispebjerg Hospital, University of Copenhagen, Copenhagen, Denmark
| | - Per Winkel
- Copenhagen Trial Unit, Centre for Clinical Intervention Research, Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
| | - Jørgen Hilden
- Department of Public Health Research, Section of Biostatistics, University of Copenhagen, Copenhagen, Denmark
| | - Christian Gluud
- Copenhagen Trial Unit, Centre for Clinical Intervention Research, Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
| | - Jørn Wetterslev
- Copenhagen Trial Unit, Centre for Clinical Intervention Research, Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
| |
Collapse
|
11
|
McClure LA, Szychowski JM, Benavente O, Hart RG, Coffey CS. A post hoc evaluation of a sample size re-estimation in the Secondary Prevention of Small Subcortical Strokes study. Clin Trials 2016; 13:537-44. [PMID: 27094488 DOI: 10.1177/1740774516643689] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/21/2022]
Abstract
BACKGROUND/AIMS The use of adaptive designs has been increasing in randomized clinical trials. Sample size re-estimation is a type of adaptation in which nuisance parameters are estimated at an interim point in the trial and the sample size re-computed based on these estimates. The Secondary Prevention of Small Subcortical Strokes study was a randomized clinical trial assessing the impact of single- versus dual-antiplatelet therapy and control of systolic blood pressure to a higher (130-149 mmHg) versus lower (<130 mmHg) target on recurrent stroke risk in a two-by-two factorial design. A sample size re-estimation was performed during the Secondary Prevention of Small Subcortical Strokes study resulting in an increase from the planned sample size of 2500-3020, and we sought to determine the impact of the sample size re-estimation on the study results. METHODS We assessed the results of the primary efficacy and safety analyses with the full 3020 patients and compared them to the results that would have been observed had randomization ended with 2500 patients. The primary efficacy outcome considered was recurrent stroke, and the primary safety outcomes were major bleeds and death. We computed incidence rates for the efficacy and safety outcomes and used Cox proportional hazards models to examine the hazard ratios for each of the two treatment interventions (i.e. the antiplatelet and blood pressure interventions). RESULTS In the antiplatelet intervention, the hazard ratio was not materially modified by increasing the sample size, nor did the conclusions regarding the efficacy of mono versus dual-therapy change: there was no difference in the effect of dual- versus monotherapy on the risk of recurrent stroke hazard ratios (n = 3020 HR (95% confidence interval): 0.92 (0.72, 1.2), p = 0.48; n = 2500 HR (95% confidence interval): 1.0 (0.78, 1.3), p = 0.85). With respect to the blood pressure intervention, increasing the sample size resulted in less certainty in the results, as the hazard ratio for higher versus lower systolic blood pressure target approached, but did not achieve, statistical significance with the larger sample (n = 3020 HR (95% confidence interval): 0.81 (0.63, 1.0), p = 0.089; n = 2500 HR (95% confidence interval): 0.89 (0.68, 1.17), p = 0.40). The results from the safety analyses were similar to 3020 and 2500 patients for both study interventions. Other trial-related factors, such as contracts, finances, and study management, were impacted as well. CONCLUSION Adaptive designs can have benefits in randomized clinical trials, but do not always result in significant findings. The impact of adaptive designs should be measured in terms of both trial results, as well as practical issues related to trial management. More post hoc analyses of study adaptations will lead to better understanding of the balance between the benefits and the costs.
Collapse
Affiliation(s)
- Leslie A McClure
- Department of Biostatistics, University of Alabama at Birmingham, Birmingham, AL, USA
| | - Jeff M Szychowski
- Department of Biostatistics, University of Alabama at Birmingham, Birmingham, AL, USA
| | - Oscar Benavente
- Division of Neurology, University of British Columbia, Vancouver, BC, Canada
| | - Robert G Hart
- Department of Medicine, McMaster University, Hamilton, ON, Canada
| | | |
Collapse
|
12
|
Wan H, Ellenberg SS, Anderson KM. Stepwise two-stage sample size adaptation. Stat Med 2015; 34:27-38. [PMID: 25252082 DOI: 10.1002/sim.6311] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2013] [Revised: 08/14/2014] [Accepted: 09/02/2014] [Indexed: 11/10/2022]
Abstract
Several adaptive design methods have been proposed to reestimate sample size using the observed treatment effect after an initial stage of a clinical trial while preserving the overall type I error at the time of the final analysis. One unfortunate property of the algorithms used in some methods is that they can be inverted to reveal the exact treatment effect at the interim analysis. We propose using a step function with an inverted U-shape of observed treatment difference for sample size reestimation to lessen the information on treatment effect revealed. This will be referred to as stepwise two-stage sample size adaptation. This method applies calculation methods used for group sequential designs. We minimize expected sample size among a class of these designs and compare efficiency with the fully optimized two-stage design, optimal two-stage group sequential design, and designs based on promising conditional power. The trade-off between efficiency versus the improved blinding of the interim treatment effect will be discussed.
Collapse
Affiliation(s)
- Hong Wan
- Shire, 735 Chesterbrook Blvd., Wayne, 19087-5637, PA, U.S.A
| | | | | |
Collapse
|
13
|
Graf AC, Bauer P, Glimm E, Koenig F. Maximum type 1 error rate inflation in multiarmed clinical trials with adaptive interim sample size modifications. Biom J 2014; 56:614-30. [PMID: 24753160 PMCID: PMC4282114 DOI: 10.1002/bimj.201300153] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2013] [Revised: 01/20/2014] [Accepted: 01/22/2014] [Indexed: 11/24/2022]
Abstract
Sample size modifications in the interim analyses of an adaptive design can inflate the type 1 error rate, if test statistics and critical boundaries are used in the final analysis as if no modification had been made. While this is already true for designs with an overall change of the sample size in a balanced treatment-control comparison, the inflation can be much larger if in addition a modification of allocation ratios is allowed as well. In this paper, we investigate adaptive designs with several treatment arms compared to a single common control group. Regarding modifications, we consider treatment arm selection as well as modifications of overall sample size and allocation ratios. The inflation is quantified for two approaches: a naive procedure that ignores not only all modifications, but also the multiplicity issue arising from the many-to-one comparison, and a Dunnett procedure that ignores modifications, but adjusts for the initially started multiple treatments. The maximum inflation of the type 1 error rate for such types of design can be calculated by searching for the "worst case" scenarios, that are sample size adaptation rules in the interim analysis that lead to the largest conditional type 1 error rate in any point of the sample space. To show the most extreme inflation, we initially assume unconstrained second stage sample size modifications leading to a large inflation of the type 1 error rate. Furthermore, we investigate the inflation when putting constraints on the second stage sample sizes. It turns out that, for example fixing the sample size of the control group, leads to designs controlling the type 1 error rate.
Collapse
Affiliation(s)
- Alexandra C Graf
- Center for Medical Statistics, Informatics and Intelligent Systems, Medical University of ViennaSpitalgasse 23, 1090, Vienna, Austria
- Competence Center for Clinical Trials, University of BremenLinzer Strasse 4, 28359, Bremen, Germany
| | - Peter Bauer
- Center for Medical Statistics, Informatics and Intelligent Systems, Medical University of ViennaSpitalgasse 23, 1090, Vienna, Austria
| | - Ekkehard Glimm
- Novartis Pharma AG, Novartis Campus4056, Basel, Switzerland
| | - Franz Koenig
- Center for Medical Statistics, Informatics and Intelligent Systems, Medical University of ViennaSpitalgasse 23, 1090, Vienna, Austria
| |
Collapse
|
14
|
Levin GP, Emerson SC, Emerson SS. An evaluation of inferential procedures for adaptive clinical trial designs with pre-specified rules for modifying the sample size. Biometrics 2014; 70:556-67. [DOI: 10.1111/biom.12168] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/01/2013] [Revised: 02/01/2014] [Accepted: 03/01/2014] [Indexed: 11/27/2022]
Affiliation(s)
- Gregory P. Levin
- Department of Biostatistics; University of Washington; Seattle, Washington 98195 U.S.A
| | - Sarah C. Emerson
- Department of Statistics; Oregon State University; Corvallis, Oregon 97331 U.S.A
| | - Scott S. Emerson
- Department of Biostatistics; University of Washington; Seattle, Washington 98195 U.S.A
| |
Collapse
|
15
|
Jakobsen JC, Gluud C, Winkel P, Lange T, Wetterslev J. The thresholds for statistical and clinical significance - a five-step procedure for evaluation of intervention effects in randomised clinical trials. BMC Med Res Methodol 2014; 14:34. [PMID: 24588900 PMCID: PMC4015863 DOI: 10.1186/1471-2288-14-34] [Citation(s) in RCA: 120] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2014] [Accepted: 02/20/2014] [Indexed: 02/06/2023] Open
Abstract
BACKGROUND Thresholds for statistical significance are insufficiently demonstrated by 95% confidence intervals or P-values when assessing results from randomised clinical trials. First, a P-value only shows the probability of getting a result assuming that the null hypothesis is true and does not reflect the probability of getting a result assuming an alternative hypothesis to the null hypothesis is true. Second, a confidence interval or a P-value showing significance may be caused by multiplicity. Third, statistical significance does not necessarily result in clinical significance. Therefore, assessment of intervention effects in randomised clinical trials deserves more rigour in order to become more valid. METHODS Several methodologies for assessing the statistical and clinical significance of intervention effects in randomised clinical trials were considered. Balancing simplicity and comprehensiveness, a simple five-step procedure was developed. RESULTS For a more valid assessment of results from a randomised clinical trial we propose the following five-steps: (1) report the confidence intervals and the exact P-values; (2) report Bayes factor for the primary outcome, being the ratio of the probability that a given trial result is compatible with a 'null' effect (corresponding to the P-value) divided by the probability that the trial result is compatible with the intervention effect hypothesised in the sample size calculation; (3) adjust the confidence intervals and the statistical significance threshold if the trial is stopped early or if interim analyses have been conducted; (4) adjust the confidence intervals and the P-values for multiplicity due to number of outcome comparisons; and (5) assess clinical significance of the trial results. CONCLUSIONS If the proposed five-step procedure is followed, this may increase the validity of assessments of intervention effects in randomised clinical trials.
Collapse
Affiliation(s)
- Janus Christian Jakobsen
- Copenhagen Trial Unit, Centre for Clinical Intervention Research, Department 7812 Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
- Emergency Department, Holbæk Hospital, Holbæk, Denmark
| | - Christian Gluud
- Copenhagen Trial Unit, Centre for Clinical Intervention Research, Department 7812 Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
| | - Per Winkel
- Copenhagen Trial Unit, Centre for Clinical Intervention Research, Department 7812 Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
| | - Theis Lange
- Department of Biostatistics, Faculty of Health Sciences, University of Copenhagen, Copenhagen, Denmark
| | - Jørn Wetterslev
- Copenhagen Trial Unit, Centre for Clinical Intervention Research, Department 7812 Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
| |
Collapse
|