76
|
Urach S, Gaasterland C, Posch M, Jilma B, Roes K, Rosenkranz G, Van der Lee JH, Ristl R. Statistical analysis of Goal Attainment Scaling endpoints in randomised trials. Stat Methods Med Res 2018; 28:1893-1910. [PMID: 29921167 PMCID: PMC6566461 DOI: 10.1177/0962280218777896] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Goal Attainment Scaling is an assessment instrument to evaluate interventions on the basis of individual, patient-specific goals. The attainment of these goals is mapped in a pre-specified way to attainment levels on an ordinal scale, which is common to all goals. This approach is patient-centred and allows one to integrate the outcomes of patients with very heterogeneous symptoms. The latter is of particular importance in clinical trials in rare diseases because it enables larger sample sizes by including a broader patient population. In this paper, we focus on the statistical analysis of Goal Attainment Scaling outcomes for the comparison of two treatments in randomised clinical trials. Building on a general statistical model, we investigate the properties of different hypothesis testing approaches. Additionally, we propose a latent variable approach to generate Goal Attainment Scaling data in a simulation study, to assess the impact of model parameters such as the number of goals per patient and their correlation, the choice of discretisation thresholds and the type of design (parallel group or cross-over). Based on our findings, we give recommendations for the design of clinical trials with a Goal Attainment Scaling endpoint. Furthermore, we discuss an application of Goal Attainment Scaling in a clinical trial in mastocytosis.
Collapse
|
77
|
Graf AC, Wassmer G, Friede T, Gera RG, Posch M. Robustness of testing procedures for confirmatory subpopulation analyses based on a continuous biomarker. Stat Methods Med Res 2018; 28:1879-1892. [PMID: 29888651 PMCID: PMC6566459 DOI: 10.1177/0962280218777538] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
With the advent of personalized medicine, clinical trials studying treatment effects in subpopulations are receiving increasing attention. The objectives of such studies are, besides demonstrating a treatment effect in the overall population, to identify subpopulations, based on biomarkers, where the treatment has a beneficial effect. Continuous biomarkers are often dichotomized using a threshold to define two subpopulations with low and high biomarker levels. If there is insufficient information on the dependence structure of the outcome on the biomarker, several thresholds may be investigated. The nested structure of such subpopulations is similar to the structure in group sequential trials. Therefore, it has been proposed to use the corresponding critical boundaries to test such nested subpopulations. We show that for biomarkers with a prognostic effect that is not adjusted for in the statistical model, the variability of the outcome may vary across subpopulations which may lead to an inflation of the family-wise type 1 error rate. Using simulations we quantify the potential inflation of testing procedures based on group sequential designs. Furthermore, alternative hypotheses tests that control the family-wise type 1 error rate under minimal assumptions are proposed. The methodological approaches are illustrated by a trial in depression.
Collapse
|
78
|
Ristl R, Xi D, Glimm E, Posch M. Optimal exact tests for multiple binary endpoints. Comput Stat Data Anal 2018. [DOI: 10.1016/j.csda.2018.01.001] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
79
|
Abstract
Blinded sample size reassessment is a popular means to control the power in clinical trials if no reliable information on nuisance parameters is available in the planning phase. We investigate how sample size reassessment based on blinded interim data affects the properties of point estimates and confidence intervals for parallel group superiority trials comparing the means of a normal endpoint. We evaluate the properties of two standard reassessment rules that are based on the sample size formula of the z-test, derive the worst case reassessment rule that maximizes the absolute mean bias and obtain an upper bound for the mean bias of the treatment effect estimate.
Collapse
|
80
|
Pearce M, Hee SW, Madan J, Posch M, Day S, Miller F, Zohar S, Stallard N. Value of information methods to design a clinical trial in a small population to optimise a health economic utility function. BMC Med Res Methodol 2018; 18:20. [PMID: 29422021 PMCID: PMC5806391 DOI: 10.1186/s12874-018-0475-0] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2017] [Accepted: 01/14/2018] [Indexed: 01/20/2023] Open
Abstract
Background Most confirmatory randomised controlled clinical trials (RCTs) are designed with specified power, usually 80% or 90%, for a hypothesis test conducted at a given significance level, usually 2.5% for a one-sided test. Approval of the experimental treatment by regulatory agencies is then based on the result of such a significance test with other information to balance the risk of adverse events against the benefit of the treatment to future patients. In the setting of a rare disease, recruiting sufficient patients to achieve conventional error rates for clinically reasonable effect sizes may be infeasible, suggesting that the decision-making process should reflect the size of the target population. Methods We considered the use of a decision-theoretic value of information (VOI) method to obtain the optimal sample size and significance level for confirmatory RCTs in a range of settings. We assume the decision maker represents society. For simplicity we assume the primary endpoint to be normally distributed with unknown mean following some normal prior distribution representing information on the anticipated effectiveness of the therapy available before the trial. The method is illustrated by an application in an RCT in haemophilia A. We explicitly specify the utility in terms of improvement in primary outcome and compare this with the costs of treating patients, both financial and in terms of potential harm, during the trial and in the future. Results The optimal sample size for the clinical trial decreases as the size of the population decreases. For non-zero cost of treating future patients, either monetary or in terms of potential harmful effects, stronger evidence is required for approval as the population size increases, though this is not the case if the costs of treating future patients are ignored. Conclusions Decision-theoretic VOI methods offer a flexible approach with both type I error rate and power (or equivalently trial sample size) depending on the size of the future population for whom the treatment under investigation is intended. This might be particularly suitable for small populations when there is considerable information about the patient population. Electronic supplementary material The online version of this article (10.1186/s12874-018-0475-0) contains supplementary material, which is available to authorized users.
Collapse
|
81
|
Abstract
AbstractIn this overview we introduce the basic ideas behind a new flexible approach in sequential designs. The different concepts based on two-stage combination tests and conditional error functions are brought together. We sketch the construction of p-values, confidence intervals, and median unbiased estimates. Finally, recursive combination tests are introduced which extend the flexibility to the choice of the number of interim analyses.
Collapse
|
82
|
Hofer MP, Hedman H, Mavris M, Koenig F, Vetter T, Posch M, Vamvakas S, Regnstrom J, Aarum S. Marketing authorisation of orphan medicines in Europe from 2000 to 2013. Drug Discov Today 2018; 23:424-433. [DOI: 10.1016/j.drudis.2017.10.012] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2017] [Revised: 09/28/2017] [Accepted: 10/13/2017] [Indexed: 01/12/2023]
|
83
|
Miller F, Zohar S, Stallard N, Madan J, Posch M, Hee SW, Pearce M, Vågerö M, Day S. Approaches to sample size calculation for clinical trials in rare diseases. Pharm Stat 2018; 17:214-230. [PMID: 29322632 DOI: 10.1002/pst.1848] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2017] [Revised: 09/05/2017] [Accepted: 12/08/2017] [Indexed: 01/27/2023]
Abstract
We discuss 3 alternative approaches to sample size calculation: traditional sample size calculation based on power to show a statistically significant effect, sample size calculation based on assurance, and sample size based on a decision-theoretic approach. These approaches are compared head-to-head for clinical trial situations in rare diseases. Specifically, we consider 3 case studies of rare diseases (Lyell disease, adult-onset Still disease, and cystic fibrosis) with the aim to plan the sample size for an upcoming clinical trial. We outline in detail the reasonable choice of parameters for these approaches for each of the 3 case studies and calculate sample sizes. We stress that the influence of the input parameters needs to be investigated in all approaches and recommend investigating different sample size approaches before deciding finally on the trial size. Highly influencing for the sample size are choice of treatment effect parameter in all approaches and the parameter for the additional cost of the new treatment in the decision-theoretic approach. These should therefore be discussed extensively.
Collapse
|
84
|
Ondra T, Jobjörnsson S, Beckman RA, Burman CF, König F, Stallard N, Posch M. Optimized adaptive enrichment designs. Stat Methods Med Res 2017; 28:2096-2111. [PMID: 29254436 PMCID: PMC6613177 DOI: 10.1177/0962280217747312] [Citation(s) in RCA: 22] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/16/2023]
Abstract
Based on a Bayesian decision theoretic approach, we optimize frequentist single-
and adaptive two-stage trial designs for the development of targeted therapies,
where in addition to an overall population, a pre-defined subgroup is
investigated. In such settings, the losses and gains of decisions can be
quantified by utility functions that account for the preferences of different
stakeholders. In particular, we optimize expected utilities from the
perspectives both of a commercial sponsor, maximizing the net present value, and
also of the society, maximizing cost-adjusted expected health benefits of a new
treatment for a specific population. We consider single-stage and adaptive
two-stage designs with partial enrichment, where the proportion of patients
recruited from the subgroup is a design parameter. For the adaptive designs, we
use a dynamic programming approach to derive optimal adaptation rules. The
proposed designs are compared to trials which are non-enriched (i.e. the
proportion of patients in the subgroup corresponds to the prevalence in the
underlying population). We show that partial enrichment designs can
substantially improve the expected utilities. Furthermore, adaptive partial
enrichment designs are more robust than single-stage designs and retain high
expected utilities even if the expected utilities are evaluated under a
different prior than the one used in the optimization. In addition, we find that
trials optimized for the sponsor utility function have smaller sample sizes
compared to trials optimized under the societal view and may include the overall
population (with patients from the complement of the subgroup) even if there is
substantial evidence that the therapy is only effective in the subgroup.
Collapse
|
85
|
Eskandary F, Regele H, Baumann L, Bond G, Kozakowski N, Wahrmann M, Hidalgo LG, Haslacher H, Kaltenecker CC, Aretin MB, Oberbauer R, Posch M, Staudenherz A, Handisurya A, Reeve J, Halloran PF, Böhmig GA. A Randomized Trial of Bortezomib in Late Antibody-Mediated Kidney Transplant Rejection. J Am Soc Nephrol 2017; 29:591-605. [PMID: 29242250 DOI: 10.1681/asn.2017070818] [Citation(s) in RCA: 178] [Impact Index Per Article: 25.4] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2017] [Accepted: 09/20/2017] [Indexed: 01/03/2023] Open
Abstract
Late antibody-mediated rejection (ABMR) is a leading cause of kidney allograft failure. Uncontrolled studies have suggested efficacy of the proteasome inhibitor bortezomib, but no systematic trial has been undertaken to support its use in ABMR. In this randomized, placebo-controlled trial (the Bortezomib in Late Antibody-Mediated Kidney Transplant Rejection [BORTEJECT] Trial), we investigated whether two cycles of bortezomib (each cycle: 1.3 mg/m2 intravenously on days 1, 4, 8, and 11) prevent GFR decline by halting the progression of late donor-specific antibody (DSA)-positive ABMR. Forty-four DSA-positive kidney transplant recipients with characteristic ABMR morphology (median time after transplant, 5.0 years; pretransplant DSA documented in 19 recipients), who were identified on cross-sectional screening of 741 patients, were randomly assigned to receive bortezomib (n=21) or placebo (n=23). The 0.5-ml/min per 1.73 m2 per year (95% confidence interval, -4.8 to 5.8) difference detected between bortezomib and placebo in eGFR slope (primary end point) was not significant (P=0.86). We detected no significant differences between bortezomib- and placebo-treated groups in median measured GFR at 24 months (33 versus 42 ml/min per 1.73 m2; P=0.31), 2-year graft survival (81% versus 96%; P=0.12), urinary protein concentration, DSA levels, or morphologic or molecular rejection phenotypes in 24-month follow-up biopsy specimens. Bortezomib, however, associated with gastrointestinal and hematologic toxicity. In conclusion, our trial failed to show that bortezomib prevents GFR loss, improves histologic or molecular disease features, or reduces DSA, despite significant toxicity. Our results reinforce the need for systematic trials to dissect the efficiency and safety of new treatments for late ABMR.
Collapse
|
86
|
Zajic P, Bauer P, Rhodes A, Moreno R, Fellinger T, Metnitz B, Stavropoulou F, Posch M, Metnitz PGH. Weekends affect mortality risk and chance of discharge in critically ill patients: a retrospective study in the Austrian registry for intensive care. CRITICAL CARE : THE OFFICIAL JOURNAL OF THE CRITICAL CARE FORUM 2017; 21:223. [PMID: 28877753 PMCID: PMC5588748 DOI: 10.1186/s13054-017-1812-0] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/30/2017] [Accepted: 08/07/2017] [Indexed: 12/24/2022]
Abstract
Background In this study, we primarily investigated whether ICU admission or ICU stay at weekends (Saturday and Sunday) is associated with a different risk of ICU mortality or chance of ICU discharge than ICU admission or ICU stay on weekdays (Monday to Friday). Secondarily, we analysed whether weekend ICU admission or ICU stay influences risk of hospital mortality or chance of hospital discharge. Methods A retrospective study was performed for all adult patients admitted to 119 ICUs participating in the benchmarking project of the Austrian Centre for Documentation and Quality Assurance in Intensive Care (ASDI) between 2012 and 2015. Readmissions to the ICU during the same hospital stay were excluded. Results In a multivariable competing risk analysis, a strong weekend effect was observed. Patients admitted to ICUs on Saturday or Sunday had a higher mortality risk after adjustment for severity of illness by Simplified Acute Physiology Score (SAPS) 3, year, month of the year, type of admission, ICU, and weekday of death or discharge. Hazard ratios (95% confidence interval) for death in the ICU following admission on a Saturday or Sunday compared with Wednesday were 1.15 (1.08–1.23) and 1.11 (1.03–1.18), respectively. Lower hazard ratios were observed for dying on a Saturday (0.93 (0.87–1.00)) or Sunday (0.85 (0.80–0.91)) compared with Wednesday. This is probably related to the reduced chance of being discharged from the ICU at the weekend (0.63 (0.62–064) for Saturday and 0.56 (0.55–0.57) for Sunday). Similar results were found for hospital mortality and hospital discharge following ICU admission. Conclusions Patients admitted to ICUs at weekends are at increased risk of death in both the ICU and the hospital even after rigorous adjustment for severity of illness. Conversely, death in the ICU and discharge from the ICU are significantly less likely at weekends. Electronic supplementary material The online version of this article (doi:10.1186/s13054-017-1812-0) contains supplementary material, which is available to authorized users.
Collapse
|
87
|
Musuamba FT, Manolis E, Holford N, Cheung S, Friberg LE, Ogungbenro K, Posch M, Yates J, Berry S, Thomas N, Corriol-Rohou S, Bornkamp B, Bretz F, Hooker AC, Van der Graaf PH, Standing JF, Hay J, Cole S, Gigante V, Karlsson K, Dumortier T, Benda N, Serone F, Das S, Brochot A, Ehmann F, Hemmings R, Rusten IS. Advanced Methods for Dose and Regimen Finding During Drug Development: Summary of the EMA/EFPIA Workshop on Dose Finding (London 4-5 December 2014). CPT-PHARMACOMETRICS & SYSTEMS PHARMACOLOGY 2017; 6:418-429. [PMID: 28722322 PMCID: PMC5529745 DOI: 10.1002/psp4.12196] [Citation(s) in RCA: 41] [Impact Index Per Article: 5.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/21/2016] [Revised: 03/27/2017] [Accepted: 03/27/2017] [Indexed: 02/05/2023]
Abstract
Inadequate dose selection for confirmatory trials is currently still one of the most challenging issues in drug development, as illustrated by high rates of late‐stage attritions in clinical development and postmarketing commitments required by regulatory institutions. In an effort to shift the current paradigm in dose and regimen selection and highlight the availability and usefulness of well‐established and regulatory‐acceptable methods, the European Medicines Agency (EMA) in collaboration with the European Federation of Pharmaceutical Industries Association (EFPIA) hosted a multistakeholder workshop on dose finding (London 4–5 December 2014). Some methodologies that could constitute a toolkit for drug developers and regulators were presented. These methods are described in the present report: they include five advanced methods for data analysis (empirical regression models, pharmacometrics models, quantitative systems pharmacology models, MCP‐Mod, and model averaging) and three methods for study design optimization (Fisher information matrix (FIM)‐based methods, clinical trial simulations, and adaptive studies). Pairwise comparisons were also discussed during the workshop; however, mostly for historical reasons. This paper discusses the added value and limitations of these methods as well as challenges for their implementation. Some applications in different therapeutic areas are also summarized, in line with the discussions at the workshop. There was agreement at the workshop on the fact that selection of dose for phase III is an estimation problem and should not be addressed via hypothesis testing. Dose selection for phase III trials should be informed by well‐designed dose‐finding studies; however, the specific choice of method(s) will depend on several aspects and it is not possible to recommend a generalized decision tree. There are many valuable methods available, the methods are not mutually exclusive, and they should be used in conjunction to ensure a scientifically rigorous understanding of the dosing rationale.
Collapse
|
88
|
Stallard N, Miller F, Day S, Hee SW, Madan J, Zohar S, Posch M. Determination of the optimal sample size for a clinical trial accounting for the population size. Biom J 2017; 59:609-625. [PMID: 27184938 PMCID: PMC5516263 DOI: 10.1002/bimj.201500228] [Citation(s) in RCA: 27] [Impact Index Per Article: 3.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2015] [Revised: 02/10/2016] [Accepted: 03/09/2016] [Indexed: 11/28/2022]
Abstract
The problem of choosing a sample size for a clinical trial is a very common one. In some settings, such as rare diseases or other small populations, the large sample sizes usually associated with the standard frequentist approach may be infeasible, suggesting that the sample size chosen should reflect the size of the population under consideration. Incorporation of the population size is possible in a decision-theoretic approach either explicitly by assuming that the population size is fixed and known, or implicitly through geometric discounting of the gain from future patients reflecting the expected population size. This paper develops such approaches. Building on previous work, an asymptotic expression is derived for the sample size for single and two-arm clinical trials in the general case of a clinical trial with a primary endpoint with a distribution of one parameter exponential family form that optimizes a utility function that quantifies the cost and gain per patient as a continuous function of this parameter. It is shown that as the size of the population, N, or expected size, N∗ in the case of geometric discounting, becomes large, the optimal trial size is O(N1/2) or O(N∗1/2). The sample size obtained from the asymptotic expression is also compared with the exact optimal sample size in examples with responses with Bernoulli and Poisson distributions, showing that the asymptotic approximations can also be reasonable in relatively small sample sizes.
Collapse
|
89
|
Hee SW, Willis A, Tudur Smith C, Day S, Miller F, Madan J, Posch M, Zohar S, Stallard N. Does the low prevalence affect the sample size of interventional clinical trials of rare diseases? An analysis of data from the aggregate analysis of clinicaltrials.gov. Orphanet J Rare Dis 2017; 12:44. [PMID: 28253932 PMCID: PMC5335492 DOI: 10.1186/s13023-017-0597-1] [Citation(s) in RCA: 30] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2016] [Accepted: 02/14/2017] [Indexed: 11/20/2022] Open
Abstract
BACKGROUND Clinical trials are typically designed using the classical frequentist framework to constrain type I and II error rates. Sample sizes required in such designs typically range from hundreds to thousands of patients which can be challenging for rare diseases. It has been shown that rare disease trials have smaller sample sizes than non-rare disease trials. Indeed some orphan drugs were approved by the European Medicines Agency based on studies with as few as 12 patients. However, some studies supporting marketing authorisation included several hundred patients. In this work, we explore the relationship between disease prevalence and other factors and the size of interventional phase 2 and 3 rare disease trials conducted in the US and/or EU. We downloaded all clinical trials from Aggregate Analysis of ClinialTrials.gov (AACT) and identified rare disease trials by cross-referencing MeSH terms in AACT with the list from Orphadata. We examined the effects of prevalence and phase of study in a multiple linear regression model adjusting for other statistically significant trial characteristics. RESULTS Of 186941 ClinicalTrials.gov trials only 1567 (0.8%) studied a single rare condition with prevalence information from Orphadata. There were 19 (1.2%) trials studying disease with prevalence <1/1,000,000, 126 (8.0%) trials with 1-9/1,000,000, 791 (50.5%) trials with 1-9/100,000 and 631 (40.3%) trials with 1-5/10,000. Of the 1567 trials, 1160 (74%) were phase 2 trials. The fitted mean sample size for the rarest disease (prevalence <1/1,000,000) in phase 2 trials was the lowest (mean, 15.7; 95% CI, 8.7-28.1) but were similar across all the other prevalence classes; mean, 26.2 (16.1-42.6), 33.8 (22.1-51.7) and 35.6 (23.3-54.3) for prevalence 1-9/1,000,000, 1-9/100,000 and 1-5/10,000, respectively. Fitted mean size of phase 3 trials of rarer diseases, <1/1,000,000 (19.2, 6.9-53.2) and 1-9/1,000,000 (33.1, 18.6-58.9), were similar to those in phase 2 but were statistically significant lower than the slightly less rare diseases, 1-9/100,000 (75.3, 48.2-117.6) and 1-5/10,000 (77.7, 49.6-121.8), trials. CONCLUSIONS We found that prevalence was associated with the size of phase 3 trials with trials of rarer diseases noticeably smaller than the less rare diseases trials where phase 3 rarer disease (prevalence <1/100,000) trials were more similar in size to those for phase 2 but were larger than those for phase 2 in the less rare disease (prevalence ≥1/100,000) trials.
Collapse
|
90
|
Eichler H, Bloechl‐Daum B, Bauer P, Bretz F, Brown J, Hampson LV, Honig P, Krams M, Leufkens H, Lim R, Lumpkin MM, Murphy MJ, Pignatti F, Posch M, Schneeweiss S, Trusheim M, Koenig F. "Threshold-crossing": A Useful Way to Establish the Counterfactual in Clinical Trials? Clin Pharmacol Ther 2016; 100:699-712. [PMID: 27650716 PMCID: PMC5114686 DOI: 10.1002/cpt.515] [Citation(s) in RCA: 53] [Impact Index Per Article: 6.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2016] [Revised: 09/15/2016] [Accepted: 09/16/2016] [Indexed: 12/15/2022]
Abstract
A central question in the assessment of benefit/harm of new treatments is: how does the average outcome on the new treatment (the factual) compare to the average outcome had patients received no treatment or a different treatment known to be effective (the counterfactual)? Randomized controlled trials (RCTs) are the standard for comparing the factual with the counterfactual. Recent developments necessitate and enable a new way of determining the counterfactual for some new medicines. For select situations, we propose a new framework for evidence generation, which we call "threshold-crossing." This framework leverages the wealth of information that is becoming available from completed RCTs and from real world data sources. Relying on formalized procedures, information gleaned from these data is used to estimate the counterfactual, enabling efficacy assessment of new drugs. We propose future (research) activities to enable "threshold-crossing" for carefully selected products and indications in which RCTs are not feasible.
Collapse
|
91
|
Ondra T, Jobjörnsson S, Beckman RA, Burman CF, König F, Stallard N, Posch M. Optimizing Trial Designs for Targeted Therapies. PLoS One 2016; 11:e0163726. [PMID: 27684573 PMCID: PMC5042421 DOI: 10.1371/journal.pone.0163726] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2016] [Accepted: 08/17/2016] [Indexed: 11/21/2022] Open
Abstract
An important objective in the development of targeted therapies is to identify the populations where the treatment under consideration has positive benefit risk balance. We consider pivotal clinical trials, where the efficacy of a treatment is tested in an overall population and/or in a pre-specified subpopulation. Based on a decision theoretic framework we derive optimized trial designs by maximizing utility functions. Features to be optimized include the sample size and the population in which the trial is performed (the full population or the targeted subgroup only) as well as the underlying multiple test procedure. The approach accounts for prior knowledge of the efficacy of the drug in the considered populations using a two dimensional prior distribution. The considered utility functions account for the costs of the clinical trial as well as the expected benefit when demonstrating efficacy in the different subpopulations. We model utility functions from a sponsor's as well as from a public health perspective, reflecting actual civil interests. Examples of optimized trial designs obtained by numerical optimization are presented for both perspectives.
Collapse
|
92
|
Urach S, Posch M. Multi-arm group sequential designs with a simultaneous stopping rule. Stat Med 2016; 35:5536-5550. [PMID: 27550822 PMCID: PMC5157767 DOI: 10.1002/sim.7077] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/18/2016] [Revised: 07/01/2016] [Accepted: 07/28/2016] [Indexed: 11/08/2022]
Abstract
Multi‐arm group sequential clinical trials are efficient designs to compare multiple treatments to a control. They allow one to test for treatment effects already in interim analyses and can have a lower average sample number than fixed sample designs. Their operating characteristics depend on the stopping rule: We consider simultaneous stopping, where the whole trial is stopped as soon as for any of the arms the null hypothesis of no treatment effect can be rejected, and separate stopping, where only recruitment to arms for which a significant treatment effect could be demonstrated is stopped, but the other arms are continued. For both stopping rules, the family‐wise error rate can be controlled by the closed testing procedure applied to group sequential tests of intersection and elementary hypotheses. The group sequential boundaries for the separate stopping rule also control the family‐wise error rate if the simultaneous stopping rule is applied. However, we show that for the simultaneous stopping rule, one can apply improved, less conservative stopping boundaries for local tests of elementary hypotheses. We derive corresponding improved Pocock and O'Brien type boundaries as well as optimized boundaries to maximize the power or average sample number and investigate the operating characteristics and small sample properties of the resulting designs. To control the power to reject at least one null hypothesis, the simultaneous stopping rule requires a lower average sample number than the separate stopping rule. This comes at the cost of a lower power to reject all null hypotheses. Some of this loss in power can be regained by applying the improved stopping boundaries for the simultaneous stopping rule. The procedures are illustrated with clinical trials in systemic sclerosis and narcolepsy. © 2016 The Authors. Statistics in Medicine Published by John Wiley & Sons Ltd.
Collapse
|
93
|
Ristl R, Frommlet F, Koch A, Posch M. Fallback tests for co-primary endpoints. Stat Med 2016; 35:2669-86. [PMID: 26919166 PMCID: PMC5069608 DOI: 10.1002/sim.6911] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2014] [Revised: 01/22/2016] [Accepted: 01/27/2016] [Indexed: 11/17/2022]
Abstract
When efficacy of a treatment is measured by co-primary endpoints, efficacy is claimed only if for each endpoint an individual statistical test is significant at level α. While such a strategy controls the family-wise type I error rate (FWER), it is often strictly conservative and allows for no inference if not all null hypotheses can be rejected. In this paper, we investigate fallback tests, which are defined as uniform improvements of the classical test for co-primary endpoints. They reject whenever the classical test rejects but allow for inference also in settings where only a subset of endpoints show a significant effect. Similarly to the fallback tests for hierarchical testing procedures, these fallback tests for co-primary endpoints allow one to continue testing even if the primary objective of the trial was not met. We propose examples of fallback tests for two and three co-primary endpoints that control the FWER in the strong sense under the assumption of multivariate normal test statistics with arbitrary correlation matrix and investigate their power in a simulation study. The fallback procedures for co-primary endpoints are illustrated with a clinical trial in a rare disease and a diagnostic trial. © 2016 The Authors. Statistics in Medicine published by John Wiley & Sons Ltd.
Collapse
|
94
|
Dörner T, Posch M, Wagner F, Hüser A, Fischer T, Mooney L, Petricoul O, Maguire P, Pal P, Doucet J, Cabanski M, Kamphausen E, Oliver S. THU0313 Double-Blind, Randomized Study of VAY736 Single Dose Treatment in Patients with Primary Sjögren's Syndrome (PSS). Ann Rheum Dis 2016. [DOI: 10.1136/annrheumdis-2016-eular.5840] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
|
95
|
Cecconi M, Hochrieser H, Chew M, Grocott M, Hoeft A, Hoste A, Jammer I, Posch M, Metnitz P, Pelosi P, Moreno R, Pearse RM, Vincent JL, Rhodes A. Preoperative abnormalities in serum sodium concentrations are associated with higher in-hospital mortality in patients undergoing major surgery. Br J Anaesth 2016; 116:63-9. [PMID: 26675950 DOI: 10.1093/bja/aev373] [Citation(s) in RCA: 31] [Impact Index Per Article: 3.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/20/2023] Open
Abstract
BACKGROUND Abnormal serum sodium concentrations are common in patients presenting for surgery. It remains unclear whether these abnormalities are independent risk factors for postoperative mortality. METHODS This is a secondary analysis of the European Surgical Outcome Study (EuSOS) that provided data describing 46 539 patients undergoing inpatient non-cardiac surgery. Patients were included in this study if they had a recorded value of preoperative serum sodium within the 28 days immediately before surgery. Data describing preoperative risk factors and serum sodium concentrations were analysed to investigate the relationship with in-hospital mortality using univariate and multivariate logistic regression techniques. RESULTS Of 35 816 (77.0%) patients from the EuSOS database, 21 943 (61.3%) had normal values of serum sodium (138-142 mmol litre(-1)) before surgery, 8538 (23.8%) had hyponatraemia (serum sodium ≤137 mmol litre(-1)) and 5335 (14.9%) had hypernatraemia (serum sodium ≥143 mmol litre(-1)). After adjustment for potential confounding factors, moderate to severe hypernatraemia (serum sodium concentration ≥150 mmol litre(-1)) was independently associated with mortality [odds ratio 3.4 (95% confidence interval 2.0-6.0), P<0.0001]. Hyponatraemia was not associated with mortality. CONCLUSIONS Preoperative abnormalities in serum sodium concentrations are common, and hypernatraemia is associated with increased mortality after surgery. Abnormalities of serum sodium concentration may be an important biomarker of perioperative risk resulting from co-morbid disease.
Collapse
|
96
|
Heinze G, Michiels S, Posch M. Preface. Stat Med 2016; 35:965. [DOI: 10.1002/sim.6846] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2015] [Accepted: 11/23/2015] [Indexed: 11/11/2022]
|
97
|
Unkel S, Röver C, Stallard N, Benda N, Posch M, Zohar S, Friede T. Systematic reviews in paediatric multiple sclerosis and Creutzfeldt-Jakob disease exemplify shortcomings in methods used to evaluate therapies in rare conditions. Orphanet J Rare Dis 2016; 11:16. [PMID: 26897367 PMCID: PMC4761188 DOI: 10.1186/s13023-016-0402-6] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2015] [Accepted: 02/12/2016] [Indexed: 01/04/2023] Open
Abstract
BACKGROUND Randomized controlled trials (RCTs) are the gold standard design of clinical research to assess interventions. However, RCTs cannot always be applied for practical or ethical reasons. To investigate the current practices in rare diseases, we review evaluations of therapeutic interventions in paediatric multiple sclerosis (MS) and Creutzfeldt-Jakob disease (CJD). In particular, we shed light on the endpoints used, the study designs implemented and the statistical methodologies applied. METHODS We conducted literature searches to identify relevant primary studies. Data on study design, objectives, endpoints, patient characteristics, randomization and masking, type of intervention, control, withdrawals and statistical methodology were extracted from the selected studies. The risk of bias and the quality of the studies were assessed. RESULTS Twelve (seven) primary studies on paediatric MS (CJD) were included in the qualitative synthesis. No double-blind, randomized placebo-controlled trial for evaluating interventions in paediatric MS has been published yet. Evidence from one open-label RCT is available. The observational studies are before-after studies or controlled studies. Three of the seven selected studies on CJD are RCTs, of which two received the maximum mark on the Oxford Quality Scale. Four trials are controlled observational studies. CONCLUSIONS Evidence from double-blind RCTs on the efficacy of treatments appears to be variable between rare diseases. With regard to paediatric conditions it remains to be seen what impact regulators will have through e.g., paediatric investigation plans. Overall, there is space for improvement by using innovative trial designs and data analysis techniques.
Collapse
|
98
|
Magirr D, Jaki T, Koenig F, Posch M. Sample Size Reassessment and Hypothesis Testing in Adaptive Survival Trials. PLoS One 2016; 11:e0146465. [PMID: 26863139 PMCID: PMC4749572 DOI: 10.1371/journal.pone.0146465] [Citation(s) in RCA: 26] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2015] [Accepted: 12/17/2015] [Indexed: 11/18/2022] Open
Abstract
Mid-study design modifications are becoming increasingly accepted in confirmatory clinical trials, so long as appropriate methods are applied such that error rates are controlled. It is therefore unfortunate that the important case of time-to-event endpoints is not easily handled by the standard theory. We analyze current methods that allow design modifications to be based on the full interim data, i.e., not only the observed event times but also secondary endpoint and safety data from patients who are yet to have an event. We show that the final test statistic may ignore a substantial subset of the observed event times. An alternative test incorporating all event times is found, where a conservative assumption must be made in order to guarantee type I error control. We examine the power of this approach using the example of a clinical trial comparing two cancer therapies.
Collapse
|
99
|
Hlavin G, Koenig F, Male C, Posch M, Bauer P. Evidence, eminence and extrapolation. Stat Med 2016; 35:2117-32. [PMID: 26753552 PMCID: PMC5066662 DOI: 10.1002/sim.6865] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2014] [Revised: 11/24/2015] [Accepted: 12/13/2015] [Indexed: 12/30/2022]
Abstract
A full independent drug development programme to demonstrate efficacy may not be ethical and/or feasible in small populations such as paediatric populations or orphan indications. Different levels of extrapolation from a larger population to smaller target populations are widely used for supporting decisions in this situation. There are guidance documents in drug regulation, where a weakening of the statistical rigour for trials in the target population is mentioned to be an option for dealing with this problem. To this end, we propose clinical trials designs, which make use of prior knowledge on efficacy for inference. We formulate a framework based on prior beliefs in order to investigate when the significance level for the test of the primary endpoint in confirmatory trials can be relaxed (and thus the sample size can be reduced) in the target population while controlling a certain posterior belief in effectiveness after rejection of the null hypothesis in the corresponding confirmatory statistical test. We show that point‐priors may be used in the argumentation because under certain constraints, they have favourable limiting properties among other types of priors. The crucial quantity to be elicited is the prior belief in the possibility of extrapolation from a larger population to the target population. We try to illustrate an existing decision tree for extrapolation to paediatric populations within our framework. © 2016 The Authors. Statistics in Medicine Published by John Wiley & Sons Ltd.
Collapse
|
100
|
Ondra T, Dmitrienko A, Friede T, Graf A, Miller F, Stallard N, Posch M. Methods for identification and confirmation of targeted subgroups in clinical trials: A systematic review. J Biopharm Stat 2016; 26:99-119. [PMID: 26378339 PMCID: PMC4732423 DOI: 10.1080/10543406.2015.1092034] [Citation(s) in RCA: 73] [Impact Index Per Article: 9.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2015] [Accepted: 08/14/2015] [Indexed: 12/30/2022]
Abstract
Important objectives in the development of stratified medicines include the identification and confirmation of subgroups of patients with a beneficial treatment effect and a positive benefit-risk balance. We report the results of a literature review on methodological approaches to the design and analysis of clinical trials investigating a potential heterogeneity of treatment effects across subgroups. The identified approaches are classified based on certain characteristics of the proposed trial designs and analysis methods. We distinguish between exploratory and confirmatory subgroup analysis, frequentist, Bayesian and decision-theoretic approaches and, last, fixed-sample, group-sequential, and adaptive designs and illustrate the available trial designs and analysis strategies with published case studies.
Collapse
|