1 / 23

Testing and Estimation Procedures in Multi-Armed Designs with Treatment Selection

Testing and Estimation Procedures in Multi-Armed Designs with Treatment Selection. Gernot Wassmer, PhD Institut für Medizinische Statistik, Informatik und Epidemiologie Universität zu Köln ADDPLAN GmbH Adaptive Design KOL Lecture Series, August 14th, 2009. Introduction.

Samuel
Download Presentation

Testing and Estimation Procedures in Multi-Armed Designs with Treatment Selection

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Testing and Estimation Procedures in Multi-Armed Designs with Treatment Selection Gernot Wassmer, PhD Institut für Medizinische Statistik, Informatik und Epidemiologie Universität zu Köln ADDPLAN GmbH Adaptive Design KOL Lecture Series, August 14th, 2009

  2. Introduction • Confirmatory adaptive designs are a generalization of group sequential designs, where - in interim analyses - confirmatory analysis is performed under control of the Type I error rate and data dependent changes of design are allowed. • Three particular applications • Sample size reassessment • Treatment arm selection • Subset selection (“enrichment designs”) • This talk shows • how to reach a test decision in an adaptive multi-armed trial with treatment selection at interim • how to calculate confidence intervals and overall p-values

  3. Confirmatory adaptive designs can be based on • the combination testing principle • the conditional error approach Combination testing principle Combination of p-values with a specific combination function (Bauer, 1989; Bauer & Köhne, 1994) Inverse normal method: The test decision is based on where the weights wkare prefixed Lehmacher & Wassmer, 1999

  4. The conditional error approach Plan a trial with reasonable (optimum) design, including sample size calculation and timing of interim analyses. Calculate the conditional Type I error rate a(x1,…,xk) at any time during the course of the trial a(x1,…,xk) = conditional probability, under H0, of rejecting H0 in one of the subsequent stages, given x1,…,xk x1,…,xk: data up to stage k Remainder of the trial can be defined as a test at level a(x1,…,xk) where the design of this test is arbitrary. Müller & Schäfer (2001): “CRP principle” Brannath, Posch & Bauer (2002): “Recursice testing principle”

  5. The situation • Considermany-to-one comparisons, e.g., G treatment arms and one control, normal case. • Throughout this talk, we consider one-sided testing. • In an interim stage a treatment arm is selected based on data observed so far. • Not only selection procedures, but also other adaptive strategies (e.g., sample size reassessment) can be performed. • Application within “Adaptive seamless designs” using the combination testing principle

  6. Sources for alpha inflation • Interim analyses • Sample size reassessment • Multiple arms The proposed adaptive procedure fulfils the regulatory requirements for the analysis of adaptive trials in that it strongly controls the prespecified Type I error rate. This procedure will be based on the application of the closed test procedure together with combination tests (e.g., Bauer & Kieser, 1999; Hellmich, 2001; Posch et al., 2005, Bretz et al., 2009). Other approaches: Thall et al., 1988; Follmann et al, 1994; Stallard and Todd, 2003; Stallard and Friede, 2008;

  7. Closed testing procedure Stage I Stage II … ? ? Simple “trick”: Test of intersection hypotheses are formally performed as tests for

  8. At the first interim analysis, consider a test statistic for e.g., the test statistic Closed testing procedure That is, compute Dunnett’s adjusted p-value for each intersection hypothesis, critical values are according to Or compute the p-value using Dunnett’s t distribution.

  9. Test decision for the second stage: is rejected if where pJis the p-value of the Dunnett test for testing qS is the second stage test statistic for the selected treatment arm, and u2 is the critical value for the second stage. This is the use of the inverse normal method for the Dunnett test situation.

  10. Simple shortcut:If the treatment arm with the largest test statistic is selected, it suffices to combine the test for H0: with the test for H0: Example S = 3 Stage II … Stage I can be rejected if all combination tests exceed the critical value u2 .

  11. Properties of the Procedure • Choice of tests for intersection hypotheses is free, i.e., you might select, e.g., Dunnett‘s test, Bonferroni-, Simes or Sidak‘s test. • The procedure may become inconsonant and, hence, conservative. I.e., you can reject the global hypothesis, but no single hypothesis (Friede and Stallard, 2008). • A hypothesis can be rejected at a later stage even it was not selected for the current stage (and not rejected before). This can happen if, e.g., the test statistic for the global hypothesis exceeds u2 in the second stage but not u1 in the first stage, and the test statistic for the de-selected hypothesis exceeds u1 in the first stage.

  12. An alternative procedure (König et al., 2008) Compute conditional error at first stage: • In the second stage, perform a • Conditional second-stage Dunnett test • Separate second-stage Dunnett test This is the application of the CRP principle (Müller & Schäfer, 2001). It assumes the variance to be known

  13. A comparison shows that • the conditional second-stage Dunnett test performs best but is hardly better if a treatment arm selection was performed (cf., Friede and Stallard, 2008) • it is identical with the conventional Dunnett test if no adaptations were performed • becomes complicated if, e.g., • allocation is not constant • variance is unknown • the inverse normal technique is not optimum but enables early stopping and more general adaptations • is straightforward if, e.g., • allocation is not constant • variance is unknown

  14. A comparison shows that • the conditional second-stage Dunnett test performs best but is hardly better if a treatment arm selection was performed (cf., Friede and Stallard, 2008) • it is identical with the conventional Dunnett test if no adaptations were performed • becomes complicated if, e.g., • allocation is not constant • variance is unknown • the inverse normal technique is not optimum but enables early stopping and more general adaptations • is straightforward if, e.g., • allocation is not constant • variance is unknown

  15. Overall p-values • Defined as smallest p-value for which the test results yield rejection of the considered (single) hypothesis • Repeated overall p-value can be calculated at any stage of the trial. • That is, • p-values account for the step-down nature of the closed testing principle and are completely consistent with the test decision.

  16. Overall confidence intervals • Confidence intervals based on stepwise testing are difficult to construct. This is a specific feature of multiple testing procedures and not of adaptive testing. • Posch et al. (2005) proposed to construct confidence intervals based on the single step adjusted overall p-values. These can also be applied for the conditional Dunnett test. • The RCIs are not, in general, consistent with the test decision. It might happen that, e.g., a hypothesis is rejected but the lower bound of the CI is smaller 0. • They can be provided for each step of the trial. • In general, they may fail to become narrower for increasing sample size (e.g., if Bonferroni or Simes intersection tests are used).

  17. Illustration • Two-stage design with G treatment arms • Selection of treatment arm with highest respone, no efficacy stop at interim • Bonferroni (or Simes) correction is used for first stage • Lower bound lbj of 95% confidence intervals for effect dj = mj - m0of selected treatment arm at second stage is calculated through It is easy to see that

  18. Summary • The adaptive proceduresfulfiltheregulatoryrequirementsfortheanalysisof adaptive trials in thattheycontroltheprespecified Type I error rate. For regulatorypurposes, theclass of envisageddecisionsafterstage 1 shouldbestated in theprotocol. • The “rules” for adaptation and stopping for futility • not need to be pre-specified • Adaptations may depend on all interim data including secondary and safety endpoints. • can make use of Bayesian principles integrating all information available, also external to the study • should be evaluated (e.g. via simulations) and preferred version recommended, e.g., in DMC charter • Software ADDPLAN MC is available for designing and analyzing these trials

  19. References • Bauer, P. (1989). Multistage testing with adaptive designs (with Discussion). Biometrie und Informatik in Medizin und Biologie 20, 130–148. • Bauer, P., Köhne, K. (1994). Evaluation of experiments with adaptive interim analyses. Biometrics 50, 1029–1041. • Bauer, P., Kieser, M. (1999). Combining different phases in the development of medical treatments within a single trial. Statistics in Medicine 18,1833–1848. • Brannath, W., Posch, M., Bauer, P., 2002: Recursive combination tests. J. Amer. Stat. Ass. 97, 236–244. • Follmann, D. A., Proschan, M. A., Geller, N. L., 1994: Monitoring pairwise comparisons in multi-armed clinical trials. Biometrics 50, 325–336. • Friede, T., Stallard, N., 2008: A comparison of methods for adaptive treatment selection. Biometrical J. 50, 767–781. • Hellmich, M., 2001: Monitoring clinical trials with multiple arms. Biometrics 57, 892–898. • König, F., Brannath, W., Bretz, F., Posch, M. (2008). Adaptive Dunnett tests for treatment selection. Statistics in Medicine 27, 1612–1625. • Lehmacher, W., Wassmer, G. (1999). Adaptive sample size calculations in group sequential trials. Biometrics 55, 1286–1290. • Müller, H.H., Schäfer, H. (2001). Adaptive group sequential designs for clinical trials, combining the advantages of adaptive and of classical group sequential approaches. Biometrics 57,886–891. • Posch, M., König, F., Branson, M., Brannath, W., Dunger-Baldauf, C., Bauer, P. (2005). Testing and estimation in flexible group sequential designs with adaptive treatment selection. Statistics in Medicine 24, 3697–3714. • Posch, M., Wassmer, G., Brannath, W. (2008). A note on repeated p-values for group sequential designs. Biometrika 95, 253-256. • Stallard, N., Friede, T. (2008). A group-sequential design for clinical trials with treatment selection. Statistics in Medicine 27, 6209–6227. • Stallard, N., Todd, S. (2003). Sequential designs for phase III clinical trials incorporating treatment selection. Statistics in Medicine 22, 689-703. • Thall, P.F., Simon, R., Ellenberg, S.S. (1988). Two-stage selection and testing designs for comparative clinical trials. Biometrika 75, 303-310.

More Related