1 / 31

Early Stopping Rules: Interim Analyses

Clinical Trials in 20 Hours. Early Stopping Rules: Interim Analyses. Elizabeth S. Garrett esg@jhu.edu Oncology Biostatistics May 8, 2002. References.

Thomas
Download Presentation

Early Stopping Rules: Interim Analyses

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Clinical Trials in 20 Hours Early Stopping Rules:Interim Analyses Elizabeth S. Garrett esg@jhu.edu Oncology Biostatistics May 8, 2002

  2. References • S. Piantadosi (1997) “Data Dependent Stopping”, in Clinical Trials: A Methodologic Perspective. Chapter10. New York:John Wiley and Sons. • Souhami and Whitehead, eds. (1994) Workshop on Early Stopping Rules in Cancer Clinical Trials, Robinson College, Cambridge, UK, 13-15 April, 1993. Statistics in Medicine, 13: 1289-1500. • Gail (1982) “Monitoring and Stopping Clinical Trials,” in Statistics in Medical Research, (eds. Mike and Stanley), New York: John Wiley and Sons. • Pocock (1993) Statistical and Ethical Issues in Monitoring Clinical Trials, Statistics in Medicine, 12: 1459-1469. • Jones & Lewis, “Data and Safety Monitoring” in Biostatistics in Clinical Trials (ed. Redmond & Colton). Clinical Trials in 20 Hours

  3. Goals for Today • Describe what interim analysis means • Understand why they need to be done • Understand why “multiple looks” are frowned upon by many • Consider some common analyses and designs that incorporate early stopping. Clinical Trials in 20 Hours

  4. Interim Analyses • Also called “data-dependent stopping” or “early stopping” • Continuing a trial: there needs to be active monitoring so that a trial is not continued simply because it was begun. • Some issues involved in stopping: • ethics • precision of results • data quality • resource availability • Usually, we use accumulated data to decide what to do • Sometimes outside information is provided to encourage us to stop a trial (e.g. a trial using same drug had very bad/good effects elsewhere) • Early stopping can be due to efficacy but also to other reasons (e.g. accrual too slow). Clinical Trials in 20 Hours

  5. Some Examples of Why a Trial May Be Terminated* • Treatments found to be convincingly different • Treatments found to be convincingly not different • Side effects or toxicities are too severe • Data quality is poor • Accrual is slow • Definitive information becomes available from an outside source making trial unnecessary or unethical • Scientific question is no longer important • Adherence to treatment is unacceptably low • Resources to perform study are lost or diminished • Study integrity has been undermined by fraud or misconduct * From S Piantadosi Clinical Trials in 20 Hours

  6. Example: Cancer Center Trials Yearly review We consider: Interim Analyses Accrual Eligibility Answering Primary Question Adheres to Statistical Plan Safety Data Quality Investigator Report Overall Scientific Progress Clinical Trials in 20 Hours

  7. Data Safety and Monitoring Committees • Most comparative/phase III clinical trials have Data Safety and Monitoring Committees • Their goal is to ensure that the trial is safe and warrants continuation. • A qualitative review of adverse events is performed. Clinical Trials in 20 Hours

  8. To Terminate: minimize size of trial minimize number of patients on inferior arm costs and economics timeliness of results To Continue: increase precision reduce errors increase power increase ability to look at subgroups gather information on secondary endpoints Opposing Pressures in Interim Analyses* * From S Piantadosi Clinical Trials in 20 Hours

  9. Statistical Considerations in Interim Analyses • Consider a safety/efficacy study (phase II) • “At this point in time, is there statistical evidence that….” • The treatment will not be as efficacious as we would hope/need it to be? • The treatment is clearly dangerous/unsafe? • The treatment is very efficacious and we should proceed to a comparative trial? Clinical Trials in 20 Hours

  10. Statistical Considerations in Interim Analyses • Consider a comparative study (phase III) • “At this point in time, is there statistical evidence that….” • One arm is clearly more effective than the other? • One arm is clearly dangerous/unsafe? • The two treatments have such similar responses that there is no possibility that we will see a significant difference by the end of the trial? Clinical Trials in 20 Hours

  11. Statistical Considerations in Interim Analyses • We use interim statistical analyses to determine the answers to these questions. • It is a tricky business: • interim analyses involve relatively few data points • inferences can be imprecise • we increase chance of errors. • if interim results are conveyed to investigators, a bias may be introduced • in general, we look for strong evidence in one or another direction. Clinical Trials in 20 Hours

  12. Example: ECMO trial* • Extra-corporeal membrane oxygenation (ECMO) versus standard treatment for newborn infants with persistent pulmonary hypertension. • N = 39 infants enrolled in study • Trial terminated after interim analysis • 4/10 deaths in standard therapy arm • 0/9 deaths in ECMO arm • p = 0.054 (one-sided) • Questions: • Is this result sufficient evidence on which to change routine practice? • Is the evidence in favor of ECMO very strong? * O’Rourke et al. , 1989, Pediatrics 84, 957 - 963 Clinical Trials in 20 Hours

  13. Example: ISIS-2 trial* • The Second International Study of Infarct Survival (ISIS-2) • Five week study of streptokinase versus placebo based on 17,187 patients with myocardial infarction. • Trial continued until • 12% death rate in placebo group • 9.2% death rate in streptokinase group • p < 0.000001 • Issues: • strong evidence in favor of streptokinase was available early on • impact would be greater with better precision on death rate, which would not be possible if trial stopped early • earlier trials of streptokinase has similar results, yet little impact. * ISIS-2 Collaborative Group, 1988, Lancet ii, 349-360 Clinical Trials in 20 Hours

  14. Statistical Schools of Thought • Three main philosophic views • Likelihood Theory • Bayesian Theory • Frequentist Theory • All differ in their approach • Frequentist is most commonly seen (but not necessarily best!) • All treat interim analyses quite differently Clinical Trials in 20 Hours

  15. Outline • Frequentist Issues: • Multiple Looks • Group Sequential Designs • Stopping Boundaries • Alpha Spending Functions • Two Stage Designs • Likelihood Issues • Bayesian Issues Clinical Trials in 20 Hours

  16. An Example of “Multiple Looks:” Why you aren’t allowed to look at your data • Consider planning a comparative trial in which two treatments are being compared for efficacy (response rate). H0: p1 = p2 H1: p1 = 0.30, p2 = 0.50 • A standard design says that for 80% power and with alpha of 0.05, you need about 100 patients per arm. • So what happens if we find p < 0.05 before all patients are enrolled? • Why can’t we look at the data a few times in the middle of the trial and conclude that one treatment is better if we see p < 0.05? Clinical Trials in 20 Hours

  17. The plots to the right show simulated data where p1 = 0.40 and p2 = 0.50 In our trial, looking to find a difference between 0.30 to 0.50, we would not expect to conclude that there is evidence for a difference. However, if we look after every 4 patients, we get the scenario where we would stop at 96 patients and conclude that there is a significant difference. Clinical Trials in 20 Hours

  18. If we look after every 10 patients, we get the scenario where we would not stop until all 200 patients were observed and would could that there is not a significant difference (p =0.40) Clinical Trials in 20 Hours

  19. If we look after every 40 patients, we get the scenario where we would don’t stop either. If we wait until the END of the trial (N = 200), then we estimate p1 to be 0.45 and p2 to be 0.52. The pvalue for testing that there is a significant difference is 0.40. Clinical Trials in 20 Hours

  20. Why would we have messed up if we looked early on? • Every time we look at the data and consider stopping, we introduce the chance of falsely rejecting the null hypothesis. • In other words, every time we look at the data, we have the chance of a type 1 error. • If we look at the data multiple times, and we use alpha of 0.05 as our criterion for significance, then we have a 5% chance of stopping each time. • Under the true null hypothesis and just 2 looks at the data, then we “approximate” the error rates as: • Probability stop at first look: 0.05 • Probability stop at second look: 0.95*0.05 = 0.0475 • Total probability of stopping  0.0975 Clinical Trials in 20 Hours

  21. Group Sequential Designs: (1) Traditional Stopping Boundaries • Formalize stopping rules and keep error rates low. • Heuristically, we set up a “statistical boundary” and stop if it is crossed. • Simplest case: perform interim tests at a very stringent level of significance and then test at the ‘nominal’ level at the end of the study if trial has not terminated early. • Examples: • Do two interim analyses and only stop if p < 0.01. Then at end of trial, use nominal 0.05 rate. Total type 1 error rate = 0.07 • Do 5 interim analyses at 0.002 level. Then do final analysis at 0.04. Total type 1 error rate = 0.05. Clinical Trials in 20 Hours

  22. Most commonly used error boundaries • Pocock (1977) Biometrika 64, 191-199 Divide type 1 error evenly across number of analyses. Good opportunity for early stopping. • O’Brien-Fleming (1979) Biometrics 35, 549-556 Use up very little error at early looks and much more error at later looks. • Fleming-Harrington-O’Brien (1984) Controlled Clinical Trials 5, 348-361 Similar to OF above, but less conservative. Jones & Lewis, “Data and Safety Monitoring” in Biostatistics in Clinical Trials (ed. Redmond & Colton). Clinical Trials in 20 Hours

  23. Nominal Pvalues for Three Most Commonly Used Boundaries NOTE: type 1 error at kth test is NOT the same as the nominal pvalue for the kth test Clinical Trials in 20 Hours

  24. Problems with These Stopping Boundaries • Pocock: Can result in awkward situation when at the final look of the study, nominal pvalue is less than 0.05 but is not sufficiently small to achieve significance based on design. Requires largest sample size to achieve specified power. • O’Brien-Fleming: May be desirable in its conservatism for keeping trials going. Some see very small boundaries at first test too extreme. • Fleming-Harrington-O’Brien: Middle ground between other two, but more similar to OF. Nice characteristic that type I error is held constant for each interim test. • Generally: Two assumptions • number of interim tests which will be conducted is specified in advance • equal number of individuals enrolled between interim analyses (not realistic, as many DSMC reviews are based on calendar time and not accrual). Clinical Trials in 20 Hours

  25. Group Sequential Designs: (2) Alpha Spending Function Stopping boundaries as continuous function of accrual Do not need to have equal spacing as assumed by previously discussed. Do not need to specify when or how many will occur Lan & DeMets (1983) Biometrika 70, 659-663 Clinical Trials in 20 Hours

  26. Two Stage Designs:Simon Two-Stage for Safety/Efficacy Trials • Different in approach, but related • A priori, decide sample size for Stage 1 and sample size for Stage 2. • If reasonable evidence of efficacy is seen by end of Stage 1, then continue to stage 2. • Minimizes overall sample size. • Can terminate at end of stage 1 either because • treatment very efficacious • treatment not at all efficacious • Usually, termination is due to lack of efficacy Clinical Trials in 20 Hours

  27. Example of Two Stage Design • Suppose we have a new treatment that we would like to investigate for efficacy. The standard therapy has a response rate of 0.25. We would be interested in this new therapy if the response rate were 0.50 or greater. • Design assumptions: • Ho: p = 0.25 • H1: p = 0.50 • For power of 80% and with overall type 1 error rate of 0.05, the sample size for stage 1 is 9 and for stage 2 is 15. • Rules: • If 2 or more responses are seen at the end of stage 1, continue to stage 2. • If 9 or more responses (total) are seen at the end of stage 2, consider new treatment efficacious. Clinical Trials in 20 Hours

  28. Likelihood Based Methods • Based purely on the likelihood function. • Used for testing hypotheses and summarizing evidence simultaneously • Fully sequential: assess evidence after each patient/observation. • SPRT: sequential probability ratio test • Wald, Sequential Analysis, 1947. New York: John Wiley and Sons. • Likelihood Ratio: R = L(p1|data)/L(p2|data) • measures evidence in favor of p1 versus p2. • If R is large, then p1 is more likely than p2. • We can sequentially calculate R and stop if evidence in favor of either p1 or p2 is as strong as we would see in the final (fixed) sample size. • We can get small pvalue and still have R close to 1! • See Royall, “Statistial Evidence: A likelihood paradigm” for more methodologic treatment. Clinical Trials in 20 Hours

  29. Suggestions for R cutoffs • R = 32: strong evidence • R = 8: weak evidence • Analogy to frequentist theory: • Ru = (1 - )/ • RL = /(1 - ) • Using  of 0.05 and  of 0.20, we get Ru= 4.75 and RL = 0.0625 = 1/16 • Using  of 0.10 and  of 0.10, we get Ru= 9 and RL = 0.11 = 1/9 Clinical Trials in 20 Hours

  30. Back to Simulated Example Clinical Trials in 20 Hours

  31. Bayesian Methods • Well suited for interim analyses • By nature, Bayesians ‘prior’ knowledge to make current decisions and designs • They allow different scenarios and “weight” of evidence. • S. Piantadosi (1997) “Data Dependent Stopping”, in Clinical Trials: A Methodologic Perspective. Chapter10. New York:John Wiley and Sons. Clinical Trials in 20 Hours

More Related