1 / 19

The USA has Evidenced-Based Drug Prevention Policy & Practice: What does it look like?

The USA has Evidenced-Based Drug Prevention Policy & Practice: What does it look like?. D.M. Gorman Department of Epidemiology & Biostatistics Texas A&M Health Science Center Research Supported by the Smith Richardson Foundation, Inc. & National Institute on Alcohol Abuse & Alcoholism

derry
Download Presentation

The USA has Evidenced-Based Drug Prevention Policy & Practice: What does it look like?

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. The USA has Evidenced-Based Drug Prevention Policy & Practice:What does it look like? D.M. Gorman Department of Epidemiology & Biostatistics Texas A&M Health Science Center Research Supported by the Smith Richardson Foundation, Inc. & National Institute on Alcohol Abuse & Alcoholism Society for the Study of Addiction, Annual Symposium (November 18, 2005)

  2. Overview • Central Question: Does the evidence in support of programs that appear on evidence-based lists come from sound evaluation practices? • Major Conclusions: While study designs are sound, the data analysis practices are not • Policy Implications: the evidence-base of evidence-based programs is no better than that of other programs

  3. Importance of “Evidenced-Based” Interventions in US Drug Prevention • US Government Performance & Results Act (1993): requires agencies to set performance goals & collect data in order to monitor these • ONDCP introduced a Performance Measures of Effectiveness System in 1998 • As part of this, federal agencies began creating lists of “model” and “exemplary” programs deemed to be “science-based”

  4. The 10 Lists Reviewed • Maryland/NIJ (1996;1998) – Drugs & Violence • Blueprints/NIJ (1997) - Violence (& Drugs) • NIDA (1997; 2003) - Drugs • CDC (2000) – Violence • Communities that Care (2000) – Drugs & Violence • DHHS (2001) – Violence • SDDFS/DoE (2001) – Drugs & Violence • NREPP/SAMHSA (2002/ongoing) – Drugs • Sociometrics (2002) - Drugs • RAND (2002/ongoing) – Drugs & Violence

  5. The Programs • 137 programs mentioned at least once • 23 mentioned ≥4 times • Curricula & therapeutic interventions best suited to lists • Focused on 9 school-based drug & violence programs mentioned most often (5-9 mentions)

  6. Most Mentioned Drug & Violence School-based Programs • Life Skills Training (9) • Seattle Social Development Project (8) • Project STAR (MPP) (8) • Athletes Training & Learning to Avoid Steroids (ATLAS) (7) • Promoting Alternative Thinking Strategies (PATHS) (7) • Project ALERT (6) • Second Step (5) • Child Development Project (5) • Olweus Bullying Prevention Partnership (5)

  7. Criteria used in Identifying Programs • Strong research design • Design (experimental or quasi-experimental) • Sample size (large) • Attrition (low) • Measures (valid & reliable) • Effect on behavioral outcomes • Duration of effect (≥ 1 year) • Replication

  8. Underlying Assumption • Science = Design + Methodology • Hence, the better the design & method, the better the “science” • And the better the science, the more confident we can be in the “science-based” results • Is there more to science than design & method?

  9. Answer: “Yes” • “To yield knowledge, the scientific method must be accompanied by a scientific world view” (Bunge, 1991) • Some features of the scientific world view • Admits fallibility of ideas • Subjects ideas to critical tests (refutation) • Accepts unfavorable data • Welcomes criticism

  10. US “Evidenced-Based” Drug Prevention • Lacks a scientific world view • Emphasis is on verification of the hypothesis that “Program X works” • What evaluation practices has the search for verifications led to?

  11. The Practices • Practice #1: Multiple Subgroup Analysis • Practice #2: Alter Outcome Variables • Practice #3: Post hoc Sample Refinement • Practice #4: Move Baseline • Practice # 5: Partial Reporting of Findings • Practice #6: Routine use of p ≤.10 & 1-tail tests

  12. Practice #1:Multiple Subgroup Analysis • Make numerous statistical comparisons between intervention & control conditions • Take those that show a statically significant difference between the groups at follow-up as evidence of program effectiveness • Capitalizes on chance occurrences

  13. Practice #2:Alter Outcome Variables • Can involve • a total change in outcome variables over the course of the evaluation • a change in the way a specific variable is constructed from study to study • Raises the question as to whether program effects are simply measurement dependent

  14. Practice #3:Post hoc Sample Refinement • Change composition of study conditions at the stage of data analysis • Example: exclude those subjects that did not receive an “ideal” dosage of the intervention • Leads to selection bias & undermines randomization

  15. Practice #4:Move Baseline • Use something other than the pretest as the baseline • Generally involves using the posttest sample size not the pretest sample size to calculate attrition rates • Serves to obscure high attrition across the course of the evaluation

  16. Practice #5: Partial Reporting of FindingPractice #6: Use p Values of .10 or 1-tail Tests • Report-out data on selective variables and/or data on selective subgroups • Practice #6: Routine use of p ≤.10 & 1-tail tests - Emphasis on statistical significance obscures the fact that many differences are of no practical significance

  17. Conclusions: Prevention “Science” • These data analysis & presentation practices serve to verify that a program “works” • Not consistent with genuine hypothesis testing (and therefore science) • Consistent with pseudoscience • Review procedures have legitimized & encouraged these practices

  18. Conclusions: Evidenced-Based Policy & Practice • Threshold for designating “evidenced-based” interventions very low • Isolated statistically significant effects receive undue attention • Ignore how effects are produced in data analysis • Preponderance of evidence should be considered in judging program effectiveness

  19. More Details • Gorman, D.M. (2005). Drug and violence prevention: rediscovering the critical rational dimension of evaluation research. Journal of Experimental Criminology, 1, 39-62. • Botvin, G.J., & Griffin, K.W. (2005). Prevention science, drug abuse prevention, and Life Skills Training: Comments on the state of the science. Journal of Experimental Criminology, 1, 63-78. • Hawkins, J.D. & Catalano, R.F. (2005). Doing prevention science: A response to D.M. Gorman and a brief history of the quasi-experimental study nested within the Seattle Social Development Project. Journal of Experimental Criminology, 1, 79-86. • Gorman, D.M. (2005). The centrality of critical rational reasoning in science: a response and further comments on the Life Skills Training program, the Seattle Social Development Project, and prevention studies. Journal of Experimental Criminology, 1: 263-275.

More Related