1 / 38

Meta-analysis

Session 2.1 – Revision of Day 1. Funded through the ESRC’s Researcher Development Initiative. Meta-analysis. Department of Education, University of Oxford. Calculating effect sizes. 2. Questions. What are the 3 primary types of effect sizes?

Download Presentation

Meta-analysis

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Session 2.1 – Revision of Day 1 Funded through the ESRC’s Researcher Development Initiative Meta-analysis Department of Education, University of Oxford

  2. Calculating effect sizes 2

  3. Questions • What are the 3 primary types of effect sizes? • What sort of information can be used to calculate effect sizes? • What software is available for calculating effect sizes?

  4. Effect size calculation • Standardized mean difference • Group contrasts • Treatment groups • Naturally occurring groups • Inherently continuous construct • Odds-ratio • Group contrasts • Treatment groups • Naturally occurring groups • Inherently dichotomous construct • Correlation coefficient • Association between variables

  5. Summary of equations from Lipsey & Wilson (2001) (for more formulae see Lipsey & Wilson Appendix B) 5 5

  6. ES_calculator.xls 6 6

  7. What do they mean? • Standardised mean difference effect sizes indicate the amount of improvement of treatment group over control, or the difference between 2 groups. • Odds ratio effect sizes indicate the likelihood of something occurring, e.g., not catching an illness after inoculation • Correlation effect sizes indicate the strength of the relationship between 2 variables

  8. Effect size as proportion in the Treatment group doing better than the average Control group person = Control = Treatment 69% of T above 79% of T above 57% of T above Effect sizes can be thought of as the average percentile standing of the average treated participant relative to the average untreated participant. 8

  9. Assumptions of the models 10

  10. Questions • What are the key statistical assumptions of the 3 meta-analytic methods?

  11. Fixed effects assumptions • Includes the entire population of studies to be considered; do not want to generalise to other studies not included (including future studies). • All of the variability between effect sizes is due to sampling error alone. Thus, the effect sizes are only weighted by the within-study variance. • assumes that the collected studies all represent random samples from the same population • Effect sizes are independent.

  12. Fixed effects In this and following formulae, we will use the symbols d and δ to refer to any measure for the observed and the true effect size, which is not necessarily the standardized mean difference. • Where • dj is the observed effect size in study j • δis the ‘true’ population effect • and ej is the residual due to sampling variance in study j

  13. Random effects assumptions • Is only a sample of studies from the entire population of studies to be considered. As a result, we do want to generalise to other studies not included in the sample (e.g., future studies). • Variability between effect sizes is due to sampling error plus variability in the population of effects. • In contrast to fixed effects models, there are 2 sources of variance • Assumes that the studies are random samples of some population in which the underlying (infinite-sample) effect sizes have a distribution rather than having a single value. • Effect sizes are independent.

  14. Random effects • Where • dj is the observed effect size in study j • δis the mean ‘true’ population effect size • uj is the deviation of the true study effect size from the mean true effect size • and ej is the residual due to sampling variance in study j

  15. Multilevel assumptions • Meta-analytic data is inherently hierarchical (i.e., effect sizes nested within studies) and has random error that must be accounted for • Effect sizes are not necessarily independent • Allows for multiple effect sizes per study • The model combines fixed and random effects (often called a mixed effects model)

  16. Multilevel modelling • Where • dj is the observed effect size in study j • 0is the mean ‘true’ population effect size • uj is the deviation of the true study effect size from the mean true effect size • and ej is the residual due to sampling variance in study j

  17. Simplifying the multilevel equation • If between-study variance = 0, the multilevel model simplifies to the fixed effects regression model • If no predictors are included the model simplifies to random effects model • If the level 2 variance = 0 , the model simplifies to the fixed effects model

  18. In practice... • Many meta-analysts use an adaptive (or “conditional”) approach IF between-study variance is found in the homogeneity test THEN use random effects model OTHERWISE use fixed effects model

  19. In practice... • Fixed effects models are very common, even though the assumption of homogeneity is “implausible” (Noortgate & Onghena, 2003) • There is a considerable lag in the uptake of new methods by applied meta-analysts • Meta-analysts need to stay on top of these developments by • Attending courses • Wide reading across disciplines

  20. What do the models involve? 21

  21. Questions • What is the first step in the analysis of meta-analytic data in fixed or random effects models? • What 2 common statistical techniques have been adapted for use in fixed and random effects meta-analytic modelling? • What common statistical technique is multilevel modelling analogous to?

  22. Conducting fixed effects meta-analysis • Usually start with a Q-test to determine the overall mean effect size and the homogeneity of the effect sizes (MeanES.sps macro) • If there is significant homogeneity, then: • 1) should probably conduct random effects analyses instead • 2) model moderators of the effect sizes (determine the source/s of variance)

  23. The homogeneity (Q) test asks whether the different effect sizes are likely to have all come from the same population (an assumption of the fixed effects model). Are the differences among the effect sizes no bigger than might be expected by chance? Q-test of the homogeneity of variance • = effect size for each study (i = 1 to k) • = mean effect size • = a weight for each study based on the sample size • However, this (chi-square) test is heavily dependent on sample size. It is almost always significant unless the numbers (studies and people in each study) are VERY small. This means that the fixed effect model will almost always be rejected in favour of a random effects model.

  24. Q-test • The Q-test is easy to conduct using the MeanES.sps macro from David Wilson’s website • MeanES ES=d /W=weight.

  25. Fixed effects: mean effect size Significant heterogeneity in the effect sizes therefore random effects more appropriate and/or moderators need to be modelled 26

  26. ANOVA • The analogue to the ANOVA homogeneity analysis is appropriate for categorical variables • Looks for systematic differences between groups of responses within a variable • Easy to implement using MetaF.sps macro • MetaF ES = d /W = Weight /GROUP = TXTYPE /MODEL = FE.

  27. Multiple regression • Multiple regression homogeneity analysis is more appropriate for continuous variables and/or when there are multiple variables to be analysed • Tests the ability of groups within each variable to predict the effect size • Can include categorical variables in multiple regression as dummy variables • Easy to implement using MetaReg.sps macro • MetaReg ES = d /W = Weight /IVS = IV1 IV2 /MODEL = FE.

  28. Random effects models • If the homogeneity test is rejected (it almost always will be), it suggests that there are larger differences than can be explained by chance variation (at the individual participant level). There is more than one “population” in the set of different studies. • The random effects model determines how much of this between-study variation can be explained by study characteristics that we have coded. • The total variance associated with the effect sizes has two components, one associated with differences within each study (participant level variation) and one between study variance:

  29. Weighting in random effects models • The weighting for each effect size consists of the within-study variance (vi) and between-study variance (vθ) • The new weighting for the random effects model (wiRE) is given by the formula: 30

  30. Weighting in random effects models • Thus, larger studies receive proportionally less weight in RE model than in FE model. • This is because a constant is added to the denominator, so the relative effect of sample size will be smaller in RE model 31

  31. Conducting random effects meta-analysis • Like the FE model, RE uses ANOVA and multiple regression to model potential moderators/predictors of the effect sizes, if the Q-test reveals significant heterogeneity • Easy to implement using MetaF.sps macro (ANOVA) or MetaReg.sps (multiple regression). • MetaF ES = d /W = Weight /GROUP = TXTYPE /MODEL = ML. • MetaReg ES = d /W = Weight /IVS = IV1 IV2 /MODEL = ML.

  32. Random effects: mean effect size Significant heterogeneity in the effect sizes therefore need to model moderators 33

  33. Conducting multilevel model analyses • Similar to multiple regression, but corrects the standard errors for the nesting of the data • Start with an intercept-only (no predictors) model, which incorporates both the outcome-level and the study-level components • This tells us the overall mean effect size • Is similar to a random effects model • Then expand the model to include predictor variables, to explain systematic variance between the study effect sizes 34

  34. Multilevel set-up • (MLwiN screenshot)

  35. Multilevel: mean effect size • Using the same simulated data set with n = 15

  36. Conclusions • Multilevel models: • build on the fixed and random effects models • account for between-study variance (like random effects) • Are similar to multiple regression, but correct the standard errors for the nesting of the data. Improved modelling of the nesting of levels within studies increases the accuracy of the estimation of standard errors on parameter estimates and the assessment of the significance of explanatory variables (Bateman and Jones, 2003). • Multilevel modelling is more precise when there is greater between-study heterogeneity • Also allows flexibility in modelling the data when one has multiple moderator variables (Raudenbush & Bryk, 2002)

  37. References • Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.). Hillsdale, NJ: Lawrence Earlbaum Associates. • Lipsey, M. W., & Wilson, D. B. (2001). Practical meta-analysis. Thousand Oaks, CA: Sage Publications. • Van den Noortgate, W., & Onghena, P. (2003). Multilevel meta-analysis: A comparison with traditional meta-analytical procedures. Educational and Psychological Measurement, 63, 765-790. • Wilson’s “meta-analysis stuff” website: http://mason.gmu.edu/~dwilsonb/ma.html • Raudenbush, S.W. and Bryk, A.S. (2002). Hierarchical Linear Models (2nd Ed.).Thousand Oaks: Sage Publications.

More Related