140 likes | 445 Views
Interpreting Effect Size Results. Cohen’s “Rules-of-Thumb” standardized mean difference effect size small = 0.20 medium = 0.50 large = 0.80 correlation coefficient small = 0.10 medium = 0.25 large = 0.40 odds-ratio small = 1.50 medium = 2.50 large = 4.30.
E N D
Interpreting Effect Size Results • Cohen’s “Rules-of-Thumb” • standardized mean difference effect size • small = 0.20 • medium = 0.50 • large = 0.80 • correlation coefficient • small = 0.10 • medium = 0.25 • large = 0.40 • odds-ratio • small = 1.50 • medium = 2.50 • large = 4.30 Practical Meta-Analysis -- D. B. Wilson
Interpreting Effect Size Results • These do not take into account the context of the intervention • They do correspond to the distribution of effects across meta-analyses found by Lipsey and Wilson (1993) Practical Meta-Analysis -- D. B. Wilson
Interpreting Effect Size Results • Rules-of-Thumb do not take into account the context of the intervention • a “small” effect may be highly meaningful for an intervention that requires few resources and imposes little on the participants • a small effect may be meaningful if the intervention is delivered to an entire population (prevention programs for school children) • small effects may be more meaningful for serious and fairly intractable problems • Cohen’s Rules-of-Thumb do, however, correspond to the distribution of effects across meta-analyses found by Lipsey and Wilson (1993) Practical Meta-Analysis -- D. B. Wilson
Translation of Effect Sizes • Original metric • Success Rates (Rosenthal and Rubin’s BESD) • Proportion of “successes” in the treatment and comparison groups assuming an overall success rate of 50% • Can be adapted to alternative overall success rates • Example using the sex offender data • Assuming a comparison group recidivism rate of 15%, the effect size of 0.45 for the cognitive-behavioral treatments translates into a recidivism rate for the treatment group of 7% Practical Meta-Analysis -- D. B. Wilson
Methodological Adequacy of Research Base • Findings must be interpreted within the bounds of the methodological quality of the research base synthesized. • Studies often cannot simply be grouped into “good” and “bad” studies. • Some methodological weaknesses may bias the overall findings, others may merely add “noise” to the distribution. Practical Meta-Analysis -- D. B. Wilson
Confounding of Study Features • Relative comparisons of effect sizes across studies are inherently correlational! • Important study features are often confounding, obscuring the interpretive meaning of observed differences • If the confounding is not severe and you have a sufficient number of studies, you can model “out” the influence of method features to clarify substantive differences Practical Meta-Analysis -- D. B. Wilson
Final Comments • Meta-analysis is a replicable and defensible method of synthesizing findings across studies • Meta-analysis often points out gaps in the research literature, providing a solid foundation for the next generation of research on that topic • Meta-analysis illustrates the importance of replication • Meta-analysis facilitates generalization of the knowledge gain through individual evaluations Practical Meta-Analysis -- D. B. Wilson
Application of Meta-Analysisto Your Own Research Areas Practical Meta-Analysis -- D. B. Wilson