140 likes | 154 Views
BHS 204-01 Methods in Behavioral Sciences I. April 25, 2003 Chapter 6 (Ray) The Logic of Hypothesis Testing. Degrees of Freedom. Degrees of freedom (df) – how many numbers can vary and still produce the observed result. Population statistics include the degrees of freedom.
E N D
BHS 204-01Methods in Behavioral Sciences I April 25, 2003 Chapter 6 (Ray) The Logic of Hypothesis Testing
Degrees of Freedom • Degrees of freedom (df) – how many numbers can vary and still produce the observed result. • Population statistics include the degrees of freedom. • Calculated differently depending upon the experimental design – based on the number of groups. • T-Test df = (Ngroup1 -1) + (Ngroup2 -1)
Reporting T-Test Results • Include a sentence that gives the direction of the result, the means, and the t-test results. • Example: • The experimental group showed significantly greater weight gain (M = 55) compared to the control group (M = 21), t(12) = 3.97, p=.0019, two-tailed. • Give the exact probability of the t value. • Underline all statistics.
When to Use a T-Test • When two independent groups are compared. • When sample sizes are small (N< 30). • When the actual population distribution is unknown (not known to be normal). • When the variances within the two groups are unequal. • When sample sizes are unequal.
Using Error Bars in Graphs • Error bars show the standard error of the mean for the observed results. • To visually assess statistical significance, see whether: • The mean (center point of error bar) for one group falls outside the error bars for the other group. • Also compare how large the error bars are for the two groups.
Figure 5.8. (p. 124)Graphic illustration of cereal experiment.
Sources of Variance • Systematic variation – differences related to the experimental manipulation. • Can also be differences related to uncontrolled variables (confounds) or systematic bias (e.g. faulty equipment or procedures). • Chance variation – nonsystematic differences. • Cannot be attributed to any factor. • Also called “error”.
F-Ratio • A comparison of the differences between groups with the differences within groups. • Between-group variance = treatment effect + chance variance. • Within-group variance = chance variance. • If there is a treatment effect, then the between-group variance should be greater than the within-group variance.
Testing the Null Hypothesis • Between-group variance (treatment effect) must be greater than within-group variance (chance variation), F > 1.0. • How much greater? • Normal curve shows that 2 SD, p <.05 is likely to be a meaningful difference. • The p value is a compromise between the likelihood of accepting a false finding and the likelihood of not accepting a true hypothesis.
Type I error – likelihood of rejecting the null when it is true and accepting the alternative when it is false (making a false claim). • This is the p value -- .05 is probability of making a Type I error. • Type II error – likelihood of accepting null when it is false and rejecting the alternative when it is true. • Probability is b, the power of a statistic is 1-b.
Reporting the F-Ratio • ANOVA is used to calculate the F-Ratio. • Example: • The experimental group showed significantly greater weight gain (M = 55) compared to the control group (M = 21), F(1, 12) = 4.75, p=.05. • Give the degrees of freedom for the numerator and denominator.
When to Use ANOVA • When there are two or more independent groups. • When the population is likely to be normally distributed. • When variance is similar within the groups compared. • When group sizes (N’s) are close to equal.
Threats to Internal Validity • It is the experimenter’s job to eliminate as many threats to internal validity as possible. • Such threats constitute sources of systematic variance that can be confused with an effect, resulting in a Type I error. • Potential threats to validity must be evaluated in the Discussion section of the research report.