1 / 31

More About Type I and Type II Errors

More About Type I and Type II Errors. O.J. Simpson trial: the situation. O.J. is assumed innocent. Evidence collected: size 12 Bruno Magli bloody footprint, bloody glove, blood spots on white Ford Bronco, the knock on the wall, DNA evidence from above, motive(?), etc….

polly
Download Presentation

More About Type I and Type II Errors

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. More About Type I and Type II Errors

  2. O.J. Simpson trial: the situation • O.J. is assumed innocent. • Evidence collected: size 12 Bruno Magli bloody footprint, bloody glove, blood spots on white Ford Bronco, the knock on the wall, DNA evidence from above, motive(?), etc…

  3. O.J. Simpson trial: jury decisions • In criminal trial: The evidence does not warrant rejecting the assumption of innocence. Behave as if O.J. is innocent. • In civil trial: The evidence warrants rejecting the assumption of innocence. Behave as if O.J. is guilty. • Was an error made in either trial?

  4. Errors in Trials If O.J. is innocent, then an error was made in the civil trial. If O.J. is guilty, then an error was made in the criminal trial.

  5. Errors in Hypothesis Testing

  6. Definitions: Types of Errors • Type I error: The null hypothesis is rejected when it is true. • Type II error: The null hypothesis is not rejected when it is false. • There is always a chance of making one of these errors. We’ll want to minimize the chance of doing so!

  7. Example: Grade inflation? Is there evidence to suggest the mean GPA of college undergraduate students exceeds 2.7? H0: μ = 2.7 HA: μ > 2.7 Data n = 36 Random sample of students s = 0.6 and Decision Rule Set significance level α = 0.05. If p-value < 0.05, reject null hypothesis.

  8. Let’s consider what our conclusion is based upon different observed sample means… If Reject null since p-value is (just barely!) smaller then 0.05.

  9. If Reject null since p-value is smaller then 0.05.

  10. If Reject null since p-value is smaller then 0.05.

  11. Alternative Decision Rule • “Reject if p-value  0.05” is equivalent to “Reject if the sample average, X-bar, is larger than 2.865” • is called “rejection region.”

  12. Type I Error

  13. Minimize chance of Type I error... • … by making significance level  small. • Common values are  = 0.01, 0.05, or 0.10. • “How small” depends on seriousness of Type I error. • Decision is not a statistical one but a practical one.

  14. P(Type I Error) in trials • Criminal trials: “Beyond a reasonable doubt”. 12 of 12 jurors must unanimously vote guilty. Significance level  set at 0.001, say. • Civil trials: “Preponderance of evidence.” 9 out of 12 jurors must vote guilty. Significance level  set at 0.10, say.

  15. Example: Serious Type I Error • New Drug A is supposed to reduce diastolic blood pressure by more than 15 mm Hg. • H0: μ = 15 versus HA: μ > 15 • Drug A can have serious side effects, so don’t want patients on it unless μ > 15. • Implication of Type I error: Expose patients to serious side effects without other benefit. • Set  = P(Type I error) to be small  0.01

  16. Example: Not so serious Type I Error • Grade inflation? • H0: μ = 2.7 vs. HA: μ > 2.7 • Type I error: claim average GPA is more than 2.7 when it really isn’t. • Implication: Instructors grade harder. Students get unhappy. • Set  = P(Type I error) at, say, 0.10.

  17. Type II Error and Power • Type II Error is made when we fail to reject the null when the alternative is true. • Want to minimize P(Type II Error). • Now, if alternative HA is true: • P(reject|HA is true) + P(not reject|HA is true) =1 • “Power” + P(Type II error) = 1 • “Power” = 1 - P(Type II error)

  18. Type II Error and Power • “Power” of a test is the probability of rejecting null when alternative is true. • “Power” = 1 - P(Type II error) • To minimize the P(Type II error), we equivalently want to maximize power. • But power depends on the value under the alternative hypothesis ...

  19. Type II Error and Power (Alternative is true)

  20. Power • Power is probability, so number between 0 and 1. • 0 is bad! • 1 is good! • Need to make power as high as possible.

  21. Maximizing Power … • The farther apart the actual mean is from the mean specified in the null, the higher the power. • The higher the significance level , the higher the P(Type I error), the higher the power. • The smaller the standard deviation, the higher the power. • The larger the sample, the higher the power.

  22. That is, factors affecting power... • Difference between value under the null and the actual value • P(Type I error) =  • Standard deviation • Sample size

  23. Strategy for designing a good hypothesis test • Use pilot study to estimate std. deviation. • Specify . Typically 0.01 to 0.10. • Decide what a meaningful difference would be between the mean in the null and the actual mean. • Decide power. Typically 0.80 to 0.99. • Use software to determine sample size.

  24. Using JMP to Determine Sample Size – DOE > Sample Size and Power

  25. Using JMP to Determine Sample Size – One Sample Mean P(Type I Error ) = a Error Std Dev = “guessimate” for standard deviation (s or s) Enter values for one or two of the quantities:1) Difference to detect d = |Ho mean – HA mean| = |m0 - m|2) Sample Size = n 3) Power = P(Reject Ho|HA true) = 1 - b

  26. Using JMP to Determine Sample Size – DOE > Sample Size and Power For a = .05, d = .20, s = .60 and leaving Power and Sample Size empty we obtain a plot of Power vs. Sample Size (n). Here we can see:

  27. Using JMP to Determine Sample Size – DOE > Sample Size and Power (JMP Demo)

  28. If sample is too small ... • … the power can be too low to identify even large meaningful differences between the null and alternative values. • Determine sample size in advance of conducting study. • Don’t believe the “fail-to-reject-results” of a study based on a small sample.

  29. If sample is really large ... • … the power can be extremely high for identifying even meaningless differences between the null and alternative values. • In addition to performing hypothesis tests, use a confidence interval to estimate the actual population value. • If a study reports a “reject result,” ask how much different?

  30. The moral of the storyas researcher • Always determine how many measurements you need to take in order to have high enough power to achieve your study goals. • If you don’t know how to determine sample size, ask a statistical consultant to help you.

  31. The moral of the storyas reviewer • When interpreting the results of a study, always take into account the sample size.

More Related