580 likes | 585 Views
Learn about survey, correlational, and quasi-experiments, advantages and disadvantages, ABA designs, inferential statistics, population, sample, descriptive and inferential statistics, and properties of distributions.
E N D
Statistics Psych 231: Research Methods in Psychology
Quiz 8 is due tomorrow @ midnight • Journal Summary 2 assignment Reminders
Sometimes you just can’t perform a fully controlled experiment • Because of the issue of interest • Limited resources (not enough subjects, observations are too costly, etc). • Surveys • Correlational • Quasi-Experiments • Developmental designs • Small-N designs • This does NOT imply that they are bad designs • Just remember the advantages and disadvantages of each Non-Experimental designs
= observation • Baseline experiments – the basic idea is to show: • When the IV occurs, you get the effect • When the IV doesn’t occur, you don’t get the effect (reversibility) • This allows other comparisons, to the original baseline as well as to the transition steady state Transition steady state Reversibility Steady state (baseline) Treatment introduced Treatment removed Small N designs
Before introducing treatment (IV), baseline needs to be stable • Measure level and trend • Level – how frequent (how intense) is behavior? • Are all the data points high or low? • Trend – does behavior seem to increase (or decrease) • Are data points “flat” or on a slope? Unstable Stable Small N designs
ABA design (baseline, treatment, baseline) Steady state (baseline) Transition steady state Reversibility • The reversibility is necessary, otherwise • something else may have caused the effect • other than the IV (e.g., history, maturation, etc.) • There are other designs as well (e.g., ABAB see figure14.6 in your textbook) ABA design
Advantages • Focus on individual performance, not fooled by group averaging effects • Focus is on big effects (small effects typically can’t be seen without using large groups) • Avoid some ethical problems – e.g., with non-treatments • Allows to look at unusual (and rare) types of subjects (e.g., case studies of amnesiacs, experts vs. novices) • Often used to supplement large N studies, with more observations on fewer subjects Small N designs
Disadvantages • Difficult to determine how generalizable the effects are • Effects may be small relative to variability of situation so NEED more observation • Some effects are by definition between subjects • Treatment leads to a lasting change, so you don’t get reversals Small N designs
Inferential statistics used to generalize back Population Sample • 2 General kinds of Statistics • Descriptive statistics • Used to describe, simplify, & organize data sets • Describing distributions of scores • Inferential statistics • Used to test claims about the population, based on data gathered from samples • Takes sampling error into account. Are the results above and beyond what you’d expect by random chance? Samples and Populations
Properties: Shape, Center, and Spread (variability) • Shape • Symmetric v. asymmetric (skew) • Unimodal v. multimodal • Center • Where most of the data in the distribution are • Mean, Median, Mode • Spread (variability) • How similar/dissimilar are the scores in the distribution? • Standard deviation (variance), Range Describing Distributions
Properties: Shape, Center, and Spread (variability) • Visual descriptions - A picture of the distribution is usually helpful* • Numerical descriptions of distributions Describing Distributions *Note: See Chapter 5 of APA style guide: Displaying Results
Divide by the total number in the population Add up all of the X’s Divide by the total number in the sample • The mean (mathematical average) is the most popular and most important measure of center. • The formula for the population mean is (a parameter): • The formula for the sample mean is (a statistic): mean Mean & Standard deviation
mean • The mean (mathematical average) is the most popular and most important measure of center. Others include median and mode. • The standard deviation is the most popular and important measure of variability. • The standard deviation measures how far off all of the individuals in the distribution are from a standard, where that standard is the mean of the distribution. • Essentially, the average of the deviations. Mean & Standard deviation
standard deviation = σ = • Working your way through the formula: • Step 1: Compute deviation scores • Step 2: Compute the SS • Step 3: Determine the variance • Take the average of the squared deviations • Divide the SS by the N • Step 4: Determine the standard deviation • Take the square root of the variance An Example: Computing Standard Deviation (population)
Main difference: • Step 1: Compute deviation scores • Step 2: Compute the SS • Step 3: Determine the variance • Take the average of the squared deviations • Divide the SS by the n-1 • Step 4: Determine the standard deviation • Take the square root of the variance • This is done because samples are biased to be less variable than the population. This “correction factor” will increase the sample’s SD (making it a better estimate of the population’s SD) An Example: Computing Standard Deviation (sample)
Inferential statistics used to generalize back Population Sample • 2 General kinds of Statistics • Descriptive statistics • Used to describe, simplify, & organize data sets • Describing distributions of scores • Inferential statistics • Used to test claims about the population, based on data gathered from samples • Takes sampling error into account. Are the results above and beyond what you’d expect by random chance? Statistics
Population Sample A Treatment X = 80% Sample B No Treatment X = 76% • Purpose: To make claims about populations based on data collected from samples • What’s the big deal? • Example Experiment: • Group A - gets treatment to improve memory • Group B - gets no treatment (control) • After treatment period test both groups for memory • Results: • Group A’s average memory score is 80% • Group B’s is 76% • Is the 4% difference a “real” difference (statistically significant) or is it just sampling error? Inferential Statistics
Step 2: Set your decision criteria • Step 3: Collect your data from your sample(s) • Step 4: Compute your test statistics • Step 5: Make a decision about your null hypothesis • “Reject H0” • “Fail to reject H0” • Step 1: State your hypotheses Testing Hypotheses
This is the hypothesis that you are testing • Step 1: State your hypotheses • Null hypothesis (H0) • Alternative hypothesis(ses) • “There are no differences (effects)” • Generally, “not all groups are equal” • You aren’t out to prove the alternative hypothesis(although it feels like this is what you want to do) • If you reject the null hypothesis, then you’re left with support for the alternative(s) (NOT proof!) Testing Hypotheses
Step 1: State your hypotheses • In our memory example experiment • Null H0: mean of Group A = mean of Group B • Alternative HA: mean of Group A ≠ mean of Group B • (Or more precisely: Group A > Group B) • It seems like our theory is that the treatment should improve memory. • That’s the alternative hypothesis. That’s NOT the one the we’ll test with inferential statistics. • Instead, we test the H0 Testing Hypotheses
Step 2: Set your decision criteria • Your alpha level will be your guide for when to: • “reject the null hypothesis” • “fail to reject the null hypothesis” • Step 1: State your hypotheses • This could be correct conclusion or the incorrect conclusion • Two different ways to go wrong • Type I error: saying that there is a difference when there really isn’t one (probability of making this error is “alpha level”) • Type II error: saying that there is not a difference when there really is one Testing Hypotheses
Real world (‘truth’) H0 is correct H0 is wrong Type I error Reject H0 Experimenter’s conclusions Fail to Reject H0 Type II error Error types
Real world (‘truth’) Defendant is innocent Defendant is guilty Type I error Find guilty Jury’s decision Type II error Find not guilty Error types: Courtroom analogy
Type I error: concluding that there is an effect (a difference between groups) when there really isn’t. • Sometimes called “significance level” • We try to minimize this (keep it low) • Pick a low level of alpha • Psychology: 0.05 and 0.01 most common • For Step 5, we compare a “p-value” of our test to the alpha level to decide whether to “reject” or “fail to reject” to H0 • Type II error: concluding that there isn’t an effect, when there really is. • Related to the Statistical Power of a test • How likely are you able to detect a difference if it is there Error types
Step 3: Collect your data from your sample(s) • Step 4: Compute your test statistics • Descriptive statistics (means, standard deviations, etc.) • Inferential statistics (t-tests, ANOVAs, etc.) • Step 5: Make a decision about your null hypothesis • Reject H0“statistically significant differences” • Fail to reject H0“not statistically significant differences” • Make this decision by comparing your test’s “p-value” against the alpha level that you picked in Step 2. • Step 1: State your hypotheses • Step 2: Set your decision criteria Testing Hypotheses
Consider the results of our class experiment ✓ • Main effect of cell phone ✓ • Main effect of site type ✓ • An Interaction between cell phone and site type 0.04 -0.78 Factorial designs Resource: Dr. Kahn's reporting stats page
“Statistically significant differences” • When you “reject your null hypothesis” • Essentially this means that the observed difference is above what you’d expect by chance • “Chance” is determined by estimating how much sampling error there is • Factors affecting “chance” • Sample size • Population variability Statistical significance
Population mean Population Distribution Sampling error (Pop mean - sample mean) x n = 1 Sampling error
Population mean Population Distribution Sample mean Sampling error (Pop mean - sample mean) x x n = 2 Sampling error
Population mean Population Distribution Sample mean x x x x x x x x x x Sampling error (Pop mean - sample mean) • Generally, as the sample size increases, the sampling error decreases n = 10 Sampling error
Small population variability Large population variability • Typically the narrower the population distribution, the narrower the range of possible samples, and the smaller the “chance” Sampling error
Population Distribution of sample means XB XC XD Avg. Sampling error “chance” XA • These two factors combine to impact the distribution of sample means. • The distribution of sample means is a distribution of all possible sample means of a particular sample size that can be drawn from the population Samples of size = n Sampling error
“A statistically significant difference” means: • the researcher is concluding that there is a difference above and beyond chance • with the probability of making a type I error at 5% (assuming an alpha level = 0.05) • Note “statistical significance” is not the same thing as theoretical significance. • Only means that there is a statistical difference • Doesn’t mean that it is an important difference Significance
Failing to reject the null hypothesis • Generally, not interested in “accepting the null hypothesis” (remember we can’t prove things only disprove them) • Usually check to see if you made a Type II error (failed to detect a difference that is really there) • Check the statistical power of your test • Sample size is too small • Effects that you’re looking for are really small • Check your controls, maybe too much variability Non-Significance
About populations Real world (‘truth’) H0 is correct H0 is wrong Type I error Reject H0 Experimenter’s conclusions Fail to Reject H0 Type II error 76% 80% XB XA • Example Experiment: • Group A - gets treatment to improve memory • Group B - gets no treatment (control) • After treatment period test both groups for memory • Results: • Group A’s average memory score is 80% • Group B’s is 76% H0: μA = μB H0: there is no difference between Grp A and Grp B • Is the 4% difference a “real” difference (statistically significant) or is it just sampling error? Two sample distributions From last time
Real world (‘truth’) H0 is correct H0 is wrong Type I error Reject H0 Experimenter’s conclusions Fail to Reject H0 Type II error H0 is true (no treatment effect) 76% 80% XB XA • Tests the question: • Are there differences between groups due to a treatment? Two possibilities in the “real world” One population Two sample distributions “Generic” statistical test
Real world (‘truth’) H0 is correct H0 is wrong Type I error Reject H0 Experimenter’s conclusions Fail to Reject H0 Type II error XB XA XA • Tests the question: • Are there differences between groups due to a treatment? Two possibilities in the “real world” H0 is false (is a treatment effect) H0 is true (no treatment effect) Two populations XB 76% 80% 76% 80% People who get the treatment change, they form a new population (the “treatment population) “Generic” statistical test
XA XB • ER: Random sampling error • ID: Individual differences (if between subjects factor) • TR: The effect of a treatment • Why might the samples be different? (What is the source of the variability between groups)? “Generic” statistical test
TR + ID + ER ID + ER XA XB • The generic test statistic - is a ratio of sources of variability • ER: Random sampling error • ID: Individual differences (if between subjects factor) • TR: The effect of a treatment Observed difference Computed test statistic = = Difference from chance “Generic” statistical test
Population Distribution of sample means XB XC XD Avg. Sampling error “chance” XA • The distribution of sample means is a distribution of all possible sample means of a particular sample size that can be drawn from the population Samples of size = n Sampling error
Test statistic TR + ID + ER ID + ER • The generic test statistic distribution • To reject the H0, you want a computed test statistics that is large • reflecting a large Treatment Effect (TR) • What’s large enough? The alpha level gives us the decision criterion Distribution of the test statistic Distribution of sample means α-level determines where these boundaries go “Generic” statistical test
The generic test statistic distribution • To reject the H0, you want a computed test statistics that is large • reflecting a large Treatment Effect (TR) • What’s large enough? The alpha level gives us the decision criterion Distribution of the test statistic Reject H0 Fail to reject H0 “Generic” statistical test
The generic test statistic distribution • To reject the H0, you want a computed test statistics that is large • reflecting a large Treatment Effect (TR) • What’s large enough? The alpha level gives us the decision criterion Distribution of the test statistic Reject H0 “One tailed test”: sometimes you know to expect a particular difference (e.g., “improve memory performance”) Fail to reject H0 “Generic” statistical test
Things that affect the computed test statistic • Size of the treatment effect • The bigger the effect, the bigger the computed test statistic • Difference expected by chance (sample error) • Sample size • Variability in the population “Generic” statistical test
“A statistically significant difference” means: • the researcher is concluding that there is a difference above and beyond chance • with the probability of making a type I error at 5% (assuming an alpha level = 0.05) • Note “statistical significance” is not the same thing as theoretical significance. • Only means that there is a statistical difference • Doesn’t mean that it is an important difference Significance
Failing to reject the null hypothesis • Generally, not interested in “accepting the null hypothesis” (remember we can’t prove things only disprove them) • Usually check to see if you made a Type II error (failed to detect a difference that is really there) • Check the statistical power of your test • Sample size is too small • Effects that you’re looking for are really small • Check your controls, maybe too much variability Non-Significance
1 factor with two groups • T-tests • Between groups: 2-independent samples • Within groups: Repeated measures samples (matched, related) • 1 factor with more than two groups • Analysis of Variance (ANOVA) (either between groups or repeated measures) • Multi-factorial • Factorial ANOVA Some inferential statistical tests
Observed difference X1 - X2 T = Diff by chance Based on sample error • Design • 2 separate experimental conditions • Degrees of freedom • Based on the size of the sample and the kind of t-test • Formula: Computation differs for between and within t-tests T-test
Reporting your results • The observed difference between conditions • Kind of t-test • Computed T-statistic • Degrees of freedom for the test • The “p-value” of the test • “The mean of the treatment group was 12 points higher than the control group. An independent samples t-test yielded a significant difference, t(24) = 5.67, p < 0.05.” • “The mean score of the post-test was 12 points higher than the pre-test. A repeated measures t-test demonstrated that this difference was significant significant, t(12) = 5.67, p < 0.05.” T-test
XA XC XB • Designs • More than two groups • 1 Factor ANOVA, Factorial ANOVA • Both Within and Between Groups Factors • Test statistic is an F-ratio • Degrees of freedom • Several to keep track of • The number of them depends on the design Analysis of Variance