320 likes | 329 Views
Learn about the principles of statistical inference, including sampling distribution, hypothesis testing, and the concept of statistical significance.
E N D
URBDP 591 A Lecture 12:Statistical Inference Objectives • Sampling Distribution • Principles of Hypothesis Testing • Statistical Significance
Inferential Statistics • Based on probability theory. • The question that inferential statistics answers is whether the difference between the sample results and the population results is too great to be due to chance alone.
Sampling Distributions • A sampling distribution is the probability distribution of sample statistics. • A sampling distribution involves sample statistics that are distributed around a population parameter. • The central limit theorem. • (1) the sampling distribution will approximate a normal distribution, • (2) the mean of the sampling distribution will be equal to the population parameter.
The Central Limit Theorem • For any population, no matter what its shape or form, the distribution of means taken from that population with a constant sample size will be normally distributed if the sample size is sufficiently large (i.e., at least 30). • The standard deviation of the sampling distribution, or standard error, is equal to the standard deviation of the population divided by the square root of the sample size
Hypothesis testing • Step 1: Set up hypothesis • you should determine whether it is 1-tailed or 2-tailed test • Step 2: Compute test statistics • Step 3: Determine p-value of the test statistic • for a pre-determined a, you can find the corresponding critical limit • Step 4: Draw conclusion • reject H0 if p-value < alpha (ie greater than the critical limit) • accept H0 if p-value > alpha (ie less than the critical limit)
Test Statistics Very general statement: obtained difference Test statistic = difference expected by chance Basic form of all tests of significance: sample statistic – hypothesized population parameter Test statistic = standard error of the distribution of the test statistic e.g., z-score for use with sample means: m - X Z = X s X X
Interpreting p-values • p-value quantifies the role of chance • Large p-value Result may be due to chance • Small p-value Result unlikely to be due to chance. Conclude that a true and statistically significant difference exists
The logic of statistical testing 1. Assume the observed effect is due to chance - (This is null hypothesis - written H0.) 2. Find the probability of obtaining the observed effect or bigger when H0 is true. - (The p value) 3. If p is small then it is implausible that the effect is due to chance and we reject H0. (We call this result statistically significant.) 4. If p is large then the effect could be due to chance and we retain H0 as plausible. We call this result statistically not significant.
A single population mean • Suppose we want to study the effect of development on bird species richness on a randomly selected number of sites (n=100). • We measure species richness (X) after development, and mean species richness is 9. • Assume X follows a Normal distribution with a S.D. of 4.
Step 1: Set up hypothesis H0 : m0 = 8 H1 : m0 8 This is a 2-tailed test.
z - m x = 0 s / n Step 2: Compute test statistics x= (x1+x2+...+x100)/100 = 9 If x ~ N(m0 ,s2), then x ~ N(m0 ,s2/n), It follows a Normal distribution with mean 0 and variance 1 If sis known to be 4, then test statistics z = (9.0 - 8.0) / (4.0 / 100) = 2.5 23
Step 3: Determine p-value • Fora=0.05, Z0.05/2 = 1.96 (from Normal table)Since z-value = 2.5 > 1.96, so p-value < 0.05 • Fora=0.01, Z0.01/2 = 2.58 (from Normal table)Since z-value = 2.5 < 2.58, so p-value > 0.01
Step 4: Draw conclusion • We reject H0 at 5% level as p-value<0.05 and conclude that bird species richness is significantly different from 8 at 5% significance level. • Notice that we have to accept H0 at 1% level as p-value>0.01 and conclude that bird species richness is not statistically different from 8 at 1% significant level.
Difference between two population means • Suppose we want to study the effect of two development patterns A and B on bird species richness. We randomly select 52 sites which will be developed with high density development (development type A), and low density development (development type B). • We measure species richness (X) after development. The means for treatments A and B are 9 and 8 respectively. • Assume Normal distribution, and the S.D. for treatments A and B are 4 and 4.5.
Step 1: Set up hypothesis H0: mA = mB H1: mAmB orH0: mA - mB = 0 H1: mA-mB 0
Step 2: Compute test statistics SE( xA - xB ) = sA2/nA + sB2/nB = 8.51 Test statistic is: z = [ ( xA - xB ) - (mA-mB)] / SE( xA - xB ) = [ (90 - 80) - 0 ] / 8.51 =1.18
Step 3: Determine p-value • Fora=0.05, Z0.05/2 = 1.96Since z-value = 1.18 < 1.96, so p-value > 0.05 Step 4: Draw conclusion • We accept H0 at 5% level as p-value>0.05, and conclude that bird species richness after the two treatments are not statistically different at 5% significance level. In other words, the effects of the two treatments are not statistically different.
One-tailed vs. two-tailed tests • If there is an effect, the effect may either be positive (get better) or negative (get worse). • Two-tailed test is to study the existence of the effect in either direction (ie. positive or negative effect). • One-tailed test is to study the existence of the effect in one direction (eg. positive effect). • Provided that we have a priori knowledge about it • Directional hypothesis
Statistical significance vs. practical importance • If the sample size is unnecessarily large... • Differences may be established as statistically significant, and yet be too small to have any practical consequences. • The optimum sample size… • is just large enough to detect differences of a size which the researcher believes to be of practical importance. This firstly involves a professional assessment of how large a difference is important, followed by a power analysis to determine the required sample size.
Statistical significance vs. practical importance • If the sample size is unnecessarily large... • Differences may be established as statistically significant, and yet be too small to have any practical consequences. • The optimum sample size… • is just large enough to detect differences of a size which the researcher believes to be of practical importance. This firstly involves a professional assessment of how large a difference is important, followed by a power analysis to determine the required sample size.