720 likes | 949 Views
Inference in the General Normal Linear Model. Hypothesis Testing Under General Linear Model. Current model: Y=X β + e How is the dependent variable, Y, influenced by changes in explanatory variables?
E N D
Inference in the General Normal Linear Model
Hypothesis Testing Under General Linear Model • Current model: Y=Xβ + e • How is the dependent variable, Y, influenced by changes in explanatory variables? • How do we test hypotheses about the magnitudes of such influences or combinations of individual coefficients (e.g., elasticities)?
Hypothesis Testing Under General Linear Model • Previously we derived the sampling property results assuming normality: • Y = Xb + e where et~N(0,s2) • → Y~N(Xb,s2IT) • bs=(X'X)-1X'Y, E(bs)=b • Cov(bs)= β=s2(X'X)-1 • bl~N(b, s2(X'X)-1) • σU2 unbiased estimate of σ2 • An estimate of Cov(βs) = βs=σU2(X'X)-1 el = y - Xβl
Hypothesis Testing Under General Linear Model • Single Parameter (βk,L) Hypothesis Test • βk,l~N(βk,Var(βk)) Σβs=σu2(X'X)-1 kth diagonal element of βs unknown true coeff. • When σ2 is known: • When σ2 not known:
Hypothesis Testing Under General Linear Model • Can obtain (1-) CI for βk: • There is a (1-α) probability that the true unknown value of β is within this range • Does this interval contain our hypothesized value? • If it does, than we can not reject H0
Hypothesis Testing Under General Linear Model • Testing More Than One LinearCombination of Estimated Coefficients • Assume we have a-priori information about the value of β • We can represent this information via a set of J-Linear hypotheses (or restrictions): • In matrix notation
Hypothesis Testing Under General Linear Model known coefficients
Hypothesis Testing Under General Linear Model • Assume we have a model with 5 parameters to be estimated • Joint hypotheses: β1=8 and β2=β3 • J=2, K=5 β2-β3=0
Hypothesis Testing Under General Linear Model • How do we obtain parameter estimates if J hypotheses are true? • Constrained (Restricted) Least Squares • bR is β that minimizes: S=(Y-Xβ)'(Y-Xβ) s.t. Rβ=r = e'e s.t. Rβ=r e.g. we act as if H0 are true • S*=(Y-Xβ)'(Y-Xβ)+λ'(r-Rβ) • λ is (J x1) Lagrangian multipliers associated with J-joint hypotheses • We want to choose β such that we minimize SSE but also satisfy the J constraints (hypotheses), βR
Hypothesis Testing Under General Linear Model • Min. S*=(Y-Xβ)'(Y-Xβ) + λ'(r-Rβ) • What and how many FOC’s? • K+J FOC’s K-FOC’s J-FOC’s
Hypothesis Testing Under General Linear Model S*=(Y-Xβ)'(Y-Xβ)+λ'(r-Rβ) • What are the FOC’s? CRM βS • Substitute these FOC into 2nd set ∂S*/∂λ = (r-RβR) = 0J →
Hypothesis Testing Under General Linear Model • The 1st FOC • Substitute the expression for λ/2 into the 1st FOC:
Hypothesis Testing Under General Linear Model • βR is the restricted LS estimator of β as well as the restricted ML estimator • Properties of Restricted Least Squares Estimator • →E(bR) b if Rb r • V(bR) ≤ V(bS) →[V(bS) - V(bR)]is positive semi-definite • diag(V(bR)) ≤ diag(V(bS)) True but Unknown Value
Hypothesis Testing Under General Linear Model • From above, if Y is multivariate normal and H0 is true • βl,R~N(β,σ2M*(X'X)-1M*') ~N(β,σ2M*(X'X)-1) • From previous results, if r-Rβ≠0 (e.g., not all H0 true), estimate of β is biased if we continue to assume r-Rβ=0 ≠0
Hypothesis Testing Under General Linear Model • The variance is the same regardless of he correctness of the restrictions and the biasedness of βR • → βR has a variance that is smaller when compared to βs which only uses the sample information.
Hypothesis Testing Under General Linear Model • Beer Consumption Example Joint Hypotheses: • Real Prices Matter? • All prices and INC by 10% • β1 + β2 + β3 + β4=0 • Equal Price Impacts? • Liquor and Other Goods • β2=β3 • Unitary Income Elasticity? • β4=1 • Data used in the analysis
Hypothesis Testing Under General Linear Model • Given the above, what does the R-matrix and r vector look like for these joint tests? • Lets develop a test statistic to test these joint hypotheses • We are going to use the Likelihood Ratio (LR) to test the joint hypotheses
Hypothesis Testing Under General Linear Model • LR=lU*/lR* • lU*=Max [l(|y1,…,yT); =(β, σ) ] = “unrestricted” maximum likelihood function • lR*=Max[l(|y1,…,yT); • =(β, σ); Rβ=r] = “restricted” maximum likelihood function • Again, because we are possibly restricting the parameter space via our null hypotheses, LR≥1
Hypothesis Testing Under General Linear Model • If lU* is large relative to lR*→data shows evidence that the restrictions (hypotheses) are not true (e.g., reject null hypothesis) • How much should LR exceed 1 before we reject H0? • We reject H0 when LR ≥ LRC where LRC is a constant chosen on the basis of the relative cost of the Type I vs. Type II errors • When implementing the LR Test you need to know the PDF of the dependent variable which determines the density of the test statistic
Hypothesis Testing Under General Linear Model • For LR test, assume Y has a normal distribution • →e~N(0,σIT) • This implies the following LR test statistic (LR*) • What are the distributional characteristics of LR*? • Will address this in a bit
Hypothesis Testing Under General Linear Model • We can derive alternative specifications of LR test statistic • LR*=(SSER-SSEU)/(Js2U) (ver. 1) • LR*=[(Rbe-r)′[R(X′X)-1R′]-1(Rbe-r)]/(Js2U)(ver. 2) • LR*=[(bR-be)′(X′X)(bR-be)]/(Js2U) (ver. 3) βe =βS=βl • What are the Distributional Characteristics of LR* (JHGLL p. 255) • LR* ~ FJ,T-K • J = # of Hypotheses • K= # of Parameters • (including intercept)
Hypothesis Testing Under General Linear Model • Proposed Test Procedure • Choose a = P(reject H0| H0 true) = P(Type-I error) • Calculate the test statistic LR* based on sample information • Find the critical value LRcrit in an F-table such that: a = P(F(J, T – K)³ LRcrit), where α = P(reject H0| H0 true) f(LR*) α = P(FJ,T-K ≥ LRcrit) LRcrit α
Hypothesis Testing Under General Linear Model • Proposed Test Procedure • Choose a = P(reject H0| H0 true) = P(Type-I error) • Calculate the test statistic LR* based on sample information • Find the critical value LRcrit in an F-table such that: a = P(F(J, T – K)³ LRcrit), where α = P(reject H0| H0 true) • Reject H0 if LR* ³ LRcrit • Don’t reject H0 if LR* < LRcrit
Hypothesis Testing Under General Linear Model • Beer Consumption Example • Does the regression do a better job in explaining variation in beer consumption than if assumed the mean response across all obs.? • Remember SSE=(T-K)σ2U • Under H0: All slope coefficients=0 • Under H0, TSS=SSE given that that there is no RSS and TSS=RSS+SSE
Hypothesis Testing Under General Linear Model SSE = 0.059972 *25 = 0.08992 R2=1- 0.08992/0.51497 TSS=SSER Mean of LN(Beer)
Hypothesis Testing Under General Linear Model • Results of our test of overall significance of regression model • Lets look at the following GAUSS Code • GAUSS command: • CDFFC(29.544,4,25)=3.799e-009 • CDFFC Computes the complement of the cdf of the F distribution (1-Fdf1,df2) • Unlikely value of F if hypothesis is true, that is no impact of exogenous variables on beer consumption • Reject the null hypothesis • An alternative look
Hypothesis Testing Under General Linear Model • Beer Consumption Example • Three joint hypotheses example • Sum of Price and Income Elasticities Sum to 0 (e.g., β1 + β2 + β3 + β4=0) • Other Liquor and Other Goods Price Elasticities are Equal (e.g., β2=β3) • Income Elasticity = 1 (e.g., β4=1) • cdffc(0.84,3,25)=0.4848
Hypothesis Testing Under General Linear Model • Location of our calculated test statistic PDF F3,25 F 0.84 area = 0.4848
Hypothesis Testing Under General Linear Model • A side note: How do you estimate the variance of an elasticity and therefore test H0 about this elasticity? • Suppose you have the following model: FDXt = β0 + β1Inct + β2 Inc2t + et • FDX= food expenditure • Inc=household income • Want to estimate the impacts of a change in income on expenditures. Use an elasticity measure evaluated at mean of the data. That is:
Hypothesis Testing Under General Linear Model FDXt = β0 + β1Inct + β2 Inc2t + et • Income Elasticity (Γ) is: • How do you calculate the variance of Γ? • We know that: Var(α′Z)= α′Var(Z)α • Z is a column vector of RV’s • α a column vector of constants • Treat β0, β1 and β2 are RV’s. The α vector is: Linear combination of Z
Hypothesis Testing Under General Linear Model • This implies var(Γ) is: (1 x 3) σ2(X'X)-1 (3 x 3) (3 x 1) Due to 0 α value (1 x 1)
Hypothesis Testing Under General Linear Model C22 • This implies: var(Γ) is: C12 2C1C2
Hypothesis Testing Under General Linear Model • Example of Estimation of Single Output Production Function • Example 6.2, Greene, pp:102-104 • SIC-33: Primary Metals • Cross-Sectional • VA=f(Labor, Capital) • 27 observations
Hypothesis Testing Under General Linear Model
Hypothesis Testing Under General Linear Model • Example of Estimation of Production Function • Cobb-Douglas specification • Translog Functional Form • Cobb-Douglas H0: β3= β4= β5=0 • GAUSS code for calculating critical values of various test statistics
Hypothesis Testing Under General Linear Model • Coefficient on lnK<0 (β2 = -1.8931) • Does this mean that its output elasticity (ψY,K ≡ ∂lnY/∂lnK) is negative? • ψY,K=β2 + β4lnK + β5lnL
Hypothesis Testing Under General Linear Model • Using mean of the natural logarithm of inputs → ψY,K=0.5425 • Note, another point of evaluation could have been at the natural logarithm of the means: • Is ψY,K in fact positive? • Var(ψY,K)=W'ΣβSW • ΣβS is the full 6x6 parameter covariance matrix (1 x 1)
Hypothesis Testing Under General Linear Model • Var(ψY,K)=W'ΣβSW=0.01259 • H0: ψY,K=0, H1: ψY,K>0 • t.025,22 = 2.074 (critical value) • = Ψstd. dev. • t=0.5425/0.1122= 4.84 • Reject H0 • Conclude that the output elasticity for capital is positive
Asymptotic Hypothesis Tests • Asymptotic Tests (t→∞) H0: Rβ = r H1: Rβ≠ r • In general, as T→∞, J*FJ,T-K≈χ2J tT-K≈N(0,1) • → If equation error term not normally distributed and have a “large” number of observations→ joint hypothesis test should be based on J*LR*~c2J (Greene:104-108, JHGLL:268-270) • Define Ver. 2 of LR*
Asymptotic Hypothesis Tests • Using the Central Limit Theorem: • If R is known:
Asymptotic Hypothesis Tests • With J=1 • That is, when you have a single linear restriction and the error term is not normally distributed inferences with respect to this single restriction (hypothesis) should be based on a normal distribution rather than a t-distribution
Asymptotic Hypothesis Tests • Proposed Asymptotic Test Procedure: • Choose α = P(Reject H0|H0 true) = P(Type I Error) • Estimate LR** =J x LR* based on the sample information • Find the critical value of such that a=P(c2J >c2J,C) • Reject H0 if LR**>c2J,C, • Do not reject H0 if LR**< c2J,C
Testing Nonlinear Restrictions • Previously limited to linear H0 • Lets extend the model to: • H0: c(β)=r • c(β) may be J nonlinear functions • When J = 1: • t-stat and F-stat is same as before • Need to obtain estimate of variance of nonlinear function, c(β)
Testing Nonlinear Restrictions • Greene (p.109) shows the asymptotic variance of a nonlinear function is: • If c(β) is linear (e.g., Rβ=r) then: σ2U(X'X)-1 Asymptotic variance
Testing Nonlinear Restrictions • Wald statistic for testing J H0: • c(β) = r • W~χ2J • For an example see Greene p.109-110 (1 x 1)
Testing Non-Nested Models • Up to this point our hypotheses have expressed hypotheses in terms of restrictions of the underlying model • For the linear specification we have y = Xβ + ε H0: Rβ = q H1: Rβ≠ q • This could be a limiting type of hypothesis testing environment • Which of two possible sets of regressors are more appropriate • Whether a linear or log-linear model is more appropriate • Lets consider the following: H0: y = Xβ + ε0 H1: y = Zγ + ε1 Can not use previous procedures to evaluate nonnested hypotheses
Testing Non-Nested Models H0: y=Xβ+ε0 H1: y=Zγ+ε1 • Greene, p. 153-159 • Usual procedure is to seek evidence in the sample to reject H0 (with a positive probability of Type I error) • Considering the above non-nested hypothesis, each competing hypothesis is given equal weight • There is no natural null hypothesis • Common for the analyst to first test using one as H0 and then switch roles and undertake a test with the other is H0 • It could happen that both or neither model is rejected
Testing Non-Nested Models H0: y=Xβ+ε0 H1: y=Zγ+ε1 • We will examine 3 methods for undertaking non-nested tests: • Cox’s method based on maximum likelihood procedures • The Encompassing Principle: Can a maintained model explain the features of (encompass) its competitors • The Comprehensive Model Approach: Contains both models as a special case • Encompassing Principle: • Model 0 encompasses Model 1 if the features of Model 1 can be explained by Model 0 but the reverse is not true
Testing Non-Nested Models H0: y=Xβ+ε0 H1: y=Zγ+ε1 • Let • X* be the set of variables in X that are not in Z • Z* be the set of variables in Z that are not in X • W be the set of variables in both • →we can combine H0 and H1 into a single supermodel y = X*β* + Z*γ* + Wδ + ε • H1 is rejected if γ*=0 using conventional F-test (joint linear test) • H0 is rejected if β*=0 using F-test • Two problems • 1st: δ remains a part of β and γ→test does not really distinguish between Model under H0 and under H1 but between model H1 and hybrid model
Testing Non-Nested Models H0: y=Xβ+ε0 H1: y=Zγ+ε1 • Two problems • 2nd: the compound model may have a large number of regressors → collinearity could be a problem • An Alternative Encompassing Model specification: • If H0 is correct, then y may be fully explained by X except for a random error term, ε • Suppose one then attempts to estimate γ by regressing y on Z → • If H0 is correct then the estimated coefficient will be the same if one were to regress Xβ on Z since ε0 is random noise under H0 • Since β must be estimated, use βS instead, XβS = f(Z) → γ0