510 likes | 588 Views
(Chapter 2.1–2.7). Lecture 2: Estimating a Slope. Recall 計量經濟學的構成. 經濟理論 數理模型 統計理論 計量模型. 現象. 經濟理論. 數理模型. 統計理論. 計量模型. 計量與 經濟理論 之差異 ?. Economic theory : qualitative results— Demand Curves Slope Downward
E N D
(Chapter 2.1–2.7) Lecture 2: Estimating a Slope
Recall 計量經濟學的構成 • 經濟理論 • 數理模型 • 統計理論 • 計量模型 現象 經濟理論 數理模型 統計理論 計量模型
計量與經濟理論之差異? • Economic theory: qualitative results— Demand Curves Slope Downward • Econometrics: quantitative results— price elasticity of demand for milk = -.75
計量與統計之差異? • Statistics: “summarize the data faithfully”; “let the data speak for themselves.” • Econometrics: “ what do we learn from economic theory AND the data at hand?”
計量能做啥事? • Estimation: What is the marginal propensity to consume of Taiwan? (結構分析) • Hypothesis Testing: Do Korean college workers’ productivity higher than Taiwan?(檢定假說) • Prediction & Forecasting: What will Personal Savings be in 2004if GDP is $14,864? And will it grow in the near future (2008)?(預期及預測)
Economists Ask: “What Changes What and How?” • Higher Income, Higher Saving • Higher Price, Lower Quantity Demanded • Higher Interest Rate, Lower Investment
6000 5000 4000 3000 2000 1000 0 24000 48000 72000 96000 Savings Versus Income • Theory Would Assume an Exact Relationship, e.g., Y =bX
Slope of the Line Is Key! • Slope is the change in savings with respect to changes in income • Slope is the derivative of savings with respect to income • If we know the slope, we’ve quantified the relationship!
Long-run Consumption Function 特點 • 向上斜 • 經原點 • 猜猜斜率?
Underlying Mean + Random Part • (憑直覺) 四大猜法 four intuitively appealing ways to estimate b
估計策略 • Min Σ (Y – Y) • Min Σ ∣ Y – Y ∣ • Min Σ (Y – Y) • 優劣點 Y為配適值 2
“Best Guess 1” Mean of Ratios: Y X 共有n個
Figure 2.4 Estimating the Slope of a Line with Two Data Points
“Best Guess 2” Ratio of Means:
“Best Guess 3” Mean of Changes in Y over Changes in X: y X
“Best Guess 4” Ordinary Least Squares: (minimizes squared residuals in sample)
Underlying Mean + Random Part • Are lines through the origin likely phenomena?
Regression’s Greatest Hits!!! • An Econometric Top 40
Two Classical Favorites!! • Friedman’s Permanent Income hypothesis: • Capital Asset Pricing Model (CAPM) :
A Golden Oldie !! • Engel on the Demand for Rye:
Four Guesses • How to Choose?
What Criteria Did We Discuss? • Pick The One That's Right • Make Mean Error Close to Zero • Minimize Mean Absolute Error • Minimize Mean Square Error
What Criteria Did We Discuss? (cont.) • Pick The One That's Right… • In every sample, a different estimator may be “right.” • Can only decide which is right if we ALREADY KNOW the right answer—which is a trivial case.
What Criteria Did We Discuss? (cont.) • Make Mean Error Close to Zero …seek unbiased guesses • IfE(bg-b) = 0, bgis right on average • If BIAS = 0, bg is an unbiased estimator of b
Checking Understanding • Question: Which estimator does better under the “minimize mean error” condition? • bg-bis always a positive number less than 2 (our guesses are always a little high), or • bg-bis always +10 or -10 (50/50 chance)
Checking Understanding (cont.) • If our guess is wrong by +10 for half the observations, and by -10 for the other half, then E(bg-b) = 0! • The second estimator is unbiased! • Mistakes in opposite directions cancel out.The first estimator is always closer to being right, but it does worse on this criterion.
What Criteria Did We Discuss? • Minimize Mean Absolute Error… • Mistakes don’t cancel out. • Implicitly treats cost of a mistake as being proportional to the mistake’s size. • Absolute values don’t go well with differentiation.
What Criteria Did We Discuss? (cont.) • Minimize Mean Square Error… • Implicitly treats cost of mistakes as disproportionately large for larger mistakes. • Squared expressions are mathematically tractable.
What Criteria Did We Discuss? (cont.) • Pick The One That’s Right… • only works trivially • Make Mean Error Close to Zero… • seek unbiased guesses • Minimize Mean Absolute Error… • mathematically tough • Minimize Mean Square Error… • more tractable mathematically
Criteria Focus Across Samples • Make Mean Error Close to Zero • Minimize Mean Absolute Error • Minimize Mean Square Error • What do the distributions of the estimators look like?
Try the Four in Many Samples • Pros will use estimators repeatedly— what track record will they have? • Idea: Let’s have the computer create many, many data sets. • We apply all our estimators to each data set.
Try the Four in Many Samples (cont.) • We use our estimates on many datasets that we created ourselves. • We know the true value of b because we picked it! • We can compare estimators. • We run “horseraces.”
Try the Four in Many Samples (cont.) • Pros will use estimators repeatedly—what track record will they have? • Which horse runs best on many tracks? • Don’t design tracks that guarantee failure. • What properties do we need our computer-generated datasets to have to avoid automatic failure for one of our estimators?
Building a Fair Racetrack Under what conditions will each estimator fail?
Why Does Viewing Many Samples Work Well? • We are interested in means: mean error, mean absolute error, mean squared error. • Drawing many (m) independent samples lets us estimate means with variance e2/m, where e2 is the variance of that mean’s error. • If m is large, our estimates will be quite precise.
How to Build a Race Track... • n= ? • How big is each sample? • b = ? • What slope are we estimating? • Set X1, X2, … , Xn • Do it once, or for each sample? • Draw e1, e2, ... , en • Must draw randomly each sample.
What to Assume About the ei ? • What do the eirepresent? • What should the eiequal on average? • What variance do we want for the ei?
n= ? How big is each sample? b= ? What slope are we estimating? Set X1 , X2 , … , Xn Do it once, or for each sample? Draw e1 , e2 , … , en Must draw randomly each sample. Form Y1 , Y2 , … , Yn Yi =bXi+ ei We create 10,000 datasets with X and Y. For each dataset, what do we want to do? Checking Understanding
Checking Understanding (cont.) • We create 10,000 datasets with X and Y • For each dataset, we use all four of our estimators to estimate bg1, bg2, bg3,and bg4 • We save the mean error, mean absolute error, and mean squared error for each estimator
What Have We Assumed? • We are creating our own data. • We get to specify the underlying “Data Generating Process” relating Y to X. • What is our Data Generating Process (DGP)?
What Is Our Data Generating Process? • E(ei ) = 0 • Var(ei ) = 2 • Cov(ei ,ek ) = 0 i ≠ k • X1 , X2 , … , Xn are fixed across samples GAUSS–MARKOV ASSUMPTIONS
What Will We Get? • We will get precise estimates of: • Mean Error of each estimator • Mean Absolute Error of each estimator • Mean Squared Error of each estimator • Distribution of each estimator • By running different racetracks (DGPs), we check the robustness of our results.
Review • We want an estimator to form a “best guess” of the slope of a line through the origin. • Yi = bXi +ei • We want an estimator that works well across many different samples: low average error, low average absolute error, low squared errors…
Review (cont.) • We have brainstormed 4 “best guesses”:
Review (cont.) • We will compare these estimators in “horseraces” across thousands of computer-generated datasets • We get to specify the underlying relationship between Y and X • We know the “right answer” that the estimators are trying to guess • We can see how each estimator does
Review (cont.) • We choose all the rules for how our data are created. • The underlying rules are the “Data Generating Process” (DGP) • We choose to use the Gauss–Markov Rules.