380 likes | 501 Views
Multiple Regression Analysis: Part 2. Interpretation and Diagnostics. Learning Objectives. Understand regression coefficients and semi-partial correlations Learn to use diagnostics to locate problems with data (relative to MRA) Understand… Assumptions Robustness
E N D
Multiple Regression Analysis: Part 2 Interpretation and Diagnostics
Learning Objectives • Understand regression coefficients and semi-partial correlations • Learn to use diagnostics to locate problems with data (relative to MRA) • Understand… • Assumptions • Robustness • Methods of dealing with violations • Enhance our interpretation of equations • Understand entry methods
Statistical Tests & Interpretation • Interpretation of regression coefficients • Standardized • Unstandardized • Intercept • Testing regression coefficients • t-statistic & interpretation • Testing R2
Output for MRA Run (coefficients) R2 = .558
Y A B X1 X2 Variance in Y Accounted for by two uncorrelated Predictors (A+B)/Y = R2, E (in Y circle) equals Error. Y E E B A X1 X2 Example #1: Small R2, A represents variance in Y accounted for by X1, B = variance in Y accounted for by X2. Example #2: Larger R2, A represents variance in Y accounted for by X1, B = variance in Y accounted for by X2.
Y Y A B A B C C X1 X2 D X1 X2 D Example #1: Small R2 Example #2: Larger R2 Variance in Y Accounted for by two correlated Predictors: sr2 and pr2 sr2 for X1 = pr2 for X1 =
A shortcoming to breaking down sr2 R2 = .120
Multicollinearity: One way it can all go bad! Y E A B C X1 X2 D
Ways to fix multicollinearity • Discarding Predictors • Combining Predictors • Using Principal Components • Parcelling • Ridge Regression
Outliers and Influential Observations:Another way it can all go bad! • Outliers on y • Outliers on x’s • Influential data points
Outliers • Outliers on y • Standardized Residuals • Studentized Residuals (df = N – k – 1) • Deleted Studentized Residuals • Outliers on x’s • Hat elements • Mahalanobis Distance
Outliers on y tcrit(21) = 2.08
Outliers on Xs (Leverage) χ2(crit) for Mahalanobis’ Distance = 7.82
Influential Observations • Cook’s Distance (cutoff ≈ 1.0) • DFFITs [cut-offs of 2 or 2*((k+1)/n)0.5] • DFBeta • Standardized DF Beta
Once more, with feeling R2 = .687
A cautionary tale:Some more ways it can all go bad! We will use X to predict y1, y2 and y3 in turn.
Homoscadasticity:Yet another way it can all go bad! • What is homoscedasticity? • Is it better to have heteroscedasticity? • The effects of violation • How to identify it • Strategies for dealing with it
Effect Size Multiple Correlation (R): SMC (R2):
Cross Validation • Why • Useful statistics and techniques • Conditions under which likelihood of cross-validation is increased
Assumptions of Regression • Sample Size • Absence of Outliers & Influential Observations • Absence of Multicollinearity and Singularity • Normality • Linearity • Homoscedasticity of Errors • Independence of Errors
Structure Coefficients • What are they? • Vs. pattern coefficients or “weights” • Why we may need both • When they would be used in MRA • Why they are not commonly used • How you get them in SPSS • CD sales example
Model Building in MRA:“Canned” procedures • Enter • Forward • Backward Selection (Deletion) • Stepwise • Hierarchical
Hierarchical – Example Predict employee satisfaction • Block 1: “Hygiene Factor” • Block 2: “Equity” • Block 3: “Organizational Commitment”
Interpretation revisited • In light of multicollinearity • Standardized or unstandardized? • Suppressor effects • Missing predictors • Correlated / uncorrelated predictors • Structure coefficients • Reliability of indicators • Mathematical maximization nature of MRA