110 likes | 237 Views
Lesson 4 - R. Chapter 4 Review More about Relationships Between Two Variables. Objectives. Identify settings in which a transformation might be necessary in order to achieve linearity. Use transformations involving powers and logarithms to linearize curved relationships.
E N D
Lesson 4 - R Chapter 4 Review More about Relationships Between Two Variables
Objectives • Identify settings in which a transformation might be necessary in order to achieve linearity. • Use transformations involving powers and logarithms to linearize curved relationships. • Explain what is meant by a two-way table, and describe its parts. • Give an example of Simpson’s paradox. • Explain what gives the best evidence for causation. • Explain the criteria for establishing causation when experimentation is not feasible.
Vocabulary • None new
Residuals • One difference between math and stat is that statistics assumes that the measurements are not exact, that there is an error or residual • The formula for the residual is always Residual = Observed – Predicted • This relationship is not just for this chapter … it is the general way of defining error in statistics • The least squares regression line minimizes the sum of the square of the residuals
The residual The model line The observed value y The predicted value y The x value of interest Residual on the Scatter Diagram
Least-Squares Regression Model Scope Scope of the model Response Areas outside the scope of the model Explanatory Linear relationship outside the scope of the modelis not guaranteed!
Things to Remember • In the Least Square Regression model: • R² gives us the % of the variation in the response variable explained by the model • the mean of the response variable depends on the linearity of the explanatory variable • the residuals must be normally distributed with constant error variance • the εi will be normally distributed (0,σ²) • we are estimating two values (b0 b1) and therefore lose 2 degrees of freedom in the t-statistic • the prediction interval for an individual response will be wider than the confidence interval for a mean response • procedures are robust
Things to Remember • In Multiple Regression models: • the adjusted R² gives us the % of the variation in the response variable explained by the model • R² is adjusted based on # of sample and number of explanatory variables in the model • multicollinearity may be a problem if a high linear correlation exists between explanatory variables • Rule: |correlation| > 0.7 then multicollinearity possible • the procedure used in our book for multiple regression modeling is called backwards step-wise regression • Rule: remove explanatory variable with highest p-value and then rerun the model check adjusted R² values
Multiple Regression Model yi = β0 + β1x1i + β2x2i + … + βkxki + εi where yi is the value of the response variable for the ith individual β0, β1, β2, , βk ,are the parameters to be estimated based on the sample data x1iis the ithobservation for the first explanatory variable,x2iis the ithobservation for the second explanatory variable and so on εi is am independent random error term that is normally distributed with mean 0 and variance = σ² i = 1, 2, 3, …, n, where n is the sample size • The adjusted R² is used in multiple regression models • The adjusted R² will decrease if a variable is added to the model that does little to explain the variation in the response variable. • The adjusted R² will increase if a variable is taken from the model that does little to explain the variation in the response variable.
ANOVA • We use the Analysis of Variance (ANOVA) method to compare three or more means • ANOVA analyzes the differences in the sample means using sums of squares and an F test • One-way ANOVA analyzes models where there is only one factor that differentiates the groups • Two-way ANOVA analyzes models where there are two factors that differentiate the groups
Summary and Homework • Summary • Homework • pg