1 / 32

Examining Relationships in Quantitative Research

Examining Relationships in Quantitative Research. 12. Learning Objectives_1. Understand and evaluate the types of relationships between variables Explain the concepts of association and covariation Discuss the differences between Pearson correlation and Spearman correlation.

Download Presentation

Examining Relationships in Quantitative Research

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Examining Relationships in Quantitative Research 12

  2. Learning Objectives_1 • Understand and evaluate the types of relationships between variables • Explain the concepts of association and covariation • Discuss the differences between Pearson correlation and Spearman correlation

  3. Learning Objectives_2 • Explain the concept of statistical significance versus practical significance • Understand when and how to use regression analysis

  4. Proctor & Gamble

  5. Describing Relationships Between Variables Presence Direction Strength of association Type

  6. Relationships between Variables • Is there a relationship between the two variables we are interested in? • How strong is the relationship? • How can that relationship be best described?

  7. Covariation and Variable Relationships • Covariation is amount of change in one variable that is consistently related to the change in another variable • A scatter diagram graphically plots the relative position of two variables using a horizontal and a vertical axis to represent the variable values

  8. Exhibit 12.1 Scatter Diagram Illustrates No Relationship

  9. Exhibit 12.2 Positive Relationship between X and Y

  10. Exhibit 12.3 Negative Relationship between X and Y

  11. Exhibit 12.4 Curvilinear Relationship between X and Y

  12. Correlation Analysis • Pearson Correlation Coefficient–statistical measure of the strength of a linear relationship between two metric variables • Varies between – 1.00 and +1.00 • The higher the correlation coefficient–the stronger the level of association • Correlation coefficient can be either positive or negative

  13. Exhibit 12.5 Strength of Correlation Coefficients

  14. Assumptions for Pearson’s Correlation Coefficient • The two variables are assumed to have been measured using interval or ratio-scaled measures • Nature of the relationship to be measured is linear • Variables to be analyzed come from a bivariate normally distributed population

  15. Exhibit 12.6 SPSS Pearson Correlation Example

  16. Substantive Significance • Coefficient of Determination (r2) is a number measuring the proportion of variation in one variable accounted for by another • The r2 measure can be thought of as a percentage and varies from 0.0 to 1.00 • The larger the size of the coefficient of determination, the stronger the linear relationship between the two variables under study

  17. How to Measure the Relationship between Variables Measured with Ordinal or Nominal Scales • Spearman Rank Order Correlation Coefficient is a statistical measure of the linear association between two variables where both have been measured using ordinal (rank order) scales

  18. Exhibit 12.7 SPSS Example Spearman Rank Order Correlation

  19. Exhibit 12.8 SPSS Median Example for Restaurant Selection Factors

  20. What is Regression Analysis? • A method for arriving at more detailed answers (predictions) than can be provided by the correlation coefficient • Assumptions • Variables are measured on interval or ratio scales • Variables come fro a normal population • Error terms are normally and independently distributed

  21. Exhibit 12.9 Straight Line Relationship in Regression

  22. Formula for a Straight Line • y = a + bX + ei • y = the dependent variable • a = the intercept • b = the slope • X = the independent variable used to predict y • ei = the error for the prediction

  23. Exhibit 12.10 Fitting the Regression Line Using the “Least Squares” Procedure

  24. Ordinary Least Squares (OLS) • OLS is a statistical procedure that estimates regression equation coefficients which produce the lowest sum of squared differences between the actual and predicted values of the dependent variable

  25. Exhibit 12.11 SPSS Results for Bivariate Regression

  26. Key Terms in Regression Analysis • Adjusted R-square • Explained variance • Unexplained variance • Regression coefficient

  27. Significance of Regression Coefficients • Answers these questions • Is there a relationship between the dependent and independent variable? • How strong is the relationship? • How much influence does the relationship hold?

  28. Multiple Regression Analysis • Multiple regression analysis is a statistical technique which analyzes the linear relationship between a dependent variable and multiple independent variables by estimating coefficients for the equation for a straight line

  29. Beta Coefficient • A beta coefficient is an estimated regression coefficient that has been recalculated to have a mean of 0 and a standard deviation of 1 in order to enable independent variables with different units of measurement to be directly compared on their association with the dependent variable

  30. Evaluating a Regression Analysis • Assess the statistical significance of the overall regression model using the F statistic and its associated probability • Evaluate the obtained r2 to see how large it is • Examine the individual regression coefficient and their t-test statistic to see which are statistically significant • Look at the beta coefficient to assess relative influence

  31. Exhibit 12.12 SPSS Example Multiple Regression

  32. Multicollinearity • Multicollinearity is a situation in which several independent variables are highly correlated with each other and can cause difficulty in estimating separate or independent regression coefficients for the correlated variables

More Related