300 likes | 319 Views
CHAPTER 10. Quantitative Data Analysis. Quantitative Research. Statistical analysis is a tool and not an end to itself. HYPOTHESIS, ERRORS AND SIGNIFICANCE. You will have to develop a range of propositions, or suppositions that tentatively explains certain facts or phenomena .
E N D
CHAPTER 10 Quantitative Data Analysis
Quantitative Research Statistical analysis is a tool and not an end to itself
HYPOTHESIS, ERRORS AND SIGNIFICANCE • You will have to develop a range of propositions, or suppositions that tentatively explains certain facts or phenomena. • The objective of the project is to provide information that you can then use to support the claim that your proposition is “true” or “false”. • The testing of hypotheses requires that you determine whether the results you have found did not occur by chance.
HYPOTHESIS • To statistically evaluate questions you need to put forward a hypothesis, which is an unproven testable proposition or supposition that tentatively explains certain facts or phenomena. • In its simplest form, the hypothesis is an “educated guess” based on the material that you have read.
HYPOTHESIS • The hypothesis is set up so that it will not be rejected (nullified) purely as a result of random error. • It is generally expressed in the negative i.e. there is no difference between “a” and “b”, or “a” and “b” are the same (i.e. a=b). • Rejecting this statement means that there is some difference between “a” and “b”, i.e. they are not equal.
Errors and Significance • When undertaking research you want to make sure that when you evaluate a hypothesis you: (a) do not randomly reject a null hypothesis that is true –Type 1 error OR (b) that you do not accept a null hypothesis that is false- Type II error. • The more certain you are that you have not committed a Type I error, the greater the chance is that you have committed Type II error!
Errors and Significance • What is the degree of error that I am willing to accept in my research? • Traditionally researchers say that they want to be 95% confident that the hypothesis is rejected, which is the same as a 5% chance of error or a probability level of 0.05 that the hypothesis is in fact correct. • 95% (5%) of the observations will fall within (outside) a statistically determined range if the null hypothesis is true (False).
Errors and Significance • Select the hypothesis and the level of significance before undertaking the actual collection of data and empirical testing, as you need to ensure that you have the right kind of data and the level of “error” you are willing to accept. • An empirical test that is statistically significant means there is a difference, it does not identify the “importance” of a relationship. • It is possible to have large differences that are not significant or small differences that are.
EMPERICALLY EXAMINING DATA There are MANY ways to empirically look at data, included those that allow us to explore: • Basic analysis • Basic Associations • Comparing groups • Basic causal relationships • Advanced Techniques
Summary of Scales Adapted from Hair et al. (2003, pp. 380–386)
Basic Analysis Descriptive analysis is a first step and allows you to get a better understanding of the data. • mean (average) response, tells you bit about the distribution of responses. • Frequency distribution, which tells you how many people, responded to each option and tell you the maximum and minimum response. • Variance, how much responses differ to the mean (will be important for statistical testing). • Confidence intervals- an estimate of the range in which the meanshould fall for a particular population.
Basic Analysis Confidence intervals- an estimate of the range in which the meanshould fall for a particular population; • You can vary the size of the interval by changing the probability that it is correct. • High confidence intervals suggest more variance than smaller intervals.
Sampling • You will not be able to collect data from everyone in your population of interest, thus you will have a sample that is representative of this group. • There may be many factors that affect your sample, demographics, when you collect the data, etc.
Cross Tabulations You examine responses across two or more variables at the same time. • Chi-squared is basis test as to whether the respondents are distributed as expected based on the two variables. • This is a kind of test of association, but again does not tell us A causes B, only that A and B are related.
Basic Associations • Correlations tell you the extent to which that two variables move together ranging from +1 to –1. • There are different types that can be used on different types of data. • The closer to 1 (positive or negative) the more directly the two items move together. • We also need to consider the significance, i.e. is the relationship random. • WE DO NOT however know that one causes the other!
Example of Chi-Squared results table for gender and the average weekly drinks
Comparing Groups • Z-test– compares portions between 2 groups. For example do the same portions of students get HD based on whether they are face-to-face or off-campus? • t-test, which compares the mean values between 2 groups. • With both there is a p-value indicating whether the difference is statically different.
Comparing Groups • Analysis of variance(Anova) do means of one variable (or interacting variables) differ across multiple groups. This does not tell you which group or groups differ. • Multiple Analysis of Variance (MANOVA)- technique has the benefit of identifying whether a set of variables differs between two groups, thus you do not have the multiple comparison problem identified earlier.
Figure 10.5 Anova tests Panel C Post Hoc Tests
Causal relationships - How does “A” influence “B” Simple linear regression - looks at how one variable (the independent variable) causes change in another (dependent variable). Multiple regression analysis looks at how several independent variables affect a dependent variable, while holding all other variables constant. Test statistics report on- • How well the regression fits the data (F). • Whether independent variables influence the dependent variable (t or beta and p value). • How much variance in the data is explained by the regression r or r2.
Advanced Techniques • Factor analysis - groups variables that move in similar ways, which are often added together (aggregated) into one variable (called a composite construct). • Cluster analysis - groups individuals based on their responses to questions. It is used in segmentation studies that differentiate groups based on how they respond to questions, rather than on simple demographic variables. The groups can then be used in other analysis. • Structurally equation modelling allows you to identify how variables relate within complex systems of relationships, moving beyond multiple regression.
Advanced Techniques Conjoint analysis or Choice modelling allows the researcher to provide respondents with different alternatives, where attributes are varied and the researcher can empirically determine the value of each attribute to the individual.
Conclusions • The analysis technique will directly relate to your research question. • Thus you need to know what you want to explore when designing your topic and then ensure you collect the right kind of data. • In many cases you will possibly use multiple methods to assist with aspects of the research issue. • In thinking about what you need to learn about, you will need to include, learning how to undertake your analysis!
Project Checklist • Will your project have a quantitative component? • Have you collected the right data for the analysis you want to do? • Who will be in charge of this? • When will it occur? • What preparation must be done? • Who will interpret the data, and how will it be interpreted?