140 likes | 156 Views
Unanswered Questions in Typical Literature Review. 1. Thoroughness How thorough was the literature search? Did it include a computer search and a hand search? In a computer search, what were the descriptors used? Which journals were searched?
E N D
Unanswered Questions inTypical Literature Review • 1. Thoroughness • How thorough was the literature search? • Did it include a computer search and a hand search? • In a computer search, what were the descriptors used? • Which journals were searched? • Were theses and dissertations searched and included?
2. Inclusion/Exclusion • On what basis were studies included or excluded from the written review? • Were theses and dissertations arbitrarily excluded? • Did the author make decisions about the inclusion or exclusion based on the • perceived internal validity of the research? • sample size? • research design? • use of appropriate statistics?
3. Conclusions • Were conclusions based on the number of studies supporting or refuting a point (called vote counting)? • Were studies weighted differently according to sample size? • meaningfulness of the results? • quality of the journal? • internal validity of the research?
Establishing Cause and Effect • The selection of a good theoretical framework • The application of an appropriate experimental design • The use of the correct statistical model and analyses • The proper selection and control of the independent variable • The appropriate selection and measurement of the dependent variable • The use of appropriate subjects • The correct interpretation of the results
Reliability and Validity • Reliability - the consistency or repeatability of a measure; is the measure reproducible • Validity - the truthfulness of a measure; validity depends on reliability and relevance
Reliability • The observed measure is the summation of the true score and the error score • The more reliable a test, the less error is involved • Reliability is defined as the proportion of observed score variance that is true score variance • Reliability is determined using Pearson’s correlation coefficient or ANOVA
Types of Reliability • Interclass reliability - Pearson’s Product Moment (Correlation of only two variables) • Test -retest reliability (stability)-determines if a single test is stable over time • Equivalence reliability - are two tests similar in measuring the same item or trait • Split halves reliability - estimates the reliable of a test based on the scores of the odd and even test items • Spearman Brown prophecy - estimates test reliability based on addition or deletion of test items
Intraclass Reliability • Reliability within an individual’s scores of more than two measures • Cronbach’s alpha estimates the reliability of tests. • Uses AVOVA to determine mean differences within and between an individual’s scores
Indices of Reliability • Index of reliability - the theoretical correlation between true and observed scores; IR= rxx • Standard Error of Measure - SEM the degree of fluctuation of an individual’s observed score from their true score SEM = s(1-rxx)
Factors Affecting Reliability • Fatigue • Practice • Ability • Time between tests • Testing Circumstances • Test Difficulty • Type of measurement • Environment
Validity • Content validity - face validity, logical validity • Criterion Validity- measures are related to specific criterion • Predictive validity • Concurrent validity • Construct validity - test validity as a measure of psychological constraints
Possible Test Items 1 • Be able to define and differentiate between reliability and validity; what types of error do each try to explain. • Know the different classes and types of reliability; be able to differentiate between the different scores that are included in reliability. • Be able to calculate the reliability of test examples and describe what type/class of reliability is being defined.
Possible Test Items 2 • Be able to define/describe/determine Cronbach’s Alpha, Reliability Index, and the Standard Error of Measurement. • Know what factors affect test reliability and how to compensate to make a more reliable test. • Be able to describe, define and differentiate between the types of validity • Be able to describe the different methods for developing a criterion for the evaluation of validity and give examples of the different types of criterion validity.