450 likes | 589 Views
Borrowing the Strength of Unidimensional Scaling to Produce Multidimensional Educational Effectiveness Profiles. Presentation at the 12 th Annual Maryland Assessment Conference College Park, MD October 18, 2012 Joseph A. Martineau Ji Zeng Michigan Department of Education. Background.
E N D
Borrowing the Strength of Unidimensional Scaling to Produce Multidimensional Educational Effectiveness Profiles Presentation at the 12th Annual Maryland Assessment Conference College Park, MD October 18, 2012 Joseph A. Martineau Ji Zeng Michigan Department of Education
Background • Prior research showing that using unidimensional measures of multidimensional achievement constructs can distort value-added • Martineau, J. A. (2006). Distorting Value Added: The Use of Longitudinal, Vertically Scaled Student Achievement Data for Value-Added Accountability. Journal of Educational and Behavioral Statistics, 31(1), 35-62. • Construct irrelevant variance can become considerable in value-added measures when a construct is multidimensional, but is modeled in value-added as unidimensional. • Common misunderstanding is that if the multiple constructs are highly correlated, value-added should not be distorted. • Correct understanding is that if value-added on the multiple constructs is highly correlated, value-added should not be distorted
Background • Prior research showing that the choice of dimension/domain within construct changes value-added significantly • Lockwood, J.R et al. (2007). The Sensitivity of Value-Added Teacher Effect Estimates to Different Mathematics Achievement Measures. Journal of Educational Measurement, 44(1), 47-67. • Depending on choices made in value-added modeling, the correlation between teacher value-added on Procedures and Problem Solving ranged from 0.01 to 0.46. • This gives a surprisingly low correlation in value-added that indicates that at least in this situation, one needs to be concerned about modeling value-added in both dimensions rather than unidimensionally. • Only work I am aware of to date that has inspected inter-construct value-added correlations.
Background • Prior research showing that commonly used factor analytic techniques underestimate the number of dimensions in a multidimensional construct • Zeng, J. (2010) . Development of a Hybrid Method for Dimensionality Identification Incorporating an Angle-Based Approach. Unpublished doctoral dissertation, University of Michigan. • Common dimensionality identifications procedures make the unwarranted assumption that all shared variance among indicator variables arise because the indicator variables measure the same construct (shared variance can also arise because the indicator variables are influenced by a common exogenous variable) • Because of this unwarranted assumption, commonly used dimensionality identification techniques underestimate the number of dimensions in a data set.
Background • Scaling constructs as multidimensional is a difficult task • Multidimensional Item Response Theory (MIRT) is time-consuming and costly to run • Replicating MIRT analyses can be challenging (there are multiple subjective decision points along the way) • Identifying the number of dimensions in MIRT can be challenging • Once the number of dimensions is identified, identifying which items load in which dimensions in MIRT can also be challenging • The factor analysis techniques underlying MIRT are techniques for data reduction, not dimension identification
Background • Short of resolving the considerable difficulties in analytically identifying dimensions within a construct (and replicating such analyses), can another approach be used? • Propose using/trusting content experts’ identifications of dimensions within constructs (e.g., the divisions agreed upon by the writers of content standards) as the best currently available identification of dimensions, for example… • Within English language proficiency, producing reading, writing, listening, and speaking scales. • Within Mathematics, producing number & operations, algebra, geometry, measurement, and data analysis/statistics scales.
Background • However, separately scaling each dimension can also be difficult and costly compared to running a traditional unidimensional IRT calibration • Confirmatory MIRT • Bi-factor IRT model • Separate unidimensional calibration and year-to-year equating of each dimension score • Another option: • Unidimensionally calibrate the total score • Unidimensionally equate the total score from year to year • Use (fixed) item parameters from the unidimensional calibration to create the multiple dimension scores as specified by content experts • Use of this method needs to be investigated • Practical necessity for Smarter Balanced Assessment Consortium
Purpose • Investigate the feasibility and validity of relying on unidimensional total score calibration as a basis for creating multidimensional profile scores… • For reporting multidimensional student achievement scores • For reporting multidimensional value-added measures • Investigate the impact of separate versus fixed calibration of multidimensional achievement scores in terms of impact on… • Student achievement scores • Value-added scores • …as compared to the impact of other common decisions in scaling, outcome selection, and value-added modeling
Methods • Decisions Modeled in the Analyses • Psychometric decisions • Choice of psychometric model • 1-PL vs. 3-PL • PCM vs. GPCM • Estimation of sub-scores • Separate calibration for each dimension vs. fixed calibration based on unidimensional parameters • Choice of outcome metric • Which sub-score is modeled • Value-added modeling decisions • Inclusion of demographics in models • Number of pre-test covariates (for covariate adjustment models)
Methods • Outcomes • Correlations in student achievement metrics compared across each psychometric choice and outcome choice • Correlations in value-added modeling compared across each choice • Classification consistency in value-added compared across each choice for • Three-category classification decisions • Based on confidence intervals around point-estimates placing programs/schools into three categories: (1) above average, (2) statistically indistinguishable from the average, and (3) below average • Four-category classification decisions • Based on sorting programs’/schools’ point estimates into quartiles, representing arbitrary cut points for classification
Methods • Data • Michigan English Language Proficiency Assessment (ELPA) • Level III (Grades 3-5) • 3391 students each with 3 measurement occasions (10,173 total scores) • Measures • Total • Reading (domain) • Writing (domain) • Listening (domain) • Speaking (domain) • Calibrated the ELPA as a unidimensional measure using both 1-PL/Partial Credit Model and 3-PL/Generalized Partial Credit Model • Created domain scores both from fixed parameters from unidimensional calibration and in separate calibrations for each domain
Methods • Data • Michigan Educational Assessment Program (MEAP) Mathematics • Grades 7 and 8 (not on a vertical scale) • Over 110,000 students per grade • Measures • Total (using items from the two domains) • Number & Operations (domain) • Algebra (domain) • Calibrated the MEAP Math tests as unidimensional measures using both 1-PL and 3-PL models • Created domain scores both from fixed parameters from unidimensional calibration and in separate calibrations for each domain
Methods • Value-added modeling the ELPA • 3-level HLM nesting test occasion within student within English language learner program to obtain program value-added
Methods • Value-added modeling the ELPA • VAMs were run in a fully-crossed design with… • All outcomes (R, W, L, S) • PCM- and GPCM-calibrated outcomes • Fixed and separately calibrated outcomes • With and without demographics in the VAMs • 32 real-data applications across design factors
Methods • Value-added modeling MEAP mathematics • 2-level HLM covarying grade-8 outcomes on grade-7 outcomes with students nested within schools
Methods • Value-added modeling MEAP mathematics • VAMs were run in a fully-crossed design with… • Both outcomes (algebra and number & operations) • 1-PL and 3-PL calibrated outcomes • Fixed and separately calibrated outcomes • With and without demographics • With either one or two pre-test covariates • 32 real-data applications across design factors
Results ELPA
Results: ELPA Student-Level Outcomes • Correlations across fixed vs. separate calibrations
Results: ELPA Student-Level Outcomes • Correlations across model choice (PCM vs. GPCM)
Results: ELPA Student-Level Outcomes • Correlations across content areas Low to moderate inter-dimension correlations However, Rasch dimensionality analysis from WINSTEPS identified the total score as a unidimensional score
Results: ELPA Program District-Level Value-Added Outcomes • Impact of fixed versus separate calibration
Results: ELPA Program District-Level Value-Added Outcomes • Correlations between Listening and Reading VA • Min = 0.228, Max = 0.397 • Mean = 0.322, SD = 0.037
Results: ELPA Program District-Level Value-Added Outcomes • Correlations between Listening and Writing VA • Min = 0.342, Max = 0.420 • Mean = 0.373, SD = 0.019
Results: ELPA Program District-Level Value-Added Outcomes • Correlations between Listening and Speaking VA • Min = -0.005, Max = 0.108 • Mean = 0.046, SD = 0.035
Results: ELPA Program District-Level Value-Added Outcomes • Correlations between Reading and Writing VA • Min = 0.335, Max = 0.491 • Mean = 0.412, SD = 0.047
Results: ELPA Program District-Level Value-Added Outcomes • Correlations between Reading and Speaking VA • Min = 0.121, Max = 0.205 • Mean = 0.151, SD = 0.026
Results: ELPA Program District-Level Value-Added Outcomes • Correlations between Speaking and Writing VA • Min = 0.150, Max = 0.246 • Mean = 0.199, SD = 0.029
Results: ELPA Program District-Level Value-Added Outcomes • Impact of choice of psychometric model
Results: ELPA Program District-Level Value-Added Outcomes • Impact of Including/Not Including Demographics
Results MEAP Mathematics
Results: MEAP Math Student-Level Outcomes • Correlations among variables based on psychometric decisions
Results: MEAP Math Student-Level Outcomes • Very high correlations based on fixed versus separate calibrations
Results: MEAP Math Student-Level Outcomes • Very high correlations based on fixed versus separate calibrations
Results: MEAP Math Student-Level Outcomes • Not as high correlations based on 1-PL versus 3-PL calibrations
Results: MEAP Math Student-Level Outcomes • Moderate to high correlations across dimensions
Results: MEAP Mathematics School-Level Value-Added Outcomes • Impact of fixed versus separate calibration
Results: MEAP Mathematics School-Level Value-Added Outcomes • Impact of choice of outcome (Algebra vs. Number)
Results: MEAP Mathematics School-Level Value-Added Outcomes • Impact of choice of psychometric model
Results: MEAP Mathematics School-Level Value-Added Outcomes • Impact of Including/Not Including Demographics
Results: MEAP Mathematics School-Level Value-Added Outcomes • Impact of covarying on one vs. two pre-test scores
Conclusions • Practically important impacts on value-added metrics and value-added classifications • Choice of psychometric model • Including/not including demographics • Including/not including multiple pre-test values • Prohibitive impacts on value-added metrics and value-added classifications • Choice of outcome (i.e., domain within construct) • Practically negligible impacts on value-added metrics and value-added classifications • Separate versus fixed calibrations of domains within construct
Conclusions, continued… • Need to pay attention to modeling domains within constructs if constructs can reasonably be considered multidimensional • Of the common psychometric and statistical modeling decisions one can make, the choice of which subscore to use as an outcome is the most influential • Because subscores give different profiles of both student achievement and program/school value-added, each subscore should be modeled to the degree possible • 4-category (i.e., quartile) classifications on value-added are appreciably impacted by every psychometric and statistical modeling choice evaluated here, but 3-category classifications are not • Discourage more than three categories • RTTT requires at least four categories
Conclusions, continued… • 3- vs. 4-category distinction is actually a proxy for • Statistical decision categories (3-categories) • Arbitrary cut point categories (4-categories) • Can leverage unidimensional calibrations of multidimensional achievement scales to create multidimensional profiles of value-added • Except where using four categories of classifications
Limitations • Inductive reasoning • Results are likely to hold in similar circumstances • Still will need to investigate feasibility of using fixed parameters from unidimensional calibration for specific circumstances if those circumstances are high stakes • This is a proof of concept • PCM and GPCM models were run using different software (WINSTEPS vs. PARSCALE)
Contact Information • Joseph A. Martineau, Ph.D. • Executive Director • Bureau of Assessment & Accountability • Michigan Department of Education • martineauj@michigan.gov • Ji Zeng, Ph.D. • Psychometrician • Bureau of Assessment & Accountability • Michigan Department of Education • zengj@michigan.gov