350 likes | 357 Views
This session explores the requirements and best practices for alternate assessments on alternate achievement standards (AA-AAS) in the context of the 1% rule and Title I Peer Review. It discusses the current state of AA-AAS and the necessary steps for improvement.
E N D
Measurement Purgatory or Best Practice? Alternate Assessment for Students with Significant Cognitive Disabilities Don Peasley, Ohio Department of Education Tom Deeter, Iowa Department of Education Rachel Quenemoen, NCEO
What is required for alternate assessments on alternate achievement standards (AA-AAS) in the context of the 1% Rule? (and last Saturday’s presession) What is required for AA-AAS in the context of Title I Peer Review? Where are we now, and where do we have to go? Overview
Aligned with the State’s grade level content standards. Yield results separately in reading/language arts and math. Designed and implemented to support use of the results to determine AYP. Alternate Assessments as defined in “1% Rule”
Clearly defined structure Guidelines for which students may participate Clearly defined scoring criteria and procedures Report format that clearly communicates student performance in terms of the academic achievement standards defined by the State Alternate Assessments should have…
Must meet the same requirements for high technical quality that apply to regular assessments under NCLB: Validity Reliability Accessibility Objectivity Consistent with nationally-recognized professional and technical standards. Alternate Assessments
Alternate assessment scored against grade-level achievement standards Alternate assessment scored against alternate achievement standards Both must support access to grade level curriculum States may use more than one alternate assessment
1. Careful stakeholder and policymaker development of desired student outcomes for the population, reflecting the best understanding of research and practice, thoughtfully aligned to same content expected for all students, at grade-level. 2. Careful development, testing, and refinement of assessment methods. 3. Scoring of evidence of grade-level content aligned student work, according to professionally accepted standards, against criteria that reflect best understanding from research and practice. 4. Standard-setting process to allow use of results in reporting and accountability systems. 5. Continuous improvement of the assessment process. Development of Alternate AssessmentsQuenemoen, Rigney, & Thurlow, 2002
Observation Interpretation The assessment triangle (Pellegrino et al., 2001) Cognition
Professional Understanding of Learning Goals • Shifting goals for students with significant cognitive disabilities since 1975 (Browder, 2001; Kearns & Kleinert, 2004) • Developmental Goals – “ready meant never” • 1980s - Functional Goals – NOW WE HAVE REFOCUSED ON: • 1990s - Academic Goals – “general curriculum” leading to developmental traps leading to a focus on GRADE LEVEL Academic Content Standards
WHAT IS LEARNING? We must ensure all students have access to and make progress in the academic grade level content and assess achievement on that contentWhat is achievement? What is proficiency?
Section I—Assessment Development A. Overview Principles guiding development Partners and process guiding development Research base on desired outcomes for this population, clarification of theory of learning – develop draft performance level descriptors Documentation of state conceptualization for (expansion/extension) alignment and access to the state grade level content standards Pros and cons of assessment methods considered Description of selected approach Draft Technical Manual Outline
Charlie DePascale, Jeff Nellhaus, Barbara Plake, Michael Beck session on Monday – nciea.org – basic information on standard-setting Depth of understanding? Differ in substance? Differ in amount? All the content? Some of the content? Any of the content? What does it mean for these students to be proficient in mathematics? In ELA? Are we avoiding developmental traps? TASK: Write draft performance level descriptors for AA-AAS
B. Test Development Protocol for alignment to grade level content standards Development of draft assessment protocol Pilot test design and results Field test design and results C. Test blueprint English Language Arts content specifications Mathematics content specifications Other (e.g., Science) content specifications
Section II—Test Administration A. Procedures for administration Decision-making process (participation, IEP team role) Local responsibility Timelines B. Training Test oversight training for administrators Educator training for those working directly with students Ethical test administration training
The Ohio Alternate Assessment is based on a Collection of Evidence COE model Designed to be a measure of student achievement aligned with Ohio’s Academic Content Standards Alternate assessment is a “snapshot” of achievement during a window of time Ohio’s Alternate Assessment for Students with Disabilities
Collection of Evidence For each academic area Entry 1 (Standard) Cover page Entry 3 (Standard) Entry 2 (Standard)
Does the student have a disability that presents “unique and significant” challenges to participation in district and state assessment regardless of the accommodations they could use? NO YES Does the student have severe motor or sensory or cognitive or emotional disabilities? Participation in regular district and state assessments with or without accommodations Ohio’s Participation Decision Framework: NO YES Continue……
Ohio’s Decision Framework: • Does the student: • Require substantial modifications to the general education curriculum (form and substance)? • AND • Require instruction focused on the application of state standards through essential life skills? • AND • Require instruction multiple levels below age/grade level? • AND • Is the student unlikely to provide valid and reliable measure of proficiency in content areas via standardized assessment even with accommodations? NO YES Participation in regular district and state assessments with or without accommodations Student participates in Alternate Assessment
Section III— Scoring and Reporting A. Scoring design Quality control Benchmarking Selecting and training scorers Scoring activities Inter-scorer reliability
Scoring Collection of Evidence scored across four domains (scoring criteria) Performance—holistic by entry Independence/Support-holistic by entry Context/Complexity—holistic by entry Settings and Interactions—for entire collection Evidence is scored independently according to professionally accepted standards by scoring contractors Ohio’s Alternate Assessment for Students with Disabilities
B. Standard-setting Documented and validated process used for standard setting (Full description in Appendix _) Performance level descriptors and exemplars for alternate achievement standards Distribution of performance across levels Comparison of performance across levels achieved in general assessment by students with disabilities in comparable implementation years C. Reporting design School/District/State Report Parent Letter/Individual Student Report
Section IV - Reliability and Validity; Other Technical Considerations A. Summary of studies for reliability, available data B. Summary of studies for validity, available data Face validity studies Concurrent validity studies Consequential validity studies C. Other technical considerations
Appendix A Documentation of development principles, partners, process, research base Appendix B Documentation of training provided, attendance, quality control Appendix C Documentation of scoring protocols, process, quality control Appendix D Formal evaluation data if available Appendix E Standard setting report Section V—Appendices
Who are the learners who take alternate assessments? How does the type and size of the population vary in terms of learner characteristics, available response repertoires, and complex medical conditions? How do the variations of who the learners are affect the assessment triangle, and ultimately technical adequacy studies? What does the literature say about how students in this (these) population(s) learn? How do current theories of learning in the typical population apply to this population?
How is technical adequacy defined? What is meant by reliability, validity? How do traditional definitions of reliability/validity apply to alternate assessments? What are technical adequacy issues in alternate assessments that can not be resolved with the current knowledge-base in large-scale assessment? What strategies can be used to resolve these issues?
What consequential validity issues (intended and unintended consequences) challenge the foundational assumptions in an alternate assessment? What is the relationship between foundational assumptions of alternate assessments and technical adequacy issues? What lessons learned from alternate assessment need to be addressed for the general assessment as well?
Define the learners, and determine how this differs across states Build consensus on a theory of learning in the academic content domains for these students Step out of our specializations and think together about these challenges Next Steps