400 likes | 414 Views
NEBRASKA STATE ACCOUNTABILITY. Leadership in Using NeSA Data. Nebraska schools have many sources of data: Classroom assessments Diagnostic district assessment Norm-referenced and now . . . Nebraska State Accountability
E N D
NEBRASKA STATE ACCOUNTABILITY Leadership in Using NeSA Data
Nebraska schools have many sources of data: • Classroom assessments • Diagnostic district assessment • Norm-referenced • and now . . . • Nebraska State Accountability • Each assessment tool has a purpose and a role in the big picture of Continuous Improvement.
Nebraska schools should use NeSA data to . . . • Provide feedback to students, parents and the community • Inform instructional decisions. • Inform curriculum development and revision. • Measure program success and effectiveness. • Promote accountability to meet state and federal requirements.
Understanding NeSA Data • What is NeSA? • How do we access and interpret NeSA data? • How do we use NeSA data?
NeSA is . . . • A criterion-referenced summative test. • A measurement of the revised Nebraska Reading Standards specific to vocabulary and comprehension. • A tool including 45 to 50 multiple-choice items. • A test administered to students online OR paper/pencil.
NeSA is . . . • Administered during the spring of the year. • Based on Tables of Specification and Performance Level Descriptors. • Built upon the best thinking of Nebraska educators, national experts, and a worthy partner – Data Recognition Corporation.
What are . . . Tables of Specification
What are . . . Performance Level Descriptors
Where do we find the content components of NeSA? • www.education.ne.gov/Assessment • Tables of Specification • Performance Level Descriptors • Accommodations Guides • Webb’s DOK documents
NeSA . . . • Produces a raw score that converts to a scale score of 0-200. • Allows for students to be classified into one of three categories: Below the Standards, Meets the Standards, Exceeds the Standards. • Provides comparability across Nebraska school buildings and districts.
What is the DRS? (Data Reporting System) • Secure Site – through portal • Public Site – NDE website http://drs.education.ne.gov
Interpreting NeSA-R Data Reports • District/building level information • Individual student level information • Subgroup information • Indicator information
Use the Reports Interpretive Guide! http://www.education.ne.gov/Assessment/documents/NESA.Read.InterpretiveGuide.pdf
PERFORMANCE LEVELS - three possible categories of student performance on NeSA ~NeSA Terminology~
How are performance levels determined? • Cut score processes: • Contrasting Group Method – 400+ teachers • Bookmark Method – 100+ teachers • State Board of Education Reviewed • Examined results of both processes • Examined NAEP and ACT results for Nebraska • Made decisions within recommended range at public meeting
RAW SCORE – the number of items a student answers ‘right’ on NeSA-R on NeSA Reports on Conversion Chart ~NeSA Terminology~
SCALE SCORE – a student’s transformed version of the raw score earned on NeSA ~NeSA Terminology~
What is the difference between a raw score and a scale score? What is a raw score? A raw score is the number of correct items. Raw scores have been typically used in classrooms as percentages: 18/20= 90% correct. ~NeSA Terminology~
What is a scale score? A scale score is a “transformation” of the number of items answered correctly to a score that can be more easily interpreted between tests and over time. The scale score maintains the rank order of students (i.e., a student who answers more items correctly gets a higher scale score). For NeSA, we selected 0-200 and will use it for all NeSA tests, including writing. ~NeSA Terminology~
Why convert raw scores to scale scores? Raw scores are converted to scale scores in order to compare scores from year to year. Raw scores should not be compared over time because items vary in difficulty level. Additionally, raw scores should not be compared across different content area tests. Scale scores add stability to data collected over time that raw scores do not provide. ~NeSA Terminology~
On score reports why is the . . . SCALE SCORE CONVERTED TO PERCENTILE RANK? The percentile rank was placed on the score reports because our Technical Advisory Committee felt that parents would want to know their child’s position in relation to other test takers. A percentile rank of 84 means the child scored better than 84% of the students who took the test that year. ~NeSA Terminology~
NeSA (CRT) vs. NRT ? • --Differences-- • Purposes: • NeSA is intended to match and measure identified standards and instruction. • NRT is not intended to measure any state’s standards. The intention is to compare students to each other. • Item Development: • NeSA items with exact match to the standards – NDE had to prove the match with an independent alignment study • NRT – No standards to match – matches inherent and previous knowledge, enriched homes, pre-skills.
NeSA (CRT) vs. NRT ? • --Similarities— • All the psychometric steps – standard setting (Bookmark, Angoff, Contrasting Group) • Reliabilities – KR 20-21 / Inter-rater Reliabilities • Descriptive Statistics (Item P-values, Dif-analysis) • Administration: • Both standardized – are generally administered the same way.
NeSA REPORTS • Individual Student Report • School Student Roster • School Indicator Summary • School Performance Level Summary • District Reading Indicator Summary • District Performance Level Summary • District Report of School Performance
Questions for Consideration • What can we learn from this report? • Do we have other data to support these results? • What are the implications of this report?
Use NeSA Data to inform . . . • Curriculum alignment process: • Are the tested indicators in our curriculum? -- Where? • When are they taught? • How are they instructed? • At what DOK (Depth of Knowledge) level? • By whom?
Use NeSA Data to inform . . . Test preparation processes. • Examine PLDs and Tables of Specification. • Do our students have opportunity to learn the tested indicators? • How are they performing on the indicators on a day-to-day basis? • Are we assessing them locally?
Use NeSA Data to inform . . . • Practice Tests • Have our students used practice tests? • Are our students familiar with the testing tools? • Are we familiar with appropriate accommodations?
NeSA results ARE an important data source! When combined with other information, these data can support curricular, instructional, and learning support decision making. --It’s all about the Continuous Improvement Process!