1 / 45

The MAAECF: Assessment Procedures, Supporting Data, Technical Considerations

E

jadon
Download Presentation

The MAAECF: Assessment Procedures, Supporting Data, Technical Considerations

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


    1. E & R Assessments 2007 1 The MAAECF: Assessment Procedures, Supporting Data, & Technical Considerations Stephen N. Elliott Andrew T. Roach Vanderbilt University Georgia State University Presented at the 8th Annual Maryland Conference On Alternate Assessment

    2. E & R Assessments 2007 2 Key Questions to be Addressed 1. What does the Mississippi Alternate Assessment of Extended Curriculum Frameworks (MAAECF) measure? 2. How do you conduct an alternate assessment using the MAAECF Rating Scale for students with significant cognitive disabilities? 3. What evidence is there to support the claim that the MAAECF yields reliable and valid results? 4. What are salient technical issues associated with this approach to alternate assessment? Note to Presenter: Simply read these 4 questions as a means of focusing participants attention and highlighting the practical purposes of the presentation. Every participant should be able to answer these 4 questions when they have completed the session.Note to Presenter: Simply read these 4 questions as a means of focusing participants attention and highlighting the practical purposes of the presentation. Every participant should be able to answer these 4 questions when they have completed the session.

    3. E & R Assessments 2007 3 Technical Issues to Consider Issue #1. Teachers as reliable judges of student performance. Issue #2. Reliability of teachers’ ratings and the concept of independence. Issue #3. Use of standards-based IEPs to drive the development and collection of evidence for a meaningful sample of items. Issue #4. Amount and quality of classroom evidence needed for valid ratings of proficiency.

    4. E & R Assessments 2007 4 Validity Evidence Plan

    5. E & R Assessments 2007 5 MAAECF : A Comprehensive, Evidence-Based Rating Scale that Uses Teachers’ Judgments to Measure Student Achievement

    6. E & R Assessments 2007 6 Assessment of the Mississippi Extended Curriculum Frameworks (MECF) Extended Content Standards - General statements that describe what students should understand and be able to do in reading/ language arts, mathematics, and science Competencies and Objectives - Specific statements of expected knowledge and skills necessary to meet a content standard requirement Note to Presenter: Read or paraphrase the following definitions of key terms. Resources to “show and tell” a little about: WI Content Standards document. Ideally have a couple of copies to pass around so participants can see examples of a content standard, a performance standard, and the 4 level proficiency framework. MECS document. Content standards - general statements that describe what students should understand and be able to do in various subject matter content areas such as mathematics or reading. Competencies and Objectives - specific statements of expected knowledge and skills necessary to meet a content standard requirement at a particular grade level. Proficiency standards - descriptive categories that describe the degree to which performance standards have been attained. In Mississippi, there are four levels of proficiency used to describe how well a student has done on a test designed to measure the state's content and performance standards. Also note that the state developed Mississippi Extended Content Standards (MECS) – The Mississippi Extended Content Standards (MECS) are organized according to a 4-level hierarchical structure with the most general level being a Content Area. This initial version of the MECS covers two content areas: Language Arts and Mathematics. In late 2005, the Extended Content Standards for the area of Science will be developed. Each Content Area is comprised of multiple Content Strands that subsume multiple Competencies, and in turn, each Competency can be further defined by specific Objectives. Figure 1 illustrates the relationship among these four levels of content. The collective MECS includes curriculum content that students with significant cognitive disabilities are expected to be exposed to during the course of their education. Note to Presenter: Read or paraphrase the following definitions of key terms. Resources to “show and tell” a little about: WI Content Standards document. Ideally have a couple of copies to pass around so participants can see examples of a content standard, a performance standard, and the 4 level proficiency framework. MECS document. Content standards - general statements that describe what students should understand and be able to do in various subject matter content areas such as mathematics or reading. Competencies and Objectives - specific statements of expected knowledge and skills necessary to meet a content standard requirement at a particular grade level. Proficiency standards - descriptive categories that describe the degree to which performance standards have been attained. In Mississippi, there are four levels of proficiency used to describe how well a student has done on a test designed to measure the state's content and performance standards. Also note that the state developed Mississippi Extended Content Standards (MECS) – The Mississippi Extended Content Standards (MECS) are organized according to a 4-level hierarchical structure with the most general level being a Content Area. This initial version of the MECS covers two content areas: Language Arts and Mathematics. In late 2005, the Extended Content Standards for the area of Science will be developed. Each Content Area is comprised of multiple Content Strands that subsume multiple Competencies, and in turn, each Competency can be further defined by specific Objectives. Figure 1 illustrates the relationship among these four levels of content. The collective MECS includes curriculum content that students with significant cognitive disabilities are expected to be exposed to during the course of their education.

    7. E & R Assessments 2007 7 MECF Organizational Structure

    8. E & R Assessments 2007 8 LA Strands and Competencies

    9. E & R Assessments 2007 9 MECF Classroom Tasks/Activities to Help Teach and Generate Evidence

    10. E & R Assessments 2007 10 Sample MAAECF Items and Rating Scale Format

    11. E & R Assessments 2007 11 Essential Features of Evidence-Based Judgment Systems like the MAAECF Unbiased items. High-quality evidence about students’ knowledge & skills. Clear and objective scoring criteria. Procedures to ensure the reliability of the scores. Meaningful methods for communicating results of the assessment. Trained judges – all educators using the MAAECF must attend a workshop + pass a qualification test!

    12. E & R Assessments 2007 12 MAAECF Honors “Teachers as Tests” The research-based concept of teachers as tests emphasizes that teachers collect a substantial amount of data about students and when they are provided a structure for collecting evidence and a method for quantifying and reporting this information, it can be used like other highly reliable test results. Key studies on the reliability of teachers’ judgments: Hoge & Coladarci (1989) Demary & Elliott (1998) Hurwitz, Elliott, & Braden (2007)

    13. E & R Assessments 2007 13 What Does the MAAECF measure? The MAAECF (regardless of level) includes more than 100 items representing prerequisite knowledge and skills in core academic areas. Rating scale items are based on the objectives specified in the Mississippi Extended Curriculum Frameworks (MECF). Note that various levels of the MAAECF exist for students in the grade clusters of 3-5, 6-8, and 12. Teacher rates proficiency using a 4-point scale (0 = Non-Existent to 3 = Accomplished). The same item level rubric is used across all grade clusters. Results are reported as performance levels (Minimal, Basic, Proficient, or Advanced) in each content area. Definitions and cut scores for these levels vary by grade. The results are on a continuum that is very similar to the general education test’s performance levels. Presenter: Note that the MAA is a rating scale completed by an individual with knowledge of the student’s IEP goals, objectives, benchmarks, curriculum, knowledge, and skills. This person is usually a teacher. The MAA borrows from the science of behavior rating scales, the use of performance assessment scoring rubrics, and the fact that teachers have collected vast amounts of information about their students and can reliably report the information when provided a meaningful structure. Resources: (1) Hold up a copy of both the Academic Content Standards and the MECS and briefly explain their relationship. Then add a copy of the MAA and physically model how they are aligned or overlap in content (knowledge and skills) covered. (2) Hold up a copy of state proficiency standards.Presenter: Note that the MAA is a rating scale completed by an individual with knowledge of the student’s IEP goals, objectives, benchmarks, curriculum, knowledge, and skills. This person is usually a teacher. The MAA borrows from the science of behavior rating scales, the use of performance assessment scoring rubrics, and the fact that teachers have collected vast amounts of information about their students and can reliably report the information when provided a meaningful structure. Resources: (1) Hold up a copy of both the Academic Content Standards and the MECS and briefly explain their relationship. Then add a copy of the MAA and physically model how they are aligned or overlap in content (knowledge and skills) covered. (2) Hold up a copy of state proficiency standards.

    14. E & R Assessments 2007 14 Item Count Per Scales & Grade Levels

    15. E & R Assessments 2007 15 Mississippi Alternate Assessment and Accountability System

    16. E & R Assessments 2007 16 MAAECF : Administration & Scoring Procedures

    17. E & R Assessments 2007 17 The MAAECF Process Once the IEP Team decides that a student is eligible for an alternate assessment, a 5-step process must be followed: Step 1: Align MAAECF items with IEP goals, objectives, or other learning objectives. Step 2: Collect performance evidence for a sample of items for each strand using Evidence Worksheets. Step 3: Analyze and rate proficiency of all items. Presenter: Note that the student’s teacher identifies MAA items that align with the student’s IEP goals, objectives, or benchmarks. For each IEP aligned MAA item, at least two forms of evidence from the student’s classroom are collected. Elaborate on ratings – reference people to pages in MAA rating form where they can view the item proficiency criteria and the overall Performance Summary continua.Presenter: Note that the student’s teacher identifies MAA items that align with the student’s IEP goals, objectives, or benchmarks. For each IEP aligned MAA item, at least two forms of evidence from the student’s classroom are collected. Elaborate on ratings – reference people to pages in MAA rating form where they can view the item proficiency criteria and the overall Performance Summary continua.

    18. E & R Assessments 2007 18 The MAAECF Process (continued) Step 4: Summarize proficiency scores and performance level decisions. A second educator completes the process by reviewing collected evidence, providing ratings on only items with evidence samples, and checking the accuracy of the performance level decisions. Step 5: Report results after the first rater and the individual who completes the reliability check have reached agreement. Presenter: Note that a second rater, usually another teacher, reviews the evidence collected to ensure that the original teacher’s rating are reliable. A proficiency rating (i.e., non-existent, emerging, developing/developed, and proficient/generalized) is assigned to each MAA item to describe student skill development. Items that are not relevant or possible given a student’s disability are rated as not applicable. An Overall Performance Level (i.e., Prerequisite Skill Level 1-4), for each domain is determined. Resource: Show an overhead of the Proficiency Scoring Criteria and also select one of the Content Area Developmental Continua to highlight the developmental and criterion-reference nature of the various Prerequisite Skills Levels. Connect the 4 PS levels to the Proficiency Levels for the MCT. It is important that you have overheads of pages 2, 3, and 4 from the MAA Rating Scale. It is also important to have overheads of a content area, such as mathematics, that is correctly completed. Finally, have an overhead of the summary score page (last page of MAA Rating Form). Presenter: Note that a second rater, usually another teacher, reviews the evidence collected to ensure that the original teacher’s rating are reliable. A proficiency rating (i.e., non-existent, emerging, developing/developed, and proficient/generalized) is assigned to each MAA item to describe student skill development. Items that are not relevant or possible given a student’s disability are rated as not applicable. An Overall Performance Level (i.e., Prerequisite Skill Level 1-4), for each domain is determined. Resource: Show an overhead of the Proficiency Scoring Criteria and also select one of the Content Area Developmental Continua to highlight the developmental and criterion-reference nature of the various Prerequisite Skills Levels. Connect the 4 PS levels to the Proficiency Levels for the MCT. It is important that you have overheads of pages 2, 3, and 4 from the MAA Rating Scale. It is also important to have overheads of a content area, such as mathematics, that is correctly completed. Finally, have an overhead of the summary score page (last page of MAA Rating Form).

    19. E & R Assessments 2007 19 Team Work for Reliable Results: Summarizing the Roles of Raters 1 and 2 Rater 1

    20. E & R Assessments 2007 20 Step 1: Alignment of Items Determine in which content areas a student will be assessed. Identify MAAECF items that align with the student’s IEP goals and objectives, or other learning goals. If there is not at least 1 IEP-aligned item for each competency (4 in LA, 5 in Math, and 4 in Science), then an item must be selected and evidence generated for that item. Presenter: Note that this alignment process requires both professional judgment and an understanding of the state’s academic standards. The goal is to identify and assess connections between MAA items and the student’s IEP goals, objectives, or benchmarks. The alignment process requires a person to be quite familiar with the student’s IEP and the items on the MAA rating scale. It is possible that an IEP goal, objective, or benchmark may align with more than one item on the rating scale. In addition, it is entirely likely than some IEP objectives will not align with any of the MAA items. Once a set of IEP-aligned items has been identified, it is the responsibility of the rater (usually the child’s main teacher) to collect examples of work samples, progress records, observations, etc. to provide a recent and representative sample of information to substantiate the rater’s evaluation.Presenter: Note that this alignment process requires both professional judgment and an understanding of the state’s academic standards. The goal is to identify and assess connections between MAA items and the student’s IEP goals, objectives, or benchmarks. The alignment process requires a person to be quite familiar with the student’s IEP and the items on the MAA rating scale. It is possible that an IEP goal, objective, or benchmark may align with more than one item on the rating scale. In addition, it is entirely likely than some IEP objectives will not align with any of the MAA items. Once a set of IEP-aligned items has been identified, it is the responsibility of the rater (usually the child’s main teacher) to collect examples of work samples, progress records, observations, etc. to provide a recent and representative sample of information to substantiate the rater’s evaluation.

    21. E & R Assessments 2007 21 Step 2: Collect Evidence of Skills Teachers must collect evidence from 2 different categories for at least 1 item for each competency. Work Samples Tests Observations Interviews Video/Photo Audio Tape Presenter: Note that teachers conducting an alternated assessment can begin doing ratings and collecting evidence to support their ratings any time after the beginning of school up April 30th, the date that the Overall Performance Summary Report must be submitted to the district assessment coordinator. Provide more guidance about evidence – how much is enough, how one “piece” can be used for multiple items, it should be indigenous to the classroom (i.e., already exists) but can be developed in reaction to items. It also should used to communicate with parents/guardians. Stress that evidence must be Recent and Representative. Also note that one piece of evidence can be used to support ratings of several items. For example, in the case of April that we will learn more about, her teacher used a videotape of April doing work in the school store to document skills for several mathematics and social studies items. Presenter: Note that teachers conducting an alternated assessment can begin doing ratings and collecting evidence to support their ratings any time after the beginning of school up April 30th, the date that the Overall Performance Summary Report must be submitted to the district assessment coordinator. Provide more guidance about evidence – how much is enough, how one “piece” can be used for multiple items, it should be indigenous to the classroom (i.e., already exists) but can be developed in reaction to items. It also should used to communicate with parents/guardians. Stress that evidence must be Recent and Representative. Also note that one piece of evidence can be used to support ratings of several items. For example, in the case of April that we will learn more about, her teacher used a videotape of April doing work in the school store to document skills for several mathematics and social studies items.

    22. E & R Assessments 2007 22 Characteristics of Good Evidence Recent (collected during the current school year & dated) Representative (several forms that are typical performances of knowledge and skills with classroom materials and instructional accommodations) Relevant (evidence that is clearly indicative of an item that is aligned with content standards, and that identifies the relevant item number(s) for the second rater and others) Reliable (increases the likelihood that 2 or more raters have the same understanding of the item by documenting the support provided to get the response from the student and indicating the overall accuracy of the student’s typical response)

    23. E & R Assessments 2007 23 Evidence Collection Worksheets

    24. E & R Assessments 2007 24 Quality Evidence Matters: Collected Evidence Should Meet or Exceed the Standard!

    25. E & R Assessments 2007 25 Step 3: Criteria for Rating All Items 0 = NON-EXISTENT (Can’t do currently) 1 = EMERGING (Aware and starting to do) 2 = PROGRESSING (Can do partially and inconsistently) 3 = ACCOMPLISHED (Can do well and consistently) ***Examine descriptions of these Proficiency Rating Levels*** Presenter: Note that this slide lists some common criteria that IEP teams have used to describe students’ progress during annual reviews of their IEPs. These same criteria can be used to communicate about students’ performances on alternate assessments in content domains. When Showing this Slide: Read each of the four criteria listed and ask participants to provide some examples of ways to use each of the criteria. Encourage them to think about a student that they know well and how these criteria could be used to communicate about the students’ skills in reading or math. Note: It is important to discriminate between NA & non-existent. Presenter: Note that this slide lists some common criteria that IEP teams have used to describe students’ progress during annual reviews of their IEPs. These same criteria can be used to communicate about students’ performances on alternate assessments in content domains. When Showing this Slide: Read each of the four criteria listed and ask participants to provide some examples of ways to use each of the criteria. Encourage them to think about a student that they know well and how these criteria could be used to communicate about the students’ skills in reading or math. Note: It is important to discriminate between NA & non-existent.

    26. E & R Assessments 2007 26 MAAECF Item Level Proficiency Rating Scale and Descriptive Criteria

    27. E & R Assessments 2007 27 Step 4: Overall Proficiency Scores and Performance Continua Students earn an Individualized Proficiency Total Score for each content area. The total scores are used to guide the determination of which of the four Overall Performance Levels best describes the student’s achievement. Minimal Basic Proficient Advanced ***Examine detailed Performance Levels*** Presenter: Note that the score continuum is a downward extension of the four performance levels (i.e., minimal performance, basic, proficient, and advanced proficient) that result fro the MCT. Elaborate on what the characteristics are that differentiate a PS1 from a PS2, etc. * Frequency/accuracy * Amount of support * Quality of work * Generalization Reference the scoring continua. Finally, note that persons taking MCT would be expected to be functioning at or above PS4. Resource Material: Use a copy of the 8 level scoring criteria that was used to develop the MAA developmental continua to illustrate the complete continua for a content area that includes the MAA+WKCE.Presenter: Note that the score continuum is a downward extension of the four performance levels (i.e., minimal performance, basic, proficient, and advanced proficient) that result fro the MCT. Elaborate on what the characteristics are that differentiate a PS1 from a PS2, etc. * Frequency/accuracy * Amount of support * Quality of work * Generalization Reference the scoring continua. Finally, note that persons taking MCT would be expected to be functioning at or above PS4. Resource Material: Use a copy of the 8 level scoring criteria that was used to develop the MAA developmental continua to illustrate the complete continua for a content area that includes the MAA+WKCE.

    28. E & R Assessments 2007 28 LA Performance Level Descriptor (with Cut Scores for Grades 3, 4, and 5)

    29. E & R Assessments 2007 29 MAAECF Item Count & Score Ranges by Performance Level & Grade Clusters

    30. E & R Assessments 2007 30 2nd Rater’s Responsibilities Review evidence of student’s knowledge and skills and then independently rate items with which they are aligned. This is done on a separate copy of the MAAECF. Review the 1st rater’s ratings of evidence-based items to determine general degree of agreement and then examine all item ratings to learn more about the student. These ratings provide additional evidence for the 2nd rater. Based on all of the evidence, independently select the overall Performance Level (Minimal, Basic, Proficient, or Advanced) that best characterizes the student’s current functioning. Double check 1st rater’s total score for the students in Language Arts, Math, and Science to ensure all items were rated and the total scores are correct. Compare their overall Performance Level determination to that of the 1st rater and decide if it is reliable using the Inter-Rater Reliability Estimate table of the MAAECF for Language Arts and Math. Settle any disagreements that result in unreliable decisions.

    31. E & R Assessments 2007 31 Inter-Rater Agreement Table

    32. E & R Assessments 2007 32

    33. E & R Assessments 2007 33 Descriptive Statistics & Key Evidence to Support the Inferences Made From MAAECF Scores About Student Achievement

    34. E & R Assessments 2007 34 Descriptive Statistics for MAAECF Math

    35. E & R Assessments 2007 35 Frequency Distributions for Math

    36. E & R Assessments 2007 36

    37. E & R Assessments 2007 37 Inter-rater Agreement & Evidence Quality Study

    38. E & R Assessments 2007 38 Quality of Evidence Samples Matter!

    39. E & R Assessments 2007 39 Factor Analytic Evidence to Support Claims about the MAAECF’s Structure

    40. E & R Assessments 2007 40 Evidence Based on Relations to Other Variables

    41. E & R Assessments 2007 41 More Evidence Based on Relations to Other Variables

    42. E & R Assessments 2007 42 Evidence based on Consequences of Testing

    43. E & R Assessments 2007 43 Item and Scale Revision Plan in Progress for 2008

    44. E & R Assessments 2007 44 MS CAARES Project Enhancements to the MAAECF

    45. E & R Assessments 2007 45 Thank You Contact Information Stephen N. Elliott Vanderbilt University 616-322-2538 steve.elliott@vanderbilt.edu Andrew T. Roach Georgia State University 404-413-8176 aroach@gsu.edu

More Related