140 likes | 268 Views
Washington 21st Century Community Learning Centers Program Evaluation. February 2011. Introduction. Provide a summary of our proposed approach to undertaking the evaluation Objectives Methods Instruments Timeline. The Evaluation Team. American Institutes for Research
E N D
Washington 21st Century Community Learning Centers Program Evaluation February 2011
Introduction • Provide a summary of our proposed approach to undertaking the evaluation • Objectives • Methods • Instruments • Timeline
The Evaluation Team • American Institutes for Research • Recent merger with Learning Point Associates • Demonstrated 21st CCLC and afterschool content knowledge • Statewide 21st CCLC and afterschool evaluation and research studies in New Jersey, South Carolina, Texas, and Wisconsin • Responsible for the development and maintenance of the Profile and Performance Information Collection System (PPICS) • Support the U.S. Department of Education in monitoring state delivery of 21st CCLC • Provider of afterschool training and technical assistance based on our Beyond the Bell toolkit and currently serve as the statewide training and technical assistance provider for 21st CCLC in Illinois • Robust methodological, statistical, and psychometric skills
The Evaluation Team • David P. Weikart Center for Youth Program Quality • Developers of the Youth and School Age PQAs • Working to build program quality systems in 20 states, including 12 statewide 21st CCLC implementations, including Washington • Rigorously tested intervention strategy for improving the quality of youth-serving programs
Timely and Highly Relevant Topics • The current afterschool literature indicates: • An uneven level of effectiveness in supporting the achievement of positive academic outcomes. • Various paths to supporting student achievement are possible. • A need to define quality criteria and to intentionally support quality improvement efforts.
Evaluation Objective One • Provide an Assessment of the Current State of 21st CCLC Program Impact • Conduct an analysis of 2009-10 regular attendee and other program data obtained from PPICS (February to April 2011) • Synthesize results from the local evaluation reports submitted for the 2009-10 programming period (February to April 2011) • Meet with program directors and local evaluators to discuss findings and ways to further align local and statewide evaluation efforts (May 2011)
Evaluation Objective Two • Support the PPICS Reporting Process and the Collection of Student Level Data • Conduct five webinars on entering data into PPICS and support the collection of PPICS related data (April 2011–June 2011) • Modify the Regular Attendee module in PPICS to collect additional information about students to support data merges with the state assessment data warehouse (July 2011–October 2011)
Evaluation Objective Three • Document the Extent to Which 21st CCLC Programs Are Meeting Local, State, and Federal Targets and Goals • Collect data and conduct analyses to assess the impact of program participation on state assessment in reading and mathematics as compared to a nonparticipant group (July 2011–October 2011) • Create a new data collection module in PPICS to collect information about the nature of local evaluation effortsand how evaluation data is being used to support program improvement efforts
Evaluation Objective Four • Identify Characteristics Associated With High-Performing Programs • Collect additional information about center and staff practices from site coordinator and staff surveys (February 2011-May 2011) • Conduct analyses to assess the relationship between both (a) the level of participation and (b) program and student characteristics and student outcomes (July 2011–October 2011) • Replicate analyses incorporating Youth PQA data with a subset of centers
Analysis of Program Impact • Evidence that students participating more frequently demonstrated better performance. • Evidence of a relationship between center and studentcharacteristics and the likelihood that students demonstrated better performance. • Evidence that students participating in 21st CCLC demonstrated better performance as compared to similar students not participating in the program.
Evaluation Objective Five • Increase the Capacity of Grantees to Meet Their Program Improvement and Evaluation Obligations • Design and prepare leading indicator reports (February 2011–June 2011) • Conduct two regional trainings on how technical assistance providers/coaches can use the leading indicator reports in their site-level work (September 2011) • Conduct six webinars on improvement strategies and techniques that are aligned with improvement priorities identified in the leading indicator reports (September 2011–October 2011)
How Grantees Will Be Impacted • Rolling PPICS deadlines • Operations, Staffing, Feeder Schools, Partners • Activities, Attendance • Objectives, Regular Attendee Data • Reporting of all student data in the Regular Attendee module of PPICS • Completion of new local evaluation module in PPICS
How Grantees Will Be Impacted • Participation in site coordinator and staff surveys • Two stage process: (1) provide information about staff and (2) staff completion of surveys • Participation in events to shape and review leading indicators • Participation in training events on how to utilize leading indicators to inform program improvement efforts
Neil Naftzger • E-Mail: NNaftzger@air.org • 1120 East Diehl Road, Suite 200 • Naperville, IL 60563-1486 • General Information: 800-356-2735 • Website: www.learningpt.org