2 Introduction Provide a summary of our proposed approach to undertaking the evaluation Objectives Methods Instruments Timeline
3 The Evaluation Team American Institutes for Research Recent merger with Learning Point Associates Demonstrated 21st CCLC and afterschool content knowledge - Statewide 21st CCLC and afterschool evaluation and research studies in New Jersey, South Carolina, Texas, and Wisconsin - Responsible for the development and maintenance of the Profile and Performance Information Collection System (PPICS) - Support the U.S. Department of Education in monitoring state delivery of 21st CCLC - Provider of afterschool training and technical assistance based on our Beyond the Bell toolkit and currently serve as the statewide training and technical assistance provider for 21st CCLC in Illinois Robust methodological, statistical, and psychometric skills
4 The Evaluation Team David P. Weikart Center for Youth Program Quality Developers of the Youth and School Age PQAs Working to build program quality systems in 20 states, including 12 statewide 21st CCLC implementations, including Washington Rigorously tested intervention strategy for improving the quality of youth-serving programs
5 Timely and Highly Relevant Topics The current afterschool literature indicates: An uneven level of effectiveness in supporting the achievement of positive academic outcomes. Various paths to supporting student achievement are possible. A need to define quality criteria and to intentionally support quality improvement efforts.
6 Evaluation Objective One Provide an Assessment of the Current State of 21st CCLC Program Impact Conduct an analysis of 2009-10 regular attendee and other program data obtained from PPICS (February to April 2011) Synthesize results from the local evaluation reports submitted for the 2009-10 programming period (February to April 2011) Meet with program directors and local evaluators to discuss findings and ways to further align local and statewide evaluation efforts (May 2011)
7 Evaluation Objective Two Support the PPICS Reporting Process and the Collection of Student Level Data Conduct five webinars on entering data into PPICS and support the collection of PPICS related data (April 2011–June 2011) Modify the Regular Attendee module in PPICS to collect additional information about students to support data merges with the state assessment data warehouse (July 2011– October 2011)
8 Evaluation Objective Three Document the Extent to Which 21st CCLC Programs Are Meeting Local, State, and Federal Targets and Goals Collect data and conduct analyses to assess the impact of program participation on state assessment in reading and mathematics as compared to a nonparticipant group (July 2011–October 2011) Create a new data collection module in PPICS to collect information about the nature of local evaluation efforts and how evaluation data is being used to support program improvement efforts
9 Evaluation Objective Four Identify Characteristics Associated With High-Performing Programs Collect additional information about center and staff practices from site coordinator and staff surveys (February 2011-May 2011) Conduct analyses to assess the relationship between both (a) the level of participation and (b) program and student characteristics and student outcomes (July 2011–October 2011) Replicate analyses incorporating Youth PQA data with a subset of centers
10 Analysis of Program Impact Evidence that students participating more frequently demonstrated better performance. Evidence of a relationship between center and student characteristics and the likelihood that students demonstrated better performance. Evidence that students participating in 21st CCLC demonstrated better performance as compared to similar students not participating in the program.
11 Evaluation Objective Five Increase the Capacity of Grantees to Meet Their Program Improvement and Evaluation Obligations Design and prepare leading indicator reports (February 2011–June 2011) Conduct two regional trainings on how technical assistance providers/coaches can use the leading indicator reports in their site-level work (September 2011) Conduct six webinars on improvement strategies and techniques that are aligned with improvement priorities identified in the leading indicator reports (September 2011–October 2011)
12 How Grantees Will Be Impacted Rolling PPICS deadlines Operations, Staffing, Feeder Schools, Partners Activities, Attendance Objectives, Regular Attendee Data Reporting of all student data in the Regular Attendee module of PPICS Completion of new local evaluation module in PPICS
13 How Grantees Will Be Impacted Participation in site coordinator and staff surveys Two stage process: (1) provide information about staff and (2) staff completion of surveys Participation in events to shape and review leading indicators Participation in training events on how to utilize leading indicators to inform program improvement efforts
14 Neil Naftzger E-Mail: NNaftzger@air.orgNNaftzger@air.org 1120 East Diehl Road, Suite 200 Naperville, IL 60563-1486 General Information: 800-356-2735 Website: www.learningpt.org