1 / 16

H&AC Promising Strategies: Evaluation & Technical Assistance

H&AC Promising Strategies: Evaluation & Technical Assistance. Nikole Lobb Dougherty, MA Jessi LaRose , MPH October 2011. Presentation Overview. Overview of H&AC Evaluation Team and Goals The H&AC Initiative Evaluation Evaluation Technical Assistance. The H&AC Evaluation Team.

jamil
Download Presentation

H&AC Promising Strategies: Evaluation & Technical Assistance

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. H&AC Promising Strategies: Evaluation & Technical Assistance Nikole Lobb Dougherty, MA Jessi LaRose, MPH October 2011

  2. Presentation Overview • Overview of H&AC Evaluation Team and Goals • The H&AC Initiative Evaluation • Evaluation Technical Assistance

  3. The H&AC Evaluation Team Saint Louis University School of Public Health Washington University George Warren Brown School of Social Work • Doug Luke, PhD • Nikole Lobb Dougherty, MA • Chris Robichaux, MS • Tanya Montgomery, MPH • Cheryl Kelly, PhD • Darcy Scharff, PhD • Jessi LaRose, MPH • Trina Ragain, MPH

  4. Evaluation Team Goals • Conduct a modest multi-site evaluation of the H&AC Initiative • Develop evaluation plan and logic model • Implement evaluation plan • Increase grantees’ evaluation capacity through intensive technical assistance and training • Work with grantees to develop individual evaluation plans • Work with grantees to implement evaluation plans

  5. The H&AC Initiative Evaluation

  6. Definition of Internal Program Evaluation and the H&AC Evaluation Internal Program Evaluation: Evaluation conducted by each individual grantee for their own program H&AC Evaluation: Evaluation of the entire H&AC initiative (i.e., across all grantees) conducted by the evaluation team at Washington University and Saint Louis University

  7. H&ACEvaluation Focus • Capacity of programs • Reach of programs • Partnerships • Policy Change • Built Environment Opportunities • Promising Practices • Sustainability • Behavior Change

  8. How will we collect data? • Healthy & Active Programs and Policies Evaluation (HAPPE) System • Quantitative data collection • Monthly and quarterly data submission throughout all three years of funding • Focuses on your activities (what, when, where and who), capacity, & partnerships • One-on-one interviews with program staff • Qualitative data collection • Conducted in the 1st and 3rd years of the project • Focuses on lesson’s learned, capacity, collaborations • Sustainability Assessment Survey • Online survey • Focuses on structures and processes related to likelihood of sustaining project elements

  9. The HAPPE System

  10. Benefits of the Evaluation H&AC Evaluation: • Have a greater understanding of the impact of the H&AC Initiative on Missouri • Identification and dissemination of promising practices to guide future obesity prevention efforts Individual Program: • Easy and organized access to your own data • Ability to generate your own reports of your data

  11. Preview of H&AC Evaluation Activities November 2011 – January 2012 • Training on the HAPPE System/Logic Model Development Spring 2012 • Qualitative interviews with project representatives

  12. Evaluation Technical Assistance

  13. Technical Assistance Goals • To increase internal evaluation capacity • To address evaluation needs of grantees • To provide evaluation resources and networking opportunities to grantees

  14. What We Can Help With • Evaluation plans (e.g., logic models, evaluation questions) • Quantitative and qualitative instruments • Data management and analysis • Database development • Reports and presentations • Evaluation resources • Networking opportunities with other evaluators • Interim reports/contract reporting tool

  15. Preview of TA activities • Evaluation capacity online needs assessment • November 2011 (a follow-up needs assessment will be conducted in the fall of 2014) • Site visits • November 2011-January 2012 • Individualized technical assistance plans and program logic model • Based on needs assessment and first site visit • Regular communication to assist with needs • Conference calls, visits, emails

  16. Contact Information Technical Assistance Contacts Jessi LaRose ericksjl@slu.edu 314.977.8280 Trina Ragain tragain@slu.edu 314.977.8189 Initiative Evaluation Contacts Nikole Lobb Dougherty Evaluation Coordinator nlobbdougherty@.wustl.edu 314.935.3741 Chris Robichaux HAPPE Coordinator crobichaux@gwbmail.wustl.edu 314. 935.3648

More Related