1 / 51

Data Analysis for Assuring the Quality of your COSF Data

Learn to interpret, analyze, and improve data quality for Compliance, Outcomes, and Service Effectiveness (COSE) reporting. Explore OSEP categories, outcomes, and measures, understanding how COSF ratings correspond to reporting categories. Utilize the ECO Calculator for federal reporting. Discover methods for using data to enhance program quality and make informed decisions through evidence, inference, and action. Engage stakeholders, identify variations, and implement effective program improvements.

btheresa
Download Presentation

Data Analysis for Assuring the Quality of your COSF Data

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Data Analysis for Assuring the Quality of your COSF Data

  2. What are these numbers??

  3. OSEP reporting requirements: the outcomes Percentage of children who demonstrated improved: • Positive social emotional skills (including positive social relationships) • Acquisition and use of knowledge and skills (including early language/ communication [and early literacy]) • Use of appropriate behaviors to meet their needs

  4. OSEP reporting categories Percentage of children who: • Did not improve functioning • Improved functioning, but not sufficient to move nearer to functioning comparable to same-aged peers • Improved functioning to a level nearer to same-aged peers but did not reach it • Improved functioning to reach a level comparable to same-aged peers e. Maintained functioning at a level comparable to same-aged peers 3 outcomes x 5 “measures” = 15 numbers

  5. Getting to progress categories from the COSF ratings

  6. Functioning

  7. Entry

  8. Entry Exit

  9. Entry Exit

  10. Key Point • The OSEP categories describe types of progress children can make between entry and exit • Two COSF ratings (entry and exit) are needed to calculate what OSEP category describes a child progress

  11. e. % of children who maintain functioning at a level comparable to same-aged peers Rated 6 or 7 at entry; AND Rated 6 or 7 at exit How changes in ratings on the COSF correspond to reporting categories a - e

  12. Entry Exit

  13. Entry Exit

  14. Entry Exit

  15. d. % of children who improve functioning to reach a level comparable to same-aged peers Rated 5 or lower at entry; AND Rated 6 or 7 at exit How changes in ratings on the COSF correspond to reporting categories a - e

  16. Entry Exit

  17. c. % of children who improved functioning to a level nearer to same aged peers, but did not reach it Rated higher at exit than entry; AND Rated 5 or below at exit How changes in ratings on the COSF correspond to reporting categories a - e

  18. Entry Exit

  19. Entry Exit

  20. b. % of children who improved functioning, but not sufficient to move nearer to same aged peers Rated 5 or lower at entry; AND Rated the same or lower at exit; AND “Yes” on the progress question (b) How changes in ratings on the COSF correspond to reporting categories a - e

  21. Entry Exit

  22. Entry Exit

  23. Entry Exit

  24. Entry Exit

  25. a. % of children who did not improve functioning Rated lower at exit than entry; OR Rated 1 at both entry and exit; AND Scored “No” on the progress question (b) How changes in ratings on the COSF correspond to reporting categories a - e

  26. Entry Exit

  27. Entry Exit

  28. The ECO Calculator can be used to translate COSF entry and exit ratings to the 5 progress categories for federal reporting

  29. Promoting quality data through data analysis

  30. Promoting quality data through data analysis • Examine the data for inconsistencies • If/when you find something strange, what might help explain it? • Is the variation because of a program data? Or because of bad data? (at this point in the implementation process, data quality issues are likely!)

  31. The validity of your data is questionable if… The overall pattern in the data looks ‘strange’ • Compared to what you expect • Compared to other data • Compared to similar states/regions/agencies

  32. COSF Ratings – Outcome 1 Entry data (fake data)

  33. COSF Ratings – Outcome 1 Entry data (fake data)

  34. Frequency on Outcome 1 – Statewide (fake data)

  35. COSF Ratings – Outcome 1 Entry data (fake data)

  36. COSF Ratings – Outcome 1 Entry data (fake data)

  37. Questions to ask when looking at data • Do the data make sense? • Am I surprised? • Do I believe the data? Some of it? All of it? • If the data are reasonable (or when they become reasonable), what might they tell us? • When we believe the data, how can we use it for program improvement?

  38. Using data for program improvement

  39. Continuous Program Improvement Reflect Are we where we want to be? Check (Collect and analyze data) Plan (vision) Program characteristics Child and family outcomes Implement

  40. Using data for program improvement = EIA • Evidence • Inference • Action

  41. Evidence Evidence refers to the numbers, such as “45% of children in category b” The numbers are not debatable

  42. Inference How do you interpret the #s? What can you conclude from the #s? Does evidence mean good news? Bad news? News we can’t interpret? To reach an inference, sometimes we analyze data in other ways (ask for more evidence)

  43. Inference Inference is debatable -- even reasonable people can reach different conclusions Stakeholders can help with putting meaning on the numbers Early on, the inference may be more a question of the quality of the data

  44. Explaining variation • Who has good outcomes = • Do outcomes vary by • Region of the state? • Amount of services received? • Type of services received? • Age at entry to service? • Level of functioning at entry? • Family outcomes? • Education level of parent?

  45. Action Given the inference from the numbers, what should be done? Recommendations or action steps Action can be debatable – and often is Another role for stakeholders Again, early on the action might have to do with improving the quality of the data

  46. Working Assumptions • There are some high quality services and programs being provided across the state • There are some children who are not getting the highest quality services • If we can find ways to improve those services/programs, these children will experience better outcomes

  47. Questions to ask of your data • Are ALL services high quality? • Are ALL children and families receiving ALL the services they should in a timely manner? • Are ALL families being supported in being involved in their child’s program? • What are the barriers to high quality services?

  48. Program improvement: Where and how • At the state level – TA, policy • At the agency level – supervision, guidance • Child level -- modify intervention

  49. Key points • Evidence refers to the numbers and the numbers by themselves are meaningless • Inference is attached by those who read (interpret) the numbers • You have the opportunity and obligation to attach meaning

  50. Tweaking the System Is there a problem? Reflect Are we where we want to be? Why is it happening? Is it working? What should be done? Check (Collect and analyze data) Plan (vision) Program characteristics Child and family outcomes Implement Is it being done?

More Related