170 likes | 182 Views
This session will describe findings and lessons learned from a research program investigating the use of games as both learning and assessment tools. Key questions around process, such as who are the right people to have on the team and how to balance conflicting views from engagement, learning, and assessment perspectives will be discussed. In addition, research-based examples will be provided of: specification of learning progressions, task design to align with learning progression stages, identification and summarization of evidence from log files, and reporting to inform instructional decision-making.
E N D
Building Engaging Games for Learning and Assessment Kristen DiCerbo, Ph.D. Vice-President, Education Research @KristenDiCerbo July 2017 Image by Ruben Alvarado
The ability to capture data from everyday events should fundamentally change how we think about learning and assessment. 4
New Ways of Thinking Item Paradigm Activity Paradigm Problem Formulation Items pose questions Activities request action Output Items have answers Activities have features Interpretation Items indicate correctness Activities provide attributes Activities provide multi-dimensional information Information Items provide focused information 5
Some Things We’ve Figured Out You need assessment experts, content experts, and game designers 6
Some Things We’ve Figured Out Evidence-Centered Design works… sort of 7
Some Things We’ve Figured Out The process must be iterative 8
Some Things We’ve Figured Out Think About Systems 9
5 Key Elements 10 1 0
Task Model 12
Evidence Model – Evidence Identification and Scoring 13
Evidence Model – Evidence Identification and Scoring Student Time Attempts Mean Space Median Space Min Space 14
Some Things We Haven’t Figured Out ■ Despite promises of “hundreds of data points,” we always end up with relatively few in our measurement models. Why is that? ■ How do we define evidence for validity and reliability? ■ How do we scale development? ▪ Costs ▪ Time ▪ Staff 17