320 likes | 565 Views
Dr. Rose Clesham. Put into context: we have looked at education systems that use high stakes external assessments for 16 and 18 year olds, in terms of:. Contextual national characteristics and factors underpinning these educational systems
E N D
Put into context:we have looked at education systems that use high stakes external assessments for 16 and 18 year olds, in terms of: • Contextual national characteristics and factors underpinning these educational systems • How their curriculum and content standards are defined and operate • How their summative national assessments are designed and their focus
85% 81% 83% 77% World economy
The Assessment Analysis • To provide comparisons between UK based GCSEs and GCEs, PISA tests, and selected high performing PISA jurisdictions in terms of: • Their assessment structures and demands in summative assessments • The use of question types • Subject specific areas of focus (eg. use of context in maths, source based focus, particular skills) • Providing examples of good assessment practice • In order to increase our global understanding, and inform the design of WCQ qualifications and assessments.
Cross Business Research Approach • 100+ people involved • Internal/external subject team workshops • Workshop subject groups agreed on and standardised subject definitions of mapping categories • 2-4 raters at item level in each subject • Results aggregated rather than a true score judgement
Content Standards • Based on Uniform Content Standards (Porter, DfE) that can be applied internationally, eg. for Physics
Cognitive Operations • Based on the focus and type of question. These are similar to, but give more detail than Assessment Objectives – and can be used to compare national and international assessments (using a variation of Porter)
Cognitive Demand • Based on considerations of how content standards, assessment questions and mark schemes are written and applied – mainly to do with complexity and linkage (using a new model that incorporates Bloom, Webb, Pollitt, Biggs and Collis)
Outcomes For example, Content Standards in Mathematics at 16
Outcomes For example, Content Standards in Mathematics at 16
Outcomes For example, Content Standards in Mathematics at 16
Outcomes For example, Content Standards in Mathematics at 16
Outcomes For example, Content Standards in Mathematics
Outcomes For example, Content Standards in Mathematics at 16
Proportion of Lower to Higher Order Cognitive Operations in Science 70% 30%
High order science cognitive operations: % differences from the mean, by country
Cognitive demand in Science 52% 44% 4%
% difference from mean of low demand items in science, by country
% difference from mean of medium demand items in science, by country
% difference from mean of high demand items in science, by country
Dendrogram showing family relationships between Maths assessments Porter’s Alignment Indices in Maths
Why does all of this matter?Economy-wide measures of routine and non-routine task input (US) Mean task input as percentiles of the 1960 task distribution The dilemma of schools: The skills that are easiest to teach and test are also the ones that are easiest to digitise, automate and outsource (Levy and Murnane)
Education and assessments needs to prepare students: • To deal with more rapid change than ever before • For jobs that have not yet been created • Using techniques that have not yet been invented • To solve problems that we don’t yet know will arise
Overall conclusions of our research: • Appropriate content representation is a key assessment issue. • In general, lower order cognitive operations dominate assessments. • Higher order cognitive operations are not well represented in assessments, for example problem solving skills in maths and applying concepts and making connections in science. • Cognitive operations and cognitive demand often work independently of each other and therefore need to be actively designed into assessments. • A range of different question types was evident. The analysis showed that although more closed question types can assess a range of cognitive operations and demands, they are less effective than more open responses. • Two fundamental points … • No country’s assessment system was perfect either in design or in outcome- however, some are clearly more designed and comprehensive than others • Good assessment doesn’t happen by chance, it has to be carefully designed and developed to assess the knowledge and skills we want and value.
Implications This work has enabled us to: • Evaluate assessments using a set of common criteria • Make empirical and standardisedjudgements • Contrast and compare • As our qualifications now move into developmental phases, we • need to ensure that they: • Reflect intended curriculum aims and content standards • Assess the knowledge and skills that will be needed for our learners • Are informed by international best practice • And comparable with best practice • And will be more reliable, valid and discriminating in terms of design, focus and . . content
Thank you rose.clesham@pearson.com