360 likes | 609 Views
Innovations in Evaluation . IPDET Workshop, Ottawa, June 14 2013 Simon Roy & Louise Mailloux. Outline. Definitions Innovations in Canada Innovations on the International Scene Discussion: What’s your experience?. A Definition….
E N D
Innovations in Evaluation IPDET Workshop, Ottawa, June14 2013 Simon Roy &Louise Mailloux
Outline • Definitions • Innovations in Canada • Innovations on the International Scene • Discussion: What’s your experience?
A Definition… • Innovations can be defined as alternative and new ways of conducting evaluations (methods, analyses, governance, etc.) • Many drivers: • Methodological challenge affecting data quality or availability • Opportunities stemming from new technologies • Influence from other disciplines/professions • HR or governance challenges
Contextual factors • Innovations are region-specific: What is innovative in one place may not be in another area • Some innovations may work in one country, but not in another
Three notable innovations in last decade In Canada • Multimode approaches in surveys • Focus on cost analyses • Professionalization of evaluation: certification of evaluators
Multi-Mode Surveys • Surveys traditionally done in single mode: mail or phone or fax • Low response rates now major problem • Evaluators have moved to surveys administered in multiple modes: respondents offered to complete it online, by phone or by mail • Advantages: Higher response rates, less bias in terms of sampling • Disadvantage: There is a bias associated to the mode
Cost Analyses • Many governments moving towards “value for money” analyses, including analysis of input, outputs and outcomes in view of the costs involved • Innovation is in the refinement of the approaches to conduct such analyses
Perspectives on Assessing Resources Utilization and the Results Chain Allocative Efficiency Operational Efficiency Economy Inputs Activities Outputs Immediate Outcomes Intermediate Outcomes Ultimate Outcomes Results Chain The analysis for economy, operational efficiency and allocative efficiency occurs along the results chain. Primary focus of analysis Informs analysis
Credentialing • Canadian Evaluators have an association: The Canadian Evaluation Society (CES) (http://www.evaluationcanada.ca/) • The CES implemented an Evaluation Credentialing Program in 2010. Evaluators can become “Credentialed Evaluators” • This is an association-led initiative. The Governments of Canada have no direct control over this credential. • It is not a requirement to conduct evaluations.
Credentialing • Canadian Evaluators can receive a credential if they meet criteria (demonstration of competency), including 2 years of evaluation experience and competencies in 5 areas (see appendix) • Expected benefits: Evaluators gain recognition. Credentials help evaluation users select evaluation provider. • About 200 credentialed evaluators to date.
Our Overall Lessons to Date • Evaluation is evolving – becoming more and more complex • Before discounting new ways, look at the advantages, especially how they can compensate for limitations of traditional approaches (traditional methods have gaps too!) • Weigh the advantages vs. disadvantages, manage them to reduce the latter. Have a backup plan.
Real Time Evaluations (RTE) Digital Data
A definition of RTE • A real-time evaluation (RTE) is an evaluation in which the primary objective is to provide feedback in a participatory way in real time (i.e. during the evaluation fieldwork) to those executing and managing a humanitarian response. Source: Real-time evaluations of humanitarian action An ALNAP Guide Pilot Version, John Cosgrave Ben Ramalingam and Tony Beck, 2009
Origins of RTEs • In the humanitarian sector, UNHCR’s Evaluation and Policy Analysis Unit (EPAU) was for several years the chief proponent of RTE • WFP, UNICEF, the Humanitarian Accountability Project, CARE, World Vision, Oxfam GB, the IFRC, FAO, WFP and others have all to some degree taken up the practice. Source: ISSUE 32 December 2005 Humanitarian Exchange Magazine Real-Time Evaluation: where does its value lie? by Maurice Herson and John Mitchell, ALNAP
RTE vs other types of evaluations • RTEs look at today to influence this week’s/month’s programming • Mid-term evaluations look at the first phase to influence programming in the second phase • Ex-post evaluations are retrospective: they look at the past to learn from it
Key Features/Methods • Semi-structured interviews • Purposefulsampling – complemented by snowballsampling in the field • Interviews with beneficiary groups important • Observation
Methodological Contraints of RTE • Limited use of statistical sampling (sample frame) • Limited use of surveys • Lack of pre-planned coordination betweenhumanitarianactors • Baseline studiesusuallyinexistent • Attribution (cause and effect) difficultgiven the multiplicity of actors Source: Brusset, E., Cosgrave, J., & MacDonald, W. (2010). Real-time evaluation in humanitarian emergencies. In L. A. Ritchie & W. MacDonald (Eds.), Enhancing disaster and emergency preparedness, response, and recovery through evaluation. New Directions for Evaluation, 126, 9–20.
Lessons - Advantages • Timeliness: RTEs bring in an external perspective, analytical capacity and knowledge at a key point in a response. • Perspective: RTEs reduce the risks that early operational choices bring about critical problems in the longer term. • Interactivity: RTEs enable programming to be influenced as it happens, allowing agencies to make key changes at an intermediate point in programming.
Lessons - Challenges • Utilisation: Weakness in the follow up on recommendations • Ownership: workers, managers, beneficiaries? • Focus: What are the key questions? • Meeting eachpartners’ needs for accountability and learning • Few RTEs in complex emergencies Source:Lessons from recent Inter Agency Real Time Evaluations (IA RTEs) Riccardo Polastro
Rationalebehindit • Explosion in the quantity and diversity of high frequency digital data e.g. mobile-banking transactions, online user-generated content such as blog posts and Tweets, online searches, satellite images, computerized data analysis. • Digital data hold the potential—as yet largely untapped— to allow decision makers to track development progress, improve social protection, and understand where existing policies and programmes require adjustment Source: Global Pulse, Big Data for Development: Challenges & Opportunities May 2012, www.unglobalpulse.org
Big Data – UN Initiative • 1) Early warning: early detection of anomalies in how populations use digital devices and services can enable faster response in times of crisis • 2) Real-time awareness: Big Data can paint a fine-grained and current representation of reality which can inform the design and targeting of programs and policies • 3) Real-time feedback: makes it possible to understand human well-being and emerging vulnerabilities, in order to better protect populations from shocks
Potential Uses and Focus • ILO, UNICEF and WFP, researching changes in social welfare, especially with regard to food and fuel prices, and employment issues The number of tweets discussing the price of rice in Indonesia in 2011 follows a similar function as the official inflation statistics for the food basket.
WhatisBig Data? "Big Data" is a popular phrase used to describe a massive volume of both structured and unstructured data that is so large that it's difficult to process with traditional database and software techniques. Types of digital data sources Data Exhaust Online Information Physical Sensors Citizen Reporting or Crowd-sourced Data
LessonsLearned to Date Privacy • Privacy is an overarching concern that has a wide range of implications vis-à-vis data acquisition, storage, retention, use and presentation • People routinely consent to the collection and use of web-generated data by simply ticking a box without fully realising how their data might be used or misused. • Do bloggers consent to have their content analyzed by publihing on the web?
LessonsLearnedto Date Access and Sharing • Much of the publicly available online data (data from the “open web”) has potential value for development, there is a great deal more valuable data that is closely held by corporations and is not accessible • “The next movement in charitable giving and corporate citizenship may be for corporations and governments to donate data, which could be used to help track diseases, avert economic crises, relieve traffic congestion, and aid development.” Source: Data Philanntropy where are we now. Andreas Pawelke and Anoush Rima TatevossianMay8, 2013
LessonsLearnedto date Analysis • “conceptualisation” (i.e. defining categories, clusters); • selection bias (representative of general population?) • “measurement” (i.e. assigning categories and clusters to unstructured data, or vice-versa) • “verification” (i.e. assess how well steps 1 and 2 fare in extracting relevant information)
Discussion:What’s happening in your organization/ country in terms of innovation in evaluation?What lessons can you share about what works and what does not work?
Thank You!Louise Mailloux – lmailloux@ggi.caSimon Roy – sroy@ggi.ca
Appendix: Competency Domains in Evaluation 1.0 Reflective Practice: competencies focus on the fundamental norms and values underlying evaluation practice and awareness of one’s evaluation expertise and needs for growth. 2.0 Technical Practice: competencies focus on the specialized aspects of evaluation, such as design, data collection, analysis, interpretation and reporting. 3.0 Situational Practice: competencies focus on the application of evaluative thinking in analyzing and attending to the unique interests, issues, and contextual circumstances in which evaluation skills are being applied. 4.0 Management Practice: competencies focus on the process of managing a project/evaluation, such as budgeting, coordinating resources and supervising. 5.0 Interpersonal Practice: competencies focus on people skills, such as communication, negotiation, conflict resolution, collaboration, and diversity.
AppendixRTE DistinguishingFeatures Source: Getting Real About Real-Time Evaluation, Clare Nolan and Fontane Lo , Non-Profit Magazine, March 29, 2012
Appendix: Types of digital data sources • (1) Data Exhaust – passively collected transactional data from people’s use of digital services like mobile phones, purchases, web searches, etc., and/or operational metrics and other real-time data collected by UN agencies, NGOs and other aid organisations to monitor their projects and programmes (e.g. stock levels, school attendance). These digital services create networked sensors of human behaviour. • (2) Online Information – web content such as news media and social media interactions (e.g. blogs, Twitter), news articles, e-commerce, job posting. This approach considers web usage and content as a sensor of human intent, sentiments, perceptions, and want. Source : www.unglobalpulse.org
Appendix: Types of digital data sources • (3) Physical Sensors – satellite or infrared imagery of changing landscapes, traffic patterns, light emissions, urban development and topographic changes, etc. This approach focuses on remote sensing of changes in human activity • (4) Citizen Reporting or Crowd-sourced Data – Information actively produced or submitted by citizens through mobile phone-based surveys, hotlines, user- generated maps, etc. While not passively produced, this is a key information source for verification and feedback Source : www.unglobalpulse.org