380 likes | 569 Views
Re-thinking the Impact of Humanitarian Aid. 24 th Biannual Meeting Berlin, 2 nd December 2008. ALNAP 24 th Biannual Aims. To help clarify key issues around use of humanitarian impact assessments
E N D
Re-thinking the Impact of Humanitarian Aid 24th Biannual Meeting Berlin, 2nd December 2008
ALNAP 24th Biannual Aims • To help clarify key issues around use of humanitarian impact assessments • To move towards a shared understanding of the limits and possibilities of humanitarian impact assessment • To use this understanding to outline a practical vision for future work in this area 24th ALNAP Biannual Meeting, December 2008
Overview • Humanitarianism ‘transformed’ and the emergence of impact assessment • Challenges for assessing impact • Conclusions 24th ALNAP Biannual Meeting, December 2008
Humanitarianism Transformed Three broad trends since the 1980s contribute to the current interest in and debate around humanitarian impact assessment: • Humanitarian aid expanded and politicised • Humanitarian aid institutionalised and professionalised • Changing nature of vulnerability and human suffering 24th ALNAP Biannual Meeting, December 2008
Increase in initiatives, across and within agencies • Cross agency efforts include SMART; HNTS; TRIAMS; Fritz Humanitarian Impact Project; the CDA Listening Project; ALNAP HPP; SPHERE • Quality Compass; ECB ‘Good Enough’ Guide to impact measurement; DEC’s new Accountability Framework etc • WFP, ECHO and UNICEF all include impact in their evaluation guidelines; ActionAid’s ALPS; Save the Children UK’s GIM 24th ALNAP Biannual Meeting, December 2008
But despite considerable progress, problems remain. The Biannual background paper has identified six key broad challenges: • Defining impact assessment • Diverse stakeholders and interests • Indicators, baselines and data • Methodologies • Collective interpretation and analysis • Capacities and incentives 24th ALNAP Biannual Meeting, December 2008
1. Defining ‘impact’ and impact assessment 24th ALNAP Biannual Meeting, December 2008
Ideal picture of impact: the MDG Goal of universal primary education by 2015 24th ALNAP Biannual Meeting, December 2008
Reality of impact is rather different… “…Universal primary education would be achieved at present rates of progress in 2079 in sub-Saharan Africa and in 2036 in the Middle East and North Africa…” Social Watch 2007 24th ALNAP Biannual Meeting, December 2008
Real world impact is complex Output Outcome Activity Impact 24th ALNAP Biannual Meeting, December 2008
and hard to discern, even a long time after the fact 24th ALNAP Biannual Meeting, December 2008
Attribution or Contribution? Private Sector Other NGOs A.N. NGO Community and Family Local partners Religious organisations Developing Country Govmts Civil Society 24th ALNAP Biannual Meeting, December 2008
A widely recognised definition of impact assessment “…Impact assessment is the systematic analysis of the lasting or significant changes – positive or negative, intended or not – in people’s lives brought about by a given action or series of actions….” Novib/Oxfam research project, reported in C. Roche, Impact Assessment for Development Agencies 24th ALNAP Biannual Meeting, December 2008
Humanitarian action has its own challenges • Lack of clarity on IA definition and purpose • Rapidly changing humanitarian contexts • No consensus on objectives of humanitarian aid • Intended impacts of interventions often unclear or overambitious 24th ALNAP Biannual Meeting, December 2008
Development Considerable Lead Time Deliberate & pro-active Will take time, be thorough, extensive with comprehensive data collection Location chosen Duration planned Beneficiary population identifiable and static IA goals may be made compatible with socio-economic ones Humanitarian Sudden onset Reactive May need to be partial in coverage Unpredictable location Uncertain duration Beneficiary population heterogeneous and dynamic Priority given to “life saving” activities sometimes difficult to reconcile with IA goals Contextual differences between normal development aid and humanitarian aid situations 24th ALNAP Biannual Meeting, December 2008
2. Diverse stakeholders, interests and objectives 24th ALNAP Biannual Meeting, December 2008
The challenge of stakeholders (illustrative) UN agencies National and local partners Donors Media Affected Population International NGOs Private Sector Red Cross / Red Crescent Military Political authorities 24th ALNAP Biannual Meeting, December 2008
Different stakeholders have different perceptions of and interests in impact • Enabling different stakeholders to express divergent views of impact is crucial to successful impact assessment • IA findings more likely to be used they meet the interests of the end users 24th ALNAP Biannual Meeting, December 2008
Different needs may not be reconcilable and achievable in a single impact assessment: Accountability or Learning? • “The purpose of most impact assessments is to demonstrate past impact and to improve future practice, and there may be tension between the two.” • Often, too much is expected: “If we continue to expect evaluation to cover most of the accountability needs of the sector, we will be disappointed” (Sandison, P. 2006). 24th ALNAP Biannual Meeting, December 2008
3. Indicators, baselines and data 24th ALNAP Biannual Meeting, December 2008
Indicators, Baselines and Data for humanitarian IA Identifying impact indicators involves value judgements about what kinds of changes are significant for whom (Roche, C. 2000) 24th ALNAP Biannual Meeting, December 2008
‘The familiar adage [“you can lead a horse to water, but you can’t make it drink”] illuminates the challenge of committing to outcomes. The desired outcome is that the horse drinks the water. Longer-term outcomes are that the horse stays healthy and works effectively. But because program staff know they can’t make a horse drink water, they focus on the things they can control: leading the horse to water, making sure the tank is full, monitoring the quality of the water, and keeping the horse within drinking distance of the water. In short, they focus on the processes of water delivery rather than the outcome of water drunk. Patton, M. 1997:157-8 24th ALNAP Biannual Meeting, December 2008
“Reports were so consistent in their criticism of agency monitoring and evaluation practices that a standard sentence could almost be inserted into all reports along the lines of: It was not possible to assess the impact of this intervention because of the lack of adequate indicators, clear objectives, baseline data and monitoring.” (ALNAP, 2003) 24th ALNAP Biannual Meeting, December 2008
Issues include… • Weak or non-existent baselines • Data is often unavailable or unreliable • Data collected is mainly quantitative • Monitoring systems focus on process and outputs 24th ALNAP Biannual Meeting, December 2008
4. Methodologies 24th ALNAP Biannual Meeting, December 2008
Wealth of tools, techniques and approaches are available • Documentary analysis • Interviews • Questionnaires (including recipient perceptions surveys) • Monitoring • Ex-post evaluation • Case studies • Participatory Rapid Appraisal (PRA) • Experimental / quasi-experimental 24th ALNAP Biannual Meeting, December 2008
Qualitative versus quantitative • Any research design is shaped by both opportunities and constraints • Quantitative methods able to tackle ‘what’ and ‘where’ questions • Qualitative methods able to answer ‘why’ and ‘how’ questions, and are good at capturing process 24th ALNAP Biannual Meeting, December 2008
Mixed methods approaches can take into account, rather than dismiss, the complexity of assessing humanitarian impact 24th ALNAP Biannual Meeting, December 2008
5. Collective interpretation and analysis 24th ALNAP Biannual Meeting, December 2008
Improved interpretation and analysis of data through engagement with affected populations and other stakeholders • Humanitarian impact should not only be about providing more and better information, but also about making sure that findings are used in ways that improve the lives of affected populations • Wider stakeholders should also be engaged in this process • “Learning partnerships” for impact assessment 24th ALNAP Biannual Meeting, December 2008
To date, participation by and accountability to affected populations has not been a key feature of impact assessments Attempts to improve this include: • ECB ‘Good Enough Guide’ • The Quality Compass • Feinstein International Center Participatory Impact Assessments (PIA) 24th ALNAP Biannual Meeting, December 2008
6. Capacities and incentives 24th ALNAP Biannual Meeting, December 2008
Capacities and Incentives for improved humanitarian impact assessment • Lack of individual and organisational capacity to do good impact assessments • TORs are often unclear • objectives are not defined clearly within the context of the intervention • stakeholder analysis is limited • timing relates to institutional priorities rather than humanitarian need • skills relating to impact assessment methodologies are lacking • Contributing factors: high staff turnover; lack of a learning culture; inadequate investment and resources 24th ALNAP Biannual Meeting, December 2008
There is a considerable lack of incentives • Institutional incentives can override humanitarian ones; too few incentives to conduct good impact assessments; results-based approaches can create perverse incentives • A number of cultural barriers and biases that hinder good quality humanitarian impact assessment 24th ALNAP Biannual Meeting, December 2008
Recap: six challenges • Defining impact assessment • Diverse stakeholders and interests • Indicators, baselines and data • Methodologies • Collective interpretation and analysis • Capacities and incentives 24th ALNAP Biannual Meeting, December 2008
Conclusions “…Taken as a whole, the humanitarian system has been poor at measuring or analysing impact, and the introduction of results-based management systems in headquarters has yet to feed through into improved analysis of impact in the field… it is arguable that there has been significant under-investment in evaluation and impact analysis…” (Hoffman C.A. et al, 2004) • Our review gives little indication that there has been much movement from the position above, articulated in 2004 • Can we move forward? How? 24th ALNAP Biannual Meeting, December 2008