the alnap meta-evaluation
DESCRIPTION
The ALNAP Meta-evaluation. Tony Beck Presentation for the IDEAS Conference, Delhi, 14 th April 2005. Outline. Background The ALNAP Quality Proforma Agency visits Findings from the agency visits Finding from the Quality Proforma. What is the ALNAP and its meta-evaluation?. - PowerPoint PPT PresentationTRANSCRIPT
The ALNAP Meta-evaluation
Tony Beck
Presentation for the IDEAS Conference, Delhi, 14th April 2005
Outline
1) Background2) The ALNAP Quality Proforma3) Agency visits4) Findings from the agency visits5) Finding from the Quality Proforma
What is the ALNAP and its meta-evaluation?
An overview of evaluation of humanitarian action quality
Identification of strengths and weaknesses
Recommendations for improvement across the sector and in individual agencies
Process
• Review of evaluation reports against a set of standards
• Visits to and interaction with agency evaluation offices
Focus:• 2001-2002: Accountability• 2003-2005: Accountability and: good practice,
dialogue, interaction
The ALNAP Quality Proforma
• ALNAP’s meta-evaluation tool
• Draws on good practice in EHA and evaluation in general
• Revised and peer reviewed in 2004
The ALNAP Quality Proforma
Made up of seven sections:1. Terms of reference2. Methods, practice and constraints3. Contextual analysis4. Analysis of intervention5. Assessing the report6. Overall comments
The ALNAP Quality Proforma
4 point rating scale
A = goodB = satisfactoryC = unsatisfactoryD = poor
Guidance notes for meta-evaluators. Eg: Consideration given to confidentiality and dignity?
Guidance: The evaluation report should detail how the overall approach and methods will protect confidentiality and promote respect for stakeholders’ dignity and self-worth.
The ALNAP Proforma
CoverageCoverage2001-2005: 197 evaluations
Process•2 meta-evaluators•Reconciliation of rating•Analysis by section
Mainstreaming of the Quality Proforma
• By ECHO to revise tor (lesson learning, protection, identification of users, prioritisation, time frame and users of recommendations etc)
• DEC Southern Africa evaluation (rated 7 agency report)
• Groupe URD (for planning of evaluations)
Agencies included in dialogue: 2003-4
CAFOD, Danida, ECHO, ICRC, OCHA, OFDA, Oxfam, SC-UK, SIDA, UNHCR, and WHO
Purpose of agency dialogue
• Agency response to initial two years of use of Quality Proforma
• To discuss Quality Proforma rating and agency strengths and weaknesses
• To discuss processes leading to good evaluation practice
• To discuss goof practice
Findings from dialogue with evaluation managers
• Areas affecting evaluation quality are not currently captured by the QP, eg
Evaluation quality depends on subtle negotiations within agenciesEvaluation funds in most cases are not being allocated for follow-upFollow-up to recommendations is complexMore agencies are using tracking matrices
Findings from dialogue with evaluation managers: the EHA market
• Main constraint to improved evaluation quality is agencies accessing available evaluators with appropriate skills
• Does the EHA market need further regulation?
Findings from the Proforma
Area of enquiry
Rating % attaining rating 2004
% attaining rating 2001-2003
TOR – Good practice in approach and method
Good orSatisfactoryUnsatisfactory or Poor
6
94
11
89
Findings from the Proforma
Area of enquiry
Rating % attaining rating 2004
% attaining rating 2001-2003
TOR – Intended users and uses
Good orSatisfactoryUnsatisfactoryor Poor
12
88
8
92
Findings from the Proforma
Area of enquiry
Rating % attaining rating 2004
% attaining rating 2001-2003
Consultation with primary stakeholders
Good orSatisfactoryUnsatisfactory or Poor
16
84
13
87
Findings from the Proforma
Area of enquiry
Rating % attaining rating 2004
% attaining rating 2001-2003
Use of the DAC criteria
Good orSatisfactoryUnsatisfactoryor Poor
52
48
50
50
Findings from the ProformaArea of enquiry
Rating % attaining rating 2004
% attaining rating 2001-2003
HR and management
Good orSatisfactoryUnsatisfactoryor Poor
50
50
51
49
Findings from the Proforma
Area of enquiry
Rating % attaining rating 2004
% attaining rating 2001-2003
Coordination Good orSatisfactoryUnsatisfactoryor Poor
52
48
50
50
Findings from the Proforma
Area of enquiry
Rating % attaining rating 2004
% attaining rating 2001-2003
Quality of evaluation of protection issues
Good orSatisfactoryUnsatisfactoryor Poor
32
68
10
90
Findings from the Proforma - 2005
• Improvement in most areas noted above of between 10 and 30 per cent
• Too early to disaggregate or suggest why this improvement has taken place
• Still a number of areas of generic weakness
Conclusions
Process:•Meta-evaluations need to include interaction with those being meta-evaluated•Agency visits have been important is discussing constraints to improved evaluation quality•Meta-evaluations need to maintain an appropriate balance between accountability functions and the need to improve evaluation quality through lesson learning
Conclusions: findings
• EHA demonstrates some areas of strength, and improvement over four years, eg use of most of the DAC criteria, analysis of HR
• Many evaluative areas need to be strengthened, eg gender, identification of use and users, participation of primary stakeholders, transparency of methodologies used