58
Evaluation of outcomes of CGIAR’s CRPs ECoP training session 25-26 September 2014 Burt Perrin La Masque [email protected] 30770 Vissec FRANCE +33 4 67 81 50 11 1

"Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Embed Size (px)

Citation preview

Page 1: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Evaluation of outcomes of CGIAR’s CRPs

ECoP training session 25-26 September 2014

Burt Perrin La Masque [email protected] 30770 Vissec FRANCE +33 4 67 81 50 11 1

Page 2: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Purpose of the training session

Consider approaches to the evaluation of outcomes Complex programmes/initiatives Focus on the CRPs

Outcomes of the session Better understanding: what’s involved in evaluation of outcomes in complex environ. Appreciation of challenges – and opportunities Ideas that you can use

2

Page 3: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Topics to explore How to plan evaluations

Complexity: what it is, implications for evaluation of research, CRPs Evaluation vs. other related activities Some tools for evaluation planning (evaluability assessment, TOC, outcome trajectories) Focus on evaluation use Evaluation designs and methods Analysis and interpretation

What this means for evaluation of outcomes of CRPs

3

Presenter
Presentation Notes
Small group exercise: identify your 3 key priorities for what you would like to gain from this training. (Responses may influence the
Page 4: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Characteristics of Evaluation

4

Page 5: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Why do evaluation?

Raison d’être of evaluation Social betterment Sensemaking

More generally, rationale for evaluation To be used! Improved policies, programmes, projects,

services, thinking

5

Page 6: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Evaluation – some key aspects

Systematic, data based, “objective” Evidence can come from multiples sources Can consider any aspect of a strategy, policy, programme, project Major focus on outcomes that follow from the intervention (i.e. attribution, cause) E - valua - tion

6

Page 7: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Different types of evaluation Ex-ante vs. ex-post Process vs. outcome Formative vs. summative Descriptive vs. judgemental Accountability vs. learning (vs. advocacy vs. pro-forma) Short-term actions vs. long-term thinking Etc.

7

Page 8: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Maximising evaluation value & use The right questions! Outcome focus Emergent – not restricted to pre-determined objectives/indicators Respects context, identifies how it interacts with what is done Identifies alignment of activities/projects/programmes with strategy/goal Assesses results orientation as well as actual results achieved

8

Presenter
Presentation Notes
More on use later – a focus on use should underly all aspects of the evaluation process.
Page 9: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

What is “complexity”? Emergent vs. predetermined outcomes Feedback loops Indirect, non linear trajectories; tipping points Unpredictability, random events Multiple components: partners, levels, causal package (complicated)

(But: try to explain complex situations as simply as possible!)

9

Page 10: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Nature of intervention and logic chain (e.g. Rogers)

Simple E.g. following a recipe Linear cause-and effect chain Complicated E.g. sending a rocket to the moon Multiple factors happening simultaneously Complex E.g. raising a child Recursive (feedback loops), emergent

outcomes that can’t be identified in advance Tipping points

10

Page 11: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Some characteristics of non-linear change (complexity science)

Cause-effect distance (outcome trajectory): long (or short) in time Depends upon a large number of intervening variables Usually several causes for any effect Change not proportional, incremental; qualitative leaps and bounds Sometimes initial ‘negative’ effects (e.g.

the J-curve) – implications for evaluation? Feedback loops

11

Page 12: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Werner Herzog:

1. “Man is a god when he dreams, but a beggar when he reflects.”

2. “Facts do not constitute the truth. There is a deeper stratum.”

Agree or not? Implications for evaluation?

12

Page 13: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Future orientation - Dilemma

“The greatest dilemma of mankind is that all knowledge is about past events and all decisions about the future.

The objective of this planning, long-term and

imperfect as it may be, is to make reasonably sure that, in the future, we may end up approximately right instead of exactly wrong.”

13

Page 14: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

14

Page 15: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Similarities, differences, complementarities

Evaluation and: Research Monitoring Audit

15

Page 16: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Evaluation vs. Research Research Primary objective: long-term knowledge generation (no

single study rarely sufficient) Theory creation/testing/revision Evidence needs: the more the better

Evaluation Reference to a particular type of situation Practical application/utilisation in some form an

essential component Evidence needs: as little as necessary to support

meaningful use (level of confidence required) But: evaluation makes use of research methodologies – from diverse disciplines

16

Page 17: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Monitoring – the concept and common definitions

Tracking progress in accordance with previously identified objectives, indicators, or targets (plan vs. reality) RBM, performance measurement, performance

indicators …

En français: “suivi” vs. “contrôle” Some other uses of the term Any ongoing activity involving data collection and

performance (usually internal, sometimes seen as self evaluation)

17

Page 18: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Monitoring and Evaluation Monitoring

Periodic, using data routinely gathered or readily obtainable, generally internal Assumes appropriateness of programme, activities, objectives, indicators Tracks progress against small number of targets/ indicators (one at a time) Usually quantitative Cannot indicate causality Difficult to use for impact assessment

Evaluation Generally episodic, often external Can question the rationale and relevance of the program and its objectives Can identify unintended as well as planned impacts and effects Can address “how” and “why” questions Can provide guidance for future directions Can use data from different sources and from a wide variety of methods 18

Page 19: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

How Monitoring and Evaluation can be complementary Ongoing monitoring Can identify questions,

issues for (in-depth) evaluation

Provide data for evaluation Nature of the intervention

Evaluation Can identify what

should be monitored in the future

19

Page 20: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

20

Monitoring, Evaluation and Impact Evaluation

Inputs Outputs Outcomes Impact

Investments (resources, staff…)

and activities

Products Intermediate achievements of

the project

Long-term, sustainable changes

Monitoring: what has been invested, done and produced, and how are we progressing towards the achievement of the objectives?

Evaluation: what occurred and what has been achieved as a result of the project?

Impact evalua-tion: what long-term, sustainable changes have been produced (e.g. poverty reduction)?

Page 21: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Evaluation vs. audit

21

Audit Compliance focus Rules and procedures Divergence: planned vs.

actual

Main attention to process Identify transgressions Standardised approach Outside scrutiny

Evaluation Outcome orientation, context and rationale, attribution Constructive guidance “Why” and “how” as well as “what” considerations Unintended as well as planned impacts and effects Wide range of potential approaches and methods

Page 22: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Evaluability assessment (including theory of change)

22

Page 23: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

What is an evaluability assessment (EA)?

Essentially, evidence-based plan for evaluation What aspects of the programme are evaluable – and when? E.g. coherent programme logic, data

availability, conducive environment …

What the programme needs to do Expected outcome trajectories TOC that includes above considerations

23

Page 24: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Elements in an EA (Involve stakeholders – build buy-in) Review/clarify programme intent; identify varying perspectives Help articulate the TOC; identify the soundness of the programme logic, including gaps Identify evaluation priorities and questions Identify evaluation implications for the program Explore feasibility of addressing potential questions (data availability, cost, other considerations) Explore alternative evaluation designs

24

Page 25: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Outcome focus: what is this?

Change that follows from the intervention in some way OECD/DAC: The likely or achieved short-

term and medium-term effects of an intervention’s outputs

Can/should consider other factors/ interventions Consider the “whys”

25

Page 26: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Outcomes (vs. process, impact) Level What is this Example: farmer training

Process Activities, outputs What was done

E.g. programme set up, implemented (as expected, or differently), needs assessment carried out, curriculum developed, outreach, training delivered

Outcomes Changes following from the programme

E.g. learning/expertise, confidence, planting practices, increased yields, new markets, increased revenues

Impact Long-term effects following from intervention Invariably in combination Raison d’être

Sustainability of short-term gains, Poverty, hunger, malnutrition reduction; natural resources sustainability

26

Page 27: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Questions for evaluation Start with the questions Choice of methods to follow How to identify questions Who can use evaluation information? What information can be used? How? Different stakeholders – different questions Consider responses to hypothetical findings Develop the theory of change How many questions?

27

Page 28: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

28

Page 29: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Three key evaluation questions

What’s happening? (planned and unplanned, little or big at any level)

Why?

So what?

29

Page 30: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

UNEG’s three evaluation questions

Are we doing the right thing?

Are we doing it right?

Are there better ways of achieving the results?

30

Page 31: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

OECD/DAC Evaluation Criteria Relevance Effectiveness Efficiency Impact Sustainability

31

• Evaluation criteria vs. evaluation questions • Breadth vs. focus • Intelligent vs. mechanical use

Page 32: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Some uses for evaluation Programme improvement Identify new policies, programme directions, strategies Programme formation Decision making at all levels Accountability Learning Identification of needs Advocacy Instilling evaluative/questioning culture

32

Page 33: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Some priorities for an EA Focus on outcomes Identify expected/potential outcomes Be open to unintended outcomes Outcome trajectories Evaluation priorities and questions Surface and question assumptions Implicit and explicit Be realistic (priorities, expectations of the programme and the evaluation) Don’t set up the programme for failure

33

Page 34: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

34

Page 35: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Theory of Change

Why a useful tool for planning an evaluation Alternative terms (intervention logic, logic model, results chain …) Linear vs. models that reflect complexity

35

Page 36: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Results chain Impact

Outcomes

Reach

Outputs

Processes

Inputs

36

Page 37: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Intervention logic model

Inputs

Activities OutputsResults/

Intermediate Outcomes

Ultimate Impacts

37

Page 38: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Generic logic model (simplified)

38

Page 39: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Generic logic model – in context

Inputs Activities Intermediate results (1)

Intermediate results (2)

Impacts

Other results Other

results

Other results

Other results

Other factors

Other factors

Other factors

Needs

Environment et context

Knowledge

Outputs

Other factors

Other interventions

Other interventions

39

Page 40: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

IMPACT ON CHILDREN

IPEC/partner

Initiatives

Targeted Interventions

Capacity building

Children

Families and communities

The enabling environment (Institutions,

policies & programmes, legislation, awareness,

mobilization…)

40

Page 41: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

41

Outline of factors affecting maternal and child health and nutrition

Fig. from Victora, Cesar G, Robert E Black, J Ties Boerma, Jennifer Bryce. (2010). Measuring impact in the Millennium Development Goal era and beyond: a new approach to large-scale effectiveness. The Lancet. Published Online July 9, 2010

Page 42: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

42

AAS Theory of Change

Page 43: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

43

AAS Theory of Change: Stakeholder engagement workshop

Page 44: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Design, analysis and method considerations

44

Page 45: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Alternative models of causality (All recognised in the physical and social sciences)

Successionist (factual) causality Counterfactual logic All but one possible explanation ruled out Generative (physical) causality Focus on underlying processes, the

“signature” Simultaneous or alternative causal strands “INUS” conditions: insufficient but necessary,

sufficient but unnecessary Non linear (e.g. “tipping point”) causality

45

Page 46: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Some considerations in choice of design (and methods)

Addresses, somehow, priority questions Simplest approach – at needed confidence Internal/external validity Face validity, construct validity Gets at “the whys” as well as “the whats” Engages stakeholders, partners Practicality (resources, time, data …)

46

Page 47: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Determining attribution – some alternatives

Experimental/quasi-experimental designs (counterfactual, randomisation) Eliminate rival plausible hypotheses Generative (physical) causality, INUS, non linear (“tipping point”) Theory of change approach “Reasonable attribution” “Contribution” vs. “cause”

47

Page 48: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Eliminate rival plausible hypotheses (Donald T. Campbell)

Identify plausible alternative explanations Plausible to multiple stakeholders Anticipate possible questions of sceptics Consider threats to both internal and to external validity Use the simplest means possible to rule out likelihood of alternative explanations

48

Page 49: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Contribution Analysis (Mayne: Using performance measures sensibly)

1. Develop the results chain 2. Assess the existing evidence on results 3. Assess the alternative explanations 4. Assemble the performance story 5. Seek out additional evidence 6. Revise and strengthen the

performance story

49

Page 50: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Further considerations for meaningful outcome evaluation

Need information about inputs and activities as well as about outcomes Check, don’t assume that what is

mandated in (Western) capitals is what actually takes place sur le terrain

Check: are data sources really accurate?

Dealing with responsiveness – a problem or a strength? (Internal vs. external validity) 50

Page 51: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Some alternative approaches Theory based Realist evaluation Most Significant Change, Success Case Method, Appreciative Inquiry Participative Outcome mapping/harvesting Anthropological Etc. etc. etc.

51

Page 52: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

To bear in mind “For every complex question, there is a simple answer – and it is wrong.” – H.L. Mencken “One cannot succeed on visible figures alone… The most important figures that one needs for management are unknown or unknowable.” – W. Edward Deming “Not every that can be counted counts, and not everything that counts can be counted.” – Einstein

52

Page 53: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

And … “Assessment of many of the most common activities in government requires soft judgment… Measurement often misses the point, sometimes causing awful distortions.” – Mintzberg “Better an approximate answer to the right question than an exact answer to the wrong question that can always be made precise.” – Tukey

53

Page 54: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Methods for data gathering: possible options

Surveys Panel studies/longitudinal (experimental/quasi-experimental) Interviews, group interviews Documentation, analysis of records Observation (quantitative, qualitative) Community members as researchers Alternative methods Multiple methods

54

Page 55: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Making evaluation useful - 1 Be strategic E.g. start with the big picture – identify questions

arising Focus on priority questions and information requirements Consider needs, preferences, of key evaluation users Don’t be limited to stated/intended effects Be realistic, don’t set programs up for failure Don’t try to do everything in one evaluation

55

Page 56: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Making evaluation useful - 2 Primary focus: how evaluation can be relevant and useful Bear the beneficiaries in mind Take into account diversity, including differing world views, logics, and values Be an (appropriate) advocate Don’t be too broad Don’t be too narrow

56

Page 57: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

How else can one practice evaluation so that it is useful?

Follow the Golden Rule “There are no golden rules.” (European Commission) Art as much as science

Be future oriented – focused on use Involve stakeholders Use multiple and complementary methods, qualitative and quantitative Recognize differences between monitoring and evaluation

57

Page 58: "Assessing Outcomes in CGIAR: Practical Approaches and Methods"

Conclusion

Primary focus: helping to make a difference (think strategically!) Requires focus of some form on outcomes What happens when, why, and so what Use evaluation to embrace complexity – as

simply as possible Questions are more important than the

“right” method Thank you / grazie / merci / gracias Burt Perrin

[email protected] 58