29
Developing State Monitoring Systems: Measuring Results

Developing State Monitoring Systems

Embed Size (px)

DESCRIPTION

Developing State Monitoring Systems

Citation preview

Page 1: Developing State Monitoring Systems

Developing State Monitoring Systems: Measuring Results

Page 2: Developing State Monitoring Systems

Presentation Overview

Traditional v. Result based M&E Approach

Government effort towards Result based M&E

J-PAL SA effort towards a result focused Monitoring System

Page 3: Developing State Monitoring Systems

The ‘M’ and the ‘E’

Evaluation gives evidence on whether targets and outcomes have been achieved. And impact evaluations seeks to ascertain whether these can be attributed to the program

Monitoring is used to continuously gauge whether the project or intervention is being implemented according to plan/targets

Page 4: Developing State Monitoring Systems

TRADITIONAL v. RESULT BASED M&E

Page 5: Developing State Monitoring Systems

Traditional M&E approach

Addresses compliance—“did they do it” question Did they mobilize the needed inputs? Did they undertake and complete the agreed activities? Did they deliver the intended outputs (the products or

services to be produced)?

So What? So what that activities have taken place? So what that the outputs from these activities have been

counted?

Is that enough to ascertain that whether the project/program/policy was a success or a failure?

Aparna Krishnan
Often this is just "did they spend the money on agreed activities" not even verify if the activities were undertaken
Page 6: Developing State Monitoring Systems

Changing Context Focus on Result

What are the results and impacts of government actions?

Governments are increasingly being called upon to demonstrate results in the face of

• Citizen accountability

• Donor focus on results

• Political climate

Additional how do we know • If policies, programs, and projects led to the desired results

and outcomes?

• How do we measure progress? How can we tell success from failure?

Page 7: Developing State Monitoring Systems

Result based M&E approach

A results-based approach can provides feedback on the actual outcomes and goals of government actions

Result-based monitoring and evaluation (M&E) is a management tool that if properly used can help Systematically track progress of project implementation, demonstrate results on the ground, and assess whether changes to the project design are needed in view of evolving circumstances (World Bank)

Results-based M&E differs from traditional implementation-focused M&E in that it moves beyond an emphasis on inputs and outputs to a greater focus on outcomes and impacts.

Page 8: Developing State Monitoring Systems

Revisiting the ToC: Immunization Incentives Example

Situation/Context Analysis: High health worker absenteeism, low value of immunization, limited income and time

GOALINPUT OUTPUT OUTCOME

Immunization

Camps Increased

Immuni-zation

Incentives for

Immunization

Camps are

reliably Open

Parents bring

children to the camps

Parents bring

children to the camps

repeatedly

Incentives are

delivered

Parents value

incentives

Incentives paid

regularly

Parents trust camps

Camp provides

immunizations

Page 9: Developing State Monitoring Systems

Increased Immunization Rates

Parents bring

children to the camps repeatedly

Immunization Camps

+ Incentives

Camps are open and incentives

are delivered

Parents bring

children to the camps

Measurement along a ToC

Reporting / Routine monitoring: On expenditures, activities, coverage (targets)

Periodic Monitoring : Of use of inputs/activities and process, intermediate outcomes

Evaluation: To assess long-term outcomes and impact through studies

After 9 months, camps were running on a

monthly basis in 90% of the

planned villages.100% incentives were delivered to

these camps

70-75% of Parents brought children to be immunized in the camps that

were open and reported receiving

incentives.

90 -95% of parents who

immunized the children during

the first round of immunization,

brought them to be immunized for the second round

At the end of the program

immunization rates were 39%

in the intervention villages as

compared to 6% in comparison

villages

After 6 months, camps were

established and equipped to run

in 90% of program villages. All health workers

were trained to offer parents the

appropriate incentives at

their visit

Page 10: Developing State Monitoring Systems

Increased Immunization Rates

Parents bring

children to the camps repeatedly

Immunization Camps

+ Incentives

Camps are open and incentives

are delivered

Parents bring

children to the camps

Measurement along a ToC

Traditional approach

After 9 months, camps were running on a

monthly basis at 90% of the

planned villages.Incentives were

delivered to these camps

70-75% of Parents brought children to be immunized in the camps that

were open and reported receiving

incentives.

90 to 95% of parents who

immunized the children during

the first round of immunization,

brought them to be immunized for the second round

At the end of the program

immunization rates was 39% in the intervention

villages as compared to 6% in comparison

villages

After 6 months, camps were

established and equipped to run

in 90% of program

villages. . All health workers were trained to

offer parents the appropriate incentives at

their visit

Result based approach

Page 11: Developing State Monitoring Systems

Result based M&E approach

The other major deviation from the traditional approach is that it moves away from scheme-wise monitoring to resource/sector-based monitoring. It aims to tracks outcomes at the State-level that multiple schemes may be targeting.

The focus of Results-based approach is therefore on better planning, targeting and allocation of resources achieve certain targets and it links such allocation to performance or results

Thus, this approach is the first step to introducing performance management and performance based budgeting

Page 12: Developing State Monitoring Systems

GOVERNMENT EFFORT TOWARDS RESULT BASED M&E

Page 13: Developing State Monitoring Systems

Government effort towards Result based M&E

Concurring with the recommendations of India’s second Administrative Reforms Commission(ARC), in 2009 the PM announced The introduction of Performance Monitoring & Evaluation System (PMES), to be implemented through a Performance Management Division (created inside) the Cabinet Secretariat.

The Result Framework document(RFD) is the essence of PMES- its a performance agreement through which targets are to be agreed upon, success indicators listed and performance evaluated.

The Process

• Generic RFD framework and guidelines were developed and an Adhoc task force set up to help Ministries/departments at the centre conduct their RFD exercise.

• The same exercise was to be replicated by states

Page 14: Developing State Monitoring Systems

The Result Framework Document

The RFD seeks to address three basic questions:

• What are ministry’s/department’s main objectives for the year?

• What actions are proposed by the department to achieve these objectives?

• How would someone know at the end of the year the degree of progress made in implementing these actions? That is, what are the relevant success indicators and their targets which can be monitored

Key idea is to enable departments to transit from an input driven approach to results/outcomes orientation

Page 15: Developing State Monitoring Systems

RFD Process ToC Process

Page 16: Developing State Monitoring Systems

Current status of RFD exercise

While the RFD exercise is a step in the right direction, its utilization remains a big question

At the national level, while a majority of Ministries have created the RFD document, they have still not started reporting against the targets and success indicators.

At the state level, this is an optional exercises , hence some states have opted not to do this. Even with states such as Chattisgarh, who have made department level RFD, these are merely lying as documents with the Line Departments and have not helped them realign the targets set in their Annual Plans.

Page 17: Developing State Monitoring Systems

J-PAL SA EFFORT TOWARDS RESULT FOCUSED MONITORING: SCHOOL BASED MONITORING IN HARYANA--- ABRC CASE STUDY

Page 18: Developing State Monitoring Systems

ABRC Case Study: Background

During the pilot evaluation of CCE and RE-LEP program in

Haryana, implementation challenges of newly launched

educational programs noticed in schools

Lack of monitoring and mentoring for teachers in

schools seen as key reason for poor program

implementation. Field level monitoring structure unclear

and not a widespread practice

ABRCs have the mandate of monitoring, but have not

carried out these roles. Time spent acting as ‘couriers’ of

information

Page 19: Developing State Monitoring Systems

Situational Context: quality of monitoring

ABRCs have the mandate of monitoring, but have not carried

out these roles. Time spent acting as ‘couriers’ of information

• Each ABRC is assigned 1 cluster to oversee – 12 to 15

schools

Preliminary results from JPAL Process Evaluation from Jan-Mar

2012:

• 13% of sampled schools have never been visited by

ABRCs in 2011-12

• 56% respondents reported ABRC visits last less than 1

hour

• 55% respondents reported never having received

feedback from ABRCs

• activities

Page 20: Developing State Monitoring Systems

Existing System of Monitoring & Mentoring in Haryana

Director of Elementary

Education DEE

Director of Secondary

Education DSE

State Project Director SPD

District Elementary

Education Officer DEEO

District Education

Officer DEO

 

District Project Coordinator DPC

 Block

Elementary Education Officer

BEEO

Block Education Officer BEO

Block Resource Coordinator BRC

Assistant Block Resource Coordinator

ABRC

Schools

• Focus on information gathering related to small number of inputs and financial flows. Focus on outputs and outcomes almost non-existent.

• Data collected in an ad-hoc manner, not timely, collected mainly from the perspective of reporting• No mechanism to use information to influence functioning of schools

NO systematic way of teaching information collection, review or flow from schools to blocks to districts

Page 21: Developing State Monitoring Systems

System of Monitoring & Mentoring in J-PAL Study Areas

Director of Elementary

Education DEE

Director of Secondary

Education DSE

State Project Director SPD

District Elementary

Education Officer DEEO

District Education

Officer DEO

 

District Project Coordinator DPC

 

Block Elementary Education Officer

BEEO

Block Education Officer BEO

Block Resource Coordinator BRC

Assistant Block Resource Coordinator ABRC

Schools

• Re-orient focus of monitoring – collect data on a variety of aspects. Information collected is focussed on teaching practices/programs .

• Set-up mechanisms to allow for use of collected information - double headed arrows depict two way information flow- emergence of a feedback loop.

• Set-up mechanisms to foster accountability for not just financial aspects but also higher order outcomes

Systematic & regular information collection , review and flow from the field to blocks to districts to HQ on a monthly basis

Page 22: Developing State Monitoring Systems

Key Changes to the System

Resources : Lack of clarity on role of monitors:

• Clarified roles and responsibilities of District, Block and Cluster officials --Fosters accountability

Lack of knowledge on monitoring tools, activities and

presentation of results

• Piloted and created monitoring tools

• Trained ABRCs on monitoring, basic data analysis and

report writing.

• Trained district and block level officials on the same as

well as performance management of ABRCs

Lack of forum to share experiences and “best-practices”

• Monthly review meetings set up for data sharing and

identification of issues

Page 23: Developing State Monitoring Systems

Data Collected On

Training and availability of materials

Attendance

Non-academic activities

Use of TLM

Teacher practices and student behavior

CCE documentation

LEP implementation

Block and district school observation

Page 24: Developing State Monitoring Systems

Findings :Teacher Practices and Student Behavior

Most of the teachers asked questions while teaching and assigned

in-class exercises

• Students responded to questions in more than 85% of

schools

• However students asked questions in only about 62% of

schools

• Students were also found repeating the teacher’s

answers in a large percentage of schools

Are the students comprehending?

Page 25: Developing State Monitoring Systems

Findings :Teacher Practices and Student Behavior

~ 50% of teachers review in-class work of students

• ABRCs corroborate in MRM that teachers do not check

homework thoroughly

Very few teachers used local context to explain concepts

Very few teachers adopted the practice of ‘group work’

No uniformity in syllabus coverage

• Some teachers have completed the syllabus already

• Many teachers have no plan – teach whatever they feel

like teaching

Should there be a prescribed Primary syllabus

schedule?

Page 26: Developing State Monitoring Systems

Examples of ABRC Presentations at MRMs Block Pehowa, Kurukshetra

1. Factual Data

Number of School where teachers were out of class

Number of Teacher out of class

Number of schools Insufficient Material

Number of schools where CCE and RE record Not maintained

CCE RE-LEP CCE RE-LEP

1st visit-Sep 2012

25 98 14 1 43 11

5th visit-Feb 2013

04 08 0 0 2 0

Nothing Letter Word Para Story

285

103121

102

167

125147

129 126

251

First Assessment Final Assessment

3. Observation of teaching practices

a) Mostly unused Teaching Method Models, charts Project based learning Workbooks Activity based learning

b) Common issue highlighted by the teacher and headmaster related to quality of education

Lack of teaching staff/extra work (official on duty, BLO, daak and construction work

Issues with RE-LEP syllabus Separate teacher for nursery class Student irregularity

2. On spot testing of learning quality

Page 27: Developing State Monitoring Systems

Pictorial Depiction of the Monitoring System

ABRCs visit school

ABRCs consolidate data from

visits

Data shared with

other ABRCs,

block and district officials

Discussions on action to

be taken

Unresolved issues to be elevated to

higher levels

To monitor and provide inputs on implementation based on previously collected data and discussions.

Collect data on inputs, outputs, outcomes

Data collated systematically, major issues identified

Course correction advice, inputs identified and discussed

Issues not resolved at district level elevated to state level officials for inputs

Page 28: Developing State Monitoring Systems

Action Taken by GoH Based on Monitoring

Course Correction:

Visits by block and district officials to resolve issues with

“problem” schools and teachers

Timely delivery of textbooks and other program related

materials

Greater focus on teacher attendance and in-class activities

• ABRCs shared best-practices in teaching

Request for re-fresher training for both programs

Page 29: Developing State Monitoring Systems

The Power of Measuring Results

If you do not measure results, you cannot tell success from failure.

If you cannot see success, you cannot reward it.

If you cannot see success, you cannot learn from it.

If you cannot recognize failure, you cannot correct it.

Source: Adapted from Osborne & Gaebler 1992