evaluation performance-monitoring

5

Click here to load reader

Upload: rahul-bhargava

Post on 22-May-2015

201 views

Category:

Business


0 download

DESCRIPTION

Performance monitoring and evaluation, high level

TRANSCRIPT

Page 1: Evaluation performance-monitoring

Performance Monitoring and EvaluationRahul Bhargava27 May 2012

Contents

Context 1

Components for a functional M&E system 2

Methodology 4

Context

Most progressive governments have institutionalized results-basedmanagement leading to performance-enhancement and effectivedelivery of progress and change. The objective of results basedmanagement is to “provide a coherent framework for strategic plan-ning and management based on learning and accountability in adecentralised environment.”1 Introducing a results-based approach 1 Note on Results Based Management,

Operations Evaluation Department,World Bank, 1997

aims to improve management effectiveness and accountability by“defining realistic expected results, monitoring progress toward theachievement of expected results, integrating lessons learned intomanagement decisions and reporting on performance.”2 2 “Results-based Management in

Canadian International DevelopmentAgency”, CIDA, January 1999

Results Based Management at UNDP, for example, is based on

• the definition of strategic goals which provide a focus for action;

• the specification of expected results which contribute to thesegoals and align programs, processes and resources behind them;

• on-going monitoring and assessment of performance, integratinglessons learnt into future planning;

• improved accountability, based on continuous feedback to improveperformance

Development programs and policies are designed to achieveoutcomes, for example, to raise incomes or improve agriculturalproductivity. Impact evaluations are a part of developing evidence-based policy. Outlining the Millennium Development Goals, ResultsFramework Documents and performance-pay incentives, make imple-menters focus on results that are set to be tracked internationally andnationally. These results are to be used to increase accountability, forbudgeting and informing policy.

Monitoring and Evaluation is used to improve the quality, effi-ciency and effectiveness of interventions.

Page 2: Evaluation performance-monitoring

Performance Monitoring and Evaluation Center of Excellence

Monitoring and evaluation (M&E) is key for the effective imple-mentation of results-based management. Within a results-orientedenvironment, the emphasis of M&E is on:

• active application of monitoring and evaluation information tothe continuous improvement of strategies, programs and otheractivities;

• monitoring of substantive development results instead of justinputs and implementation processes;

• monitoring and evaluation of results as they emerge instead of asan ex-post activity;

• conducting monitoring and evaluation as joint exercises withGovernment departments

Components for a functional M&E system

The World Bank identified twelve components of a working monitor-ing and evaluation system following international peer review. Thisapproach was formally adopted by UNAIDS and partners, for theirM&E capacity building efforts in 2007, to support the measurementand management of the HIV/AIDS epidemic.

The twelve components of a functional M&E system are3 3 Goergens, Marelize and Kusek,Jody Zall. Making Monitoring andEvaluation Systems Work: A CapacityDevelopment Tool Kit. World BankPublications. 2010.

2

Page 3: Evaluation performance-monitoring

Performance Monitoring and Evaluation Center of Excellence

1. Structure and organisationalalignment for M&E systems

2. Human capacity forM&E systems

3. M&Epartnerships

4. M&Eplans

5. CostedM&E workplans

6. Advocacycommunicationand culture forM&E systems

12. Usinginformation toimprove results

7. Routinemonitoring

8. Periodicsurveys

9. Databasesuseful toM&Esystems

10. Supportivesupervision anddata auditing

11. Evaluationand research

Components relating to “people, partnerships and planning”

1. Structure and organizational alignment for M&E systems

2. Human capacity for M&E systems

3. M&E partnerships

4. M&E plans

5. Costed M&E work plans

6. Advocacy, communication, and culture for M&E systems

Components relating to “collecting, capturing and verifying data”

3

Page 4: Evaluation performance-monitoring

Performance Monitoring and Evaluation Center of Excellence

7. Routine monitoring

8. Periodic surveys

9. Databases useful to M&E systems

10. Supportive supervision and data auditing

11. Evaluation and research

Final component about “using data for decision-making”

12. Using information to improve results

As suggested by the authors, these Components may be used as anorganizing framework for planning a M&E system’s staff, resources,support and funding requirements. It may be used as a reference forconducting assessments of a national M&E system, akin to the RFDframework, such that individual components may be assessed andto divide responsibilities at a country level, as a framework withinwhich all partners can work together.

Methodology

Monitoring is a continuous process that is used to inform programimplementation and day-to-day management. It usually tracks per-formance against expected results, facilitates comparisons acrossprograms and allows for the reviewing of trends over time. Inputs,activities, outputs and occasionally outcomes, such as toward na-tional and international development goals, are tracked.

Evaluations, meanwhile, are periodic objective assessments of com-pleted projects, programs or policy. They set out to answer specificquestions about design, implementation and results or outcomes. Tojustify them, programs should be,

Innovative To test a novel approach;

Replicable To decide on whether to scale up in a different setting,geography or context;

Strategically relevant To review flagship initiatives;

Untested Globally or in context;

Influence policy

Cost-effectiveness of programs can be determined following im-pact evaluations. Specifically, questions regarding the cost-benefitbalance of a given program and comparisons of the cost-effectiveness

4

Page 5: Evaluation performance-monitoring

Performance Monitoring and Evaluation Center of Excellence

of implementation alternatives can be answered based on the evi-dence.

Impact evaluations should be approached pragmatically, thatis, the methods should fit the operational context, not vice versa.This is achieved at the outset of programs, by designing prospectiveimpact evaluations into the project’s implementation. Evaluationdesigns that fit the political and operational context are as importantas the method itself. Where policy makers and civil society demandresults and accountability from public programs, impact evaluationsprovide credible evidence on performance and on whether a programachieved its desired outcome.4 4 Gertler, Paul (2010): Impact Evaluation

in Practice. Herndon, VA, USA: WorldBank Publications.

There are caveats, however. Often there is greater emphasis oncontrolling inputs, say, funds utilized or literature distributed, thanon assessing whether a program has achieved a goal.

Attribution is the hallmark of impact evaluations. They assess theimprovements in the well-being of persons that can be attributedto specific projects, programs or policy. It follows, that executedcorrectly, impact evaluations should be carried out within a logi-cal framework that set out causal pathways by which a programproduces outputs and influences outcomes.5 5 Ibid.

For example, the Government of Mexico recognized the needto monitor and evaluate the roll out of the innovative conditionalcash transfer program called “Progresa” in the 1990s. Its objectivewas to provide short-term support to create incentives to invest inchildren’s human capital, primarily conditional on regular attendanceat school and visiting health centres. Impact evaluation was built intothe program’s scale-up and replication. External evaluators found,in 2001, that the program targeted the poor well, improved schoolenrollment by an average of 0.7 additional years of schooling, andbrought down illness by 23 percent among children and 19 percentfewer sick or disability days among adults. The program reducedthe probability of stunting by 1 centimeter per year for childrenbetween the ages of 12 and 36 months. The evidence contributedto the decision by the new administration, following a presidentialelection, to expand the program by proving upper-middle schoolscholarships and health programs for adolescents. Other socialassistance programs, such as a large and well-targeted tortilla subsidyprogram, were scaled back.

5