2. The role of evaluation

Evaluation plays an essential role in development co-operation. It is a vital tool for maximising the impact of development co-operation efforts. At its core, development evaluation is a process that critically examines a policy, programme or project to better understand whether, and how, it has delivered the intended results. Institutions have put in place unique governance arrangements to ensure evaluation systems provide relevant, objective, and impartial insight into the performance of development efforts, while connecting country-level data and evidence with decision makers, who are often removed from day-to-day implementation and results.

“The Department uses evaluation to assess its policies, strategies, programmes, projects and other initiatives to generate evidence that provides both accountability for public funding and learning to inform strategic and operational decision making.”
Ireland, Department of Foreign Affairs  
        

Evaluation helps answer critical questions about interventions and their results. These questions are captured in the six criteria defined by the OECD-DAC (Figure 2.1). Evaluation questions include: Is the intervention doing the right things? How well does the intervention fit? Is it achieving its objectives? Is it implemented coherently and efficiently? Is it having positive impacts that last? (OECD, 2021[1]).

In the current global context, evaluation is more critical than ever. Eight years ago, countries around the world adopted the 2030 Agenda for Sustainable Development. As we pass the halfway point in implementing this ambitious agenda, it is clear that progress has not been fast enough. The global community remains off track to achieve the Sustainable Development Goals (SDGs) by 2030. Persistent poverty and inequality, deadly conflicts, gender inequality and the climate crisis all threaten people. These challenges have been further exacerbated in recent years by the COVID-19 pandemic, growing disaster risks from climate change, rising food and fuel prices, and violent conflicts that have led to record levels of people being displaced.

The development financing landscape puts further stress on development initiatives. Aid budgets face downward pressure in many countries, while development organisations are pushed to demonstrate results and value for money. In this context, it is important to think critically about development co-operation systems and structures to ensure that each development dollar is spent on relevant priorities and has maximum impact.

Accountability and learning are the dual objectives of development evaluation. Accountability in development co-operation, between governments and development partners, as well as towards citizens, civil society and other development stakeholders, is essential for effective development activities and maximising impact. Accountability in development co-operation means ensuring all resources are used efficiently and as intended. It also goes beyond this, ensuring that results are achieved and co-operation delivered in a way that supports inclusive, green and sustainable development.

“Evaluations contribute to the ongoing streamlining of innovation of the development co-operation programme as a whole by making recommendations for the improvement of future development interventions.”
Czechia, Department of Development Cooperation and Humanitarian Aid  
        

Learning is an equally important evaluation objective. Development evaluations provide important data and evidence on why an approach has worked or has not, as well as whether results are contributing to overall development goals and whether they are sustainable. When evaluation findings are used to inform future planning and resourcing decisions, the quality of co-operation is improved. In this vein, evaluation findings, including lessons learned and success factors, are used alongside complementary research to design future policies, programmes and projects.

In line with past findings, participating organisations highlighted that both accountability and learning objectives remain central. The 2016 study noted that organisations aimed to find a balance between accountability and learning (OECD, 2016[2]). Similarly, all respondents in 2022 noted that the purpose of evaluations includes both accountability and learning dimensions. However, interviews confirm an increasing focus on the learning function of evaluation (further details can be found in Chapter 5), as well as on results reporting.

The recent focus on learning reflects challenges in systematically using evaluation findings to inform decision making and strengthen overall effectiveness. During the 2022 data collection process, including written submission and interviews, participants highlighted a strong focus on the use of evaluation findings to drive organisational learning and to support evidence-based decision making. This is not to say that one objective is more important than the other. Rather, it indicates that more work can be done to facilitate the use of evaluation findings (see Chapter 5).

The overall number of centralised and decentralised evaluations undertaken by participating organisations has increased in the past decade.1 This demonstrates the significant value participating organisations place on evaluative evidence and reflects a growing appetite for evaluations from programme staff and other audiences. The number of evaluations conducted in the 2017-2021 period increased 82 percentage points over the 2011-2016 period (Figure 2.2). Over 2017-2021, the number of evaluations undertaken remained relatively steady (Figure 2.3), despite a slight drop in 2020 – likely due to the COVID-19 pandemic – and a jump in 2021 when travel restrictions were eased in most member countries allowing for evaluation units to catch up on some planned evaluations.

The number of evaluations conducted by reporting institutions varies greatly across organisations. Of the responding evaluation units, only a handful conduct more than 50 centralised evaluations (Figure 2.4,Figure 2.5). Conversely, 27 respondents conducted 10 evaluations or fewer (see Annex C). This reflects the significant diversity in the mandates, roles, and resourcing of evaluation units and the size and composition of the portfolios covered, rather than the overall size of the organisation per se. For example, KfW’s evaluation unit is mandated to evaluate the programmes and projects that it implements on behalf of BMZ, while DEval has a crosscutting function, covering all German development assistance (Box 2.1). In contrast, in the Netherlands, one evaluation unit is mandated to cover all Dutch development assistance, as well as foreign policy and trade activities that impact partner countries.

The age and size of bilateral organisations do not systematically affect the number of evaluations conducted, which is instead driven by the role evaluation plays within an organisation and the institutional set-up. Analysis was conducted to determine whether the age of the organisation and the amount of official development assistance (ODA) provided (both in terms of total amount and as a share of gross national income – GNI) influences the number of evaluations undertaken. No such relationship was found. This reflects the significant diversity in how development organisations set up and use their evaluation function. For example, some participating organisations conduct many centralised evaluations themselves. Other organisations focus more on decentralised evaluations, with centralised units providing an oversight or quality assurance function, or only conducting a smaller number of strategic evaluations.

Multilateral organisations conduct more evaluations overall than bilateral organisations, and implementing agencies conduct more evaluations than those with a policy-focused role. On average, in 2021, multilateral organisations conducted 123 evaluations (centralised and decentralised), compared to 35 for bilateral organisations. The Asian Infrastructure Investment Bank, which is a relatively new institution, has just begun conducting evaluations and is therefore not included in the overall dataset for this study (Box 2.2). When comparing the number of evaluations undertaken by organisations with different roles (primarily policy, primarily implementation or a dual role), analysis found that implementing agencies conduct more evaluations, on average (Figure 2.6). As can be seen in the individual profiles (Annex C), 2020 – and to a lesser extent 2021 – saw a dip in the number of evaluations conducted, due to pandemic-related challenges including travel restrictions.

Most organisations use both centralised and decentralised evaluations. Centralised evaluations, undertaken by units in headquarters, often focus on high-level policies, strategies or themes. Decentralised evaluations are undertaken by evaluation units of implementing agencies and by programmatic units or country offices, and often focus on specific sectors, programmes or projects.

There is significant variety in the types and methodologies of evaluations conducted. Across participating organisations, centralised evaluation units conduct many different types of evaluations, including policy/strategy evaluations, sector programme evaluations, country programme evaluations, project evaluations, process evaluations, thematic evaluations, cluster evaluations, impact evaluations; and syntheses or meta-evaluations (Box 2.3).

As was found in 2016, the links between centralised evaluation units and colleagues conducting decentralised evaluations are often weak. This is usually because there is no comprehensive evaluation workplan that includes all evaluations undertaken by the organisation. In some cases, there is no full accounting of all decentralised evaluations undertaken. However, some centralised units, for example in FCDO, do provide guidance, advice and quality assurance support to other parts of the organisation, allowing for a more complete understanding of the evaluation portfolio across departments.

“A large majority of the decentralised evaluations are deemed to be useful. However, insights from decentralised evaluations are not systematically exploited for organisational learning but [instead] remain on the individual level.”
Finland, Development Evaluation Unit   
        

About half of evaluations conducted in 2021 were decentralised, though due to limited reporting this is likely an underestimate. Of the total 2 321 evaluations reported by participating organisations, 760 (33 %) were conducted or commissioned by centralised evaluation units. Some 1 032 decentralised evaluations were conducted (44 %). However, as noted, this study focuses on centralised evaluation units in headquarters. While data were collected on both centralised and decentralised evaluations, detailed information on evaluations undertaken by specific programmatic units, country offices or projects was not specifically included in the data collection and therefore is likely undercounted in these figures.

References

[3] OECD (2023), Glossary of Key Terms in Evaluation and Results-based Management for Sustainable Development (Second Edition), OECD Publishing, Paris, https://doi.org/10.1787/632da462-en-fr-es.

[1] OECD (2021), Applying Evaluation Criteria Thoughtfully, OECD Publishing, Paris, https://doi.org/10.1787/543e84ed-en.

[2] OECD (2016), Evaluation Systems in Development Co-operation: 2016 Review, OECD Publishing, Paris, https://doi.org/10.1787/9789264262065-en.

Note

← 1. Participants were asked to report on the total number of evaluations conducted by their organisations in 2021 – both centralised and decentralised. However, kindly note that not all participating centralised evaluation units reported on the number of decentralised evaluations, and several have provided only an estimate.

Legal and rights

This document, as well as any data and map included herein, are without prejudice to the status of or sovereignty over any territory, to the delimitation of international frontiers and boundaries and to the name of any territory, city or area. Extracts from publications may be subject to additional disclaimers, which are set out in the complete version of the publication, available at the link provided.

© OECD 2023

The use of this work, whether digital or print, is governed by the Terms and Conditions to be found at https://www.oecd.org/termsandconditions.