OUTCOME MAPPING
BUILDING LEARNING AND REFLECTION
INTO DEVELOPMENT
PROGRAMS
Building Learning and Reflection into Development Programs
Foreword by Michael Quinn Patton
INTERNATIONAL DEVELOPMENT RESEARCH CENTRE
Ottawa • Cairo •
Dakar • Montevideo • Nairobi • New Delhi • Singapore
© International Development Research Centre 2001
Published by the International Development
Research Centre
PO Box 8500, Ottawa, ON, Canada K1G 3H9
http://www.idrc.ca
National Library of Canada cataloguing in publication data
Earl, Sarah, 1971–
Outcome mapping : building learning and reflection into development programs
Includes bibliographical references.
ISBN
0-88936-959-3
1. Economic development projects — Evaluation.
2. Technical assistance — Developing countries — Evaluation.
3.
International cooperation.
I. Carden, Fred.
II. Smutylo, Terry.
III. International Development Research Centre
(Canada)
IV. Title.
HD75.9E72 2001 338.91 C2001-980277-3
All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, or otherwise, without the prior permission of the International Development Research Centre. Mention of a proprietary name does not constitute endorsement of the product and is given only for information.
IDRC Books endeavours to produce environmentally friendly publications. All paper used is recycled as well as recyclable. All inks and coatings are vegetable-based products. The full catalogue of IDRC Books is available at http://www.idrc.ca/booktique/.
Foreword— Michael Quinn Patton |
||
Preface |
||
1. |
Outcome Mapping: The Theory |
|
What Is Outcome Mapping? |
||
Three Stages of Outcome Mapping |
||
Why Not Impact? |
||
How Can Outcome Mapping Be Used? |
||
How Outcome Mapping Differs from Other Logic Models |
||
When Is Outcome Mapping Best Used? |
||
Is Outcome Mapping Appropriate for You? |
||
2. |
Outcome Mapping: The Workshop Approach |
|
Overview of the Steps |
||
Workshop Outputs |
||
Who Should Participate? |
||
Who Should Facilitate? |
||
Workshop Materials |
||
Presentation Methods |
||
Preparing for an Outcome Mapping Workshop |
||
Conducting a Historical Scan |
||
Developing a Common Understanding of Evaluation |
||
Guiding Principles for Evaluation |
||
3. |
Stage 1: Intentional Design |
|
Introduction to Stage 1 |
||
Step 1. Describe the Vision |
||
Step 2. Identify the Mission |
||
Step 3. Identify the Boundary Partners |
||
Step 4. Identify the Outcome Challenge |
||
Design Worksheet 1: Program Framework |
||
Step 5. Develop Graduated Progress Markers |
||
Design Worksheet 2: Progress Markers |
||
Step 6. Complete a Strategy Map for Each Outcome Challenge |
||
Design Worksheet 3: Strategy Map |
||
Step 7. Articulate Organizational Practices |
||
Design Worksheet 4: Organizational Practices |
4. |
Stage 2: Outcome & Performance Monitoring |
|
Introduction to Stage 2 |
||
Monitoring Three Parallel Processes |
||
Will Outcome Mapping Provide the Appropriate Monitoring System? |
||
Step 8. Setting Monitoring Priorities |
||
Monitoring Worksheet 1: Monitoring Plan |
||
Step 9. Develop an Outcome Journal |
||
Monitoring Worksheet 2: Outcome Journal |
||
Step 10. Customize a Strategy Journal |
||
Monitoring Worksheet 3: Strategy Journal |
||
Step 11. Customize a Performance Journal |
||
Monitoring Worksheet 4: Performance Journal |
||
Monitoring Worksheet 5: Program Response |
||
Monitoring Worksheet 6: Reviewing the Logic of the Program |
||
5. |
Stage 3: Evaluation Planning |
|
Introduction to Stage 3 |
||
Step 12. Evaluation Plan |
||
Evaluation Worksheet 1: Evaluation Plan |
||
Appendix A: Sample Intentional Design Framework |
||
Appendix B: Overview of Evaluation Methods |
||
Appendix C: Glossary |
||
Appendix D: Terms in French, English, and Spanish |
||
References |
||
About the Authors |
||
The Publisher |
Imagine a map . . . drawn from your memory instead of from the atlas. It is made of strong places stitched together by the vivid threads of transforming journeys. It contains all the things you learned from the land and shows where you learned them . . . . Think of this map as a living thing, not a chart but a tissue of stories that grows half-consciously with each experience. It tells where and who you are with respect to the earth, and in times of stress or disorientation it gives you the bearings you need in order to move on. We all carry such maps within us as sentient and reflective beings, and we depend upon them unthinkingly, as we do upon language or thought . . . . And it is part of wisdom, to consider this ecological aspect of our identity.
– John Tallmadge, Meeting the Tree of Life (1997: IX)
Maps are cognitive guides. They locate us, helping us to figure out where we are now in relation to where we’ve been, and to plan where we’re going. It is altogether appropriate, then, that IDRC’s Evaluation Unit has chosen the metaphor of mapping to guide those interested in development on the sometimes confusing, even frightening journey through the hazardous territory of outcomes.
The language can be daunting: outcomes, impacts, goals, objectives, purposes, mission, and outputs — and these terms just scratch the surface. The questions can overwhelm. What’s the difference between evaluation and monitoring? How do short-term changes relate to intermediate changes and long-term results? What kinds of results count as outcomes? How can the need for accountability be balanced against the need for learning? Then there’s the attribution problem. To what extent and in what ways can one establish a causal linkage between activities, outputs, outcomes, and impacts? Who gets credit for results? What kinds of evidence are credible? What’s the unit of analysis? What role does stakeholder involvement play in all this?
The territory of developmental change and evaluation is vast, complex, and ever-changing. Trying to manoeuver through that territory one is likely to encounter deep ravines of uncertainty, mountains of data, and side-tracks that lead nowhere. It sure would help to have a map to the
territory. This manual on outcome mapping can’t provide the specific map you need for your own territory, for each territory is unique and presents its own challenges, but this manual will tell you how to create your own map. It will guide you through the language forest of terminology. It will show you how to navigate the winding river of a results chain. It will help you figure out where the boundaries are of the territory you are exploring and provide assistance in identifying “boundary partners” to accompany you on the outcomes journey. It will show you how to construct a strategy map and figure out progress markers.
A vision of useful and meaningful evaluation — evaluation in support of learning — undergirds this manual and is essential to understanding its important contributions. A sophisticated understanding of contemporary evaluation issues informs what may appear as simple mapping exercises provided here. One of the strengths of the manual is that it cuts through the complex evaluation literature, extracting deceptively straightforward and commonsensical wisdom from the many divisions and debates within the evaluation community based on a deep knowledge of, and explicit value premises related to, development. The staff of IDRC’s Evaluation Unit have long been working to support learning as a primary outcome of development program evaluation. They have observed that longer term outcomes and impacts often occur a long way downstream from program implementation and may not take the form anticipated. These longer term outcomes depend on responsiveness to context-specific factors, creating diversity across initiatives. The outcomes examined include the depth and breadth of involvement by many stakeholders, processes that become results in and of themselves when done in ways that are sustainable. These characteristics make it difficult for external agencies to identify and attribute specific outcomes to specific components of their programs or to aggregate and compare results across initiatives.
Outcome Mapping offers a methodology that can be used to create planning, monitoring, and evaluation mechanisms enabling organizations to document, learn from, and report on their achievements. It is designed to assist in understanding an organization’s results, while recognizing that contributions by other actors are essential to achieving the kinds of sustainable, large-scale improvements in human and ecological well-being toward which the organization is working. The innovations introduced in Outcome Mapping provide ways of overcoming some of the barriers to
learning faced by evaluators and development partners. Attribution and measuring downstream results are dealt with through a more direct focus on transformations in the actions of the main actors. The methodology has also shown promise for across-portfolio learning in that it facilitates standardization of indicators without losing the richness in each case’s story, thus combining quantitative and qualitative approaches.
An old hiking adage warns that “the map is not the territory.” True enough. You need to keep your eyes open and watch for unexpected outcroppings and changes in the terrain. But without a map, you can get so lost in the territory that it’s hard to even figure out where you started much less find your way to an outcome. Outcome Mapping provides not only a guide to essential evaluation map-making, but also a guide to learning and increased effectiveness, and affirmation that being attentive along the journey is as important as, and critical to, arriving at a destination.
Michael Quinn Patton
17 September 2001
This page intentionally left blank.
The International Development Research Centre’s (IDRC) conceptual and practical work over the past few years with donors, Southern research institutions, program staff, and evaluation experts has brought to the fore a fundamental problem with existing approaches to reporting on development impacts. When referring to “impact,” development organizations usually mean significant and lasting changes in the well-being of large numbers of intended beneficiaries. These changes are the results for which donors expect accountability. This is problematic because the complexity and fluidity of development processes mean that achieving such impacts requires the involvement of a variety of actors, often over a considerable period of time. When large-scale change — or impact — manifests itself, it is often the product of a confluence of events over which no single agency has control or can realistically claim full credit.
In response to this problem, several of IDRC’s programs and its Evaluation Unit have been working with Dr Barry Kibel, of the Pacific Institute for Research and Evaluation, to adapt his Outcome Engineering approach to the development research context. Outcome Engineering was developed to help Dr Kibel’s clients in the American social service sector meet their reporting needs while improving their performance. Although individuals being treated by social service providers in the United States face different constraints and require different types of support than international applied research institutions, the conceptual and practical problems associated with assessing results have proven to be quite similar. Some of the main adaptations have related to modifying the basic unit of analysis from the individual to groups, organizations, consortia, and networks. Adapting the methodology has been greatly enhanced through methodological collaboration with the West African Rural Foundation (Senegal) and testing with the Nagaland Empowerment of People Through Economic Development Project (India) and the International Model Forest Network Secretariat (Canada). The result is a methodology, called “Outcome Mapping,” that characterizes and assesses the contributions made by development projects, programs, or organizations to the achievement of outcomes. This methodology is applicable during the design stage or during a midterm or ex-post assessment.
This manual is intended as an introduction to the theory and concepts of Outcome Mapping and as a guide to conducting an Outcome Mapping workshop. Although Outcome Mapping may be appropriate in various contexts, it has primarily been tested by development research organizations and programs working in Canada, Africa, Latin America, and Asia. This manual reflects that perspective and Outcome Mapping may have to be adapted to be used with groups other than our constituency of researchers, scientific organizations, government officials, policymakers, and NGOs (for example, communities).
Outcome Mapping has been developed in organizations where monitoring and evaluation are primarily intended to help with program learning and improvement. The tenor of a program’s approach to Outcome Mapping will necessarily be influenced by its own and its donor’s perspectives on monitoring and evaluation. Outcome Mapping will only be as empowering, participatory, and learning-oriented as the program that implements it. Outcome Mapping takes into consideration the threats and anxiety that can be associated with planning, monitoring and evaluation, especially in a donor/recipient relationship. It offers a participatory methodology that can help programs develop a system that can meet both accountability and learning needs.
Section 1 presents the theory underpinning Outcome Mapping — its purpose and uses, as well as how it differs from other approaches to monitoring and evaluation in the development field, such as logic models. Section 2 presents an overview of the workshop approach to Outcome Mapping — including the steps of the workshop, as well as how to select participants and facilitators. Sections 3, 4, and 5 outline each of the stages of an Outcome Mapping workshop, suggest a process that can be followed by the facilitator, and provide examples of the finished “products.”
Outcome Mapping is a dynamic methodology that is currently being tested at the project, program, and organizational levels. Its conceptual development has been a collaborative effort between IDRC’s Evaluation Unit, program initiatives, secretariats, and partner organizations. We would especially like to thank a number of organizations who have been instrumental in field testing the approach: the Sustainable Use of Biodiversity (team leader, Wardie Leppan) and the Alternative Approaches to Natural Resource Management (team leader, Simon Carter) program initiatives of IDRC, Fadel Diamé and the staff of the
West African Rural Foundation, K. Kevichusa and the staff of the Nagaland Empowerment of People Through Economic Development Project team, Sonia Salas and the Condesan Arracacha project, Jim Armstrong and the Governance Network, and Fred Johnson and the International Model Forest Network Secretariat staff. For their valuable comments on the content and structure of this manual, we thank Marie-Hélène Adrien, Charles Lusthaus, Fiona Mackenzie, Nancy MacPherson, Greg Mason, John Mayne, Alex Moiseev, Carroll Salomon, and Ian Smillie. Outcome Mapping remains a work in progress, and so we look forward to receiving your comments and suggestions toward its improvement. You can reach us at the address noted below. Your comments will be valued and will enrich our work.
Evaluation Unit
International Development Research Centre
PO Box 8500
Ottawa, Ontario
Canada K1G 3H9
Phone: (+1 613) 236-6163 (ext. 2350)
Fax: (+1 613) 563-0815
E-mail: evaluation@idrc.ca
You can learn more about the Evaluation Unit’s work with Outcome Mapping on our Web site at www.idrc.ca/evaluation.
Sarah Earl
Fred Carden
Terry Smutylo
This page intentionally left blank.
1 |
OUTCOME MAPPING: |
---|
Outcome Mapping focuses on one specific type of result: outcomes as behavioural change. Outcomes are defined as changes in the behaviour, relationships, activities, or actions of the people, groups, and organizations with whom a program works directly. These outcomes can be logically linked to a program’s activities, although they are not necessarily directly caused by them. These changes are aimed at contributing to specific aspects of human and ecological well-being by providing partners with new tools, techniques, and resources to contribute to the development process. Boundary partners are those individuals, groups, and organizations with whom the program interacts directly and with whom the program anticipates opportunities for influence. Most activities will involve multiple outcomes because they have multiple boundary partners. By using Outcome Mapping, a program is not claiming the achievement of development impacts; rather, the focus is on its contributions to outcomes. These outcomes, in turn, enhance the possibility of development impacts — but the relationship is not necessarily a direct one of cause and effect. Ultimately, all organizations engaged in international development want their work to contribute to long-term development impacts. However, this is rarely accomplished by the work of a single actor (especially an external donor agency). The complexity of the development process makes it extremely difficult to assess impact (especially for an external donor agency seeking attribution). Furthermore, focusing assessment on long-term development impacts does not necessarily provide the kind of information and feedback that programs require to improve their performance. For these reasons, Outcome Mapping focuses on outcomes instead of impact, while recognizing the importance of impact as the ultimate goal toward which programs work.
Outcome Mapping assumes that the boundary partners control change and that, as external agents, development programs only facilitate the
process by providing access to new resources, ideas, or opportunities for a certain period of time. A focus on the behaviour of the boundary partners does not mean that the program decides how, when, and why those partners will change. In fact, by focusing on changes in behaviour, Outcome Mapping makes explicit something that has been accepted by development practitioners for a long time: the most successful programs are those that devolve power and responsibility to endogenous actors.
As development is essentially about people relating to each other and their environments, the focus of Outcome Mapping is on people. The originality of the methodology is its shift away from assessing the development impact of a program (defined as changes in state — for example, policy relevance, poverty alleviation, or reduced conflict) and toward changes in the behaviours, relationships, actions or activities of the people, groups, and organizations with whom a development program works directly. This shift significantly alters the way a program understands its goals and assesses its performance and results. Outcome Mapping establishes a vision of the human, social, and environmental betterment to which the program hopes to contribute and then focuses monitoring and evaluation on factors and actors within that program’s direct sphere of influence. The program’s contributions to development are planned and assessed based on its influence on the partners with whom it is working to effect change. At its essence, development is accomplished by, and for, people. Therefore, this is the central concept of Outcome Mapping.
Outcome Mapping does not belittle the importance of changes in state (such as cleaner water or a stronger economy) but instead argues that for each change in state there are correlating changes in behaviour. Many programs can better plan for and assess their contributions to development by focusing on behaviour. This is particularly true for programs that focus on capacity building. For example, a program’s objective may be to provide communities with access to cleaner water by installing purification filters. Traditionally, the method of evaluating the results of this program would be to count the number of filters installed and measure changes in the level of contaminants in the water before and after the filters were installed. A focus on changes in behaviour begins instead from the premise that water does not remain clean without people being able to maintain its quality over time. The program’s outcomes are therefore evaluated in terms of whether those responsible for water purity in the
communities not only have, but use, the appropriate tools, skills, and knowledge to monitor the contaminant levels, change filters, or bring in experts when required. Outcome Mapping provides a method for development programs to plan for and assess the capacities that they are helping to build in the people, groups, and organizations who will ultimately be responsible for improving the well-being of their communities. Outcome Mapping does not attempt to replace the more traditional forms of evaluation, which focus on changes in conditions or in the state of well-being. Instead, Outcome Mapping supplements other forms of evaluation by focusing specifically on related behavioural change.
Outcome Mapping is divided into three stages. The first stage, Intentional Design, helps a program establish consensus on the macro level changes it will help to bring about and plan the strategies it will use. It helps answer four questions: Why? (What is the vision to which the program wants to contribute?); Who? (Who are the program’s boundary partners?); What? (What are the changes that are being sought?); and How? (How will the program contribute to the change process?). The second stage, Outcome and Performance Monitoring, provides a framework for the ongoing monitoring of the program’s actions and the boundary partners’ progress toward the achievement of outcomes. It is based largely on systematized self-assessment. It provides the following data collection tools for elements identified in the Intentional Design stage: an “Outcome Journal” (progress markers); a “Strategy Journal” (strategy maps); and a “Performance Journal” (organizational practices). The third stage, Evaluation Planning, helps the program identify evaluation priorities and develop an evaluation plan. Figure 1 (next page) illustrates the three stages of Outcome Mapping.
The process for identifying the macro-level changes and designing the monitoring framework and evaluation plan is intended to be participatory and, wherever feasible, can involve the full range of stakeholders, including boundary partners. Outcome Mapping is based on principles of participation and purposefully includes those implementing the program in the design and data collection so as to encourage ownership and use of findings. It is intended to be used as a consciousness-raising, consensus-building, and empowerment tool for those working directly in the development program.
Outcome Mapping introduces monitoring and evaluation considerations at the planning stage of a program. It also unites process and outcome evaluation by collecting data on the program’s success in implementing its strategies and the results being achieved by its boundary partners. Separating process and outcome evaluation is misleading, because it implies that an organization achieves outcomes as a direct result of imple-
menting a program. But obtaining results in international development is not that straightforward. Focusing monitoring and evaluation on the program’s boundary partners makes it possible to obtain useful feedback about the program’s performance and results within its sphere of influence. The logic underpinning this is that the resources for monitoring and evaluation are limited and are best spent on studies that will allow for better understanding of the influences of a program’s work, so as to improve its performance. Also, having a system in place that regularly collects credible data on external results and internal performance can help a program meet its accountability requirements to its donor or management.
Rather than attempting a “rolling-up” of evaluation information from a number of disparate activities, Outcome Mapping provides a method for a program to set overall intentions and strategies, monitor its contributions to outcomes, and target priority areas for detailed evaluation studies. As a whole, Outcome Mapping provides a program with a continuous system for thinking holistically and strategically about how it intends to achieve results. It also provides the tools for a program to tell its performance story. It does so by monitoring in three key areas: changes in the behaviour of partners; the program’s strategies; and the way in which it functions as an organizational unit. By actively engaging the team in the monitoring and evaluation process, Outcome Mapping empowers them to articulate, with accurate and reliable data, both what they do to support outcomes and the depth of change in their partners. In essence, it tries to implant the passion and enthusiasm of programming into the assessment process. Outcome Mapping moves away from the notion that monitoring and evaluation are done to a program, and, instead, actively engages the team in the design of a monitoring framework and evaluation plan and promotes self-assessment.
As they are currently applied, the concepts of “attribution” and “impact” can limit the potential of programs to learn from evaluations of development efforts. In light of shrinking international development aid dollars and the need to optimize what is left, donors are increasingly basing funding decisions on their recipients’ abilities to demonstrate “impact.” In development terms, this typically means providing evidence that a
particular program has brought about a sustainable improvement in the environment or in the well-being of a large number of targeted beneficiaries. Methodologically, this requires isolating the key factors that caused the desired results and attributing them to a particular agency or set of activities.
For development agencies, this means identifying and measuring the net, positive effects as they result directly from the activities that those agencies support. In the literature, there are few good examples where this has been done. IDRC has long struggled with this challenge in relation to development research — a struggle made more difficult by IDRC’s style of program delivery. Research results improve peoples’ lives via long, busy, discontinuous pathways. Tracing the connections is at best unreliable and at worst impossible. This disconnect between important upstream contributions and downstream goals has been recognized in evaluation circles for a long time. In his 1967 book, Evaluative Research, Edward A. Suchman stated (p. 55.):
The extent to which immediate and intermediate goals can be divorced from ultimate goals as valid in themselves poses a difficult question. Certainly there is a tremendous amount of activity, perhaps the largest portion of all public service work, devoted to the successful attainment of immediate and intermediate goals which appear to have only a very indirect bearing upon ultimate goals.
While the push to measure, demonstrate, and be accountable for development impact is most obvious within the donor community, it has also made its way into recipient agencies and communities through requirements such as “logical framework analysis” (LFA) or “results-based management” (RBM) for planning and reporting on activities to donors. Consequently, the search for impact has become an accepted and dominant part of the development discourse. Nonetheless when donors and recipients try to be accountable for achieving impact, they are severely limiting their potential for understanding how and why impact occurs. The drive to claim credit interferes with the creation of knowledge. As one colleague has expressed it, the singular focus on results yields “clueless feedback.”
There are a number of ways in which the practices of donor agencies relating to impact contradict the understanding of socially and environmentally sustainable development built up over 30 years of experience, research, and evaluation.
Linear, “cause and effect” thinking contradicts the understanding of development as a complex process that occurs in open systems. Pressure to demonstrate, measure, and be accountable for impact has led donors to conceptualize, implement and evaluate programs using tools and methods which seek a linear cause and effect relationship between a problem and the identified “solution” to that problem. However, experience tells us that development is a complex process that takes place in circumstances where a program cannot be isolated from the various actors with which it will interact (for example, other donors, partner organizations, government departments, communities, organizations, and groups within the community). Nor can it be insulated from the factors by which it will be influenced (these include social, political, cultural, economic, historical, and environmental factors). While it is necessary to simplify to some extent in order to plan and implement programs, methodologically one needs to acknowledge the contextual reality of which programs are a part.
Bureaucratized programming contradicts the relationships, vision, and values of socially sustainable development. Donor efforts to measure, demonstrate, and be accountable for development impact have led to a reliance on logic-based approaches — requirements that a “logical framework” or “results chain” be prepared in the planning, management, and evaluation of each program. The state to which this practice has evolved has fostered a bureaucratization in programming. In his book, Impact Assessment for Development Agencies, Chris Roche identifies a trend in the way we conceptualize and plan for development. In this trend, “agreements and partnerships based on shared values have been replaced by bureaucratic trust based on plans, budgets, and accounts” (Roche 1999). This emphasis on plans, budgets, and accounts has implications for: how partners are selected; the ways in which relationships with partners develop; the kind of programs that are initiated; the ways in which those programs are designed and implemented; and the type of management and reporting that is expected of program partners. Overall, this shift in approach leads to structures that foster the belief among managers that appropriately completed and updated planning and reporting documents greatly increase the quality of development work, as well as increasing managers’ influence over the achievement of results. While this may be true in a very limited sense, experience suggests that this approach reduces the likelihood of strong partnerships and stakeholder involvement. Shared vision, values, and commitment, risk-taking, innovation, and the seeking
out of new partners are all incompatible with the domination of bureaucratic instruments. Passion and a sense of purpose can easily be lost and, with them, opportunities for achieving sustained benefits.
Attribution of impact to donors contradicts the multiple endogenous contributions and conditions necessary for sustainable development. Even though it is commonly understood that socially sustainable development occurs when a variety of actors and factors converge to allow local organizations to take ownership and responsibility for planning and implementing programming, donors feel that it is necessary to be accountable for impact. They feel pressured to attribute to their interventions the changes that have occurred. In seeking to demonstrate how programs have led to development impact, one does not examine or learn about how the contributions or involvement of others (including donors, recipients, other organizations, and participants), and the environment in which those interventions occur, fit together to create development results.
In particular, efforts to account for donor contributions to development impact ignore a key factor of socially sustainable development: the need for endogenous organizations and communities to take ownership of program components. Local ownership requires the devolution of planning, decision making, and other elements from external actors to internal actors. Yet, if one seeks to demonstrate how one’s contribution led to development impact, one intentionally elevates that contribution above those of the other actors and factors that have played a part in developmental change. Using the simplified “results chain” model, Figure 2 illustrates the shifts in influence as events move along the “results chain” towards impact.
Typically, at the “input” end of the process, the externally supported program has a great deal of control over decisions and events (such as budget and design, choice of partners, location, and timing). At this stage, the local partners and beneficiaries have the least influence. However, once funding flows, activities start, and local participants become increasingly active, the balance of influence should begin to change. If the program progresses as intended, local actors become more committed and their roles become more prominent. For the outcomes of the program to be relevant and lead to long-term, large-scale, sustainable benefits (in other words, to create impact), local ownership needs to become effective and dominant. This model suggests that the more successful the program, as it moves towards the impact stage, the more its exogenous influence is
supplanted by endogenous activities and institutions. Thus a paradox exists for external agencies under pressure to take credit for results at the “impact” stage; for it is here, where their influence, if they have been successful, is low and decreasing relative to that of other actors.
The notion of program impact contradicts the ongoing confluence of initiatives and conditions that affect social and environmental well-being. The search for impact leads donors to look for and expect results at the end of each program. The intention is often to alter plans for further funding or implementation based on whether or not results are being achieved at certain points within the funding cycle. While it is appropriate to look for development results, in many cases these may not occur until some time after the program is completed. The desire to show measurable or actionable results can shorten the vision of an initiative to aim at goals that are attainable in the short term with very low risk.
Even in cases where it is realistic to look for development impact at the end of a single initiative, the story does not end once “impact” arrives. In IDRC’s experience, whether the desired results are achieved or not, change continues. Conditions will perpetually be affected by combinations of social and natural events. Development does not occur in one, three, or five-year periods with a clear beginning, middle, and end. It is realistic, therefore, to expect that individual programs will make a difference that is incremental and cumulative rather than singular and
dramatic. It is also reasonable to expect that even once the desired outcome has been reached, it may erode due to subsequent influences either within or entirely independent of the program. Aware of this reality, many development agencies seek to build endogenous capacity to maintain the ability to respond to ongoing changes.
Outcome Mapping deals with the problem of how to attribute impact by increasing the value and attention placed on results achieved “upstream” from impact. It does this by focusing on the changes that are clearly within a program’s sphere of influence. While, at first glance, this appears to suggest concentrating on easier, less important, short-term achievements, in fact it does the opposite. It focuses attention on incremental, often subtle changes, without which the large-scale, more prominent achievements in human well-being cannot be attained or sustained.
Outcome Mapping focuses planning, monitoring, and evaluation on targeted behaviours, actions, and relationships within a program’s sphere of influence, as well as on learning how to increase a program’s effectiveness in relation to its ultimate goals. A corresponding change in reporting requirements would be for donors to make their recipients accountable for demonstrating that they are progressing towards impact and improving in their effectiveness — but not accountable for the impact itself. In this shift to accountability for learning, the connection between the results of a program’s activities and the desired impact becomes rational, not empirical. The intended “impact” of the program is its guiding light and directional beacon, a test of its relevance — it is not the yardstick against which performance is measured. Thus the threat of failing to discover “hidden attribution” is eliminated when feedback on performance concentrates on improving rather than on proving, on understanding rather than on reporting, and on creating knowledge rather than on taking credit.
Outcome Mapping is presented in this manual as an integrated planning, monitoring, and evaluation approach that is best used at the beginning of a program once the main focus of the program has been decided. Outcome Mapping is particularly effective when used from the planning stage, as it helps a program to focus on supporting specific changes in its
partners. With some adaptations, its various elements and tools can be used separately or in conjunction with other processes (for example, a SWOT, a situational analysis, or an LFA). This manual does not provide the reader with instructions on how to adapt Outcome Mapping, but instead assumes it is being used in its totality. Outcome Mapping helps the program to clarify what it wants to accomplish, with whom, and how. It then offers a method for monitoring changes in the boundary partners and in the program as an organization, and encourages the program to look regularly at how it can improve its performance. It can also be used as an end-of-program assessment tool when the purpose of the evaluation is to study the program as a whole.
In Outcome Mapping, planning, monitoring, and evaluation are not discrete events but are designed to be cyclical, with one feeding into the other. It is impossible to plan for all eventualities, therefore a successful program is one that assesses and adapts to changing situations in an intelligent way, based on thoughtful reflection. Planning is done based on the best knowledge available, and the program uses monitoring and evaluation as reflective tools to assess change and choose appropriate actions.
Outcome Mapping can help a program tell its performance story by providing a framework for it to design and articulate its logic; record internal and external monitoring data; indicate cases of positive performance and areas for improvement; evaluate intended and unexpected results; gather data on its contribution to changes in its partners; and establish priorities and an evaluation plan. The framework helps the program to ask a series of questions, outlined below.
Designing and articulating the program’s logic:
Recording internal and external monitoring data:
Indicating cases of positive performance and areas for improvement:
Evaluating intended and unexpected results:
Gathering data on the contribution that a program made to bringing about changes in its partners:
Establishing evaluation priorities and an evaluation plan:
Outcome Mapping can help a program to be more strategic about the actors it targets, the changes it expects to see, and the means it employs. It is designed to make the program more effective in terms of the results to which it contributes.
Outcome Mapping provides a way to model what a program intends to do; however, it differs from most traditional logic models in a number of ways. Most significantly, Outcome Mapping recognizes that different boundary partners operate within different logic and responsibility systems. Outcome Mapping is not based on a cause–effect framework; rather, it recognizes that multiple, nonlinear events lead to change. It does not attempt to attribute outcomes to any single intervention or series of interventions. Instead, it looks at the logical links between interventions and behavioural change. By doing this, Outcome Mapping assumes only that a contribution has been made, and never attempts attribution.
In operational terms, this means that instead of attempting to monitor and evaluate all elements of the program with one set of tools, Outcome Mapping defines three distinct but highly interrelated sets of activities and changes, and offers tools to monitor each one. Thus, in addition to monitoring changes in boundary partners, it also monitors the program’s strategies and organizational practices to enhance understanding of how the program has contributed to change. See Figure 3, below. As already discussed, if a program is to make a significant contribution, it needs to grow and develop so that it can constantly be improving its ability to work with its partners.
Outcome Mapping has expanded the “reach” concept into a performance area by moving from a conceptual understanding of intermediate impacts to a focus on the change process in boundary partners (using outcome challenge statements and progress markers).1 Although other logic models may include changes in behaviour as a part of their framework, Outcome Mapping focuses exclusively on changes in the behaviour of those individuals, groups, and organizations with whom a program works directly. This does not mean that other social, environmental, economic, or political changes should not be measured, but that is not the
________
1 The
“reach” concept was originally introduced to IDRC through Steve Montague in the
early 1990s as an output measure of the scope of a project or program. The
concept evolved to be understood as a dimension of impact, because it dealt with
the uptake or adoption of research findings. For further information, see Sander
(1998).
focus of Outcome Mapping. It monitors and evaluates whether a program has contributed to changes in behaviours in a way that would be logically consistent with supporting development changes in the future. By focusing on this level of change, it assesses the work of the program under review, and is not limited to the end state of the context in which the boundary partners operate. Therefore, programs get credit not only for being present when a major development change occurs, but for their ongoing contributions to that change.
Outcome Mapping recognizes challenges beyond the scope of the program, but limits performance assessment to the program’s direct sphere of influence. In this way, it pushes the program toward increased risk-taking, with the view that risk-taking is a necessary condition for encouraging development. Most logical frameworks tend to limit vision and scope by expressly linking a program’s modest activities to a development impact. This significantly reduces the potential for significant contributions to development by forcing the program to link each activity causally to a measurable result.
By linking organizational assessment with the monitoring of changes in boundary partners, Outcome Mapping recognizes that the program will also have to change during the course of an initiative. If the program is unwilling or incapable of recognizing and adapting to the roles of its boundary partners, then it will eventually defeat its very purpose for being. Outcome Mapping encourages a program to think of itself as a dynamic organization whose goals, methods, and relationships with partners need to be reconsidered and adjusted regularly.
Once a program has established its strategic directions or primary program areas, Outcome Mapping helps sort out who is going to be affected, in what ways, and through what activities on the program’s part. It then permits a program to design a monitoring and evaluation system to help document and manage outcomes.
Once programming areas are chosen, Outcome Mapping can be used at two levels. At the project level, it is particularly effective in planning, monitoring, and evaluating large projects. Small projects rarely require
the level of detail that Outcome Mapping provides. It is best used from the start of a project, but can also be used as an assessment tool (for external or self-assessment) either during or at the end of a project. If used after a project is underway, it will usually require a reformulation of vision, mission, boundary partners, and outcome challenge statements, as few project statements can be directly translated into behavioural terms without consultation among the key participants.
Outcome Mapping can also be used at the program level, but it often requires that the program organize its activities in terms of its sub-programming areas. For example, a program on urban agriculture could be divided into waste-water reuse and confined space agriculture areas. At too high a level of generality, it is difficult to identify who will change and how they will change, thereby reducing the likelihood of success. A program must be sufficiently specific for the group to be able to identify key groups who will be influenced. At both levels, Outcome Mapping requires that groups be able to name identifiable partners whose behaviours should be influenced by the activities of the project or program.
Outcome Mapping is not necessarily the appropriate methodology for all programs, as it may require a change in perspective and approach that is not possible. The program needs to be able to focus on outcomes —defined as changes in the behaviours, relationships, activities, or actions of the people, groups, and organizations with whom it works directly. Outcome Mapping also involves a commitment to change on the part of the program, and self-assessment is an integral element of the methodology. In order to assess whether Outcome Mapping is appropriate for a program, the group should consider the following issues.
Outcome Mapping is best used once a program has made a decision about its strategic directions or primary program areas. It does not provide a means to prioritize or select programming niches. Outcome Mapping helps sort out what, in operational terms, the strategic plan means: who is going to be affected, in what ways, and through what program activities. It then permits a program to design a monitoring and evaluation system to help track intended outcomes.
Outcome Mapping can provide information for a particular type of evaluation study — one that looks at the program’s performance or the outcomes achieved by its partners. It is not intended for a technical evaluation to assess the relevance of the programming area or an evaluation of the cost-effectiveness of one approach compared with another.
Outcome Mapping depends largely on self-assessment data generated systematically by the program team and the boundary partners. It requires a commitment to participatory and learning-based approaches to monitoring and evaluation. Although the approach can be modified to include an external assessment, the program should ensure that the data collected meets its formal reporting requirements and that its managers or donors consider it sufficiently credible. This should be done before starting the process to ensure its utility.
Ideally, the team shares an understanding of the purpose of their work, works collaboratively, and values each other’s opinions. If this is not the case and there is disagreement among team members, Outcome Mapping is not the magic bullet that will solve all the problems. It can, however, provide an opportunity for the group to discuss and negotiate viewpoints systematically and move towards consensus.
The program has to be willing to commit the financial, human, and time resources necessary to design and implement a monitoring and evaluation system. A design workshop takes approximately three days. The monitoring system will take one staff member about one day per monitoring session. A few hours will be required from each of the other staff members to contribute data.
2 |
OUTCOME MAPPING: THE |
---|
The tools and methods of Outcome Mapping as presented here are designed for use in a facilitated three-day workshop. This chapter outlines the basic steps and outputs of the workshop, discusses who should participate, and offers suggestions on how to facilitate the workshop and use the tools provided in the manual.
There are three stages and twelve steps to Outcome Mapping. They take the program from reaching consensus about the macro-level changes it would like to support to developing a monitoring framework and an evaluation plan. The 12 steps are shown on the next page. The process is participatory so that, once members of the program articulate their intentions, decide on strategies, develop a monitoring system, and identify evaluation priorities and data collection needs together, they will share ownership and commitment for the entire enterprise and can integrate it into their daily programming activities and management.
The twelve steps are the elements of an Outcome Mapping design workshop. The first stage, Intentional Design, helps a program clarify and establish consensus regarding the macro-level changes it would like to support. Intentional design is carried out only once a program has chosen its strategic directions and wants to chart its goals, partners, activities, and progress towards anticipated results. After clarifying the changes the program intends to help bring about, activities are chosen that maximize the likelihood of success. The Intentional Design stage helps answer four questions: Why? (vision statement); Who? (boundary partners); What? (outcome challenges and progress markers); and How? (mission, strategy maps, organizational practices).
The second stage, Outcome & Performance Monitoring, helps a program clarify its monitoring and evaluation priorities. It provides a framework for ongoing monitoring of the program’s actions in support of its boundary partners’ progress towards the achievement of outcomes, the strategies the program has employed to foster those changes, and its organizational practices. The program uses progress markers — a set of graduated indicators of the behavioural change identified in the intentional design stage — to clarify directions with boundary partners and to monitor outcomes. It uses the strategy maps and organizational practices to create a performance monitoring framework. If completed regularly, this framework gives the program the opportunity and tools both to reflect on and improve its performance, and to collect data on the results of its work with its boundary partners.
Whereas, using the monitoring framework in Stage Two, the program gathers information that is broad in coverage, the evaluations planned in Stage Three assess a strategy, issue, or relationship in greater depth. This stage, Evaluation Planning, helps the program set evaluation priorities so that it can target evaluation resources and activities where they will be
most useful. An evaluation plan outlines the main elements of the evaluations to be conducted and, finally, an evaluation design is presented. It should be noted that Outcome Mapping provides a method to frame, organize, and collect data, but it does not analyze the information. The program will still need to interpret the data in order to make it useful for learning and improvement or to share its experiences or results with others. Although all elements of Outcome Mapping (such as progress markers and strategy maps) can be used independently, this manual is set up as though the program were using the full approach from the outset, as established through a workshop process.
The outputs of an Outcome Mapping design workshop include:
The workshop method presented in this manual is most appropriate for a group of 18 to 25 participants. If there are more participants, the facilitator will have to adjust the process to include more small group and plenary
activities. Testing of Outcome Mapping to date has been with people who are educated and are comfortable and confident expressing themselves orally in front of a group. The groups have also been relatively free of conflict, as it has been tested with people who share a common cause (for example, those working in an organizational or program setting). If this is not the case, the facilitator will most likely have to adjust the process to fit the culture, skills, and dynamics of the group.
Outcome Mapping, as a participatory process, can ideally involve the full range of stakeholders, including staff from the program, boundary partners, donors, and ultimate beneficiaries. It is, however, premised on the belief that responsibility for planning, monitoring, and evaluation should rest primarily with program staff. This is because, ultimately, they are responsible for the program, need to learn how to better serve their boundary partners, and must fulfill their reporting requirements. This does not mean that boundary partners cannot or should not be incorporated into the process. However, there are no concrete rules about who should participate in each stage, and this will need to be considered by the program on an iterative basis.
The planning, monitoring, and evaluation of a program benefits from the participation of boundary partners. Participation in the Outcome Mapping design workshop by boundary partners may be an opportunity for frank and open dialogue or negotiation about the purpose and relevance of the program. It may also offer a way in which decisions can be made consultatively about how the program will be implemented and assessed. Nonetheless, genuine participation is not simple in the context of an externally funded development program. Donor agencies wanting to use participatory approaches need to be aware of the power imbalance that necessarily exists between the agency managing the program and any beneficiaries or partner groups who are expected to participate. Participation should be requested in a spirit of equitable collaboration, and the complexity of existing relationships should be considered in each instance. Whether the boundary partner is a recipient of funding or not, the program should also consider what that partner will get out of participating. Is it useful for the partner? The program could ask itself at various stages, “How can our program be open, consultative, and transparent with our boundary partners while being realistic about their desire and ability to participate fully?”
There may be cases where the participation of boundary partners in a program’s Outcome Mapping workshop may not be necessary or useful. Participation of boundary partners needs to be considered on a case-by-case basis. When deciding who should participate in the Outcome Mapping workshop, the program should consider:
A facilitator who is familiar with the Outcome Mapping methodology and comfortable with the concepts of evaluation for learning, participatory research, and organizational change should lead the workshop. This may be either an internal or external facilitator. The role of the facilitator is to engage participants, ensure that all opinions are heard and considered, note areas of consensus and differences, and keep the discussion moving so that the group can complete its task in the time allotted. Therefore, the individual should be skilled in consensus building, be able to determine the power dynamics in a group, and be able to ensure that the diverse opinions of the group are captured.
There are advantages and disadvantages to both internal and external facilitators. An outsider can ask the basic questions that will force the group to reflect on, and be very clear about, the purpose of the program. Using an external facilitator can also allow the entire group to concentrate on the substance of the meeting. Nonetheless, external facilitators can be costly and, unless the outsider understands the substance of the program-
ming area, the group may have to provide detailed background information and explanation to enable effective facilitation. On the other hand, having an internal facilitator means that one member of the group will not be able to contribute fully to the substantive discussions. A good option is to team an internal team member with an external facilitator. This can be especially useful if the program plans to facilitate Outcome Mapping sessions itself at a later date.
This manual presents the elements of the Outcome Mapping methodology as they proceed from design to monitoring to evaluation. In Sections Three, Four, and Five, each step of Outcome Mapping is discussed, a group process is outlined, and an example of the finished “products” is provided.
This manual offers suggestions and tips for running the workshop, but these are not intended to be a rigid set of activities that must be followed lockstep. There are many different ways to facilitate groups, and the processes provided in this manual are the ones that, based on IDRC’s experiences, have worked well in development research organizations. The facilitator should be flexible and employ the methods that work best in their specific context.
If available, an LCD projector can be used throughout the twelve steps of Outcome Mapping to project the group’s thoughts and ideas onto a screen as the facilitator enters them into the computer. This immediate feedback promotes a common language among the participants and allows them to easily review and revise their vision, mission, list of boundary partners, outcome challenge statements, progress markers, strategy maps, organizational practices, monitoring journals, and evaluation plan. It also makes the workshop documentation available immediately to the participants. It is helpful if the information is printed and distributed to participants at each step so that they can consult it throughout the process. This approach can also slow down the discussion slightly and give participants an opportunity to reflect while the facilita-
tor is typing. The disadvantage of this facilitation method is that if the group is not used to the technology, participants can be distracted by it. If this is the case, flip charts are a better alternative — although the facilitator should ensure that the material generated at each step remains displayed in the room so that it can be consulted throughout the process.
Simultaneously facilitating and typing information into a laptop is difficult when working with a large group. If the facilitator cannot see the entire group while seated at the computer, then (s)he will have trouble watching and responding to the group. The facilitator should have experience and be very comfortable with using the laptop and LCD projector, otherwise the session can be held up by technical problems, which is very frustrating.
Although the setup of the room is dependent on its shape and the number of participants, the participants should all be able to view one another and speak easily. There should be a large empty wall where information can be posted either on small cards or on flip chart paper. Generally, a U-shaped arrangement works well. The facilitator should have the room prepared before the group arrives.
The items required to facilitate an Outcome Mapping design workshop using the process proposed in this manual include
Outcome Mapping requires a change in thinking about what constitutes results. Therefore it is advisable to have an introductory presentation on the theory of the methodology prior to the actual design workshop. Some programs will want more background on the methodology than others. Nonetheless, explaining it prior to the workshop can make the steps easier to move through. As the facilitator plans the agenda and timing of the workshop, (s)he should assemble and adapt the materials that seem most appropriate to the needs of the group.
The following two activities may be used before the workshop or as it begins, depending on the needs of participants and the stage of their program:
Purpose: When beginning a program planning process, it can be useful for a group to review the program’s history, its achievements, and the events and issues that have influenced its development to date. If the program has been active and is entering a new phase, this "warm-up" exercise can help contextualize the events and influencing factors for both new and seasoned staff. It can also give a group the opportunity to develop a shared understanding of the past and a common language with which to discuss the future.
Process |
Approximate Time |
3 hours |
---|
ICA (Institute of Cultural Affairs) Canada (n.d.) has developed a participatory approach for a group to conduct a quick historical scan. As many members of the program team as possible should participate in this brainstorming activity, outlined below.
Depending on the length of time the group has to cover, the time period can be divided various ways (monthly, quarterly, semiannually, annually, biannually, and so on).
Purpose: Before beginning the Outcome Mapping steps, engaging the group in a short discussion on evaluation can encourage them to create a common language for the duration of the workshop and ensure that everyone understands the evaluation approach being proposed.
Process |
Approximate Time |
30 min. |
---|
Everyone has preconceived ideas about evaluation — some positive, some negative. Although it is possible that some participants think of evaluation as useful, many more probably have negative perceptions, are fearful of the process, or view evaluation as having limited utility. The facilitator needs to be aware of these opinions and build on, or respond to, them throughout the workshop. Because Outcome Mapping integrates monitoring and evaluation issues from a program’s earliest planning stage, the staff will not be passive recipients who are fed information and reports. Rather, they will be active contributors from the outset. Therefore the facilitator needs to create a participatory environment in which everyone feels comfortable, able, and interested in contributing. This will be accomplished primarily through actions rather than words. That is, the facilitator’s actions will demonstrate that (s)he is an active listener who remains neutral. However, it can be useful to have the group engage in a short discussion on the topic before beginning the Outcome Mapping steps. This will allow the group to create a common language for the workshop. Michael Quinn Patton suggests one technique for organizing this type of discussion (Patton 1997):
At IDRC, evaluation is viewed as an integral part of good project and program management. Corporate and program learning and improvement drives evaluation activities, with collegial participation by stakeholders as a key ingredient. IDRC has chosen to use evaluation first as a corporate learning tool, believing this to be the best way to strengthen its accountability function. The following principles, which guide evaluation at IDRC, are embedded in the process of Outcome Mapping and argue for the relevance of evaluation as an integral component of a program’s learning system:
This page intentionally left blank.
3 |
STAGE 1: |
---|
Intentional design implies that a program frames its activities based on the changes it intends to help bring about and that its actions are purposely chosen so as to maximize the effectiveness of its contributions to development. During the intentional design stage, workshop participants define the development goals of the program. The logic of the program is articulated by following seven steps to outline the vision, mission, boundary partners, outcome challenges, graduated progress markers, strategy maps, and organizational practices.
The vision statement describes why the program is engaged in development and provides an inspirational focus. Workshop participants are encouraged to be truly visionary in Step 1, establishing a vivid beacon to motivate staff and highlight the ultimate purpose of their day-to-day work. The boundary partners (those with whom the program will work directly) are identified so that they can contribute to the vision. The outcome challenge statements and progress markers identify the results that the program would like to see its boundary partners achieve. The mission, strategy maps, and organizational practices describe how the program’s contributions to the vision will be framed, by focusing on what it will do. By working its way through the seven steps, the program articulates how it will contribute to change within the complex system in which it operates.
Defining the elements of the intentional design stage is easiest if the members of the program already have a shared understanding of the ultimate purpose of their work. This consensus may be the result of a formal process or of informal discussions. Regardless, it makes the articulation of a vision and mission quite straightforward. If there is not consensus, and the group is using Outcome Mapping to negotiate various perspectives, options, and scenarios, then the intentional design stage may take longer to develop. Nonetheless, appropriate effort should be devoted to this “visioning” stage because, if properly articulated in the beginning, the vision’s elements should remain relevant for the duration of the program (three to five years).
The vision reflects the large-scale development-related changes that the program hopes to encourage. It describes economic, political, social, or environmental changes that the program hopes to help bring about, as well as broad behavioural changes in key boundary partners. The vision is related to the program’s objectives but goes deeper, is broader in scope, and is longer-term. The ultimate achievement of the vision lies beyond the program’s capability; however, its activities should contribute to and facilitate that end. It is the program’s contribution toward the vision (through its boundary partners) that will be measured in an evaluation — not whether the vision was achieved. The vision represents the ideal that the program wants to support and should be sufficiently broad and
inspirational to remain relevant over time, despite changing circumstances. The facilitator should return to the vision statement over the course of the planning exercise to ensure that the program’s activities are consistent with its intent.
Depending on the complexity of its programming, a program can use Outcome Mapping at different levels. For example, an organization or large program might want to articulate a vision and mission for its overall work and for each of its programming areas. A relatively contained program, on the other hand, would probably only need to develop one vision and mission statement.
The vision statement can be recorded on Design Worksheet 1 on page 51.
Process |
Approximate Time |
2 hours |
---|
PURGING THE JARGON
CHART
| |||||
---|---|---|---|---|---|
0 | rural peasants | innovatively | sustainable development | profitable | vitality |
| |||||
1 | rural poor | with | attention to imagination | money-making environment | prosperity |
| |||||
2 | farmers | taking risks | protection of resources | worth doing | health |
| |||||
3 | farm workers | creatively | long-term development | job-generating and income-level | energy producing |
| |||||
4 | villagers | in new ways | holistic development | competitive | optimism |
|
Source: Kibel, B.; Baretto, T.; Dieng, M.; Ndiaye, A.; Carden, F.; Earl, S. July 1999. Draft Outcome Engineering Manual.
The mission statement describes how the program intends to support the vision. It states the areas in which the program will work toward the vision, but does not list all the activities in which the program will engage. Rather, it is an ideal statement about how the program will contribute. It represents what the program wants to grow into as it supports the achievement of the vision.
When drafting its mission statement, the group should consider not only how the program will support the achievement of outcomes by its boundary partners, but also how it will keep itself effective, efficient, and relevant. The manner in which it operates, uses information, and handles change plays a key role in determining how well it can carry out its activities. This is discussed in greater detail in Step 7 (organizational practices).
The mission statement can be recorded on Design Worksheet 1 on page 51.
Process |
Approximate Time |
2 hours |
---|
Boundary partners are those individuals, groups, or organizations with whom the program interacts directly and with whom the program can anticipate opportunities for influence. These actors are called boundary partners because, even though the program will work with them to effect change, it does not control them. The power to influence development rests with them. The program is on the boundary of their world. The program tries to facilitate the process by providing access to new resources, ideas, or opportunities for a certain period of time. A single boundary partner may include multiple individuals, groups, or organizations if a similar change is being sought in all (for example, research centres or women’s NGOs). Figure 4 illustrates the relationship of the program to its boundary partners, and the relationship of those actors to the broader environment in which development occurs.
When listing the program’s boundary partners, the focus should be on the actors with whom it works directly. If the program cannot directly influence an actor, the group needs to determine who it can influence who will, in turn, influence that actor. The actor who can be influenced is then included as a boundary partner instead. In this way, the program maintains a focus on its sphere of influence, but with a broader vision. For example, a rural development NGO may not be able to influence the Minister of Finance directly, but it can influence farmer organizations, which can then lobby the Ministry to effect change. Therefore, the farmer organizations would be included in the NGO’s list of boundary partners, but the Minister of Finance would not.
Generally, a program does not have more than four or five types of boundary partners (although each boundary partner can include multiple individuals, groups, or organizations). When deciding how to group the individuals, groups, and organizations with whom the program works, the crucial feature is that the program truly wants to encourage changes in the behaviour, relationships, activities, or actions of that partner. For example, a rural development NGO may be working with five different farmer organizations in five provinces, but, if the changes that it is trying to help bring about in those organizations are the same, then they are grouped together as a single type of boundary partner.
If there are other actors that the program needs to work with but does not necessarily want to change, they can be listed separately under “strategic partners” so that they can later be considered when developing strategies. Other donors would most likely fit into this category. The program may want, or need, an alliance with them to achieve its objectives, but it is not trying to change their behaviour. Strategic partners are considered in terms of their contribution to the mission.
Figure 5 presents a typology of the actors whom IDRC-supported programs influence to act in ways consistent with sustainable and equitable development. In other words, these are the actors among whom
IDRC tries to encourage outcomes that contribute to its vision. Rather than being linear, the activities and relationships are dialogical, therefore the arrows go in both directions. There are both actions and reactions, because in each situation there may be cooperation, resistance, or negotiation, resulting in mutual influence and learning. The program may try to influence the applied research community through intermediary institutions or it may work directly with the research community. This typology is provided only as an example. Other organizations would have a different configuration of boundary partners.
Intermediary institution: The institution that represents IDRC in the field (for example, an NGO that is implementing a small-grants program).
Applied research community: The research organizations or individual researchers who are implementing the programming activities in the field. For IDRC, this often implies a government, NGO, or university department that is primarily responsible for the research activities.
Development agents or system gatekeepers: The individuals, groups, or organizations who have the power to promote, block, or otherwise influence how the ultimate beneficiaries are reached and affected. For development research, this group often includes the intended users of the research findings.
Beneficiaries: The individuals or groups for whom, or with whom, the program is ultimately working to help improve their economic, social, political, or environmental well-being.
The program’s boundary partners can be listed on Design Worksheet 1 on page 51.
Process |
Approximate Time |
1 hour |
---|
If the group is having trouble identifying people, organizations, or groups, the facilitator can present the IDRC typology (if it is appropriate) to try to stimulate ideas. The facilitator asks: “Among which actors does the program want to encourage change so that they can contribute to the vision?” Other questions that can guide the process are: “Who are the ultimate beneficiaries? Who can the program influence most directly? Who can help or hinder its work?”
If the group has only identified one boundary partner, the facilitator asks whether change in that boundary partner needs to be supported by others. It may be that the program is only working directly with one type of person, group, or organization, but the facilitator should verify that they are approaching the change process holistically and have not left anyone out.
If the group still has a lot of boundary partners, the
facilitator goes over the concept of “boundary partner” with them and asks
them to
consider whether everyone on the list is a
boundary partner or if some actors have a different relationship with the
program. The facilitator asks, “Do any of them belong together because you are
hoping to see a similar change in them or because they play a similar role?”
If so, these are grouped together. The facilitator asks whether the program
can realistically expect to influence that quantity of boundary partners or if
they need to focus.
If the group wants to better articulate the logic of the influence it wants to have through its boundary partners, it can sketch out its boundary partners’ boundary partners. These are the people, organizations, or groups with whom the program’s boundary partners are working and whom they are trying to influence to contribute to social change and poverty alleviation. The facilitator asks, “Who are your boundary partners going to influence to contribute to the achievement of the vision?” These individuals, organizations and groups are listed on a flip chart.
This page intentionally left blank.
Once the boundary partners have been identified, an outcome challenge statement is developed for each of them. Outcomes are the effects of the program “being there,” with a focus on how actors behave as a result of being reached. An outcome challenge describes how the behaviour, relationships, activities, or actions of an individual, group, or institution will change if the program is extremely successful. Outcome challenges are phrased in a way that emphasises behavioural change. They should be idealistic but realistic. This is done for two reasons: it stresses that development is done by, and for, people; and it illustrates that, although the program can influence the achievement of outcomes, it cannot control them. The program contributes to the change, but ultimate responsibility and power for change rests with the boundary partners themselves.
Outcome challenges are phrased so that they capture how the actor would be behaving and relating to others if the program had achieved its full potential as a facilitator of change. The group is encouraged to think about how it can intentionally contribute to the most profound transformation possible. The “challenge” is for the program to help bring about these changes. Because changes in people, groups, and organizations cannot be understood in isolation from one another, the outcome challenge incorporates multiple changes within a single statement rather than breaking them up into separate statements. A set of progress markers will be identified in Step 5 that outlines the progressive levels of change leading to the achievement of the outcome challenge.
In order to keep the discussion as relevant as possible to the vision and mission, outcome challenges tend not to be quantified (for example, in terms of the percentage increase in boundary partners who have changed) or tied to a specific completion date. This also helps, at both the monitoring and evaluation stages, to avoid simply focusing on indicators that are easier to achieve and measure at the expense of more profound qualitative changes. Outcome challenges should not include information about the strategies or activities that the program intends to employ, but rather should describe the qualities of change in the boundary partners’ behaviours. Outcome challenges are usually phrased as “The [program] intends to see [boundary partner] who [description of behaviours in the active present tense].”
The outcome challenges can be listed on Design Worksheet 1 on page 58.
Process |
Approximate Time per boundary partner |
30 min. |
---|
This process works best with relatively small groups (up to 20 people), because it requires that all members of the group go up to the front and write on a flip chart. If there are too many participants, a bottleneck is created at the flip chart and the process can take too much time.
The following process is repeated for each boundary partner.
It is also important to make sure that the program being
designed fits with the implementing organization. The facilitator asks, “Have
the necessary links and connections between the program and your orga-
nization’s mandate been established?” If the goals of the
program and the organization are not compatible, then the group should decide
whether that is acceptable, and whether they need to redefine certain elements
or get buy-in from someone in the organization.
DESIGN WORKSHEET 1: PROGRAM FRAMEWORK | |
---|---|
| |
Vision: | |
| |
Mission: | |
| |
Boundary Partner 1: | Outcome Challenge 1: |
| |
Boundary Partner 2: | Outcome Challenge 2: |
| |
Boundary Partner 3: | Outcome Challenge 3: |
| |
Boundary Partner 4: | Outcome Challenge 4: |
This page intentionally left blank.
Graduated progress markers are identified for each of the outcome challenges that the program is helping to bring about. They show the complexity of the change process associated with each boundary partner and represent the information that the program can gather in order to monitor achievements toward the desired outcome. For a detailed study of progress, the program can conduct an in-depth evaluation, combining the data from the progress markers with information on their context (gathered using other evaluation tools and methods).
A set of progress markers represents a change model for the boundary partner that illuminates the depth and complexity of change being sought. The progress markers should advance in degree from the minimum one would expect to see the boundary partner doing as an early response to the program’s basic activities, to what it would like to see them doing, to what it would love to see them doing if the program were having a profound influence. For instance, progress markers that indicate reactive participation by the boundary partner are relatively easy to achieve and are listed first, under “expect to see”; whereas those that indicate more active learning or engagement are listed second, under “like to see”; and those that are truly transformative are listed third, under “love to see.” In this way, the program will be able to trace what has been accomplished, while being reminded of what still needs to be achieved. The “love to see” progress markers should be set sufficiently high to represent profound change. These often come directly from the outcome challenge statement. The program will know it is not setting sufficiently challenging progress markers if the “love to see” markers are frequently met by its boundary partners early in the life of the program. Progress markers are generally framed as follows: “Who? Is doing what? How?”
Individually, progress markers can be considered as sample indicators of behavioural change, but their real strength rests in their utility as a set. Cumulatively, they illustrate the complexity and logic of the change process. This is something that no single indicator can accomplish.
Progress markers can also be used to describe how the boundary partner is interacting with its own boundary partners, thereby reflecting the indirect influence of the program. A program selects its boundary partners because of the influence it thinks they can have on development, and the progress markers permit this influence to be tracked.
Progress markers are intended as a way for the program to understand and react to the change process in which the boundary partner is engaged. It is only logical that no successful development program strives to change its partners in ways they do not want. The progress markers need to be mutually agreeable, and worthwhile to both the program and the boundary partner. If there are fundamental differences, these need to be resolved. Inherently, the program only facilitates change by helping to remove constraints and providing access to new information and opportunities for the boundary partners. The progress markers can be negotiated with the boundary partners before, during, or after the design workshop to ensure that they are relevant and appropriate.
Some progress markers may be attached to a date (that is, there may be an expectation that they will be reached by a certain point), but this is not obligatory and should not constrain the program’s work. Although time lines are a reality in a development program, they are often externally imposed, may not be intrinsically relevant to the vision, and may even distract from the achievements most relevant to the vision. Reaching the “deadline,” or the “target,” therefore, should not be a primary focus of the program. The purpose of the program is to foster sustained change in boundary partners, and the purpose of the progress markers is to monitor the achievements that contribute to that outcome.
In Stage Two, Outcome and Performance Monitoring, a monitoring system can be devised for these progress markers, using an Outcome Journal to collect data. In order to manage the amount of data that needs to be collected, the program should limit the number of progress markers it sets for each outcome challenge to no more than
Limiting the number of progress markers in this way also makes it possible to quantify results when monitoring the change process, if this is what the program would like to do. This is discussed in Step 9.
The progress markers for each outcome challenge can be listed on Design Worksheet 2 on page 59.
Process |
Approximate Time per boundary partner |
45 min. |
---|
The following process is followed for each boundary partner to be monitored.
The group discusses the changes in behaviour that have been outlined and arranges them in the order of the change process from simplest to deepest. Ideally, there should be no more than 15 progress markers in the list; otherwise, there will be too much data to collect. If the group identifies too many progress markers, the facilitator asks the group to identify those that best indicate the boundary partner’s engagement and show that change is occurring.
DESIGN WORKSHEET 2: PROGRAM MARKERS | |
---|---|
| |
Outcome
Challenge: | |
| |
EXPECT TO SEE_______________________________________________________________ | |
[Boundary Partner] | |
| |
1 | |
| |
2 | |
| |
3 | |
| |
4 | |
| |
LIKE TO SEE_______________________________________________________________ | |
[Boundary Partner] | |
| |
5 | |
| |
6 | |
| |
7 | |
| |
8 | |
| |
9 | |
| |
10 | |
| |
11 | |
| |
12 | |
| |
LOVE TO SEE_______________________________________________________________ | |
[Boundary Partner] | |
| |
13 | |
| |
14 | |
| |
15 |
This page intentionally left blank.
The purpose of the strategy map is to identify the strategies used by the program to contribute to the achievement of an outcome. A strategy map is created for each outcome challenge. Matrix 1, on page 63, provides a method for dividing strategies into six types. Each of the six cells represents a different type of relationship between the program and the boundary partner it is attempting to influence. For most outcome challenges, a mixed set of strategies will be used because this has a greater potential for success, with one or two being dominant and the remainder being supportive.
Plotting the strategies using this 2 x 3 strategy map will
The three strategy types in the first row are labeled “I” because they are aimed directly at specific individuals, groups, or organizations. The three strategy types on the bottom row are labeled “E” because they are aimed at the environment in which the individuals, groups, or organizations operate. “E” strategies are meant to influence the boundary partners indirectly by altering the setting in which they operate.
Both the “I” and “E” strategy types are further divided into three categories: those that are causal (I-1 and E-1); those relying on persuasion (I- 2 and E-2); and those that are based on building supportive networks (I- 3 and E-3). Primary control and responsibility for the consequences only rests with the program when I-1 and E-1 strategies are employed. With I- 2, I-3, E-2, and E-3 strategies, the program tries to facilitate change, but the ultimate responsibility rests with the individual, group, or institution (i.e., with the boundary partners). The further the program moves away from causal activities, the less control it has. This has important implications for assessing contributions toward the achievement of outcomes.
The purpose of the strategy map exercise is not simply to ensure that all the boxes have something in them. Indeed, some boxes may appropriately be left empty, depending on the nature of the program’s work. The appropriateness of strategies largely depends on the type of changes that the program wants to encourage in its boundary partner.
Once the group has completed a strategy map, they can look it over and analyze the total approach. This can help determine whether they are working too much or not enough in certain areas, what relative influence they can expect (given the strategies being employed), and whether they are spreading themselves too thin to have a substantial effect.
In Stage Two, Outcome and Performance Monitoring, a monitoring system can be devised for these strategies, using a strategy journal to collect data.
MATRIX 1: STRATEGY MAP | |||
---|---|---|---|
| |||
STRATEGY | CASUAL | PERSUASIVE | SUPPORTIVE |
| |||
I-1 | I-2 | I-3 | |
| |||
Aimed at a Specific Individual or Group |
|
|
|
| |||
E-1 | E-2 | E-3 | |
| |||
Aimed at Individual’s or Group’s Environment |
|
|
|
The strategy map can be filled out using Design Worksheet 3 on page 67.
Process |
Approximate Time per boundary partner |
1 hour |
---|
This process is followed for each outcome challenge and set of associated progress markers.
Before showing the group the strategy map, the facilitator verifies that they have thought through all six strategy types by asking them the following questions about each:
DESIGN WORKSHEET 3: STRATEGY MAP | |||
---|---|---|---|
| |||
Outcome Challenge: | |||
| |||
STRATEGY | CASUAL | PERSUASIVE | SUPPORTIVE |
| |||
I-1 | I-2 | I-3 | |
| |||
Strategies and Activities Aimed at a Specific Individual or Group | |||
| |||
E-1 | E-2 | E-3 | |
| |||
Strategies and Activities Aimed at Individual’s or Group’s Environment |
This page intentionally left blank.
Now that the nature of the work is more clearly defined, the program needs to spend some time looking at how it is going to operate to fulfill its mission. The purpose of this step is to identify the organizational practices that the program will use to be effective. Taken together, these organizational
practices describe a well-performing organization that has the potential to sustain change interventions over time.2 Some of the practices relate directly to the activities being undertaken, whereas others relate to the “care and feeding” of the program so that it can thrive. Collecting and reviewing data on organizational practices contributes in two ways. First, it provides an opportunity to reflect on what is going on internally in the organization, and how that situation could be improved upon so that the program can operate more effectively. In essence, this is about maintaining the relevance of the program, as well as its capacity to stay on top of developments in its field. Second, unintended results often occur in relation to organizational practices, and the opportunity should be taken to document these. For example, through keeping in touch with clients previously served, one may find out about contributions that the program has made outside its intended scope of influence. Unintended results may also occur when making contact with key informants, obtaining the support of the next highest power, and sharing wisdom with others. Unintended outcomes may be important to the survival of a program and may also contribute to new program directions.
In this step, the program reviews the outcome challenges and identifies the organizational practices that will best help it to contribute to the intended change in its boundary partners. In Stage Two, Outcome and Performance Monitoring, a monitoring process can be devised for these practices, using a performance journal to collect data.
EIGHT ORGANIZATIONAL PRACTICES | |
---|---|
| |
1. Prospecting for new ideas, opportunities, and resources | 5. Checking up on those already served to add value |
2. Seeking feedback from key informants | 6. Sharing your best wisdom with the world |
3. Obtaining the support of your next highest power | 7. Experimenting to remain innovative |
4. Assessing and (re)designing products, services, systems, and procedures | 8. Engaging in organizational reflection |
__________
2 A
well-performing organization is defined as one that is efficient, effective,
relevant, and sustainable (Lusthaus et al. 1999).
The eight practices are not discrete, nor are they operationalized in a linear fashion. Cumulatively, the practices offer a program a method for considering more broadly the ways in which it works. Each organizational practice represents activities that enable a program to remain relevant, innovative, sustainable, and connected to its environment. The practices are defined below.
Members of the program should scan information sources, both locally and globally, for new opportunities that could help them to fulfill their mission. They should be looking for relationships to ongoing activities, as well as entirely new opportunities. These could include new regions in which to work (or that could be linked to existing efforts in other regions) or new groups that could be engaged in the change activity.
In the program’s work with its key boundary partners and other organizations that support the delivery of its program (such as other funders), it will likely maintain frequent and regular communication. However, the program should also keep in mind other actors in the field who are not regularly involved, but who have expertise and a robust knowledge of the field of activity. These informants should be people who are willing to be honest and open, and to share both the good and bad impressions made by the program’s work with its boundary partners. They can provide feedback that recipients may be unable or unwilling to offer.
The program should think about how it is going to present its ideas to higher-level decision bodies in its organization (such as senior management or a board of governors). Good ideas must be presented in an appealing manner that can be understood by decision makers in the context of broader organizational objectives. This involves not only creating opportunities to obtain their support, but also engaging them in activities that strengthen their ownership and commitment to the program. This should be an ongoing process in order to maintain support and interest over time.
Ongoing review of systems and services is critical to ensuring continued relevance. Once a product or service has been put in place, the program needs to monitor it and to make modifications so that it continues to meet (emerging) needs and demands.
Program staff need to obtain feedback (both positive and negative) from boundary partners. They may find that an earlier project involving the same systems, methods, or products is no longer effective or has created other problems. The program needs to work with its boundary partners to address and fix such problems, and to build the learning from these activities into new work.
The program needs to put dissemination procedures in place in order to share its learning. The sharing should be both internal and with external colleagues and networks.
Space needs to be created for a program to explore new directions, tools, and partnerships. This may involve exploring literature not normally associated with the program’s work, discussing ideas with those not normally consulted, or taking on challenges beyond its established boundaries. This could include providing informal and individual time for experimentation, organizing group activities, and paying attention to personal and group physical, spiritual, and mental health.
Program members should be viewing each intervention as an opportunity to learn. Efforts should be made to create time for reflection on program performance and direction and on resource allocation.
The facilitator can capture these actions on Design Worksheet 4 on page 79.
Process |
Approximate Time per boundary partner |
30 min. |
---|
DESIGN WORKSHEET 4: ORGANIZATIONAL PRACTICES | |
---|---|
| |
Key Actions | |
| |
Practice 1. Prospecting for new ideas, opportunities, and resources |
|
| |
Practice 2. Seeking feedback from key informants |
|
| |
Practice 3. Obtaining the support of your next highest power |
|
| |
Practice 4. Assessing and (re)designing products, services, systems, and procedures |
|
| |
Practice 5. Checking up on those already served to add value |
|
| |
Practice 6. Sharing your best wisdom with the world |
|
| |
Practice 7. Experimenting to remain innovative |
|
| |
Practice 8. Engaging in organizational reflection |
4 |
STAGE 2: OUTCOME & |
---|
During the Outcome and Performance Monitoring stage of an Outcome Mapping design workshop, the participants can develop a framework to monitor: the progress of boundary partners towards the achievement of outcomes; the program’s strategies to support outcomes; and the organizational practices used. First, however, they set monitoring priorities. Based on these priorities, data collection sheets are developed to track outcomes, strategies, and/or organizational practices. An outcome journal (described in Step 9) is suggested for collecting data on the boundary partners’ achievement of progress markers. A strategy journal (described in Step 10) is suggested for collecting data on the program’s actions in support of the boundary partner. A performance journal (described in Step 11) is suggested for collecting data on the organizational practices being employed by the program to remain relevant, innovative, sustainable, and connected to its environment.
Outcome Mapping helps the program design these data collection instruments, but their value and usefulness will depend on their integration into the program’s ongoing management and reporting processes and on the commitment of program members to collect data regularly and reflect on their work honestly. Outcome Mapping cannot interpret the data collected for the program. It is up to the program team to determine what the information means in terms of the program’s performance and what needs to be changed, given the team composition; organizational priorities; political, economic, or social contexts; and available resources.
With regular use of these self-assessment tools, the team can reflect upon, and enhance, its performance throughout the implementation of the program. Staff will have the means to ask themselves regularly, “How can we get better at what we are doing so as to increase our contributions to outcomes?” Some of the benefits of regular monitoring meetings will come from the products of those meetings (information gathered that can be used to meet reporting requirements, to feed into evaluations, or to develop promotional materials), while other benefits (such as team building and consensus building) will come from the process of meeting to discuss successes and failures.
Although data is gathered on the program’s actions and on changes in its boundary partners, there is no attempt to imply a causal relationship
between the two. It is assumed that the program is only one of the many influences on boundary partners. The program can make a logical argument regarding its contributions to changes in its boundary partners, but cannot claim sole credit. By combining the data on external outcomes with data on internal performance, the program will be able to tell learning stories illustrating how it has improved its work in order to encourage the deepest levels of transformation possible in its boundary partners. It can also document the extent to which the partners have moved in the desired directions.
Outcome Mapping recognizes that development is a complex process that is made up of parallel processes. It not only builds an understanding of changes in the development setting in which a program is working, but also monitors and assesses the strategies and activities of a program. Further, it monitors and assesses how a program is operating as an organizational unit. Is the program itself learning and adapting to new conditions and changes in its partners? As such, Outcome Mapping assesses a program holistically and is based on the premise that a program needs to know not only about development results, but also about the processes by which they were attained, as well as about its own internal effectiveness. It is through the combination of information and knowledge in these three areas that a program can build a better understanding of what it is achieving and how it can improve its levels of success.
Through monitoring these three elements of a program — (1) the changes in the behaviors, actions, activities, and relationships of the people, groups, and organizations with whom a program works directly; (2) the strategies that a program employs to encourage change in its partners; and (3) the functioning of a program as an organizational unit — Outcome Mapping unites process and outcome evaluation. Therefore, Outcome Mapping is well suited to the complex functioning and long-term aspects of international development programs where outcomes are intermeshed and cannot be easily or usefully segregated from each other. It provides a program with the information it requires to change along with its partners. Outcome Mapping encourages a program to link itself explicitly to processes of transformation. It focuses on how a program
facilitates change rather than on how it causes change, and looks to assess contribution rather than attribution. Looking at how the three elements interrelate and the context in which change occurs is essential to program learning. A program does not operate in isolation of other factors and actors; therefore, it cannot plan and assess as though it does. Systems thinking is not simple, however, as Peter Senge (1990, p. 15) points out:
Seeing interrelationships, not things, and processes, not snapshots. Most of us have been conditioned throughout our lives to focus on things and to see the world in static images. This leads us to linear explanations of systemic phenomenon.
International development programs are particularly prone to excluding themselves from the system in which development change occurs. By separating themselves from development processes (i.e., something "we" help "them" accomplish) and explaining change using linear reasoning, programs lose the opportunity to explore their full potential as change agents. Outcome Mapping encourages a program to think of itself as part of the change process and to embrace complex reasoning and multiple logic systems (Earl and Carden 2001) .
A sustainable monitoring system must be simple to grasp, be light to implement, provide useful information for learning, and help meet reporting requirements. Monitoring should help the program see its work more holistically. There is no denying that sustaining a monitoring system requires effort and tenacity. Before developing a monitoring system, the group should discuss its readiness to implement the Outcome Mapping monitoring system. Some of the challenges associated with implementing the proposed system are generic to all monitoring systems, while others are specific to Outcome Mapping.
First and foremost is the fact that all monitoring takes time, resources, commitment, and effort. There is no getting around this — it is true of any system selected. If a process is well organized, it should not be difficult; however, it does take work to gather and interpret the data. Outcome
Mapping offers a system to gather data and encourage reflection on
The program should consider the time and resources it is prepared to spend on the system before embarking on its design. There should be a clear indication from the group about the amount of time and the level of effort that they will realistically be able to put into monitoring. Any system designed should remain within those confines. Monitoring should not be considered in isolation from the other work that the program is doing. Therefore, other reporting and information tracking and sharing activities already being used should be explored, so as to avoid duplication and to link the various monitoring tools to existing processes and practices.
Outcome Mapping encourages the program team to get together, either face-to-face or in an electronic meeting, to collect data and reflect on the program’s work. This makes it possible to plan for the future. The frequency of these meetings depends on the group’s work schedule, although every two to three months is suggested. If too much time elapses between monitoring sessions, people will forget details and only remember highlights. On the other hand, if meetings are held too frequently, the group may be discouraged by the perceived slowness of change. Each program must establish a monitoring cycle that fits its goals, schedule, and work patterns.
Outcome Mapping was designed primarily as a learning tool for a program that conducts its own monitoring. Although an external monitor can be used to collect data, it is difficult, if not impossible, to contract out learning and appropriate feedback mechanisms. These are not discussed in this manual. One staff member can be assigned to prepare for the meetings and collate the data. Depending on the size and complexity of the program, this should not be an enormous task — it should take approximately one half day per meeting. Having one person
responsible will help ensure that the process is not abandoned and that the information from the meetings is recorded and archived systematically. Setting a preassigned day for the meeting helps make it more automatic and routine as well. Once the framework has been established and someone has been assigned the task of gathering information at the meeting, the data can be gathered quite quickly. The group can use the meeting not only to provide data on the recent past, but also to brainstorm ideas and assign tasks for the future. In this way, monitoring becomes an integral part of the program’s ongoing management process.
If members of the program cannot meet face-to-face, alternative electronic methods can be developed. For example, each person on the team can fill out the forms separately and submit them by a certain date, and one staff member can be responsible for collation. In this case, however, the final product should be circulated for everyone’s information. At some regular interval, however, there should be an opportunity to discuss the information being amassed.
Outcome Mapping is structured around a philosophy of learning and open exchange. The primary purpose of monitoring in Outcome Mapping is to carve out a space where a program can reflect on its experiences in order to feed this knowledge into its planning cycle. A program using Outcome Mapping must be willing to adapt to changing circumstances and learn from its past experiences. If there is not already an environment for sharing experiences and honestly reflecting on performance, Outcome Mapping cannot magically create one. However, it can encourage a more learning-oriented environment by providing a structure for collecting data and for organizing monitoring and evaluation processes.
Outcome Mapping begins from the premise that the easiest and most reliable place to gather data is from those implementing the program. Most of the data collected in the Outcome and Performance Monitoring stage is self-assessment data generated by the program. There are strengths and weaknesses to this approach, and the program should decide whether it is appropriate, given their context and needs, before beginning to design the system. No data is completely objective, whether it is generated internally or externally. Generally, self-assessment data is reliable if it is collected systematically over time. One common concern about self-assessment is that program staff can white-wash failures. Although this is always a possibility, IDRC’s experience is that program
staff are often more critical of themselves than an external evaluator would be. Furthermore, the self-assessment data collected through monitoring can, if required, be fed into an external evaluation mid-way through the program or after it ends. Regular self-assessment can also help a program to participate more fully in a formal evaluation process by helping the program communicate better with external evaluators.
Self-assessment can also help a program manage its activities better. As C.L. Usher (1995, cited by Fetterman 2001) explains:
By developing the capacity to monitor and assess their own performance, program managers and staff can risk the mistakes that often occur with innovation. This is because they can detect problems and make midcourse corrections before the results of errors due to planning or execution become widely apparent and costly. Having the capacity and responsibility to obtain such information about program operations and impact thus empowers managers and staff to explore new ways to enhance their performance.
Besides those listed above, there are other organizational factors, listed in the table below, that affect the readiness of a program to implement its monitoring system. These should be discussed by the group before it gets into the details of designing a data collection system.
FACTORS THAT CAN AFFECT READINESS TO MONITOR | |
---|---|
| |
YOU MUST HAVE:
|
THESE ARE MIXED BLESSINGS:
|
| |
IT IS NICE TO HAVE:
|
THESE CAN BE MAJOR BARRIERS:
|
Source: Adapted from Lusthaus et al. (1999).
This page intentionally left blank.
Although it is tempting to gather information on a whole range of issues, this is not practical. Monitoring and evaluation priorities need to be set so as not to waste human and financial resources. The first task for the group is to define what they will monitor on an ongoing basis and what they will evaluate in depth in the future. Outcome Mapping identifies three types of information that can be monitored on an ongoing basis, either to demonstrate results or to help improve performance: the organizational practices being employed by the program to remain relevant, innovative, and viable; progress toward the outcomes being achieved by boundary partners; and the strategies that the program is employing to encourage change in its boundary partners. In each element of the monitoring system, the type of information to be gathered is deliberately limited in order to keep the effort manageable and sustainable. Three data collection tools that can be used for monitoring (the outcome journal, strategy journal, and performance journal) are defined in Steps 9, 10, and 11.
The program may choose to implement a “light” monitoring system in which they meet regularly, either face-to-face or electronically, to discuss their boundary partners’ achievements and their own strategies or practices. Alternatively, they may choose to implement a “heavier” monitoring system in which they document the data. There are advantages and disadvantages to each approach, and the program needs to determine what works best for its situation.
The information on the monitoring priorities can be recorded on Monitoring Worksheet 1 on page 86.
Process |
Approximate Time per boundary partner |
1 hour |
---|
Although it is often true that what gets monitored is what gets managed, it is not always possible or useful to monitor everything about a program or its partners. Therefore, the number of boundary partners, or types of strategies, or practices to be looked at should be limited. Outcome Mapping can be modified to consider an entire program broadly, or it can focus on a particular partner or strategy. It is up to the program to determine its priorities.
The facilitator asks whether the group wants to monitor strategies, organizational practices, and/or its boundary partners’ achievement of progress markers. (S)he asks the group to be as specific as possible about who will use the information; in what documents, presentations, or meetings it will be presented; and when it would be needed in order to be useful.
MONITORING WORKSHEET 1: MONITORING PLAN | |||||||
---|---|---|---|---|---|---|---|
| |||||||
Monitoring Priority | Who Will Use the Info.? | Purpose of the Info.? | When is the Info. Needed? | Who Will Collect the Info.? | How Often Will It Be Collected? | How Will It Be Collected? | Proposed Monitoring Tool |
| |||||||
Boundary Partner’s Achievement of Outcomes | Outcome Journal | ||||||
| |||||||
Program’s Strategy(ies) | Strategy Journal | ||||||
| |||||||
Program’ Organizational Practices | Peformance Journal |
To track progress over time, an outcome journal is established for each boundary partner that the program has identified as a priority. It includes the graduated progress markers set out in Step 5, a description of the level of change as low, medium, or high, and a place to record who among the boundary partners exhibited the change. Information explaining the reasons for the change, the people and circumstances that contributed to the change, evidence of the change, a record of unanticipated change, and lessons for the program is also recorded in order to keep a running track of the context for future analysis or evaluation.
The progress markers are graduated, and some of them, particularly those listed under “like to see” and “love to see,” describe a complex behaviour that is difficult to categorize as “having occurred” or “not having occurred.” Although many of the progress markers could be the subject of an in-depth evaluation themselves, this is not their intended purpose. The progress markers are not developed as a lockstep description of how the change process must occur; rather, they describe the major milestones that would indicate progress towards the achievement of the outcome challenge. If the program or the boundary partner feel that they are exhibiting the change in behaviour, activities, actions, or relationships described by the progress markers, then that should be noted. At another stage, they may exhibit the achievement of the same progress marker in a different way, and that is perfectly acceptable and should also be recorded. The data collected on the progress markers as a set should reflect the complexity of the change in the people, groups, or organizations, and will invariably need to be contextualized and explained in order to be useful. The purpose of the progress markers in monitoring is to systematize the collection of data on the boundary partners’ accomplishments. These details should be viewed as the richness of the results, not as check-marks to be obtained.
The progress markers articulate the results that the program has helped to achieve. They do so by tracking and discussing trends in the behaviours of the boundary partners. Although there is not a cause-and-effect relationship between the program’s actions and changes in the boundary partner, by compiling information using the outcome journal, the program will better understand how its actions do, or do not, influence its boundary partners. With this information, the program will be able to improve its own performance and encourage its boundary partners to achieve deeper levels of transformation. The program will also be creating records of observed changes. These records can periodically be synthesized to tell the story of influence and change relative to areas of interest or achievement.
Progress markers should not be viewed as unchangeable during monitoring. If the program does not see changes in its boundary partners over a period of time, then it must ask itself whether the problem is with the progress markers or the strategies being employed to encourage change. Are the progress markers still appropriate (in other words, has the context in which the boundary partner operates changed to the point where the progress markers no longer reflect a desirable or required change)? Does the program need to be doing something else to encourage that change? Has the boundary partner changed direction? If the progress markers are no longer appropriate indicators of change, then they should be revised to reflect the new conditions. In this way, the program will be gathering information on the changing context, and will have information about unexpected results in the boundary partner.
The outcome journals can be built using Monitoring Worksheet 2 on page 94.
Process |
Approximate Time per boundary partner |
30 min. |
---|
This process is followed for each boundary partner whose achievements will be monitored over time.
If there is a finite number of individuals, groups, or
organizations with which the program can work (for example, if 25
organizations comprise one type of boundary partner) then a rating scheme
based on percentages is appropriate (for example, where “high” is 80–100%,
“medium” is 50–79%, and “low” is 0–49%). If the number of individuals, groups,
or organizations is not finite, then the program can set an optimum number
that it hopes to work with (for example, five African research institutions
over three
years). If the program is concerned with the
depth of the boundary partner’s response, then the values should reflect the
quality or consistency of the change (for example, where “high” means that
specific actions have been taken in favour of the outcome, “medium” means that
the partner is passive or that there is nothing negative to report, and “low”
means that actions have been against the behaviour or the goal has been
ignored). In each case, the program needs to be able to identify a measurement
with which it feels comfortable and which is defensible to anyone looking at
the monitoring data.
If regular face-to-face or electronic meetings are selected, the facilitator asks: “Who will be responsible for collecting and collating the data? How often will monitoring meetings be held? How can the boundary partners participate in the monitoring process? In what format does the data need to be kept in order to be easily attainable and useful? How can information be constructively fed back to the team and the boundary partners?” The facilitator captures the decisions on a flip chart.
If interviews or focus groups with boundary partners are selected, then the group will need to develop the questionnaires to be used. The group must also answer the following questions: Who will be responsible for organizing, conducting, and collating the data from the focus groups or interviews? How often will the focus groups or interviews be conducted? In what format does the data need to be kept in order to be easily attainable and useful? How can information be fed back to the team and the boundary partners constructively? The facilitator captures the decisions on a flip chart.
These questions are listed on Monitoring Worksheet 5 on page 110.
MONITORING WORKSHEET 2: OUTCOME JOURNAL | ||
| ||
Work Dating from/to: | ||
| ||
Contributors to Monitoring Update: | ||
| ||
Outcome Challenge | ||
| ||
Low
= Medium= High= |
||
| ||
EXPECT TO SEE | WHO? | |
| ||
LMH | ||
| ||
OOO | 1 | |
| ||
OOO | 2 | |
| ||
OOO | 3 | |
| ||
OOO | 4 | |
| ||
LIKE TO SEE | ||
| ||
OOO | 5 | |
| ||
OOO | 6 | |
| ||
OOO | 7 | |
| ||
OOO | 8 | |
| ||
OOO | 9 | |
| ||
OOO | 10 | |
| ||
OO | 11 | |
| ||
OOO | 12 | |
| ||
LOVE TO SEE | ||
| ||
OOO | 13 | |
| ||
OOO | 14 | |
| ||
OOO | 15 |
MONITORING WORKSHEET 2: OUTCOME JOURNAL |
---|
|
Description of Change: |
|
Contributing Factors & Actors: |
|
Sources of Evidence: |
|
Unanticipated Change:
(include description, contributing factors, sources of evidence) |
|
Lessons / Required Program Changes / Reactions: |
This page intentionally left blank.
Outcome Mapping is based on the premise that the program has to be prepared to change along with its boundary partners. It will need to get better and better at its job in order to respond to its boundary partners’ changing needs. In order to provide the program with a systematic way to monitor its actions in support of its boundary partners (so that it can think strategically about its contributions and modify its actions as required), Outcome Mapping provides a monitoring tool called the strategy journal.
The strategy journal records data on the strategies being employed to encourage change in the boundary partners. It is filled out during the program’s regular monitoring meetings. Although it can be customized to include specific elements that the program wants to monitor, the generic format includes the resources allocated (inputs), the activities undertaken, a judgement on their effectiveness, the outputs, and any required follow-up.
If gathered regularly and systematically, this information will enable the program to gauge whether it is making optimum contributions to the achievement of outcomes and modify its actions accordingly. To make optimum use of the strategy journal as a learning and management tool, the program, when completing the journal, should not just ask, “How well have we done?” It should also ask, “How can we improve over the next few months?” This dual function is intended to help a program build its own capacity to be effective and relevant.
Together with the information collected in the outcome journal in Step 9 and the performance journal in Step 11, the program will have a systematized set of data about its operations and the results being achieved by its boundary partners. Ideally, the program should be able to make a logical connection between its strategies and its boundary partners’ achievement of outcomes — but, again, the relationship is not causal. Analyzing and interpreting the internal and external monitoring data will require the program to reflect on the environment in which it and its partners are operating and contextualize its achievements and failures.
The strategy journal can be built using Monitoring Worksheet 3 on page 101.
Process |
Approximate Time |
1 hour |
---|
Even though there is no information to put into the strategy journal during the design workshop, the facilitator helps the group customize it as required and devise a monitoring process. This will familiarize them with the strategy journal so that they will feel comfortable using it during their monitoring meetings. Some programs may prefer simply to use the strategy journal as a guide for a conversation, while others may want to gather detailed information and evidence. There are no guidelines on which method is better, because it is completely dependent on the needs of the program (in terms of its objectives for the uses of the data). The facilitator asks the group, “How much and what kind of information do you need to collect about your strategies in order to meet your learning needs and reporting requirements?”
The following process is followed for each strategy that the program wants to monitor.
These questions are listed on Monitoring Worksheet 5 on page 110.
If regular face-to-face or electronic meetings are selected, the facilitator asks: “Who will be responsible for collecting and collating the data? How often will monitoring meetings be held? How can the boundary partners participate in the monitoring process? In what format does the data need to be kept in order to be easily attainable and useful? How can information be fed back to the team and the boundary partners constructively?” The facilitator captures the answers on a flip chart.
MONITORING WORKSHEET 3: STRATEGY JOURNAL | |
---|---|
| |
Work Dating from/to: | |
| |
Contributors to Monitoring Update: | |
| |
Strategy to be Monitored: | Strategy Type: |
| |
Description of Activities
(What did you do? With whom? When?) |
|
| |
Effectiveness (How did it influence change in the boundary partner(s)) |
|
| |
Outputs | |
| |
Required Program Follow-up or Changes | |
| |
Lessons | |
| |
Date of Next Monitoring Meeting: |
This page intentionally left blank.
Outcome Mapping is based on the premise that the program has to be prepared to change along with its boundary partners. It will need to get better and better at its job in order to respond to its boundary partners’ changing needs. In order to provide the program with a systematic way to monitor its actions (so that it can think strategically about its contributions and modify its actions as required), Outcome Mapping provides a monitoring tool called the performance journal.
The performance journal records data on how the program is operating as an organization to fulfill its mission. A single performance journal is created for the program and filled out during the regular monitoring meetings. It includes information on the organizational practices being employed by the program to remain relevant, sustainable, and connected to its environment. Data on these organizational practices can be gathered through quantitative indicators, qualitative examples, or a combination of the two. This learning can then be fed into future work plans.
If gathered regularly and systematically, this information will enable the program to gauge whether it is making optimum contributions to the achievement of outcomes and to modify its actions accordingly. To make maximum use of the performance journal as a learning and management tool, the program, when completing this journal, should not just ask, “How well have we done?” It should also ask, “How can we improve?” This dual function is intended to help a program build its own capacity to be effective.
Together with the information collected in the outcome journal in Step 9 and the strategy journal in Step 10, the program will have a systematized set of data — a map — about its operations and the results being achieved by its boundary partners. Ideally, the program should be able to make a logical connection between its strategies and practices and its boundary partners’ achievement of outcomes — but the relationship is not causal. Analyzing and interpreting the internal and external monitoring data will require the program to reflect on the environment in which it and its partners are operating and contextualize its achievements and failures.
The performance journal can be built using Monitoring Worksheet 4 on page 108.
Process |
Approximate Time |
1 hour |
---|
Even though there is no information to put into the performance journal during the design workshop, the facilitator helps the group customize it and devise a monitoring process. This will familiarize them with the performance journal so that they will feel comfortable using it themselves during their monitoring meetings. It is also an opportunity to talk abut the actions that can be taken by the team to improve the collective capacity to contribute to change. Some programs may prefer to use the performance journal simply as a guide for a conversation, while others may want to gather detailed information and evidence. There are no guidelines on which method is better, because it is completely dependent on the needs of the program (in terms of its objectives for the monitoring and for the uses of the data). The facilitator asks the group, “How much and what kind of information do you need to collect about your strategies and organizational practices to meet your learning needs and reporting requirements?”
The example used to represent each practice is subjective, but should reflect the action that group members feel has had the greatest effect. A program does not have to do something in each of the eight practice areas in every period, but regularly reviewing all areas may remind the program of an area it is neglecting. The number of key actions should be limited to two or three, otherwise there will be too much data to collect regularly. The facilitator encourages the group to look at both positive and negative examples during their monitoring meetings.
These questions are listed on Monitoring Worksheet 5 on page 110.
If regular face-to-face or electronic meetings are selected, the facilitator asks: “Who will be responsible for collecting and collating the data? How often will monitoring meetings be held? How can the boundary partners participate in the monitoring process? In what format does the data need to be kept in order to be easily attainable and useful? How can information be fed back to the team and the boundary partners constructively?” The facilitator captures the answers on a flip chart.
MONITORING WORKSHEET 4: PERFORMANCE JOURNAL |
---|
|
Work Dating from/to: |
|
Contributors to Monitoring Update: |
|
Practice 1. Prospecting for New Ideas, Opportunities, and Resources |
|
Example or
Indicators: |
Sources of Evidence: |
Lessons: |
|
Practice 2. Seeking Feedback from Key Informants |
|
Example or
Indicators: |
Sources of Evidence: |
Lessons: |
|
Practice 3. Obtaining the Support of Your Next Highest Power |
|
Example or
Indicators: |
Sources of Evidence: |
Lessons: |
|
Practice 4. Assessing and (Re)designing Products, Services, Systems, and Procedures |
|
Example or
Indicators: |
Sources of Evidence: |
Lessons: |
MONITORING WORKSHEET 4: PERFORMANCE JOURNAL |
---|
|
Practice 5. Checking Up on those Already Served to Add Value |
|
Example or
Indicators: |
Sources of Evidence: |
Lessons: |
|
Practice 6. Sharing Your Best Wisdom With the World |
|
Example or
Indicators: |
Sources of Evidence: |
Lessons: |
|
Practice 7. Experimenting to Remain Innovative |
|
Example or
Indicators: |
Sources of Evidence: |
Lessons: |
|
Practice 8. Engaging in Organizational Reflection |
|
Example or
Indicators: |
Sources of Evidence: |
Lessons: |
|
Date of Next Monitoring Meeting: |
MONITORING WORKSHEET 5: PROGRAM RESPONSE | ||
---|---|---|
| ||
Responsible Person | Timing | |
| ||
What should we keep doing? | ||
| ||
What do we need to change in order to improve? | ||
| ||
What strategies/practices do we need to add? | ||
| ||
What strategies/practices do we need to drop (those that have produced no results, or require too much efforts or too many resources to produce results)? | ||
| ||
Has any issue come up that we need to evaluate in greater depth? What? When? Why? How? |
MONITORING WORKSHEET 6: REVIEWING THE LOGIC OF THE PROGRAM | |
---|---|
| |
It is important to review the logic of the program periodically to ensure that it remains relevant. Based on practical experience, the program looks at whether new boundary partners have been added; whether others have been dropped; and whether the vision, mission, outcome challenges, and progress markers still make sense. The changes in program logic can then be made to the documentation (for example, Design Worksheet 1 may be revised). | |
This can be done as often as the program feels is necessary, and can be as heavy or light as the group wishes. It is advisable, however, to incorporate external opinions into the process (such as the views of the boundary partners, experts in the region or in the programming area, managers, donors, and others). The program should also reflect on its monitoring data. | |
This group activity is not intended to be used in the Outcome Mapping design workshop. This worksheet is intended for the program to use after having collected a substantial amount of data in the journals. The process is as follows: | |
| |
1. Read the Vision Statement | Does this still reflect the program’s dream? |
| |
2. Read the Mission Statement | Is this the greatest contribution our program can make? Have we been doing this? Why? Why not? Should we add anything or take anything away? |
| |
3. Review Boundary Partners | Is this who we are working with directly? Do we need to work with anyone else? |
| |
4. Review Outcomes | Do these accurately describe the ideal way that our boundary partners could act to contribute to the achievement of the vision? |
| |
5. Review Progress Markers | Was the change process we set out accurate and useful? What now needs to be added or taken out? |
| |
6. Review Strategies | What did we plan to do? Have we implemented these activities? Why? Why not? |
| |
7. Review Organizational Practices | Are we doing everything we can to maintain our capacity to support our partners? |
This page intentionally left blank.
5 |
STAGE 3: |
---|
At the Evaluation Planning stage, Outcome Mapping provides a method for the program to identify its evaluation priorities and develop an evaluation plan. The purpose of both monitoring and evaluation is to encourage a program to base its management and programming decisions on systematically collected data rather than on perceptions and assumptions. Using the outcome, strategy, and performance journals for monitoring, the program can gather information that is broad in coverage rather than deep in detail. By conducting an evaluation, the program can choose a strategy, issue, or relationship to study and assess in depth. There will never be sufficient time and resources for a program to evaluate everything, therefore priorities should be set and choices made. Regardless of the evaluation issue selected, however, the program first needs to devise an evaluation plan carefully so that resources are allocated wisely and the evaluation findings are useful. The program needs to be clear about the identification of clients for the in-depth evaluation and ensure their participation in the process. The Evaluation Planning stage offers a process by which a program can do this. The program will not necessary complete the evaluation plan during the Outcome Mapping workshop; it can be developed at any point, whenever the program is preparing to begin an evaluation process.
The evaluation plan provides a short description of the main elements of the evaluation to be conducted by the program. It outlines the evaluation issue, the way findings will be used, the questions, the information sources, the evaluation methods, the evaluation team, the dates for the evaluation, and the approximate cost. The information in the evaluation plan will guide the evaluation design and, if the program has decided to use an external evaluator, it can be used to set the terms of reference for the contract. Discussing the various elements of the evaluation plan will help the program to plan an evaluation that will provide useful findings. Whether the evaluation is a formal requirement or not, the program should ensure that it is relevant to its needs, in order not to waste human and financial resources.
Even if an evaluation is being done primarily to meet an accountability requirement for an external donor or manager, it can also be used to generate new knowledge, support learning, question assumptions, plan and motivate future activities, or build the analytical capacity of those involved. However, the program needs to plan for utilization, because utilization does not necessarily follow naturally from the results. Michael Quinn Patton argues that, in some ways, assessment is the most straightforward element of an evaluation, and that the politics of getting findings used is the greatest challenge (Patton 1997).
The information needs of the primary user of the evaluation findings are paramount. The primary user must attend the session when the group is developing the evaluation plan. Getting the client involved in the evaluation process from the planning phase will focus data collection activities on the critical issues and prevent the waste of human and financial resources. Regular involvement of the client throughout the phases of data collection and analysis will test the validity of the findings and increase the likelihood of their utilization. Utilization is the ultimate purpose of evaluation, therefore this “front-end” work should be given due attention, whether the program is conducting a self-assessment or responding to external pressure.
EXAMPLE EVALUATION PLAN | ||||||
---|---|---|---|---|---|---|
| ||||||
Evaluation Issue: Results Achieved by Recipient Research Institutions | ||||||
| ||||||
Who Will Use the Evaluation? How? When? | Questions | Info.Sources | Evaluation Methods | Who Will Conduct and Manage the Evaluation? | Date (Start & Finish) | Cost |
| ||||||
Program management to:
|
What is the profile of the research institutions engaged? How has their capacity to conduct programs changed? What interactions were most/least useful in increasing capacity? Why? What influence has the institution had on research users and other researchers? |
Progress Markers & Outcome Journal Trip reports Strategy Journal |
Focus groups with research institutions Document review Site visit Key informant interviews with program staff |
Senior Consultant with regional and sectoral expertise Managed internally by: A. Gomez A. Beluda |
Six months (to be completed by May 1, 2000) | 35K |
The program will not necessarily complete the evaluation plan during the Outcome Mapping workshop. It can be developed at the point where the program is preparing to begin an evaluation process.
All the information for the evaluation plan can be captured on Evaluation Worksheet 1 on page 124.
Process |
Approximate Time per boundary partner |
2 hours |
---|
The facilitator asks the group, “What do managers and funders want or need to know about the program? What do we need to report on? What do you currently not know that if you did know would help you be more effective? What information could make a difference to what you do? What areas or issues does the program need to know more about in order to improve? What are your knowledge gaps? What are your partners’ knowledge gaps? Is there an issue that you have been monitoring that should be studied in greater depth? What can be done to help the program’s partners fulfill their learning and accountability requirements?”
The group should also consider what it wants to get out of both the evaluation process and the evaluation product. For example, even if the primary purpose of the evaluation findings is to fulfill reporting requirements, the program can plan the evaluation process so that its staff or partners are regularly involved in order to build their evaluation capacity. It is often the learning that takes place in the course of the evaluation process which is most used, rather than the report prepared at the end. This makes the participation of key actors even more important.
Certain factors are recognized to enhance the likelihood that evaluation findings will be utilized. These include both organizational factors and factors related to the evaluation, as listed below.
ORGANIZATIONAL FACTORS | FACTORS RELATED TO THE EVALUATION |
---|---|
| |
|
|
Source: El-Husseiny and Earl (1996).
Further elaboration on utilization strategies is provided in Michael Quinn Patton’s “Utilization Focussed Flowchart” (Patton 1997, pp. 378–379). The group can also consider dissemination strategies for the intended users of the evaluation findings. The facilitator asks, “How can you most effectively communicate the evaluation findings to the intended users? What format, style, or product would they find most credible as a basis for decisions?”
If multiple suggestions are offered and there is disagreement about which evaluation questions to select, the facilitator provides the group with mock data or answers to each question. The facilitator asks the group, “If you asked question X and received this information back, would it inform your program and help you to improve? What would it change in your program? What would you have to do in response?” (Patton 1997). If, after going through this process, there is still disagreement about which questions to focus on, the facilitator asks if the questions can be sequenced (with some questions answered in this evaluation, and some in another). If at all possible, however, the facilitator should try to avoid compromising the questions to the point where they have all been watered down and nobody’s question is answered.
The design and monitoring stages of Outcome Mapping include elements and tools that can be used in an ex-post evaluation to study a strategy, an outcome, the results achieved by a particular boundary partner, or an internal performance issue in greater depth. Much of the data will already have been collected if the program has been using outcome journals, strategy journals, or performance journals for monitoring. If not, the various tools can be adapted for use in a number of ways:
__________
3 For an
example of an ex-post evaluation using Outcome Mapping, see Armstrong et al.
(2000).
4 For an example of a self-assessment using Outcome Mapping, see NEPED Project (1999).
When using the strategy maps for an ex-post evaluation, rather than getting the group to identify the entire strategy map, the facilitator can instead query participants only about the dominant strategy employed by the program to promote the achievement of each outcome. The disadvantage of limiting the discussion to the dominant strategy, however, is that the group will not be able to study the full range of their activities in order to identify gaps and better articulate what it is they do.
EVALUATION WORKSHEET 1: EVALUATION PLAN | ||||||
---|---|---|---|---|---|---|
| ||||||
Evaluation Issue: | ||||||
| ||||||
Who Will Use the Evaluation? How? When? | Questions | Info.Sources | Evaluation Methods | Who Will Conduct and Manage the Evaluation? | Date (Start & Finish) | Cost |
| ||||||
|
A |
APPENDIX: |
---|
This page intentionally left blank.
B |
APPENDIX: |
---|
METHOD | USE WHEN . . . |
---|---|
| |
QUESTIONNAIRE SURVEY Involves a printed or electronic list of questions Is distributed to a predetermined group of individuals Individuals complete and return questionnaire |
SURFACE-MAIL OR FAXED SURVEY: The target population is large (more than 200) You require a large amount of categorical data You require quantitative data and statistical analyses You want to examine the responses of designated subgroups (male/female, for example) The target population is geographically disperse You want to clarify your team’s objectives by involving team members in a questionnaire-development exercise You have access to people who can process and analyze this type of data accurately E-MAIL OR WEB PAGE SURVEY: You have the appropriate software and knowledge of this method Your respondents have the technological capabilities to receive, read, and return the questionnaire Time is of the essence |
| |
FACE-TO-FACE INTERVIEW |
|
Involves a printed or electronic list of questions Is distributed to a predetermined group of individuals Individuals complete and return questionnaire |
You need to incorporate the views of key people (key-informant interview) The target population is small (less than 50) Your information needs call for depth rather than breadth You have reason to believe that people will not return a questionnaire |
METHOD | USE WHEN . . . |
---|---|
| |
TELEPHONE INTERVIEWS Like a face-to-face interview, but it is conducted over the telephone Interviewer records responses |
ONE-TO-ONE TELEPHONE INTERVIEWS: The target population is geographically dispersed Telephone interviews are feasible (cost, trust of respondent . . .) |
TELECONFERENCE INTERVIEWS: The target population is geographically dispersed Equipment is in place | |
| |
GROUP TECHNIQUE (INTERVIEW, FACILITATED WORKSHOP, FOCUS GROUP) |
|
Involves group discussion of predetermined issues or topic Group members share certain common characteristics Facilitator or moderator leads the group Assistant moderator usually records responses Can be conducted in person or through teleconferencing if available |
You need rich description to understand client needs Group synergy is necessary to uncover underlying feelings You have access to a skilled facilitator and data has been recorded You want to learn what the stakeholders want through the power of group observation (using a one-way mirror or video) |
| |
DOCUMENT REVIEW |
|
Involves identification of written or electronic documents containing information or issues to be explored Researchers review documents and identify relevant information Researchers keep track of the information retrieved from documents |
The relevant documents exist and are accessible You need a historical perspective on the issue You are not familiar with the organization’s history You need hard data on selected aspects of the organization |
C |
APPENDIX: |
---|
Boundary Partners | Those individuals, groups, or organizations with whom the program interacts directly and with whom the program can anticipate some opportunities for influence. |
| |
Development Impact | Significant and lasting changes in the well-being of large numbers of intended beneficiaries. |
| |
Evaluation | A process by which a strategy, issue, or relationship is studied and assessed in-depth. |
| |
Evaluation Plan | A short description of the main elements of an evaluation study to be conducted. |
| |
Evaluation Planning Stage | The third stage of Outcome Mapping. It helps the program identify evaluation priorities and develop an evaluation plan. |
| |
Facilitator | The person who leads a group through the Outcome Mapping design workshop. |
| |
Inputs | Resources that are invested into a program in order to encourage results through the relevant activities. |
| |
Intentional Design | The planning stage of Outcome Mapping, where a program reaches consensus on the macro level changes it would like to help bring about and plans strategies to provide appropriate support. |
Mission | An ideal description of how the program intends to support the achievement of the vision. It states with whom the program will work and the areas in which it will work, but does not list all the activities in which the program will engage. |
| |
Monitoring | A process by which data is systematically and regularly collected about a program over time. |
| |
Organizational Practices | Eight separate practices by which a program remains relevant, innovative, sustainable, and connected to its environment. |
| |
Outcome | Changes in the behaviour, relationships, activities, and/or actions of a boundary partner that can be logically linked to a program (although they are not necessarily directly caused by it). |
| |
Outcome Challenge | Description of the ideal changes in the behaviour, relationships, activities, and/or actions of a boundary partner. It is the program’s challenge to help bring about the changes. |
| |
Outcome & Performance Monitoring Stage | The second stage of Outcome Mapping. It provides a framework for the ongoing monitoring of the program’s actions in support of the outcomes and the boundary partners’ progress towards the achievement of outcomes. It is based largely on systematized self-assessment. |
| |
Outcome Journal | A data collection tool for monitoring the progress of a boundary partner in achieving progress markers over time. |
Outputs | Directly achievable and observable, though not necessarily short-term, products of a program. |
| |
Performance Journal | A data collection tool for monitoring how well the program is carrying out its organizational practices. |
| |
Program | A group of related projects and activities with a specified set of resources (human, capital, and financial) directed to the achivement of a set of common goals within a specified period of time. |
| |
Progress Markers | A set of graduated indicators of changed behaviours for a boundary partner that focus on the depth or quality of change. |
| |
Reach | Describes how actors were touched by their interaction with the activities and/or results of the research program. |
| |
Results | The external effects (outputs, outcomes, reach, and impact) of a program. |
| |
Strategy Journal | A data collection tool for monitoring the strategies a program uses to encourage change in the boundary partner. |
| |
Strategy Map | A matrix that categorizes six strategy types (causal, persuasive, and supportive), which a program employs to influence its boundary partner. Strategies are aimed at either the boundary partner or the environment in which the boundary partner operates. |
| |
Vision | A description of the large-scale development changes (economic, political, social, or environmental) to which the program hopes to contribute. |
This page intentionally left blank.
D |
APPENDIX: |
---|
FRENCH | ENGLISH | SPANISH |
| ||
Cartographie des incidences | Outcome Mapping | Mapeo de Alcances |
| ||
Stade 1. Définition des intentions | Stage 1. Intentional Design | Etapa 1. Diseño intencional |
| ||
Étape 1. Vision | Step 1. Vision | Paso 1. Visión |
| ||
Étape 2. Mission | Step 2. Mission | Paso 2. Misión |
| ||
Étape 3. Partenaires limitrophes | Step 3. Boundary Partners | Paso 3. Socios directos |
| ||
Étape 4. Incidences visées | Step 4. Outcome Challenge | Paso 4. Alcances deseados |
| ||
Étape 5. Marqueurs de progrès | Step 5. Progress Markers | Paso 5. Señales de progreso |
| ||
Étape 6. Grilles stratégiques | Step 6. Strategy Map | Paso 6. Mapa de estrategias |
| ||
Étape 7. Pratiques organisationnelles | Step 7. Organizational Practices | Paso 7. Prácticas de la organización |
| ||
Stade 2. Suivi des incidences et du rendement | Stage 2. Outcome and Performance Monitoring | Etapa 2. Seguimiento de alcances y desempeño |
| ||
Étape 8. Priorités du suivi | Step 8. Monitoring Priorities | Paso 8. Prioridades para el seguimiento |
| ||
Étape 9. Journal des incidences | Step 9. Outcome Journals | Paso 9. Diario de alcances |
| ||
Étape 10. Journal des stratégies | Step 10. Strategy Journal | Paso 10. Diario de estrategias |
| ||
Étape 11. Journal du rendement | Step 11. Performance Journal | Paso 11. Diario de desempeño |
| ||
Stade 3. Planification de l`évaluation | Stage 3. Evaluation Planning | Etapa 3. Planificación de la evaluación |
| ||
Étape 12. Plan d’évaluation | Step 12. Evaluation Plan | Paso 12. Plan de evaluación |
This page intentionally left blank.
Armstrong, J.; Carden, F.; Coe, A.; Earl, S. 2000. IMFNS (International Model Forest Network Secretariat) outcomes assessment. Evaluation Unit, International Development Research Centre. Ottawa, ON, Canada. http://www.idrc.ca/evaluation/finalreport.htm
Earl, S.; Carden, F. 2001. Learning from complexity: IDRC’s experience with outcome mapping. Development in Practice (in press).
El-Husseiny, N.; Earl, S. 1996. Enhancing the use of evaluation findings: results of a survey. Evaluation Unit, International Development Research Centre, Ottawa, ON, Canada.
Fetterman, D.M. 2001. Foundations of empowerment evaluation. Sage Publications, Thousands Oaks, CA, USA. pp. 62–63.
ICA (Institute of Cultural Affairs) Canada. n.d. The ToP historical scan. ICA Canada, Toronto, ON, Canada.
Kibel, B.M. 1999. Success stories as hard data: an introduction to results mapping. Kluwer, New York, NY, USA.
Lusthaus, C.; Adrien, M.-H.; Anderson, G.; Carden, F. 1999. Enhancing organizational performance: a toolbox for self-assessment. International Development Research Centre, Ottawa. ON, Canada.
NEPED (Nagaland Environmental Projection and Economic Development) Project. 1999. Nagaland Environmental Projection and Economic Development Project: a self-assessment using outcome mapping. Evaluation Unit, International Development Research Centre, Ottawa, ON, Canada. http://www.idrc.ca/evaluation/nagaland.htm
Patton, M.Q. 1997. Utilization-focused evaluation: the new century text. Sage Publications, Thousand Oaks, CA, USA.
Roche, C. 1999. Impact assessment for development agencies. Oxfam Publishing, Oxford, UK.
Sander, C. 1998. Development research impact: reach. Paper presented at the ICRAF International Workshop on Assessing Impacts in Natural Resource Management Research, 27–29 April 1998, Nairobi, Kenya. Evaluation Unit, International Development Research Centre, Ottawa, ON, Canada. http://www.idrc.ca/evaluation/reach_e.pdf
Senge, P. 1990. The leader’s new work: building learning organizations. Sloan Management Review, 32(1), 7–23.
Smutylo, T. 2001. Crouching impact, hidden attribution: overcoming threats to learning in development programs. Draft Learning Methodology Paper prepared for the Block Island Workshop on Across Portfolio Learning, 22–24 May 2001. Evaluation Unit, International Development Research Centre, Ottawa, ON, Canada. http://www.idrc.ca/evaluation/crouching_impact.pdf
Tallmadge, J. 1997. Meeting the tree of life: a teacher’s path. University of Utah Press, Utah, USA.
Suchman, E.A. 1967. Evaluative research: principles and practice in public service and social actions programs. Russell Sage Foundation, New York, NY, USA.
Usher, C.L. 1995. Improving evaluability through self-evaluation. Evaluation Practice, 16(1), 59–68.
W.K. Kellogg Foundation. 1998. W.K. Kellogg Foundation evaluation handbook. Kellogg Publications, Battle Creak, MI. USA. http://www. wkkf.org/documents/wkkf/evaluationhandbook/evalhandbook.pdf
Sarah Earl holds a masters degree in Russian politics and development from Carleton University and an MA in Russian history from the University of Toronto. She has carried out research and worked in Eastern Europe and the former Soviet Union, and since 1998 has worked for IDRC’s Evaluation Unit.
Fred Carden holds a PhD from the Université de Montréal and a master’s degree in environmental studies from York University. He has taught and carried out research at York University, the Cooperative College of Tanzania, the Bandung Institute of Technology (Indonesia), and the University of Indonesia. Dr Carden is coauthor of Enhancing Organizational Performance (IDRC 1999) and senior program specialist in IDRC’s Evaluation Unit.
Terry Smutylo has been the Director of IDRC’s Evaluation Unit since its creation in 1992. He holds a master’s degree in African studies from the University of Ghana and an undergraduate degree in sociology from the University of Toronto. Mr Smutylo has worked extensively throughout the developing world and has been involved in conducting evaluations, providing evaluation training, and facilitating workshops in Canada, the United States, Europe, Asia, Africa, and Latin America.
The International Development Research Centre (IDRC) is a public corporation created by the Parliament of Canada in 1970 to help developing countries use science and knowledge to find practical, long-term solutions to the social, economic, and environmental problems they face. Support is directed toward developing an indigenous research capacity to sustain policies and technologies developing countries need to build healthier, more equitable, and more prosperous societies.
IDRC Books publishes research results and scholarly studies on global and regional issues related to sustainable and equitable development. As a specialist in development literature, IDRC Books contributes to the body of knowledge on these issues to further the cause of global understanding and equity. IDRC publications are sold through its head office in Ottawa, Canada, as well as by IDRC’s agents and distributors around the world. The full catalogue is available at http://www.idrc.ca/booktique/.
This page intentionally left blank.