Appendix 2 United States Agency for International Development

Appendix 2
United States Agency for International Development
Performance Monitoring and Evaluation TIPS
NUMBER 1
2011 Printing
PERFORMANCE MONITORING & EVALUATION
TIPS
CONDUCTING A PARTICIPATORY EVALUATION
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to
performance monitoring and evaluation. This publication is a supplemental reference to the
Automated Directive Service (ADS) Chapter 203.
WHAT IS DIRECT
OBSERVATION ?
USAID is promoting
participation in all aspects of its development
work.
Participatory evaluation provides for active involvement in the evaluation process of those
with a stake in the program: providers, partners, customers (beneficiaries), and any other
interested parties. Participation typically takes
place throughout all phases of the evaluation:
planning and design; gathering and analyzing the
data; identifying the evaluation findings, conclusions, and recommendations; disseminating results; and preparing an action plan to improve
program performance.
This TIPS outlines how
to conduct a participatory
evaluation.
CHARACTERISTICS OF
PARTICIPATORY
EVALUATION
1
Participatory evaluations typically share several cally, rapid appraisal techniques are used to decharacteristics that set them apart from trad- termine what happened and why.
tional evaluation approaches. These include:
Use of facilitators. Participants actually conParticipant focus and ownership. Partici- duct the evaluation, not outside evaluators as is
patory evaluations are primarily oriented to traditional. However, one or more outside exthe information needs of program stakehold- perts usually serve as facilitator—that is, proers rather than of the donor agency. The donor vide supporting roles as mentor, trainer, group
agency simply helps the participants conduct processor, negotiator, and/or methodologist.
their own evaluations, thus building their ownership and commitment to the results and facilitating their follow-up action.
WHY CONDUCT A
PARTICIPATORY
Scope of participation. The range of particiEVALUATION?
pants included and the roles they play may vary.
For example, some evaluations may target only
program providers or beneficiaries, while othExperience has shown that participatory evaluers may include the full array of stakeholders.
ations improve program performance. Listening
Participant negotiations. Participating to and learning from program beneficiaries, field
groups meet to communicate and negotiate to staff, and other stakeholders who know why a
reach a consensus on evaluation findings, solve program is or is not working is critical to makproblems, and make plans to improve perfor- ing improvements. Also, the more these insiders are involved in identifying evaluation quesmance.
tions and in gathering and analyzing data, the
Diversity of views.Views of all participants are more likely they are to use the information to
sought and recognized. More powerful stake- improve performance. Participatory evaluation
holders allow participation of the less powerful. empowers program providers and beneficiaries
to act on the knowledge gained.
Learning process. The process is a learning experience for participants. Emphasis is on Advantages to participatory evaluations are
identifying lessons learned that will help partici- that they:
pants improve program implementation, as well
as on assessing whether targets were achieved. • Examine relevant issues by involving key
players in evaluation design
Flexible design. While some preliminary
planning for the evaluation may be necessary,
design issues are decided (as much as possible)
in the participatory process. Generally, evaluation questions and data collection and analysis
methods are determined by the participants,
not by outside evaluators.
• Promote participants’ learning about the
program and its performance and enhance
their understanding of other stakeholders’
points of view
• Improve participants’ evaluation skills
• Mobilize stakeholders, enhance teamwork,
and build shared commitment to act on evalua-
Empirical orientation. Good participatory
evaluations are based on empirical data. Typi2
ment exists among stakeholders that a collaborative approach is likely to fail.
tion recommendations
• Increase likelihood that evaluation informaStep 2: Decide on the degree of particition will be used to improve performance
pation. What groups will participate and what
roles will they play? Participation may be broad,
But there may be disadvantages. For example,
with a wide array of program staff, beneficiaries,
participatory evaluations may
partners, and others. It may, alternatively, tar• Be viewed as less objective because program get one or two of these groups. For example,
if the aim is to uncover what hinders program
staff, customers, and other stakeholders
implementation, field staff may need to be inwith possible vested interests participate
volved. If the issue is a program’s effect on local communities, beneficiaries may be the most
• Be less useful in addressing highly technical
appropriate participants. If the aim is to know
aspects
if all stakeholders understand a program’s goals
and view progress similarly, broad participation
• Require considerable time and resources to
identify and involve a wide array of stakehold- may be best. Roles may range from serving as
a resource or informant to participating fully in
ers
some or all phases of the evaluation.
• Take participating staff away from ongoing
Step 3: Prepare the evaluation scope of
activities
work. Consider the evaluation approach—the
• Be dominated and misused by some stake- basic methods, schedule, logistics, and funding.
Special attention should go to defining roles of
holders to further their own interests
the outside facilitator and participating stakeholders. As much as possible, decisions such as
STEPS IN CONDUCTING A the evaluation questions to be addressed and
the development of data collection instruments
PARTICIPATORY
and analysis plans should be left to the particiEVALUATION
patory process rather than be predetermined
in the scope of work.
Step 1: Decide if a participatory evaluation approach is appropriate. Participatory
evaluations are especially useful when there are
questions about implementation difficulties or
program effects on beneficiaries, or when information is wanted on stakeholders’ knowledge
of program goals or their views of progress.
Traditional evaluation approaches may be more
suitable when there is a need for independent
outside judgment, when specialized information
is needed that only technical experts can provide, when key stakeholders don’t have time to
participate, or when such serious lack of agree-
Step 4: Conduct the team planning meeting. Typically, the participatory evaluation process begins with a workshop of the facilitator
and participants. The purpose is to build consensus on the aim of the evaluation; refine the
scope of work and clarify roles and responsibilities of the participants and facilitator; review
the schedule, logistical arrangements, and agenda; and train participants in basic data collection and analysis. Assisted by the facilitator, participants identify the evaluation questions they
want answered. The approach taken to identify
questions may be open ended or may stipulate
3
broad areas of inquiry. Participants then select
appropriate methods and develop data-gathering instruments and analysis plans needed to
answer the questions.
and interpreting them help participants build a
common body of knowledge. Once the analysis
is complete, facilitators work with participants
to reach consensus on findings, conclusions, and
recommendations. Facilitators may need to negotiate among stakeholder groups if disagreements emerge. Developing a common understanding of the results, on the basis of empirical
evidence, becomes the cornerstone for group
commitment to a plan of action.
Step 5: Conduct the evaluation. Participatory evaluations seek to maximize stakeholders’ involvement in conducting the evaluation
in order to promote learning. Participants define the questions, consider the data collection
skills, methods, and commitment of time and labor required. Participatory evaluations usually
use rapid appraisal techniques, which are simpler, quicker, and less costly than conventional
sample surveys. They include methods such as
those in the box below. Typically, facilitators are
skilled in these methods, and they help train
and guide other participants in their use.
Step 7: Prepare an action plan. Facilitators
work with participants to prepare an action
plan to improve program performance. The
knowledge shared by participants about a program’s strengths and weaknesses is turned into
action. Empowered by knowledge, participants
become agents of change and apply the lessons
they have learned to improve performance.
Step 6: Analyze the data and build consensus on results. Once the data are gathered, participatory approaches to analyzing
WHAT’S DIFFERENT ABOUT PARTICIPATORY
EVALUATIONS?
Traditional Evaluation
Participatory Evaluation
• participant focus and ownership of
evaluation
• donor focus and ownership of evaluation
• broad range of stakeholders participate
• stakeholders often don’t participate
• focus is on accountability
• focus is on learning
• predetermined design
• flexible design
• formal methods
• rapid appraisal methods
• outsiders are evaluators
• outsiders are facilitators
4
may be selected through probability or
nonprobability sampling techniques, or
through “convenience” sampling (interviewing stakeholders at locations where
they’re likely to be, such as a clinic for
a survey on health care programs). The
major advantage of minisurveys is that
the datacan be collected and analyzed
within a few days. It is the only rapid appraisal method that generates quantitative data.
Rapid Appraisal Methods
Key informant interviews. This involves interviewing 15 to 35 individuals
selected for their knowledge and experience in a topic of interest. Interviews are
qualitative, in-depth, and semistructured.
They rely on interview guides that list
topics or open-ended questions. The interviewer subtly probes the informant to
elicit information, opinions, and experiences.
Case studies. Case studies record
anedotes that illustrate a program’s
shortcomings or accomplishments. They
tell about incidents or concrete events,
often from one person’s experience.
Focus group interviews. In these,
8 to 12 carefully selected participants
freely discuss issues, ideas, and experiences among themselves. A moderator introduces the subject, keeps the
discussion going, and tries to prevent
domination of the discussion by a few
participants. Focus groups should be
homogeneous, with participants of similar backgrounds as much as possible.
Village imaging. This involves
groups of villagers drawing maps or diagrams to identify and visualize problems
and solutions.
Selected Further Reading
Community group interviews.
Aaker, Jerry and Jennifer Shumaker. 1994.
Looking Back and Looking Forward: A Participatory Approach to Evaluation. Heifer Project
International. P.O. Box 808, Little Rock, AK
72203.
These take place at public meetings
open to all community members. The primary interaction is between the participants and the interviewer, who presides
over the meeting and asks questions,
following a carefully prepared questionnaire.
Aubel, Judi. 1994. Participatory Program Evaluation: A Manual for Involving Program Stakeholders in the Evaluation Process. Catholic
Relief Services. USCC, 1011 First Avenue, New
York, NY 10022.
Direct observation. Using a detailed
observation form, observers record what
they see and hear at a program site. The
information may be about physical surroundings or about ongoing activities,
processes, or discussions.
Minisurveys. These are usually
Freeman, Jim. Participatory Evaluations: Making
Projects Work, 1994. Dialogue on Development Technical Paper No. TP94/2. International
Centre, The University of Calgary.
based on a structured questionnaire with
a limited number of mostly closeended
questions. They are usually administered to 25 to 50 people. Respondents
Feurstein, Marie-Therese. 1991. Partners inEvaluation: Evaluating Development and Community Programmes with Participants. TALC,
5
Box 49, St. Albans, Herts AL1 4AX, United
Kingdom.
Guba, Egon and Yvonna Lincoln. 1989. Fourth
Generation Evaluation. Sage Publications.
Pfohl, Jake. 1986. Participatory Evaluation: A
User’s Guide. PACT Publications. 777 United
Nations Plaza, New York, NY 10017.
Rugh, Jim. 1986. Self-Evaluation: Ideas for
Participatory Evaluation of Rural Community
Development Projects. World Neighbors Publication.
6
1996, Number 2
Performance Monitoring and Evaluation
TIPS
USAID C enter for D evelopment I nformation and E valuation
CONDUCTING KEY INFORMANT INTERVIEWS
What Are Key Informant Interviews?
USAID reengineering
emphasizes listening
to and consulting
with customers, partners and other stakeholders as we undertake development
activities.
Rapid appraisal techniques offer systematic ways of getting
such information
quickly and at low
cost. This Tips advises how to conduct
one such method—
key informant interviews.
They are qualitative, in-depth interviews of 15 to 35 people selected
for their first-hand knowledge about a topic of interst. The interviews are loosely structured, relying on a list of issues to be discussed. Key informant interviews resemble a conversation among
acquaintances, allowing a free flow of ideas and information. Interviewers frame questions spontaneously, probe for information and
takes notes, which are elaborated on later.
When Are Key Informant Interviews Appropriate?
This method is useful in all phases of development activities—
identification, planning, implementation, and evaluation. For example, it can provide information on the setting for a planned activity that might influence project design. Or, it could reveal why
intended beneficiaries aren’t using services offered by a project.
Specifically, it is useful in the following situations:
1. When qualitative, descriptive information is sufficient for decision-making.
2. When there is a need to understand motivation, behavior, and
perspectives of our customers and partners. In-depth interviews
of program planners and managers, service providers, host
government officials, and beneficiaries concerning their attitudes
and behaviors about a USAID activity can help explain its
successes and shortcomings.
3. When a main purpose is to generate recommendations. Key
informants can help formulate recommendations that can improve a program’s performance.
4. When quantitative data collected through other methods need to
be interpreted. Key informant interviews can provide the how
and why of what happened. If, for example, a sample survey
showed farmers were failing to make loan repayments, key
informant interviews could uncover the reasons.
PN-ABS-541
2
5. When preliminary information is needed to
design a comprehensive quantitative study.
Key informant interviews can help frame the
issues before the survey is undertaken.
Advantages and Limitations
Advantages of key informant interviews include:
•
they provide information directly from
knowledgeable people
•
they provide flexibility to explore new ideas
and issues not anticipated during planning
•
they are inexpensive and simple to conduct
Some disadvantages:
•
they are not appropriate if quantitative data are
needed
•
they may be biased if informants are not
carefully selected
•
•
they are susceptible to interviewer biases
Step 3. Select key informants.
The number should not normally exceed 35. It is
preferable to start with fewer (say, 25), since often
more people end up being interviewed than is
initially planned.
Key informants should be selected for their specialized knowledge and unique perspectives on a
topic. Planners should take care to select informants with various points of view.
Selection consists of two tasks: First, identify the
groups and organizations from which key informants should be drawn—for example, host government agencies, project implementing agencies,
contractors, beneficiaries. It is best to include all
major stakeholders so that divergent interests and
perceptions can be captured.
Second, select a few people from each category
after consulting with people familiar with the
groups under consideration. In addition, each
informant may be asked to suggest other people
who may be interviewed.
Step 4. Conduct interviews.
it may be difficult to prove validity of
findings
Once the decision has been made to conduct key
informant interviews, following the step-by-step
advice outlined below will help ensure highquality information.
Establish rapport. Begin with an explanation of
the purpose of the interview, the intended uses of
the information and assurances of confidentiality.
Often informants will want assurances that the
interview has been approved by relevant officials.
Except when interviewing technical experts,
questioners should avoid jargon.
Steps in Conducting the Interviews
Step 1. Formulate study questions.
These relate to specific concerns of the study.
Study questions generally should be limited to five
or fewer.
Step 2. Prepare a short interview guide.
Key informant interviews do not use rigid questionnaires, which inhibit free discussion. However,
interviewers must have an idea of what questions
to ask. The guide should list major topics and
issues to be covered under each study question.
Because the purpose is to explore a few issues in
depth, guides are usually limited to 12 items.
Different guides may be necessary for interviewing different groups of informants.
Sequence questions. Start with factual questions.
Questions requiring opinions and judgments
should follow. In general, begin with the present
and move to questions about the past or future.
Phrase questions carefully to elicit detailed information. Avoid questions that can be answered by a
simple yes or no. For example, questions such as
“Please tell me about the vaccination campaign?”
are better than “Do you know about the vaccination campaign?”
Use probing techniques. Encourage informants to
detail the basis for their conclusions and recommendations. For example, an informant’s comment, such as “The water program has really
changed things around here,” can be probed for
more details, such as “What changes have you
noticed?” “Who seems to have benefitted most?”
“Can you give me some specific examples?”
3
Maintain a neutral attitude. Interviewers should be
sympathetic listeners and avoid giving the impression of having strong views on the subject under
discussion. Neutrality is essential because some
informants, trying to be polite, will say what they
think the interviewer wants to hear.
Minimize translation difficulties. Sometimes it is
necessary to use a translator, which can change the
dynamics and add difficulties. For example,
differences in status between the translator and
informant may inhibit the conversation. Often
information is lost during translation. Difficulties
can be minimized by using translators who are not
known to the informants, briefing translators on
the purposes of the study to reduce misunderstandings, and having translators repeat the informant’s
comments verbatim.
Step 5. Take adequate notes.
Interviewers should take notes and develop them
in detail immediately after each interview to
ensure accuracy. Use a set of common subheadings
for interview texts, selected with an eye to the
major issues being explored. Common subheadings ease data analysis.
Step 6. Analyze interview data.
Interview summary sheets. At the end of each
interview, prepare a 1-2 page interview summary
sheet reducing information into manageable
themes, issues, and recommendations. Each
summary should provide information about the
key informant’s position, reason for inclusion in
the list of informants, main points made, implications of these observations, and any insights or
ideas the interviewer had during the interview.
Descriptive codes. Coding involves a systematic
recording of data. While numeric codes are not
appropriate, descriptive codes can help organize
responses. These codes may cover key themes,
concepts, questions, or ideas, such as
sustainability, impact on income, and participation
of women. A usual practice is to note the codes or
categories on the left-hand margins of the interview text. Then a summary lists the page numbers
where each item (code) appears. For example,
women’s participation might be given the code
“wom–par,” and the summary sheet might indicate
it is discussed on pages 7, 13, 21, 46, and 67 of the
interview text.
Categories and subcategories for coding (based on
key study questions, hypotheses, or conceptual
frameworks) can be developed before interviews
begin, or after the interviews are completed.
Precoding saves time, but the categories may not
be appropriate. Postcoding helps ensure empirically relevant categories, but is time consuming. A
compromise is to begin developing coding categories after 8 to 10 interviews, as it becomes apparent which categories are relevant.
Storage and retrieval. The next step is to develop a
simple storage and retrieval system. Access to a
computer program that sorts text is very helpful.
Relevant parts of interview text can then be organized according to the codes. The same effect can
be accomplished without computers by preparing
folders for each category, cutting relevant comments from the interview and pasting them onto
index cards according to the coding scheme, then
filing them in the appropriate folder. Each index
card should have an identification mark so the
comment can be attributed to its source.
Presentation of data. Visual displays such as
tables, boxes, and figures can condense information, present it in a clear format, and highlight
underlying relationships and trends. This helps
communicate findings to decision-makers more
clearly, quickly, and easily. Three examples below
and on page 4 illustrate how data from key informant interviews might be displayed.
Table 1. Problems Encountered in
Obtaining Credit
Male Farmers
1. Collateral
requirements
2. Burdensome
paperwork
Female Farmers
1. Collateral
requirements
2. Burdensome
paperwork
3. Long delays in 3. Long delays in
getting loans
getting loans
4. Land registered under
male's name
5. Difficulty getting to
bank location
4
Table 2. Impacts on Income of a
Microenterprise Activity
“In a survey I did of the participants last year, I
found that a majority felt their living conditions have improved.”
—university professor
Assess reliability of key informants. Assess informants’ knowledgeability, credibility, impartiality,
willingness to respond, and presence of outsiders
who may have inhibited their responses. Greater
weight can be given to information provided by
more reliable informants.
“I have doubled my crop and profits this year
as a result of the loan I got.”
—participant
Check interviewer or investigator bias. One’s own
biases as an investigator should be examined,
including tendencies to concentrate on information
that confirms preconceived notions and hypotheses, seek consistency too early and overlook
evidence inconsistent with earlier findings, and be
partial to the opinions of elite key informants.
“I believe that women have not benefitted as
much as men because it is more difficult for us
to get loans.”
—female participant
Check for negative evidence. Make a conscious
effort to look for evidence that questions preliminary findings. This brings out issues that may have
been overlooked.
Table 3. Recommendations for
Improving Training
Recommendation
Number of
Informants
Develop need-based training
courses
39
Develop more objective selection
procedures
20
Plan job placement after training
11
Get feedback from informants. Ask the key informants for feedback on major findings. A summary
report of the findings might be shared with them,
along with a request for written comments. Often a
more practical approach is to invite them to a
meeting where key findings are presented and ask
for their feedback.
Selected Further Reading
These tips are drawn from Conducting Key Informant Interviews in Developing Countries, by
Krishna Kumar (AID Program Design and Evaluation Methodology Report No. 13. December 1986.
PN-AAX-226).
Step 7. Check for reliability and validity.
Key informant interviews are susceptible to error,
bias, and misinterpretation, which can lead to
flawed findings and recommendations.
Check representativeness of key informants. Take
a second look at the key informant list to ensure no
significant groups were overlooked.
U.S. Agency for International Development
For further information on this topic, contact Annette
Binnendijk, CDIE Senior Evaluation Advisor, via
phone (703) 875-4235), fax (703) 875-4866), or e-mail.
Copies of TIPS can be ordered from the Development
Information Services Clearinghouse by calling (703)
351-4006 or by faxing (703) 351-4039. Please refer to
the PN number. To order via the Internet, address a
request to [email protected]
Washington, D.C. 20523
2
ND
NUMBER 3
EDITION, 2010
PERFORMANCE MONITORING & EVALUATION
TIPS
PREPARING AN EVALUATION STATEMENT OF WORK
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to
performance management and evaluation. This publication is a supplemental reference to the
Automated Directive System (ADS) Chapter 203.
PARTICIPATION IS KEY
Use a participatory process to ensure
resulting information will be relevant
and useful. Include a range of staff
and partners that have an interest in
the evaluation to:
 Participate in planning meetings
and review the SOW;
 Elicit input on potential evaluation
questions; and
 Prioritize and narrow the list of
questions as a group.
WHAT IS AN
EVALUATION
STATEMENT OF
WORK (SOW)?
The statement of work (SOW) is
viewed as the single most critical
document in the development of
a good evaluation. The SOW
states (1) the purpose of an
evaluation, (2) the questions that
must be answered, (3) the
expected quality of the evaluation
results, (4) the expertise needed
to do the job and (5) the time
frame and budget available to
support the task.
WHY IS THE SOW IMPORTANT?
The SOW is important because it
is a basic road map of all the
elements of a well-crafted
evaluation. It is the substance of
a
contract
with
external
evaluators, as well as the
framework for guiding an internal
evaluation team. It contains the
information that anyone who
implements the evaluation needs
to know about the purpose of the
1
evaluation, the background and
history of the program being
evaluated,
and
the
issues/questions that must be
addressed. Writing a SOW is
about managing the first phase of
the evaluation process. Ideally,
the writer of the SOW will also
exercise management oversight
of the evaluation process.
PREPARATION – KEY
ISSUES
BALANCING FOUR
DIMENSIONS
A well drafted SOW is a critical
first step in ensuring the
credibility and utility of the final
evaluation report.
Four key
dimensions of the SOW are
interrelated and should be
balanced against one another
(see Figure 1):
 The number and complexity of
the evaluation questions that
need to be addressed;
 Adequacy of the time allotted
to obtain the answers;
 Availability of funding (budget)
to support the level of
evaluation design and rigor
required; and
 Availability
of the expertise
needed to complete the job.
The development of the SOW is
an iterative process in which the
writer has to revisit, and
sometimes adjust, each of these
dimensions.
Finding
the
appropriate balance is the main
challenge faced in developing any
SOW.
essential
that
evaluation
planning form an integral part
of the initial program or project
design. This includes factoring
in baseline data collection,
possible comparison or „control‟
site
selection,
and
the
preliminary design of data
collection
protocols
and
instruments. Decisions about
evaluation design must be
reflected in implementation
planning and in the budget.
 There
will always be unanticipated
problems
and
opportunities
that
emerge
during an evaluation.
It is
helpful to build-in ways to
accommodate
necessary
changes.
 The writer of the SOW is, in
essence, the architect of the
evaluation. It is important to
commit adequate time and
energy to the task.
 Adequate time is required to
ADVANCE PLANNING
It is a truism that good planning
is a necessary – but not the only –
condition for success in any
enterprise. The SOW preparation
process is itself an exercise in
careful and thorough planning.
The writer must consider several
principles when beginning the
process.
gather information and to build
productive relationships with
stakeholders (such as program
sponsors,
participants,
or
partners) as well as the
evaluation team, once selected.
 The sooner that information can
be made available to the
evaluation team, the more
efficient they can be in
providing credible answers to
the
important
questions
outlined in the SOW.
 The quality of the evaluation is
dependent on providing quality
guidance in the SOW.
 As USAID and other donors
place more emphasis on
rigorous impact evaluation, it is
WHO SHOULD BE INVOLVED?
Participation in all or some part of
the evaluation is an important
decision for the development of
the SOW. USAID and evaluation
experts strongly recommend that
evaluations maximize stakeholder
participation, especially in the
initial
planning
process.
Stakeholders may encompass a
wide array of persons and
institutions,
including
policy
makers,
program
managers,
implementing
partners,
host
country
organizations,
and
beneficiaries.
In some cases,
stakeholders
may
also
be
involved
throughout
the
evaluation
and
with
the
dissemination of results. The
benefits
of
stakeholder
participation
include
the
following:
 Learning
across a broader
group of decision-makers, thus
increasing the likelihood that
the evaluation findings will be
used to improve development
effectiveness;
 Acceptance of the purpose and
process of evaluation by those
concerned;
 A more inclusive and better
focused list of questions to be
answered;
 Increased
acceptance
and
ownership of the process,
findings and conclusions; and
 Increased possibility that the
evaluation will be used by
decision makers and other
stakeholders.
USAID operates in an increasingly
complex implementation world
2
FIGURE 2. ELEMENTS OF A
GOOD EVALUATION SOW
1. Describe the activity, program, or
process to be evaluated
2. Provide a brief background on the
development hypothesis and its
implementation
3. State the purpose and use of the
evaluation
4. Clarify the evaluation questions
5. Identify the evaluation method(s)
6. Identify existing performance
information sources, with special
attention to monitoring data
7. Specify the deliverables(s) and the
timeline
8. Identify the composition of the
evaluation team (one team
member should be an evaluation
specialist) and participation of
customers and partners
9. Address schedule and logistics
10. Clarify requirements for reporting
and dissemination
11. Include a budget
with many players, including
other USG agencies such as the
Departments of State, Defense,
Justice and others. If the activity
engages other players, it is
important to include them in the
process.
Within USAID, there are useful
synergies that can emerge when
the SOW development process is
inclusive. For example, a SOW
that focuses on civil society
advocacy might benefit from
input by those who are experts in
rule of law.
Participation by host government
and local organizational leaders
and beneficiaries is less common
among
USAID
supported
evaluations. It requires sensitivity
and
careful
management;
however,
the
benefits
to
development practitioners can be
substantial.
Participation of USAID managers
in evaluations is an increasingly
common practice and produces
many benefits. To ensure against
bias or conflict of interest, the
USAID manager‟s role can be
limited to participating in the fact
finding phase and contributing to
the analysis. However, the final
responsibility
for
analysis,
conclusions
and
recommendations will rest with
the independent members and
team leader.
THE ELEMENTS OF A
GOOD EVALUATION
SOW
1. DESCRIBE THE ACTIVITY,
PROGRAM, OR PROCESS TO BE
EVALUATED
Be as specific and complete as
possible in describing what is to
be evaluated.
The more
information provided at the
outset, the more time the
evaluation team will have to
develop the data needed to
answer the SOW questions.
If the USAID manager does not
have the time and resources to
bring together all the relevant
information needed to inform the
evaluation in advance, the SOW
might require the evaluation
team to submit a document
review as a first deliverable. This
will, of course, add to the amount
of time and budget needed in the
evaluation contract.
3
2. PROVIDE A BRIEF
BACKGROUND
Give a brief description of the
context, history and current status
of the activities or programs,
names of implementing agencies
and organizations involved, and
other information to help the
evaluation
team
understand
background and context.
In
addition, this section should state
the development hypothesis(es)
and clearly describe the program
(or project) theory that underlies
the program‟s design.
USAID
activities,
programs
and
strategies, as well as most
policies, are based on a set of “ifthen” propositions that predict
how a set of interventions will
produce intended results.
A
development
hypothesis
is
generally represented in a results
framework (or sometimes a
logical framework at the project
level) and identifies the causal
relationships
among
various
objectives sought by the program
(see TIPS 13: Building a Results
Framework). That is, if one or
more objectives are achieved,
then the next higher order
objective will be achieved.
Whether
the
development
hypothesis is the correct one, or
whether it remains valid at the
time of the evaluation, is an
important question for most
evaluation SOWs to consider.
3. STATE THE PURPOSE AND
USE OF THE EVALUATION
Why is an evaluation needed?
The clearer the purpose, the more
likely it is that the evaluation will
produce credible and useful
findings,
conclusions
and
recommendations.
In defining
the purpose, several questions
should be considered.
 Who wants the information?
Will higher level decision
makers be part of the intended
audience?
 What do they want to know?
 For
what purpose will
information be used?
the
 When will it be needed?
 How accurate must it be?
ADS 203.3.6.1 identifies a number
of triggers that may inform the
purpose and use of an evaluation,
as follows:
 A key management decision is
required for which there is
inadequate information;
 Performance
information
indicates an unexpected result
(positive or negative) that
should be explained (such as
gender differential results);
 Customer,
partner, or other
informed feedback suggests
that there are implementation
problems, unmet needs, or
unintended consequences or
impacts;
 Issues of impact, sustainability,
cost-effectiveness, or relevance
arise;
 The validity of the development
hypotheses
or
critical
assumptions is questioned, for
example, due to unanticipated
changes in the host country
environment; and
 Periodic portfolio reviews have
identified key questions that
need to be answered or require
consensus.
4. CLARIFY THE EVALUATION
QUESTIONS
The core element of an
evaluation SOW is the list of
questions
posed
for
the
evaluation.
One of the most
common
problems
with
evaluation SOWs is that they
contain a long list of poorly
defined or “difficult to answer”
questions given the time, budget
and resources provided. While a
participatory process ensures
wide ranging input into the initial
list of questions, it is equally
important to reduce this list to a
manageable number of key
questions. Keeping in mind the
relationship between budget,
time, and expertise needed, every
potential question should be
thoughtfully examined by asking
a number of questions.
 Is this question of essential
importance to the purpose and
the users of the evaluation?
 Is this question clear, precise
and „researchable‟?
 What level of reliability and
validity is expected in answering
the question?
 Does determining an answer to
the question require a certain
kind
of
experience
and
expertise?
 Are we prepared to provide the
management
commitment,
time and budget to secure a
credible
answer
to
this
question?
4
If these questions can be
answered yes, then the team
probably has a good list of
questions that will inform the
evaluation team and drive the
evaluation process to a successful
result.
5. IDENTIFY EVALUATION
METHODS
The SOW manager has to decide
whether the evaluation design
and methodology should be
specified in the SOW. 1
This
depends on whether the writer
has expertise, or has internal
access to evaluation research
knowledge and experience. If so,
and the writer is confident of the
„on the ground‟ conditions that
will allow for different evaluation
designs, then it is appropriate to
include specific requirements in
the SOW.
If the USAID SOW manager does
not have the kind of evaluation
experience needed, especially for
more formal and rigorous
evaluations, it is good practice to:
1) require that the team (or
bidders, if it is contracted out)
include a description of (or
approach for developing) the
proposed research design and
methodology, or 2) require a
detailed design and evaluation
plan to be submitted as a first
deliverable. In this way, the SOW
manager benefits from external
evaluation expertise. In either
case,
the
design
and
methodology should not be
finalized until the team has an
opportunity to gather detailed
1
See USAID ADS 203.3.6.4 on
Evaluation Methodologies;
information and discuss
issues with USAID.
final
The selection of the design and
data collection methods must be
a function of the type of
evaluation and the level of
statistical and quantitative data
confidence needed. If the project
is selected for a rigorous impact
evaluation, then the design and
methods used will be more
sophisticated and technically
complex. If external assistance is
necessary, the evaluation SOW
will be issued as part of the initial
RFP/RFA (Request for Proposal or
Request
for
Application)
solicitation process. All methods
and evaluation designs should be
as
rigorous
as
reasonably
possible. In some cases, a rapid
appraisal
is
sufficient
and
appropriate (see TIPS 5: Using
Rapid Appraisal Methods). At the
other extreme, planning for a
sophisticated
and
complex
evaluation
process
requires
greater up-front investment in
baselines, outcome monitoring
processes,
and
carefully
constructed
experimental
or
quasi-experimental designs.
6. IDENTIFY EXISTING
PERFORMANCE INFORMATION
Identify
the
existence
and
availability
of
relevant
performance information sources,
such as performance monitoring
systems
and/or
previous
evaluation reports. Including a
summary of the types of data
available, the timeframe, and an
indication of their quality and
reliability will help the evaluation
team to build on what is already
available.
7. SPECIFY DELIVERABLES
AND TIMELINE
The SOW must specify the
products, the time frame, and the
content of each deliverable that is
required
to
complete
the
evaluation contract. Some SOWs
simply require delivery of a draft
evaluation report by a certain
date. In other cases, a contract
may require several deliverables,
such as a detailed evaluation
design, a work plan, a document
review, and the evaluation report.
The most important deliverable is
the final evaluation report. TIPS
17: Constructing an Evaluation
Report provides a suggested
outline of an evaluation report
that may be adapted and
incorporated directly into this
section.
The evaluation report should
differentiate between findings,
conclusions,
and
recommendations, as outlined in
Figure 3. As evaluators move
beyond
the
facts,
greater
interpretation is required.
By
ensuring that the final report is
organized
in
this
manner,
decision makers can clearly
understand the facts on which the
evaluation is based. In addition,
it
facilitates
greater
understanding of where there
might
be
disagreements
concerning the interpretation of
those facts.
While individuals
may
disagree
on
recommendations, they should
not disagree on the basic facts.
5
Another consideration is whether
a section on “lessons learned”
should be included in the final
report. A good evaluation will
produce knowledge about best
practices, point out what works,
what does not, and contribute to
the more general fund of tested
experience on which other
program
designers
and
implementers can draw.
Because unforeseen obstacles
may emerge, it is helpful to be as
realistic as possible about what
can be accomplished within a
given time frame. Also, include
some wording that allows USAID
and the evaluation team to adjust
schedules in consultation with the
USAID manager should this be
necessary.
8. DISCUSS THE COMPOSITION
OF THE EVALUATION TEAM
USAID evaluation guidance for
team
selection
strongly
recommends that at least one
team member have credentials
and experience in evaluation
design and methods. The team
leader must have strong team
management skills, and sufficient
experience
with
evaluation
standards and practices to ensure
a credible product.
The
appropriate team leader is a
person with whom the SOW
manager can develop a working
partnership as the team moves
through the evaluation research
design and planning process.
He/she must also be a person
who can deal effectively with
senior U.S. and host country
officials and other leaders.
Experience with USAID is often an
important factor, particularly for
management
focused
evaluations, and in formative
evaluations designed to establish
the basis for a future USAID
program or the redesign of an
existing program.
If the
evaluation entails a high level of
complexity, survey research and
other sophisticated methods, it
may be useful to add a data
collection and analysis expert to
the team.
Generally, evaluation skills will be
supplemented with additional
subject matter experts. As the
level of research competence
increases in many countries
where USAID has programs, it
makes good sense to include
local
collaborators,
whether
survey
research
firms
or
independents, to be full members
of the evaluation team.
9. ADDRESS SCHEDULING,
LOGISTICS AND OTHER
SUPPORT
Good scheduling and effective
local support contributes greatly
to the efficiency of the evaluation
team. This section defines the
time frame and the support
structure needed to answer the
evaluation questions at the
required level of validity. For
evaluations involving complex
designs and sophisticated survey
research data collection methods,
the schedule must allow enough
time, for example, to develop
sample frames, prepare and
pretest
survey
instruments,
training interviewers, and analyze
data. New data collection and
analysis
technologies
can
accelerate this process, but need
to be provided for in the budget.
In some cases, an advance trip to
the field by the team leader
and/or methodology expert may
be justified where extensive
pretesting
and
revision
of
instruments is required or when
preparing for an evaluation in
difficult or complex operational
environments.
Adequate
logistical
and
administrative support is also
essential. USAID often works in
countries with poor infrastructure,
frequently in conflict/post-conflict
environments where security is an
issue. If the SOW requires the
team to make site visits to distant
or
difficult
locations,
such
planning must be incorporated
into the SOW.
6
Particularly overseas, teams often
rely on local sources for
administrative support, including
scheduling of appointments,
finding
translators
and
interpreters,
and
arranging
transportation. In many countries
where foreign assistance experts
have been active, local consulting
firms have developed this kind of
expertise. Good interpreters are
in high demand, and are essential
to any evaluation team‟s success,
especially when using qualitative
data collection methods.
10. CLARIFY REQUIREMENTS
FOR REPORTING AND
DISSEMINATION
Most evaluations involve several
phases of work, especially for
more complex designs.
The
SOW can set up the relationship
between the evaluation team, the
USAID manager and other
stakeholders. If a working group
was established to help define
the SOW questions, continue to
use the group as a forum for
interim reports and briefings
provided by the evaluation team.
The SOW should specify the
timing and details for each
briefing session.
Examples of
what might be specified include:
 Due dates for draft and final
reports;
 Dates for oral briefings (such as
a mid-term and final briefing);
 Number of copies needed;
 Language requirements, where
applicable;
 Formats and page limits;
 Requirements for datasets, if
primary
data
collected;
has
been
 A requirement to submit all
evaluations to the Development
Experience Clearing house for
archiving
this
is
the
responsibility of the evaluation
contractor; and
 Other
needs
communicating, marketing
disseminating results that
the
responsibility
of
evaluation team.
for
and
are
the
The SOW should specify when
working drafts are to be
submitted for review, the time
frame allowed for USAID review
and comment, and the time
frame to revise and submit the
final report.
11. INCLUDE A BUDGET
With the budget section, the
SOW comes full circle. As stated,
budget considerations have to be
part of the decision making
process from the beginning.
The budget is a product of the
questions
asked,
human
resources needed, logistical and
administrative support required,
and the time needed to produce
a high quality, rigorous and
useful evaluation report in the
most efficient and timely manner.
It is essential for contractors to
understand the quality, validity
and rigor required so they can
develop a responsive budget that
will meet the standards set forth
in the SOW.
For more information:
TIPS publications are available online at [insert website].
Acknowledgements:
Our thanks to those whose experience and insights helped shape this publication including USAID‟s
Office of Management Policy, Budget and Performance (MPBP). This publication was written by Richard
Blue, Ph.D. of Management Systems International.
Comments regarding this publication can be directed to:
Gerald Britan, Ph.D.
Tel: (202) 712-1158
[email protected]
Contracted under RAN-M-00-04-00049-A-FY0S-84
Integrated Managing for Results II
7
1996, Number 4
Performance Monitoring and Evaluation
TIPS
USAID Center for Development Information and Evaluation
USING DIRECT OBSERVATION TECHNIQUES
What is Direct Observation?
Most evaluation teams conduct some fieldwork, observing what's actually going on at
assistance activity sites. Often, this is done informally, without much thought to the
quality of data collection. Direct observation techniques allow for a more systematic,
structured process, using well-designed observation record forms.
Advantages and Limitations
USAID's
reengineering guidance encourages
the use of rapid, low
cost methods for
collecting information on the performance of our development activities.
The main advantage of direct observation is that an event, institution, facility, or
process can be studied in its natural setting, thereby providing a richer understanding
of the subject.
For example, an evaluation team that visits microenterprises is likely to better
understand their nature, problems, and successes after directly observing their
products, technologies, employees, and processes, than by relying solely on
documents or key informant interviews. Another advantage is that it may reveal
conditions, problems, or patterns many informants may be unaware of or unable to
describe adequately.
On the negative side, direct observation is susceptible to observer bias. The very act
of observation also can affect the behavior being studied.
When Is Direct Observation Useful?
Direct observation,
the subject of this
Tips, is one such
method.
Direct observation may be useful:
When performance monitoring data indicate results are not being
accomplished as planned, and when implementation problems are suspected,
but not understood. Direct observation can help identify whether the process
is poorly implemented or required inputs are absent.
When details of an activity's process need to be assessed, such as whether
tasks are being implementing according to standards required for
effectiveness.
When an inventory of physical facilities and inputs is needed and not
available from existing sources.
PN-ABY-208
2
When interview methods are unlikely to elicit
needed information accurately or reliably, either
because the respondents don't know or may be
reluctant to say.
When preparing direct observation forms, consider the
following:
1. Identify in advance the possible response categories for
each item, so that the observer can answer with a simple
Steps in Using Direct Observation
The quality of direct observation can be improved by
following these steps.
Step 1. Determine the focus
Because of typical time and resource constraints, direct
observation has to be selective, looking at a few activities,
events, or phenomena that are central to the evaluation
questions.
For example, suppose an evaluation team intends to study a
few health clinics providing immunization services for
children. Obviously, the team can assess a variety of
areas—physical facilities and surroundings, immunization
activities of health workers, recordkeeping and managerial
services, and community interactions. The team should
narrow its focus to one or two areas likely to generate the
most useful information and insights.
Next, break down each activity, event, or phenomena into
subcomponents. For example, if the team decides to look at
immunization activities of health workers, prepare a list of
the tasks to observe, such as preparation of vaccine,
consultation with mothers, and vaccine administration.
Each task may be further divided into subtasks; for
example, administering vaccine likely includes preparing
the recommended doses, using the correct administration
technique, using sterile syringes, and protecting vaccine
from heat and light during use.
If the team also wants to assess physical facilities and
surroundings, it will prepare an inventory of items to be
observed.
OBSERVATION OF GROWTH
MONITORING SESSION
Name of the Observer
Date
Time
Place
Was the scale set to 0 at the beginning of the growth
session?
Yes______ No ______
How was age determined?
By asking______
From growth chart_______
Other_______
When the child was weighed, was it stripped to
practical limit?
Yes______ No______
Was the weight read correctly?
Yes______No______
Process by which weight and age transferred to record
Health Worker wrote it_____
Someone else wrote it______
Other______
Did Health Worker interpret results for the mother?
Yes_______No_______
Step 2. Develop direct observation forms
The observation record form should list the items to be
observed and provide spaces to record observations. These
forms are similar to survey questionnaires, but
investigators record their own observations, not
respondents' answers.
Observation record forms help standardize the observation
process and ensure that all important items are covered.
They also facilitate better aggregation of data gathered
from various sites or by various investigators. An excerpt
from a direct observation form used in a study of primary
health care in the Philippines provides an illustration below.
yes or no, or by checking the appropriate answer. Closed
response categories help minimize observer variation, and
therefore improve the quality of data.
2. Limit the number of items in a form. Forms should
normally not exceed 40–50 items. If nessary, it is better to
use two or more smaller forms than a single large one that
runs several pages.
3
3. Provide adequate space to record additional observations
for which response categories were not determined.
4. Use of computer software designed to create forms can
be very helpful. It facilitates a neat, unconfusing form that
can be easily completed.
People and organizations follow daily routines associated
with set times. For example, credit institutions may accept
loan applications in the morning; farmers in tropical
climates may go to their fields early in the morning and
return home by noon. Observation periods should reflect
work rhythms.
Step 3. Select the sites
Step 5. Conduct the field observation
Once the forms are ready, the next step is to decide where
the observations will be carried out and whether it will be
based on one or more sites.
Establish rapport. Before embarking on direct observation,
a certain level of rapport should be established with the
people, community, or organization to be studied. The
presence of outside observers, especially if officials or
experts, may generate some anxiety among those being
observed. Often informal, friendly conversations can
reduce anxiety levels.
A single site observation may be justified if a site can be
treated as a typical case or if it is unique. Consider a
situation in which all five agricultural extension centers
established by an assistance activity have not been
performing well. Here, observation at a single site may be
justified as a typical case. A single site observation may
also be justified when the case is unique; for example, if
only one of five centers had been having major problems,
and the purpose of the evaluation is trying to discover why.
However, single site observations should be avoided
generally, because cases the team assumes to be typical or
unique may not be. As a rule, several sites are necessary to
obtain a reasonable understanding of a situation.
In most cases, teams select sites based on experts' advice.
The investigator develops criteria for selecting sites, then
relies on the judgment of knowledgeable people. For
example, if a team evaluating a family planning project
decides to observe three clinics—one highly successful,
one moderately successful, and one struggling clinic—it
may request USAID staff, local experts, or other
informants to suggest a few clinics for each category. The
team will then choose three after examining their
recommendations. Using more than one expert reduces
individual bias in selection.
Alternatively, sites can be selected based on data from
performance monitoring. For example, activity sites
(clinics, schools, credit institutions) can be ranked from
best to worst based on performance measures, and then a
sample drawn from them.
Step 4. Decide on the best timing
Timing is critical in direct observation, especially when
events are to be observed as they occur. Wrong timing can
distort findings. For example, rural credit
Also, let them know the purpose of the observation is not to
report on individuals' performance, but to find out what
kind of problems in general are being encountered.
Allow sufficient time for direct observation. Brief visits can
be deceptive partly because people tend to behave
differently in the presence of observers. It is not
uncommon, for example, for health workers to become
more caring or for extension workers to be more
persuasive when being watched. However, if observers
stay for relatively longer periods, people become less selfconscious and gradually start behaving naturally. It is
essential to stay at least two or three days on a site to
gather valid, reliable data.
Use a team approach. If possible, two observers should
observe together. A team can develop more
comprehensive, higher quality data, and avoid individual
bias.
Train observers. If many sites are to be observed,
nonexperts can be trained as observers, especially if
observation forms are clear, straightforward, and mostly
closed-ended.
Step 6. Complete forms
Take notes as inconspicuously as possible. The best time
for recording is during observation. However, this is not
always feasible because it may make some people selfconscious or disturb the situation. In these cases, recording
should take place as soon as possible after observation.
Step 7. Analyze the data
organizations receive most loan applications during the
planting season, when farmers wish to purchase
agricultural inputs. If credit institutions are observed during
the nonplanting season, an inaccurate picture of loan
processing may result.
Data from close-ended questions from the observation
form can be analyzed using basic procedures such as
frequency counts and cross-tabulations. Statistical software
packages such as SAS or SPSS facilitate such statistical
analysis and data display.
4
Analysis of any open-ended interview questions can also
provide extra richness of understanding and insights. Here,
use of database management software with text storage
capabilities, such as dBase, can be useful.
Step 8. Check for reliability and validity.
Direct observation techniques are susceptible to error and
bias that can affect reliability and validity. These can be
minimized by following some of the procedures suggested,
such as checking the representativeness of the sample of
Direct Observation of Primary
Health Care Services in the Philippines
An example of structured direct observation was an
effort to identify deficiencies in the primary health
care system in the Philippines. It was part of a
larger, multicountry research project, the Primary
Health Care Operations Research Project (PRICOR).
The evaluators prepared direct observation forms
covering the activities, tasks, and subtasks health
workers must carry out in health clinics to
accomplish clinical objectives. These forms were
closed-ended and in most cases observations could
simply be checked to save time. The team looked at
18 health units from a "typical" province, including
samples of units that were high, medium and low
performers in terms of key child survival outcome
indicators.
The evaluation team identified and quantified many
problems that required immediate government
attention. For example, in 40 percent of the cases
where followup treatment was required at home,
health workers failed to tell mothers the timing and
amount of medication required. In 90 percent of
cases, health workers failed to explain to mothers the
results of child weighing and growth plotting, thus
missing the opportunity to involve mothers in the
nutritional care of their child. Moreover, numerous
errors were made in weighing and plotting.
This case illustrates that use of closed-ended
observation instruments promotes the reliability and
consistency of data. The findings are thus more
credible and likely to influence program managers to
make needed improvements.
sites selected; using closed-ended, unambiguous response
categories on the observation forms, recording observations
promptly, and using teams of observers at each site.
Selected Further Reading
Information in this Tips is based on "Rapid Data Collection
Methods for Field Assessments" by Krishna Kumar, in
Team Planning Notebook for Field-Based Program
Assessments (USAID PPC/CDIE, 1991).
For more on direct observation techniques applied to the
Philippines health care system, see Stewart N. Blumenfeld,
Manuel Roxas, and Maricor de los Santos, "Systematic
Observation in the Analysis of Primary Health Care
Services," in Rapid Appraisal Methods, edited by Krishna
Kumar (The World Bank:1993)
CDIE's Tips series provide advice and suggestions to
USAID managers on how to plan and conduct
performance monitoring and evaluation activities.
They are supplemental references to the reengineering
automated directives system (ADS), chapter 203. For
further information, contact Annette Binnendijk, CDIE
Senior Evaluation Advisor, phone (703) 875–4235, fax
(703) 875–4866, or e-mail. Tips can be ordered from
the Development Information Services Clearinghouse
by calling (703) 351-4006 or by faxing (703) 351–4039.
Please refer to the PN number. To order via Internet,
address requests to
[email protected]
NUMBER 5
2ND EDITION, 2010
PERFORMANCE MONITORING & EVALUATION
TIPS
USING RAPID APPRAISAL METHODS
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to performance
monitoring and evaluation. This publication is a supplemental reference to the Automated Directive
System (ADS) Chapter 203.
WHAT IS RAPID
APPRAISAL?
Rapid Appraisal (RA) is an approach
that draws on multiple evaluation
methods and techniques to quickly,
yet systematically, collect data when
time in the field is limited. RA
practices are also useful when there
are budget constraints or limited
availability of reliable secondary
data. For example, time and budget
limitations may preclude the option
of using representative sample
surveys.
BENEFITS – WHEN TO USE
RAPID APPRAISAL
METHODS
Rapid appraisals are quick and can
be done at relatively low cost.
Rapid appraisal methods can help
gather, analyze, and report relevant
information for decision-makers
within days or weeks. This is not
possible with sample surveys. RAs
can be used in the following cases:
• for formative evaluations, to make
mid-course corrections in project
design or implementation when
customer or partner feedback
indicates a problem (See ADS
203.3.6.1);
• when a key management decision
is required and there is inadequate
information;
• for performance monitoring, when
data are collected and the
techniques are repeated over time
for measurement purposes;
• to better understand the issues
behind performance monitoring
data; and
• for project pre-design assessment.
LIMITATIONS – WHEN
RAPID APPRAISALS ARE
NOT APPROPRIATE
Findings from rapid appraisals may
have limited reliability and validity,
and cannot be generalized to the
larger population. Accordingly,
1
rapid appraisal should not be the
sole basis for summative or impact
evaluations. Data can be biased and
inaccurate unless multiple methods
are used to strengthen the validity
of findings and careful preparation is
undertaken prior to beginning field
work.
WHEN ARE RAPID
APPRAISAL
METHODS
APPROPRIATE?
Choosing between rapid appraisal
methods for an assessment or more
time-consuming methods, such as
sample surveys, should depend on
balancing several factors, listed
below.
• Purpose of the study. The
importance and nature of the
decision depending on it.
• Confidence in results. The
accuracy, reliability, and validity of
findings needed for management
decisions.
• Time frame. When a decision
must be made.
• Resource constraints (budget).
• Evaluation questions to be
answered. (see TIPS 3: Preparing
an Evaluation Statement of Work)
USE IN TYPES OF
EVALUATION
Rapid appraisal methods are often
used in formative evaluations.
Findings are strengthened when
evaluators
use
triangulation
(employing more than one data
collection method) as a check on
the validity of findings from any one
method.
Rapid appraisal methods are also
used in the context of summative
evaluations. The data from rapid
appraisal methods and techniques
complement the use of quantitative
methods such as surveys based on
representative
sampling.
For
example, a randomized survey of
small holder farmers may tell you
that farmers have a difficult time
selling their goods at market, but
may not have provide you with the
details of why this is occurring. A
researcher
could
then
use
interviews
with
farmers
to
determine the details necessary to
construct a more complete theory
of why it is difficult for small holder
farmers to sell their goods.
KEY PRINCIPLES
FOR ENSURING
USEFUL RAPID
APPRAISAL DATA
COLLECTION
No set of rules dictates which
methods and techniques should be
used in a given field situation;
however, a number of key principles
can be followed to ensure the
collection of useful data in a rapid
appraisal.
• Preparation is key. As in any
evaluation, the evaluation design
and selection of methods must
begin
with
a
thorough
understanding of the evaluation
questions and the client’s needs
for evaluative information. The
client’s intended uses of data must
guide the evaluation design and
the types of methods that are
used.
• Triangulation increases the validity
of findings. To lessen bias and
strengthen the validity of findings
from rapid appraisal methods and
techniques, it is imperative to use
multiple methods. In this way,
data collected using one method
can be compared to that collected
using other methods, thus giving a
researcher the ability to generate
valid and reliable findings. If, for
example, data collected using Key
Informant Interviews reveal the
same findings as data collected
from Direct Observation and
Focus Group Interviews, there is
less chance that the findings from
the first method were due to
researcher bias or due to the
findings being outliers. Table 1
summarizes
common
rapid
appraisal methods and suggests
how findings from any one
method can be strengthened by
the use of other methods.
COMMON RAPID
APPRAISAL
METHODS
INTERVIEWS
This method involves one-on-one
interviews with individuals or key
informants selected for their
knowledge or diverse views.
Interviews are qualitative, in-depth
and semi-structured.
Interview
guides are usually used and
2
EVALUATION METHODS
COMMONLY USED IN RAPID
APPRAISAL
• Interviews
• Community Discussions
• Exit Polling
• Transect Walks (see p. 3)
• Focus Groups
• Minisurveys
• Community Mapping
• Secondary Data Collection
• Group Discussions
• Customer Service Surveys
• Direct Observation
questions may be further framed
during the interview, using subtle
probing techniques.
Individual
interviews may be used to gain
information on a general topic but
cannot provide the in-depth inside
knowledge on evaluation topics that
key informants may provide.
MINISURVEYS
A minisurvey consists of interviews
with between five to fifty individuals,
usually
selected
using
nonprobability sampling (sampling in
which respondents are chosen based
on their understanding of issues
related to a purpose or specific
questions, usually used when sample
sizes are small and time or access to
areas is limited).
Structured
questionnaires are used with a
limited number of close-ended
questions.
Minisurveys generate
quantitative data that can often be
collected and analyzed quickly.
FOCUS GROUPS
The focus group is a gathering of a
homogeneous body of five to twelve
participants to discuss issues and
experiences among themselves.
These are used to test an idea or to
get a reaction on specific topics. A
moderator introduces the topic,
stimulates
and
focuses
the
discussion, and prevents domination
of discussion by a few, while another
documents
the
evaluator
conversation.
THE ROLE OF TECHNOLOGY
IN RAPID APPRAISAL
Certain equipment and technologies
can aid the rapid collection of data
and help to decrease the incidence of
errors. These include, for example,
hand held computers or personal
digital assistants (PDAs) for data
input,
cellular
phones,
digital
recording devices for interviews,
videotaping and photography, and the
use of geographic information systems
(GIS) data and aerial photographs.
GROUP DISCUSSIONS
This method involves the selection
of approximately five participants
who are knowledgeable about a
given topic and are comfortable
enough with one another to freely
discuss the issue as a group. The
moderator introduces the topic and
keeps the discussion going while
another evaluator records the
discussion. Participants talk among
each other rather than respond
directly to the moderator.
COMMUNITY DISCUSSIONS
This method takes place at a public
meeting that is open to all
community members; it can be
successfully moderated with as
many as 100 or more people. The
primary interaction is between the
participants while the moderator
leads the discussion and asks
questions following a carefully
prepared interview guide.
DIRECT OBSERVATION
Teams of observers record what
they hear and see at a program site
using a detailed observation form.
Observation may be of the physical
surrounding or of ongoing activities,
processes, or interactions.
COLLECTING SECONDARY
DATA
This method involves the on-site
collection of existing secondary
data, such as export sales, loan
information, health service statistics,
etc. These data are an important
augmentation
to
information
collected using qualitative methods
such as interviews, focus groups, and
community discussions.
The
3
evaluator must be able to quickly
determine the validity and reliability
of the data. (see TIPS 12: Indicator
and Data Quality)
TRANSECT WALKS
The transect walk is a participatory
approach in which the evaluator
asks a selected community member
to walk with him or her, for
example, through the center of
town, from one end of a village to
the other, or through a market.
The evaluator asks the individual,
usually a key informant, to point out
and
discuss
important
sites,
neighborhoods, businesses, etc., and
to discuss related issues.
COMMUNITY MAPPING
Community mapping is a technique
that requires the participation of
residents on a program site. It can
be used to help locate natural
resources, routes, service delivery
points, regional markets, trouble
spots, etc., on a map of the area, or
to use residents’ feedback to drive
the development of a map that
includes such information.
COMMON RAPID APPRAISAL METHODS
Table 1
Method
Useful for
Providing
Example
Advantages
Limitations
Further
References
INDIVIDUAL INTERVIEWS
Interviews
− A general overview of
the topic from
someone who has a
broad knowledge and
in-depth experience
and understanding
(key informant) or indepth information on
a very specific topic or
subtopic (individual)
− Suggestions and
recommendations to
improve key aspects
of a program
Minisurveys
− Quantitative data on
narrowly focused
questions, for a
relatively
homogeneous
population, when
representative
sampling is not
possible or required
Key informant:
Interview with
program
implementation
director
− Provides in-depth, − Susceptible to
inside information
interviewer and
on specific issues
selection biases
from the
− Individual
individuals
interviews lack the
perspective and
broader
experience
understanding and
− Flexibility permits
insight that a key
exploring
informant can
unanticipated
provide
topics
Interview with
director of a regional
trade association
Individual:
Interview with an
activity manager within
an overall
− Easy to administer
development program
− Low cost
Interview with a local
entrepreneur trying to
enter export trade
− A customer service
assessment
− Quantitative data
from multiple
respondents
− Rapid exit interviews
after voting
− Low cost
TIPS No. 2,
Conducting Key
Informant Interviews
K. Kumar, Conducting
Key Informant Surveys
in Developing
Countries, 1986
Bamberger, Rugh, and
Mabry, Real World
Evaluation, 2006
UNICEF Website: M&E
Training Modules:
Overview of RAP
Techniques
− Findings are less
generalizable than
those from sample
surveys unless the
universe of the
population is
surveyed
TIPS No. 9,
Conducting Customer
Service Assessments
K. Kumar, Conducting
Mini Surveys in
Developing Countries,
1990
Bamberger, Rugh, and
Mabry, RealWorld
Evaluation, 2006 on
purposeful sampling
− Quick data on
attitudes, beliefs,
behaviors of
beneficiaries or
partners
GROUP INTERVIEWS
Focus Groups
− Customer views on
services, products,
benefits
− Information on
implementation
problems
− Suggestions and
recommendations for
improving specific
activities
− Discussion on
− Group discussion
experience related
may reduce
to a specific program
inhibitions,
intervention
allowing free
exchange of ideas
− Effects of a new
business regulation
− Low cost
or proposed price
changes
4
− Discussion may be
dominated by a
few individuals
unless the process
is facilitated/
managed well
TIPS No. 10,
Conducting Focus
Group Interviews
K. Kumar, Conducting
Group Interviews in
Developing Countries,
1987
T. Greenbaum,
Moderating Focus
Groups: A Practical
Guide for Group
Facilitation, 2000
Group
Discussions
Community
Discussions
− Understanding of
issues from different
perspectives and
experiences of
participants from a
specific subpopulation
− Discussion with
young women on
access to prenatal
and infant care
− Small group size
allows full
participation
− Discussion with
entrepreneurs about
export regulations
− Understanding of an
− A Town Hall
issue or topic from a
meeting
wide range of
participants from key
evaluation sites within
a village, town, city, or
city neighborhood
− Findings cannot be Bamberger, Rugh, and
Mabry, RealWorld
generalized to a
Evaluation, 2006
larger population
UNICEF Website: M&E
Training Modules:
Community Meetings
− Allows good
understanding of
specific topics
− Low cost
− Yields a wide
range of opinions
on issues
important to
participants
− Findings cannot be
generalized to
larger population
or to
subpopulations of
concern
Bamberger, Rugh, and
Mabry, RealWorld
Evaluation, 2006
− Observer bias
unless two to
three evaluators
observe same
place or activity
TIPS No. 4, Using
Direct Observation
Techniques
− Must be able to
determine
reliability and
validity of data
TIPS No. 12,
Guidelines for
Indicator and Data
Quality
− A great deal of
information can be − Larger groups
obtained at one
difficult to
point of time
moderate
UNICEF Website: M&E
Training Modules:
Community Meetings
ADDITIONAL COMMONLY USED TECHNIQUES
Direct
Observation
− Visual data on physical − Market place to
− Confirms data
infrastructure,
observe goods being
from interviews
supplies, conditions
bought and sold,
− Low cost
who is involved,
− Information about an
sales interactions
agency’s or business’s
delivery systems,
services
WFP Website:
Monitoring & Evaluation
Guidelines: What Is
Direct Observation and
When Should It Be Used?
− Insights into behaviors
or events
Collecting
Secondary
Data
− Validity to findings
gathered from
interviews and group
discussions
− Microenterprise
bank loan info.
− Value and volume of
exports
− Quick, low cost
way of obtaining
important
quantitative data
− Number of people
served by a health
clinic, social service
provider
PARTICIPATORY TECHNIQUES
Transect
Walks
Community
Mapping
− Important visual and
locational information
and a deeper
understanding of
situations and issues
− Walk with key
informant from one
end of a village or
urban neighborhood
to another, through
a market place, etc.
− Info. on locations
important for data
collection that could
be difficult to find
− Insiders viewpoint − Susceptible to
interviewer and
− Quick way to find
selection biases
out location of
places of interest
to the evaluator
− Low cost
− Map of village and
− Important
− Rough locational
surrounding area
locational data
information
with locations of
when there are no
markets, water and
detailed maps of
fuel sources, conflict
the program site
− Quick comprehension
areas, etc.
on spatial location of
services/resources in a
region which can give
insight to access issues
5
Bamberger, Rugh, and
Mabry, Real World
Evaluation, 2006
UNICEF Website: M&E
Training Modules:
Overview of RAP
Techniques
Bamberger, Rugh, and
Mabry, Real World
Evaluation, 2006
UNICEF Website: M&E
Training Modules:
Overview of RAP
Techniques
References Cited
M. Bamberger, J. Rugh, and L. Mabry, Real World Evaluation. Working Under Budget, Time, Data, and Political
Constraints. Sage Publications, Thousand Oaks, CA, 2006.
T. Greenbaum, Moderating Focus Groups: A Practical Guide for Group Facilitation. Sage Publications, Thousand Oaks,
CA, 2000.
K. Kumar, “Conducting Mini Surveys in Developing Countries,” USAID Program Design and Evaluation Methodology
Report No. 15, 1990 (revised 2006).
K. Kumar, “Conducting Group Interviews in Developing Countries,” USAID Program Design and Evaluation
Methodology Report No. 8, 1987.
K. Kumar, “Conducting Key Informant Interviews in Developing Countries,” USAID Program Design and Evaluation
Methodology Report No. 13, 1989.
For more information:
TIPS publications are available online at [insert website].
Acknowledgements:
Our thanks to those whose experience and insights helped shape this publication including USAID’s Office of
Management Policy, Budget and Performance (MPBP). This publication was authored by Patricia Vondal, PhD., of
Management Systems International.
Comments regarding this publication can be directed to:
Gerald Britan, Ph.D.
Tel: (202) 712-1158
[email protected]
Contracted under RAN-M-00-04-00049-A-FY0S-84
Integrated Managing for Results II
6
NUMBER 6
2ND EDITION, 2010
PERFORMANCE MONITORING & EVALUATION
TIPS
SELECTING PERFORMANCE INDICATORS
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to
performance monitoring and evaluation. This publication is a supplemental reference to the
Automated Directive System (ADS) Chapter 203.
WHAT ARE
PERFORMANCE
INDICATORS?
Performance indicators define a
measure of change for the
results identified in a Results
Framework (RF). When wellchosen, they convey whether
key objectives are achieved in a
meaningful
way
for
performance
management.
While a result (such as an
Assistance Objective or an
Intermediate Result) identifies
what we hope to accomplish,
indicators tell us by what
standard that result will be
measured.
Targets define
whether there will be an
expected increase or decrease,
and by what magnitude.1
Indicators may be quantitative
or
qualitative
in
nature.
Quantitative
indicators
are
numerical: an example is a
person’s height or weight. On
the other hand, qualitative
indicators require subjective
evaluation. Qualitative data are
sometimes
reported
in
numerical form, but those
numbers do not have arithmetic
meaning on their own. Some
examples are a score on an
institutional capacity index or
progress along a milestone
scale.
When developing
quantitative
or
qualitative
indicators, the important point
is that the indicator be
1
For further information, see TIPS 13:
Building a Results Framework and TIPS
8: Baselines and Targets.
1
Selecting an optimal set of indicators
to track progress against key results
lies at the heart of an effective
performance management system.
This TIPS provides guidance on how to
select effective performance
indicators.
constructed in a way that
permits
consistent
measurement over time.
USAID has developed many
performance indicators over the
years. Some examples include
the dollar value of nontraditional
exports,
private
investment as a percentage of
gross
domestic
product,
contraceptive prevalence rates,
child mortality rates, and
progress on a legislative reform
index.
WHY ARE
PERFORMANCE
INDICATORS
IMPORTANT?
FOR WHAT RESULTS
ARE PERFORMANCE
INDICATORS
REQUIRED?
Performance indicators provide
objective evidence that an
intended change is occurring.
Performance indicators lie at
the heart of developing an
effective
performance
management system – they
define the data to be collected
and enable actual results
achieved to be compared with
planned results over time.
Hence,
they
are
an
indispensable management tool
for making evidence-based
decisions
about
program
strategies
and
activities.
Performance indicators can also
be used:
THE PROGRAM LEVEL
USAID’s ADS requires that at
least one indicator be chosen
for each result in the Results
Framework in order to measure
progress (see ADS 203.3.3.1)2.
This includes the Assistance
Objective (the highest-level
objective
in
the
Results
Framework)
as
well
as
supporting Intermediate Results
(IRs)3. These indicators should
be included in the Mission or
Office
Performance
Management Plan (PMP) (see
TIPS 8: Preparing a PMP).





To assist managers in
focusing
on
the
achievement
of
development results.
To
provide
objective
evidence that results are
being achieved.
To orient and motivate staff
and
partners
toward
achieving results.
To communicate USAID
achievements
to
host
country counterparts, other
partners, and customers.
To more effectively report
results achieved to USAID's
stakeholders, including the
U.S. Congress, Office of
Management and Budget,
and citizens.
PROJECT LEVEL
AO teams are required to
collect data regularly for
projects and activities, including
inputs, outputs, and processes,
to ensure they are progressing
as
expected
and
are
contributing to relevant IRs and
AOs. These indicators should
be included in a project-level
monitoring
and
evaluation
2
For further discussion of AOs and IRs
(which are also termed impact and
outcomes respectively in other
systems) refer to TIPS 13: Building a
Results Framework.
3
Note that some results frameworks
incorporate IRs from other partners if
those results are important for USAID
to achieve the AO. This is discussed in
further detail in TIPS 13: Building a
Results Framework. If these IRs are
included, then it is recommended that
they be monitored, although less
rigorous standards apply.
2
(M&E) plan. The M&E plan
should be integrated in project
management and reporting
systems (e.g., quarterly, semiannual, or annual reports).
TYPES OF
INDICATORS IN
USAID SYSTEMS
Several different types of
indicators are used in USAID
systems.
It is important to
understand the different roles
and
functions
of
these
indicators so that managers can
construct
a
performance
management
system
that
effectively
meets
internal
management
and
Agency
reporting needs.
CUSTOM INDICATORS
Custom
Indicators
are
performance indicators that
reflect progress within each
unique country or program
context. While they are useful
for managers on the ground,
they
often
cannot
be
aggregated across a number of
programs
like
standard
indicators.
Example:
Progress on a
milestone scale reflecting
legal
reform
and
implementation to ensure
credible elections, as follows:
 Draft law is developed in
consultation with nongovernmental
organizations (NGOs) and
political parties.
 Public input is elicited.
PARTICIPATION IS ESSENTIAL
Experience suggests that
participatory approaches are an
essential aspect of developing and
maintaining effective performance
management systems. Collaboration
with development partners
(including host country institutions,
civil society organizations (CSOs),
and implementing partners) as well
as customers has important benefits.
It allows you to draw on the
experience of others, obtains buy-in
to achieving results and meeting
targets, and provides an opportunity
to ensure that systems are as
streamlined and practical as possible.
 Draft law is modified based
on feedback.
 The secretariat presents
the draft to the Assembly.
 The law is passed by the
Assembly.
 The
appropriate
government
body
completes internal policies
or
regulations
to
implement the law.
The example above would differ
for each country depending on
its unique process for legal
reform.
STANDARD INDICATORS
Standard indicators are used
primarily for Agency reporting
purposes. Standard indicators
produce data that can be
aggregated
across
many
programs. Optimally, standard
indicators meet both Agency
reporting and on-the-ground
management needs. However,
in
many
cases,
standard
indicators do not substitute for
performance
(or
custom
indicators) because they are
designed to meet different
needs. There is often a tension
between measuring a standard
across many programs and
selecting indicators that best
reflect true program results and
that can be used for internal
management purposes.
Example: Number of Laws or
Amendments
to
Ensure
Credible Elections Adopted
with
USG
Technical
Assistance.
In comparing the standard
indicator above with the
previous example of a custom
indicator, it becomes clear that
the custom indictor is more
likely to be useful as a
management tool, because it
provides greater specificity and
is more sensitive to change.
Standard indicators also tend to
measure change at the output
level, because they are precisely
the types of measures that are,
at face value, more easily
aggregated
across
many
programs, as the following
example demonstrates.
Example:
The number of
people trained in policy and
regulatory practices.
CONTEXTUAL INDICATORS
Contextual indicators are used
to understand the broader
environment
in
which
a
program operates, to track
assumptions, or to examine
externalities that may affect
success, failure, or progress.
3
INDICATORS AND DATA—SO
WHAT’S THE DIFFERENCE?
Indicators define the particular
characteristic or dimension that will
be used to measure change. Height
is an example of an indicator.
The data are the actual
measurements or factual information
that result from the indicator. Five
feet seven inches is an example of
data.
They do not represent program
performance,
because
the
indicator measures very highlevel change.
Example:
Score on the
Freedom House Index or
Gross
Domestic
Product
(GDP).
This sort of indicator may be
important
to
track
to
understand the context for
USAID programming (e.g. a
severe drop in GDP is likely to
affect
economic
growth
programming), but represents a
level of change that is outside
the manageable interest of
program managers. In most
cases, it would be difficult to
say that USAID programming
has affected the overall level of
freedom within a country or
GDP (given the size of most
USAID programs in comparison
to the host country economy,
for example).
WHAT ARE USAID’S
CRITERIA FOR
SELECTING
INDICATORS?
USAID policies (ADS 203.3.4.2)
identify seven key criteria to
guide
the
selection
of
performance indicators:







Direct
Objective
Useful for Management
Attributable
Practical
Adequate
Disaggregated, as necessary
These criteria are designed to
assist managers in selecting
optimal indicators. The extent
to
which
performance
indicators meet each of the
criteria must be consistent with
the requirements of good
management.
As managers
consider these criteria, they
should use a healthy measure
of
common
sense
and
reasonableness.
While we
always
want
the
―best‖
indicators, there are inevitably
trade-offs
among
various
criteria. For example, data for
the most direct or objective
indicators of a given result
might be very expensive to
collect or might be available
too infrequently. Table 1
includes a summary checklist
that can be used during the
selection process to assess
these trade-offs.
Two
overarching
factors
determine the extent to which
performance indicators function
as useful tools for managers
and decision-makers:


The degree to which
performance
indicators
accurately
reflect
the
process or phenomenon
they are being used to
measure.
The level of comparability of
performance indicators over
time: that is, can we
measure
results
in
a
consistent and comparable
manner over time?
1. DIRECT
An indicator is direct to the
extent that it clearly measures
the intended result.
This
criterion is, in many ways, the
most important. While this may
appear to be a simple concept,
it is one of the more common
problems
with
indicators.
Indicators should either be
widely accepted for use by
specialists in a subject area,
exhibit readily understandable
face validity (i.e., be intuitively
understandable),
or
be
supported
by
research.
Managers should place greater
confidence in indicators that are
direct. Consider the following
example:
Result:
Increased
Transparency of Key Public
Sector Institutions
4
Indirect Indicator: Passage
of
the
Freedom
of
Information Act (FOIA)
Direct Indicator: Progress
on
a
milestone
scale
demonstrating
enactment
and enforcement of policies
that require open hearings
The passage of FOIA, while an
important step, does not
actually measure whether a
target institution is more
transparent.
The
better
example outlined above is a
more direct measure.
Level
Another dimension of whether
an indicator is direct relates to
whether it measures the right
level of the objective. A
common problem is that there
is often a mismatch between
the stated result and the
indicator. The indicator should
not measure a higher or lower
level than the result.
For example, if a program
measures
improved
management practices through
the real value of agricultural
production, the indicator is
measuring a higher-level effect
than is stated (see Figure 1).
Understanding levels is rooted
in
understanding
the
development
hypothesis
inherent
in
the
Results
Framework (see TIPS 13:
Building a Results Framework).
Tracking indicators at each level
facilitates better understanding
and analysis of whether the
Figure 1. Levels
RESULT
INDICATOR
Increased
Production
Real value of
agricultural
production.
Improved
Management
Practices
Number and
percent of
farmers using a
new technology.
Improved
Knowledge
and
Awareness
Number and
percent of
farmers who can
identify five out
of eight steps
for
implementing a
new technology.
development hypothesis is
working.
For example, if
farmers are aware of how to
implement a new technology,
but the number or percent that
actually use the technology is
not increasing, there may be
other issues that need to be
addressed.
Perhaps the
technology is not readily
available in the community, or
there is not enough access to
credit. This flags the issue for
managers and provides an
opportunity
to
make
programmatic adjustments.
Proxy Indicators
Proxy indicators are linked to
the result by one or more
assumptions. They are often
used when the most direct
indicator is not practical (e.g.,
data collection is too costly or
the
program
is
being
implemented in a conflict zone).
When proxies are used, the
relationship
between
the
indicator and the result should
be well-understood and clearly
articulated.
The
more
assumptions the indicator is
based upon, the weaker the
indicator.
Consider
the
following examples:
Result: Increased Household
Income
Proxy Indicator:
Dollar
value
of
household
expenditures
The proxy indicator above
makes the assumption that an
increase in income will result in
increased
household
expenditures; this assumption is
well-grounded in research.
Result: Increased Access to
Justice
Proxy Indicator: Number of
new courts opened
The indicator above is based on
the assumption that physical
access to new courts is the
fundamental
development
problem—as
opposed
to
corruption, the costs associated
with using the court system, or
lack of knowledge of how to
obtain legal assistance and/or
use court systems. Proxies can
be used when assumptions are
clear and when there is research
to support that assumption.
2. OBJECTIVE
An indicator is objective if it is
unambiguous about 1) what is
being measured and 2) what
5
data are being collected. In
other words, two people should
be able to collect performance
information for the same
indicator and come to the same
conclusion.
Objectivity is
critical to collecting comparable
data over time, yet it is one of
the most common problems
noted in audits.
As a result,
pay particular attention to the
definition of the indicator to
ensure that each term is clearly
defined, as the following
examples demonstrate:
Poor Indicator: Number of
successful firms
Objective Indicator:
Number of firms with an
annual increase in revenues
of at least 5%
The better example outlines the
exact
criteria
for
how
―successful‖ is defined and
ensures that changes in the
data are not attributable to
differences in what is being
counted.
Objectivity can be particularly
challenging when constructing
qualitative indicators.
Good
qualitative indicators permit
regular, systematic judgment
about progress and reduce
subjectivity (to the extent
possible).
This means that
there must be clear criteria or
protocols for data collection.
3. USEFUL FOR
MANAGEMENT
An indicator is useful to the
extent that it provides a
meaningful measure of change
over time for management
decision-making. One aspect of
usefulness is to ensure that the
indicator is measuring the ―right
change‖ in order to achieve
development results.
For
example, the number of
meetings between Civil Society
Organizations
(CSOs)
and
government is something that
can be counted but does not
necessarily reflect meaningful
change. By selecting indicators,
managers are defining program
success in concrete ways.
Managers
will
focus
on
achieving targets for those
indicators, so it is important to
consider the intended and
unintended incentives
that
performance indicators create.
As a result, the system may
need to be fine-tuned to ensure
that incentives are focused on
achieving true results.
A second dimension is whether
the indictor measures a rate of
change that is useful for
management purposes.
This
means that the indicator is
constructed so that change can
be monitored at a rate that
facilitates management actions
(such as corrections and
improvements).
Consider the
following examples:
Result:
Targeted legal
reform
to
promote
investment
Less Useful for
Management: Number of
laws passed to promote
direct investment.
More Useful for
Management: Progress
toward targeted legal reform
based on the following
stages:
Stage 1. Interested groups
propose that legislation is
needed on issue.
Stage 2. Issue is introduced
in the relevant legislative
committee/executive
ministry.
Stage 3.
Legislation is
drafted
by
relevant
committee
or
executive
ministry.
Stage 4.
Legislation is
debated by the legislature.
Stage 5.
Legislation is
passed by full approval
process needed in legislature.
Stage 6.
Legislation is
approved by the executive
branch (where necessary).
Stage 7.
Implementing
actions are taken.
Stage 8. No immediate need
identified for amendments to
the law.
The less useful example may be
useful for reporting; however, it
is so general that it does not
provide a good way to track
progress
for
performance
management. The process of
passing or implementing laws is
a long-term one, so that over
the course of a year or two the
AO team may only be able to
report that one or two such
laws have passed when, in
reality, a high degree of effort is
6
invested in the process. In this
case, the more useful example
better articulates the important
steps that must occur for a law
to be passed and implemented
and facilitates management
decision-making. If there is a
problem in meeting interim
milestones, then corrections
can be made along the way.
4. ATTRIBUTABLE
An indicator is attributable if it
can be plausibly associated with
USAID interventions.
The
concept
of
―plausible
association‖ has been used in
USAID for some time. It does
not mean that X input equals Y
output. Rather, it is based on
the idea that a case can be
made to other development
practitioners that the program
has
materially
affected
identified change.
It is
important to consider the logic
behind what is proposed to
ensure attribution. If a Mission
is piloting a project in three
schools, but claims national
level
impact
in
school
completion, this would not pass
the
common
sense
test.
Consider
the
following
examples:
Result: Improved Budgeting
Capacity
Less Attributable: Budget
allocation for the Ministry of
Justice (MOJ)
More Attributable:
The
extent to which the budget
produced by the MOJ meets
established criteria for good
budgeting
If the program works with the
Ministry of Justice to improve
budgeting
capacity
(by
providing technical assistance
on budget analysis), the quality
of the budget submitted by the
MOJ may improve. However, it
is often difficult to attribute
changes in the overall budget
allocation
to
USAID
interventions, because there are
a number of externalities that
affect a country’s final budget –
much like in the U.S.
For
example, in tough economic
times, the budget for all
government institutions may
decrease. A crisis may emerge
that requires the host country
to reallocate resources. The
better example above is more
attributable (and directly linked)
to USAID’s intervention.
5. PRACTICAL
A practical indicator is one for
which data can be collected on a
timely basis and at a reasonable
cost. There are two dimensions
that determine whether an
indicator is practical. The first is
time and the second is cost.
Time
Consider whether resulting data
are available with enough
frequency for management
purposes (i.e., timely enough to
correspond
to
USAID
performance management and
reporting purposes). Second,
examine whether data are
current when available.
If
reliable data are available each
year, but the data are a year
old, then it may be problematic.
Cost
Performance indicators should
provide data to managers at a
cost that is reasonable and
appropriate as compared with
the management utility of the
data. As a very general rule of
thumb, it is suggested that
between 5% and 10% of
program or project resources
be allocated for monitoring and
evaluation (M&E) purposes.
However, it is also important to
consider priorities and program
context.
A program would
likely be willing to invest more
resources in measuring changes
that are central to decisionmaking and less resources in
measuring more tangential
results.
A more mature
program may have to invest
more in demonstrating higherlevel changes or impacts as
compared to a new program.
6. ADEQUATE
Taken as a group, the indicator
(or set of indicators) should be
sufficient to measure the stated
result. In other words, they
should be the minimum
number necessary and costeffective
for
performance
management. The number of
indicators
required
to
adequately measure a result
depends on 1) the complexity
of the result being measured, 2)
the amount of information
needed to make reasonably
confident decisions, and 3) the
7
level of resources available.
Too many indicators create
information
overload
and
become overly burdensome to
maintain. Too few indicators
are also problematic, because
the data may only provide a
partial or misleading picture of
performance. The following
demonstrates
how
one
indicator can be adequate to
measure the stated objective:
Result: Increased Traditional
Exports in Targeted Sectors
Adequate Indicator: Value
of traditional exports in
targeted sectors
In
contrast,
an
objective
focusing on improved maternal
health may require two or three
indicators to be adequate. A
general rule of thumb is to
select between two and three
performance indicators per
result. If many more indicators
are needed to adequately cover
the result, then it may signify
that the objective is not
properly focused.
7. DISAGGREGATED, AS
NECESSARY
The disaggregation of data by
gender, age, location, or some
other dimension is often
important
from
both
a
management and reporting
point of view. Development
programs
often
affect
population
cohorts
or
institutions in different ways.
For example, it might be
important to know to what
extent youth (up to age 25) or
adults (25 and older) are
participating
in
vocational
training, or in which districts
schools
have
improved.
Disaggregated data help track
whether or not specific groups
participate in and benefit from
activities intended to include
them.
In particular, USAID policies
(ADS 203.3.4.3) require that
performance
management
systems and evaluations at the
AO and project or activity levels
include
gender-sensitive
indicators
and
sexdisaggregated data if the
activities or their anticipated
results involve or affect women
and men differently. If so, this
difference
would
be
an
important factor in managing
for sustainable program impact.
Consider the following example:
Result: Increased Access to
Credit
Gender-Sensitive Indicator:
Value of loans disbursed,
disaggregated
by
male/female.
WHAT IS THE
PROCESS FOR
SELECTING
PERFORMANCE
INDICATORS?
Selecting
appropriate
and
useful performance indicators
requires
careful
thought,
iterative refining, collaboration,
and consensus-building. The
following describes a series of
steps
to
select
optimal
performance
indicators4.
Although presented as discrete
steps, in practice some of these
can be effectively undertaken
simultaneously or in a more
iterative manner. These steps
may be applied as a part of a
larger process to develop a new
PMP, or in part, when teams
have to modify individual
indicators.

STEP 1. DEVELOP A
PARTICIPATORY PROCESS
FOR IDENTIFYING
PERFORMANCE INDICATORS
The most effective way to
identify indicators is to set up a
process
that
elicits
the
participation and feedback of a
number
of
partners
and
stakeholders.
This allows
managers to:
A common way to begin the
process is to hold working
sessions. Start by reviewing the
Results Framework.
Next,
identify indicators for the
Assistance
Objective,
then
move down to the Intermediate
Results. In some cases, the AO
team establishes the first round
of indicators and then provides
them to other partners for
input.
In other cases, key
partners may be included in the
working sessions.



Draw on different areas of
expertise.
Ensure
that
indicators
measure the right changes
and represent part of a
larger approach to achieve
development impact.
Build
commitment
and
understanding
of
the
linkage between indicators
and
results.
This
will
increase the utility of the
performance management
system
among
key
stakeholders.
This process focuses on presenting
greater detail related specifically to
indicator selection. Refer to TIPS 7:
Preparing a PMP for a broader set of
steps on how to develop a full PMP.
4
8

Build
capacity
for
performance management
among partners, such as
NGOs and partner country
institutions.
Ensure that systems are as
practical and streamlined as
possible.
Often
development partners can
provide excellent insight on
the
practical
issues
associated with indicators
and data collection.
It is important to task the group
with identifying the set of
minimal indicators necessary
and sufficient to manage the
program effectively. That is, the
group must go through a
process of prioritization in order
to narrow down the list. While
participatory processes may
take more time at the front end,
they almost always result in
more coherent and effective
system.
STEP 2. CLARIFY THE RESULT
Carefully define the result
desired. Good performance
indicators are based on clearly
articulated
and
focused
objectives. Review the precise
wording and intention of the
objective.
Determine what
exactly is meant by the result.
For example, if the result is
―improved
business
environment,‖ what does that
mean? What specific aspects of
the business environment will
be improved? Optimally, the
result should be stated with as
much specificity as possible. If
the result is broad (and the
team doesn’t have the latitude
to change the objective), then
the team might further define
its meaning.
Example:
One AO team
further defined their IR,
―Improved
Business
Environment,‖ as follows:
Making it easier to do
business in terms of resolving
disputes, obtaining licenses
from the government, and
promoting investment.

An identified set of key
policies are in place to
support investment.
Key
policies
include
laws,
regulations,
and
policies
related to the simplification of
investment
procedures,
bankruptcy, and starting a
business.

As the team gains greater
clarity and consensus on what
results are sought, ideas for
potential indicators begin to
emerge.
Be clear about what type of
change is implied. What is
expected
to
change—a
situation, a condition, the level
of knowledge, an attitude, or a
behavior?
For
example,
changing a country's voting
law(s) is very different from
changing citizens' awareness of
their right to vote (which is
different from voting). Each
type of change is measured by
different types of performance
indicators.
Identify more precisely the
specific targets for change. Who
or what are the specific targets
for the change? For example, if
individuals, which individuals?
For an economic growth
program designed to increase
exports, does the program
target all exporters or only
exporters of non-traditional
agricultural products? This is
known as identifying the ―unit
of analysis‖ for the performance
indicator.
STEP 3: IDENTIFY POSSIBLE
INDICATORS
Usually there are many possible
indicators for a particular result,
but some are more appropriate
and useful than others.
In
selecting indicators, don’t settle
too quickly on the first ideas
that come most conveniently or
obviously to mind. Create an
initial list of possible indicators,
using the following approaches:

Conduct a brainstorming
session with colleagues to
draw upon the expertise of
9
the full Assistance Objective
Team. Ask, ―how will we
know if the result is
achieved?‖

Consider other resources.
Many organizations have
databases or indicator lists
for various sectors available
on the internet.

Consult
experts.

Review the PMPs and
indicators
of
previous
programs
or
similar
programs in other Missions.
with
technical
STEP 4. ASSESS THE BEST
CANDIDATE INDICATORS,
USING THE INDICATOR
CRITERIA
Next, from the initial list, select
the
best
candidates
as
indicators.
The seven basic
criteria that can be used to
judge
an
indicator’s
appropriateness and utility
described in the previous
section are summarized in
Table 1. When assessing and
comparing possible indicators,
it is helpful to use this type of
checklist
to
guide
the
assessment
process.
Remember that there will be
trade-offs between the criteria.
For example, the optimal
indicator may not be the most
cost-effective to select.
STEP 5. SELECT THE “BEST”
PERFORMANCE INDICATORS
Select the best indicators to
incorporate in the performance
management system.
They
should be the optimum set of
measures that are useful to
management and can be
obtained at reasonable cost.
Be Strategic and Streamline
Where Possible. In recent years,
there has been a substantial
increase in the number of
indicators used to monitor and
track programs. It is important
to remember that there are
costs, in terms of time and
money, to collect data for each
indicator. AO teams should:

Select indicators based on
strategic thinking about
what must truly be achieved
for program success.

Review
indicators
to
determine whether any final
narrowing can be done. Are
some indicators not useful?
If so, discard them.

Use
participatory
approaches in order to
discuss
and
establish
priorities
that
help
managers focus on key
indicators that are necessary
and sufficient.
Ensure that the rationale for
indicator selection is recorded in
the PMP.
There are rarely
perfect
indicators
in
the
development environment—it
is more often a case of
weighing different criteria and
making the optimal choices for
a particular program.
It is
important to ensure that the
rationale behind these choices
is recorded in the PMP so that
new staff, implementers, or
auditors understand why each
indicator was selected.
STEP 6. FINE TUNE WHEN
NECESSARY
Indicators are part of a larger
system that is ultimately
designed to assist managers in
achieving development impact.
On the one hand, indicators
must remain comparable over
time but, on the other hand,
some refinements will invariably
be needed to ensure the system
is as effective as possible. (Of
course, there is no value in
continuing to collect bad data,
for example.) As a result, these
two issues need to be balanced.
Remember that indicator issues
are often flags for other
10
underlying problems. If a large
number of indicators are
frequently changed, this may
signify a problem with program
management or focus. At the
other end of the continuum, if
no indicators were to change
over a long period of time, it is
possible that a program is not
adapting and evolving as
necessary. In our experience,
some refinements are inevitable
as data are collected and
lessons learned. After some
rounds of data collection are
completed, it is often useful to
discuss indicator issues and
refinements among AO team
members and/or with partners
and
implementers.
In
particular, the period following
portfolio reviews is a good time
to refine PMPs if necessary.
TABLE 1. INDICATOR SELECTION CRITERIA CHECKLIST
Criteria
1. Direct
Definition
Checklist
Direct. The indicator clearly represents the
intended result. An outsider or an expert
in the field would agree that the indicator
is a logical measure for the stated result.
 Level. The indicator reflects the right
level; that is, it does not measure a
higher or lower level than the stated
result.
 Proxies. The indicator is a proxy
measure. If the indicator is a proxy, note
what assumptions the proxy is based
upon.
2. Objective
The indicator is clear and unambiguous
about what is being measured.
3. Useful for
Management
The indicator is useful for management
decision-making.
4. Attributable
The indicator can be plausibly associated
with USAID interventions.
5. Practical
Time. Data are produced with enough
frequency for management purposes (i.e.
timely enough to correspond to USAID
performance management and reporting
purposes). Data are current when
available.
Cost. Data are worth the cost to USAID
managers.
6. Adequate
The indicators, taken as a group, are
sufficient to measure the stated result. All
major aspects of the result are measured.
7. Disaggregated,
as necessary
The indicators are appropriately
disaggregated by gender, age, location, or
some other dimension that is important for
programming. In particular, gender
disaggregation has been considered as
required (see ADS 203.3.4.3).
11
Comments
For more information:
TIPS publications are available online at [insert website].
Acknowledgements:
Our thanks to those whose experience and insights helped shape this publication, including Gerry Britan
and Subhi Mehdi of USAID’s Office of Management Policy, Budget and Performance (MPBP). This
publication was updated by Michelle Adams-Matson of Management Systems International.
Comments can be directed to:
Gerald Britan, Ph.D.
Tel: (202) 712-1158
[email protected]
Contracted under RAN-M-00-04-00049-A-FY0S-84
Integrated Managing for Results II
12
1996, Number 7
Performance Monitoring and Evaluation
TIPS
USAID Center for Development Information and Evaluation
PREPARING A PERFORMANCE MONITORING PLAN
What Is a Performance Monitoring Plan?
A performance monitoring plan (PMP) is a tool USAID operating units use to
plan and manage the collection of performance data. Sometimes the plan also
includes plans for data analysis, reporting, and use.
USAID's
reengineering
guidance
requires
operating units
to prepare a
Performance
Monitoring Plan
for the
systematic and
timely collection
of performance
data.
This Tips offers
advice for
preparing such a
plan.
Reengineering guidance requires operating units to prepare PMPs once their
strategic plans are approved. At a minimum, PMPs should include:
a detailed definition of each performance indicator
the source, method, frequency and schedule of data collection, and
the office, team, or individual responsible for ensuring data are
available on schedule
As part of the PMP process, it is also advisable (but not mandated) for
operating units to plan for:
how the performance data will be analyzed, and
how it will be reported, reviewed, and used to inform decisions
While PMPs are required, they are for the operating unit's own use. Review by
central or regional bureaus is not mandated, although some bureaus encourage
sharing PMPs. PMPs should be updated as needed to ensure plans, schedules,
and assignments remain current.
Why Are PMPs Important?
A performance monitoring plan is a critical tool for planning, managing, and
documenting data collection. It contributes to the effectiveness of the
performance monitoring system by assuring that comparable data will be
collected on a regular and timely basis. These are essential to the operation of a
credible and useful performance-based management approach.
PMPs promote the collection of comparable data by sufficiently documenting
indicator definitions, sources, and methods of data collection. This enables
operating units to collect comparable data over time even when key personnel
change.
PMPs support timely collection of data by documenting the frequency and
schedule of data collection as well as by assigning responsibilities. Operating
units should also consider developing plans for data analysis, reporting, and
review efforts as part of the PMP process. It makes sense to
PN-ABY-215
2
Use a Participatory Approach
The Agency's reengineering directives require that operating units involve USAID's partners, customers, and
stakeholders in planning approaches to monitoring performance. Experience indicates the value of collaborating
with relevant host government officials, implementing agency staff, contractors and grantees, other donors, and
customer groups, when preparing PMPs. They typically have the most familiarity with the quality, availability,
think through data collection, analysis, reporting, and
review as an integrated process. This will help keep the
performance monitoring system on track and ensure
performance data informs decision-making. While there
are strong arguments for including such integrated plans
in the PMP document, this is not mandated in the
reengineering guidance. Some operating units may wish
to prepare these plans separately.
Elements of a PMP
The following elements should be considered for
inclusion in a performance monitoring plan. Elements
1- 5 are required in the reengineering guidance, whereas
6 -9 are suggested as useful practices.
I. Plans for Data Collection (Required)
In its strategic plan, an operating unit will have identified
a few preliminary performance indicators for each of its
strategic objectives, strategic support objectives, and
special objectives (referred to below simply as SOs), and
USAID-supported intermediate results (IRs). In most
cases, preliminary baselines and targets will also have
been provided in the strategic plan. The PMP builds on
this initial information, verifying or modifying the
performance indicators, baselines and targets, and
documenting decisions.
PMPs are required to include information outlined below
(elements 1-5) on each performance indicator that has
been identified in the Strategic Plan for SOs and IRs.
Plans should also address how critical assumptions and
results supported by partners (such as the host
government, other donors, NGOs) will be monitored,
although the same standards and requirements for
developing indicators and collecting data do not apply.
Furthermore, it is useful to include in the PMP lowerlevel indicators of inputs, outputs, and processes at the
activity level, and how they will be monitored and
linked to IRs and SOs.
1. Performance Indicators and Their Definitions
Each performance indicator needs a detailed definition.
Be precise about all technical elements of the indicator
statement. As an illustration, consider the indicator,
number of small enterprises receiving loans from the
private banking system. How are small enterprises
defined -- all enterprises with 20 or fewer employees, or
50 or 100? What types of institutions are considered part
of the private banking sector -- credit unions,
government-private sector joint-venture financial
institutions?
Include in the definition the unit of measurement. For
example, an indicator on the value of exports might be
otherwise well defined, but it is also important to know
whether the value will be measured in current or constant
terms and in U.S. dollars or local currency.
The definition should be detailed enough to ensure that
different people at different times, given the task of
collecting data for a given indicator, would collect
identical types of data.
2. Data Source
Identify the data source for each performance indicator.
The source is the entity from which the data are obtained,
usually the organization that conducts the data collection
effort. Data sources may include government
departments, international organizations, other donors,
NGOs, private firms, USAID offices, contractors, or
activity implementing agencies.
Be as specific about the source as possible, so the same
source can be used routinely. Switching data sources for
the same indicator over time can lead to inconsistencies
and misinterpretations and should be avoided. For
example, switching from estimates of infant mortality
rates based on national sample surveys to estimates based
on hospital registration statistics can lead to false
impressions of change.
3
Plans may refer to needs and means for strengthening the
capacity of a particular data source to collect needed data
on a regular basis, or for building special data collection
efforts into USAID activities.
3. Method of Data Collection
Specify the method or approach to data collection for
each indicator. Note whether it is primary data collection
or is based on existing secondary data.
For primary data collection, consider:
the unit of analysis (individuals, families,
communities, clinics, wells)
data disaggregation needs (by gender, age, ethnic
groups, location)
sampling techniques for selecting cases (random
sampling, purposive sampling); and
techniques or instruments for acquiring data on
these selected cases (structured questionnaires,
direct observation forms, scales to weigh infants)
For indicators based on secondary data, give the method
of calculating the specific indicator data point and the
sources of data.
Note issues of data quality and reliability. For example,
using secondary data from existing sources cuts costs and
efforts, but its quality may not be as reliable.
Provide sufficient detail on the data collection or
calculation method to enable it to be replicated.
4. Frequency and Schedule of Data Collection
Performance monitoring systems must gather
comparable data periodically to measure progress. But
depending on the performance indicator, it may make
sense to collect data on a quarterly, annual, or less
frequent basis. For example, because of the expense and
because changes are slow, fertility rate data from sample
surveys may only be collected every few years whereas
data on contraceptive distributions and sales from clinics'
record systems may be gathered quarterly. PMPs can
also usefully provide the schedules (dates) for data
collection efforts.
When planning the frequency and scheduling of data
collection, an important factor to consider is
management's needs for timely information for decisionmaking.
5. Responsibilities for Acquiring Data
For each performance indicator, the responsibility the
operating unit for the timely acquisition of data from
their source should be clearly assigned to a particular
office, team, or individual.
II. Plans for Data Analysis, Reporting,
Review, and Use
An effective performance monitoring system needs to
plan not only for the collection of data, but also for data
analysis, reporting, review, and use. It may not be
possible to include everything in one document at one
time, but units should take the time early on for careful
planning of all these aspects in an integrated fashion.
6. Data Analysis Plans
To the extent possible, plan in advance how performance
data for individual indicators or groups of related
indicators will be analyzed. Identify data analysis
techniques and data presentation formats to be used.
Consider if and how the following aspects of data
analysis will be undertaken:
Comparing disaggregated data. For indicators with
disaggregated data, plan how it will be compared,
displayed, and analyzed.
Comparing current performance against multiple
criteria. For each indicator, plan how actual performance
data will be compared with a) past performance, b)
planned or targeted performance or
c) other relevant benchmarks.
Analyzing relationships among performance indicators.
Plan how internal analyses of the performance data will
examine interrelationships. For example
How will a set of indicators (if there are more
than one) for a particular SO or IR be analyzed
to reveal progress? What if only some of the
indicators reveal progress?
How will cause-effect relationships among SOs
and IRs within a results framework be analyzed?
How will USAID activities be linked to
achieving IRs and SOs?
Analyzing cost-effectiveness. When practical and
feasible, plan for using performance data to compare
systematically alternative program approaches in terms
of costs as well as results. The Government Performance
and Results Act (GPRA) encourages this.
4
7. Plans for Complementary Evaluations
Reengineering stresses that evaluations should be
conducted only if there is a clear management need. It
may not always be possible or desirable to predict years
in advance when or why they will be needed.
Nevertheless, operating units may find it useful to plan
on a regular basis what evaluation efforts are needed to
complement information from the performance
monitoring system. The operating unit's internal
performance reviews, to be held periodically during the
year, may be a good time for such evaluation planning.
For example, if the reviews reveal that certain
performance targets are not being met, and if the reasons
why are unclear, then planning evaluations to investigate
why would be in order.
8. Plans for Communicating and Using Performance
Information
Planning how performance information will be reported,
reviewed, and used is critical for effective managing for
results. For example, plan, schedule, and assign
responsibilities for internal and external reviews,
briefings, and reports. Clarify what, how and when
management decisions will consider performance
information. Specifically, plan for the following:
Operating unit performance reviews. Reengineering
guidance requires operating units to conduct internal
reviews of performance information at regular intervals
during the year to assess progress toward achieving SOs
and IRs. In addition, activity-level reviews should be
planned regularly by SO teams to assess if activities'
inputs, outputs, and processes are supporting
achievement of IRs and SOs.
USAID/Washington reviews and the R4 Report.
Reengineering requires operating units to prepare and
submit to USAID/Washington an annual Results Review
and Resource Request (R4) report, which is the basis for
a joint review with USAID/W of performance and
resource requirements. Help plan R4 preparation by
scheduling tasks and making assignments.
External reviews, reports, and briefings. Plan for
reporting and disseminating performance information to
key external audiences, such as host government
counterparts, collaborating NGOs, other partners, donors,
customer groups, and stakeholders. Communication
techniques may include reports, oral briefings,
videotapes, memos, newspaper articles.
Influencing management decisions. The ultimate aim of
performance monitoring systems is to promote
performance-based decision-making. To the extent
possible, plan in advance what management decisionmaking processes should be influenced by performance
information. For example, budget discussions,
programming decisions, evaluation designs/scopes of
work, office retreats, management contracts, and
personnel appraisals often benefit from the consideration
of performance information.
9. Budget
Estimate roughly the costs to the operating unit of
collecting, analyzing, and reporting performance data for
a specific indicator (or set of related indicators). Identify
the source of funds.
If adequate data are already available from secondary
sources, costs may be minimal. If primary data must be
collected at the operating unit's expense, costs can vary
depending on scope, method, and frequency of data
collection. Sample surveys may cost more than
$100,000, whereas rapid appraisal methods can be
conducted for much less. However, often these low-cost
methods do not provide quantitative data that are
sufficiently reliable or representative.
Reengineering guidance gives a range of 3 to 10 percent
of the total budget for an SO as a reasonable level to
spend on performance monitoring and evaluation.
CDIE's Tips series provides advice and
suggestions to USAID managers on how to
plan and conduct performance monitoring
and evaluation activities effectively. They
are supplemental references to the
reengineering automated directives system
(ADS), chapter 203. For further information, contact Annette Binnendijk, CDIE
Senior Evaluation Advisor, via phone
(703) 875-4235, fax (703) 875-4866, or email. Copies of TIPS can be ordered from
the Development Information Services
Clearinghouse by calling (703) 351-4006 or
by faxing (703) 351-4039. Please refer to
the PN number. To order via Internet,
address requests to
[email protected]
NUMBER 8
2ND EDITION, 2010
PERFORMANCE MONITORING & EVALUATION
TIPS
BASELINES AND TARGETS
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to performance
monitoring and evaluation. This publication is a supplemental reference to the Automated Directive
System (ADS) Chapter 203.
INTRODUCTION
The achievement of planned results
is at the heart of USAID’s
performance management system. In
order to understand where we, as
project managers, are going, we
need to understand where we have
been. Establishing quality baselines
and setting ambitious, yet achievable,
targets are essential for the
successful management of foreign
assistance programs.
WHAT ARE
BASELINES AND
TARGETS?
A baseline is the value of a
performance indicator before the
implementation of projects or
activities, while a target is the
specific, planned level of result to be
achieved
within
an
explicit
timeframe (see ADS 203.3.4.5).
Targets are set for indicators at the
Assistance
Objective
(AO),
Intermediate Result (IR), and output
levels.
WHY ARE
BASELINES
IMPORTANT?
Baselines help managers determine
progress in achieving outputs and
outcomes. They also help identify
the extent to which change has
happened at each level of result.
USAID ADS 203.3.3 requires a PMP
for each AO. Program managers
should provide baseline and target
values for every indicator in the
PMP.
Lack of baseline data not only
presents challenges for management
decision-making purposes, but also
hinders evaluation efforts.
For
example, it is generally not possible
to conduct a rigorous impact
1
evaluation without solid baseline
data (see TIPS 19: Rigorous Impact
Evaluation).
ESTABLISHING THE
BASELINE
Four common scenarios provide the
context for establishing baseline
data:
1. BASELINE IS
ESTABLISHED
If baseline data exist prior to the
start of a project or activity,
additional data collected over the
life of the project must be collected
in a consistent manner in order to
facilitate comparisons. For example,
consider the drop-out rate for girls 16
and under. If baseline data are
obtained from the Ministry of
Education, the project should
continue to collect these data from
this same source, ensuring that the
data
collection
remains the same.
methodology
Data may also be obtained from a
prior
implementing
partner’s
project, provided that the data
collection protocols, instruments,
and scoring procedures can be
replicated. For example, a policy
index might be used to measure
progress of legislation (see TIPS 14:
Monitoring the Policy Reform
Process). If these activities become
a part of a new project, program
managers should consider the
benefit
of using
the
same
instrument.
In cases where baseline data exist
from primary or secondary sources,
it is important that the data meet
USAID’s data quality standards for
validity,
reliability,
precision,
integrity, and timeliness (see TIPS
12: Data Quality Standards).
2. BASELINES MUST BE
COLLECTED
In cases where there are no existing
data with which to establish a
baseline,
USAID
and/or
its
implementing partners will have to
collect it if the required data are not
already being collected by, for
example,
a
host-country
government,
an
international
organization, or another donor.
Primary data collection can be
expensive, particularly if data are
collected through a formal survey or
Participation of key stakeholders in
setting targets helps establish a
common understanding about what
the project will accomplish and
when. USAID staff, implementing
partners, host country governments,
other donors, and civil society
partners, among others, should
attend working sessions at the outset
of program implementation to
review baseline data and other
information to set interim and final
targets.
a new index. Program managers
should consider this cost and
incorporate it into program or
project planning.
Ideally, data should be collected
prior to the initiation of the
program. If this is not feasible,
baselines should be collected as
soon as possible. For example, an
implementing partner may collect
perception data on the level of
corruption in targeted municipalities
for USAID’s PMP sixty days after
approval of a project’s work plan; in
another case, a score on an
advocacy capacity index may not be
collected until Community Service
Organizations (CSOs) are awarded
grants. If baseline data cannot be
collected until later in the course of
implementing an activity, the AO
Team should document when and
how the baseline data will be
collected (ADS 203.3.4.5).
3. BASELINES ARE
ESTABLISHED ON A
ROLLING BASIS
In some cases, it is possible to
collect baseline data on a rolling
basis as implementation proceeds.
For example, imagine that a health
project is being rolled out
sequentially across three provinces
over a three-year period. Data
collected in the first province will
serve as baseline for Year One; data
collected in the second province will
serve as baseline for the second
province in Year Two; and data
collected in the third province will
serve as baseline for that province in
Year Three.
The achievement of results requires
the joint action of many
stakeholders. Manageable interest
means we, as program managers,
have sufficient reason to believe that
the achievement of our planned
results can be significantly influenced
by interventions of USAID’s
program and staff resources. When
setting targets, take into account the
achievement of how other actors
will affect outcomes and what it
means for USAID to achieve
success.
program is the number of grants
awarded, the baseline is zero.
WHY ARE TARGETS
IMPORTANT?
Beyond
meeting
USAID
requirements, performance targets
are important for several reasons.
They help justify a program by
describing in concrete terms what
USAID’s investment will produce.
Targets orient stakeholders to the
tasks to be accomplished and
motivate individuals involved in a
program to do their best to ensure
the targets are met. Targets also
help to establish clear expectations
for USAID staff, implementing
partners, and key stakeholders.
Once a program is underway, they
serve as the guideposts for
monitoring whether progress is
being made on schedule and at the
levels originally envisioned. Lastly,
targets promote transparency and
accountability by making available
information on whether results have
been achieved or not over time.
4. BASELINE IS ZERO
For some indicators, baselines will
be zero. For example, if a new
program focuses on building the
teaching skills of teachers, the
baseline for the indicator “the
number of teachers trained” is zero.
Similarly, if an output of a new
2
A natural tension exists between the
need to set realistic targets and the
value,
from
a
motivational
perspective, of setting targets
ambitious enough to ensure that
staff and stakeholders will stretch to
meet them; when motivated, people
can often achieve more than they
imagine. Targets that are easily
achievable are not useful for
management and reporting purposes
since they are, in essence, pro forma.
AO Teams should plan ahead for
the analysis and interpretation of
actual
data
against
their
performance
targets
(ADS
203.3.4.5).
USING TARGETS
FOR
PERFORMANCE
MANAGEMENT IN A
LEARNING
ORGANIZATION
Targets can be important tools for
effective program management.
However, the extent to which
targets are or are not met should
not be the only criterion for judging
the success or failure of a program.
Targets are essentially flags for
managers; if the targets are wildly
exceeded
or
well-below
expectations, the program manager
should ask, “Why?”
Consider an economic growth
project. If a country experiences an
unanticipated downturn in its
economy,
the
underlying
FIGURE 1. PORTFOLIO
REVIEWS AND
PERFORMANCE TARGETS
To prepare for Portfolio Reviews,
AO Teams should conduct analysis of
program data, including achievement
of planned targets. ADS 203.3.7.2
provides illustrative questions for
these reviews:
• Are the desired results being
achieved?
• Are the results within USAID’s
manageable interest?
• Will planned targets be met?
• Is the performance management
system currently in place adequate
to capture data on the achievement
of results?
assumptions upon which that
project was designed may be
affected. If the project does not
meet targets, then it is important for
managers to focus on understanding
1) why targets were not met, and 2)
whether the project can be adjusted
to allow for an effective response to
changed circumstances.
In this
scenario, program managers may
need to reexamine the focus or
priorities of the project and make
related adjustments in indicators
and/or targets.
Senior
managers,
staff,
and
implementing
partners
should
review performance information and
targets as part of on-going project
management responsibilities and in
Portfolio Reviews (see Figure 1.)
TYPES OF TARGETS
FINAL AND INTERIM
TARGETS
A final target is the planned value of
a performance indicator at the end
of the AO or project. For AOs, the
final targets are often set three to
five years away, while for IRs they
are often set one to three years
away. Interim targets should be set
for the key points of time in
between the baseline and final target
in cases where change is expected
and data can be collected.
QUANTITATIVE AND
QUALITATIVE TARGETS
Targets may be either quantitative
or qualitative, depending on the
nature of the associated indicator.
Targets for quantitative indicators are
numerical, whereas targets and for
qualitative indicators are descriptive.
To
facilitate
comparison
of
baselines, targets, and performance
data for descriptive data, and to
maintain
data
quality,
some
indicators convert qualitative data
into a quantitative measure (see
Figure 2).
Nonetheless, baseline
and target data for quantitative and
3
FIGURE 2. TARGET
SETTING FOR
QUANTITATIVE AND
QUALITATIVE INDICATORS
- WHAT’S THE
DIFFERENCE?
Quantitative indicators and targets
are numerical. Examples include
the dropout rate, the value of
revenues, or number of children
vaccinated.
Qualitative indicators and targets
are descriptive. However,
descriptions must be based on a set
of pre-determined criteria. It is
much easier to establish baselines
and set targets when qualitative
data are converted into a
quantitative measure. For example,
the Advocacy Index is used to
measure the capacity of a target
organization, based on agreed-upon
standards that are rated and
scored. Other examples include
scales, indexes, and scorecards (see
Figure 3).
qualitative indicators must be
collected using the same instrument
so that change can be captured and
progress towards results measured
accurately (see TIPS 6: Selecting
Performance Indicators).
EXPRESSING
TARGETS
As with performance indicators,
targets can be expressed differently.
There are several possible ways to
structure
targets
to
answer
questions about the quantity of
expected change:
• Absolute level of achievement –
e.g., 75% of all trainees obtained
jobs by the end of the program or
7,000 people were employed by
the end of the program.
• Change in level of achievement –
e.g., math test scores for students
in grade nine increased by 10% in
Year One, or math test scores for
students in grade nine increased
FIGURE 3. SETTING TARGETS FOR QUALITATIVE MEASURES
For the IR Improvements in the Quality of Maternal and Child Health Services, a service delivery scale was used as the
indicator to measure progress. The scale, as shown below, transforms qualitative information about services into a rating
system against which targets can be set:
0 points = Service not offered
1 point = Offers routine antenatal care
1 point = Offers recognition and appropriate management of high risk pregnancies
1 point = Offers routine deliveries
1 point = Offers appropriate management of complicated deliveries
1 point = Offers post-partum care
1 point = Offers neonatal care
Score
= Total number of service delivery points
Illustrative Target: Increase average score to 5 by the end of year.
by three points in Year One.
Yields
per
hectare
under
improved management practices
increased by 25% or yields per
hectare increased by 100 bushels
from 2010 to 2013.
• Change in relation to the scale of
the problem – e.g., 35% of total
births in target area attended by
skilled health personnel by the end
of year two, or the proportion of
households with access to reliable
potable water increased by 50% by
2013.
• Creation
or
provision
of
something new – e.g., 4,000 doses
of tetanus vaccine distributed in
Year One, or a law permitting
non-government organizations to
generate income is passed by
2012.
Other targets may be concerned
with the quality of expected results.
Such targets can relate to indicators
measuring customer satisfaction,
public opinion, responsiveness rates,
enrollment rates, complaints, or
failure rates.
For example, the
average customer satisfaction score
for registration of a business license
(based on a seven-point scale)
increases to six by the end of the
program, or the percentage of
mothers who return six months
after delivery for postnatal care
increases to 20% by 2011.
Targets relating to cost efficiency or
producing outcomes at the least
expense are typically measured in
terms of unit costs. Examples of
such targets might include: cost of
providing
a
couple-year-ofprotection is reduced to $10 by
1999 or per-student costs of a
training program are reduced by
20% between 2010 and 2013.
demonstrate that:
DISAGGREGATING
TARGETS
A gender-sensitive indicator can be
defined as an indicator that captures
gender-related changes in society
over time. For example, a program
may focus on increasing enrollment
of children in secondary education.
Program managers may not only
want to look at increasing
enrollment rates, but also at the gap
between girls and boys. One way to
measure performance would be to
When a program’s progress is
measured in terms of its effects on
different segments of the population,
disaggregated targets can provide
USAID with nuanced information
that may not be obvious in the
aggregate. For example, a program
may seek to increase the number of
micro-enterprise loans received by
businesses in select rural provinces.
By disaggregating targets, program
inputs can be directed to reach a
particular target group.
Targets can be disaggregated along a
number of dimensions including
gender, location, income level,
occupation, administration level
(e.g., national vs. local), and social
groups.
For USAID programs, performance
management systems must include
gender-sensitive indicators and sexdisaggregated data when the
technical analyses supporting the
AO or project to be undertaken
4
• The different roles and status of
women and men affect the
activities differently; and
• The anticipated results of the
work would affect women and
men differently.
FIGURE 4. AN EXAMPLE OF
DISAGGREGATED TARGETS
FOR GENDER SENSITIVE
INDICATORS
Indicator: Number of children
graduating from secondary school;
percent gap between boys and girls.
B=boys; G=girls
Year
2010
(baseline)
Planned
2011
175
120B; 55G
50.0%
200
120B; 80G
25.0%
200
115B; 92G
2012
2013
Actual
145
115B; 30G
58.6%
160
120 B; 40G
56.3%
200
130 B; 70G
30.0%
205
110B; 95G
disaggregate the total number of
girls and boys attending school at
the beginning and at the end of the
school year (see Figure 4). Another
indicator might look at the quality of
the participation levels of girls vs.
boys with a target of increasing the
amount of time girls engage in
classroom discussions by two hours
per week.
Gender-sensitive indicators can use
qualitative
or
quantitative
methodologies to assess impact
directly on beneficiaries. They can
also be used to assess the
differential impacts of policies,
programs, or practices supported by
USAID on women and men (ADS
201.3.4.3).
Program managers should think
carefully about disaggregates prior
to collecting baseline data and
setting targets.
Expanding the
number of disaggregates can
increase the time and costs
associated with data collection and
analysis.
SETTING TARGETS
Targets
should
be
realistic,
evidence-based,
and
ambitious.
Setting meaningful targets provides
staff, implementing partners, and
stakeholders with benchmarks to
document
progress
toward
achieving results. Targets need to
take
into
account
program
resources,
the
implementation
period, and the development
hypothesis implicit in the results
framework.
PROGRAM RESOURCES
The level of funding, human
resources, material goods, and
institutional capacity contribute to
determining project outputs and
affecting change at different levels of
results and the AO. Increases or
decreases in planned program
resources should be considered
when setting targets.
ASSISTANCE OBJECTIVES
AND RESULTS
FRAMEWORKS
Performance
targets
represent
commitments that USAID AO
Teams make about the level and
timing of results to be achieved by a
program. Determining targets is
easier
when
objectives
and
indicators are within USAID’s
manageable interest.
Where a
result sits in the causal chain, critical
assumptions, and other contributors
to achievement of the AO will affect
targets.
Other key considerations include:
1. Historical Trends: Perhaps
even
more
important
than
examining a single baseline value, is
understanding
the
underlying
historical trend in the indicator
value over time. What pattern of
change has been evident in the past
five to ten years on the performance
indicator? Is there a trend, upward
FIGURE 5. PROGRESS IS NOT ALWAYS A STRAIGHT LINE
While it is easy to establish annual targets by picking an acceptable final
performance level and dividing expected progress evenly in the years between,
such straight-line thinking about progress is often inconsistent with the way
development programs really work. More often than not, no real progress – in
terms of measureable impacts or results – is evident during the start-up period.
Then, in the first stage of implementation, which may take the form of a pilot
test, some but not much progress is made, while the program team adjusts its
approaches. During the final two or three years of the program, all of this early
work comes to fruition. Progress leaps upward, and then rides a steady path at
the end of the program period. If plotted on a graph, it would look like “stair
steps,” not a straight line
5
or downward, that can be drawn
from existing reports, records, or
statistics? Trends are not always a
straight line; there may be a period
during which a program plateaus
before improvements are seen (see
Figure 5).
2. Expert Judgments: Another
option is to solicit expert opinions
as to what is possible or feasible
with respect to a particular indicator
and country setting. Experts should
be knowledgeable about the
program area as well as local
conditions. Experts will be familiar
with what is and what is not possible
from a technical and practical
standpoint – an important input for
any target-setting exercise.
3. Research Findings: Similarly,
reviewing development literature,
especially research and evaluation
findings, may help in choosing
realistic targets. In some program
areas, such as population and health,
extensive research findings on
development trends are already
widely available and what is possible
to achieve may be well-known. In
other areas, such as democracy,
research on performance indicators
and trends may be scarce.
4. Stakeholder Expectations:
While targets should be defined on
the basis of an objective assessment
of what can be accomplished given
certain conditions and resources, it
is useful to get input from
stakeholders regarding what they
want, need, and expect from USAID
activities.
What
are
the
expectations of progress? Soliciting
expectations may involve formal
interviews, rapid appraisals, or
informal conversations. Not only
end users should be surveyed;
intermediate
actors
(e.g.,
implementing agency staff) can be
especially useful in developing
realistic targets.
5. Achievement of Similar
Programs: Benchmarking is the
FIGURE 6. BENCHMARKING
One increasingly popular way of
setting targets and comparing
performance is to look at the
achievement of another program or
process by one or a collection of
high-performing organizations.
USAID is contributing to the
development of benchmarks for
programs such as water governance
(http://www.rewab.net), financial
management (www.fdirisk.com) and
health care systems
(www.healthsystems2020.org) Targets
may be set to reflect this “best in the
business” experience, provided of
course that consideration is given to
the comparability of country
conditions, resource availability, and
other factors likely to influence the
performance levels that can be
achieved.
process of comparing or checking
the progress of other similar
programs. It may be useful to
analyze progress of other USAID
Missions or offices, or other
development agencies and partners,
to understand the rate of change
that can be expected in similar
circumstances.
APPROACHES FOR
TARGET SETTING
There is no single best approach to
use when setting targets; the
process is an art and a science.
Although much depends on available
information, the experience and
knowledge of AO Team members
will add to the thinking behind
performance target.
Alternative
approaches include the following:
1. Projecting a future trend, then
adding the “valued added” by USAID
activities.
Probably the most
rigorous and credible approach, this
involves estimating the future trend
without USAID’s program, and then
adding whatever gains can be
expected as a result of USAID’s
efforts. This is no simple task, as
projecting the future can be very
tricky. The task is made somewhat
easier if historical data are available
and can be used to establish a trend
line.
2. Establishing a final performance
target for the end of the planning
period, and then planning the
progress from the baseline level.
This approach involves deciding on
the program’s performance target
for the final year, and then defining a
path of progress for the years in
between.
Final targets may be
judged on benchmarking techniques
or on judgments of experts,
program staff, customers, or
partners about the expectations of
what can be reasonably achieved
within the planning period. When
setting interim targets, remember
that progress is not always a straight
line. All targets, both final and
interim, should be based on a careful
analysis of what is realistic to
achieve, given the stage of program
implementation,
resource
availability,
country
conditions,
technical constraints, etc.
6
3. Setting annual performance
targets. Similar to the previous
approach, judgments are made
about what can be achieved each
year, instead of starting with a final
performance level and working
backwards. In both cases, consider
variations in performance, e.g.,
seasons and timing of activities and
expected results.
DOCUMENT AND
FILE
Typically, USAID project, baselines,
targets, and actual data are kept in a
data table for analysis either in the
PMP, as a separate document, or
electronically.
Furthermore, it is important to
document in the PMP how targets
were selected and why target values
were chosen.
Documentation
serves as a future reference for:
• Explaining
a
methodology.
target-setting
• Analyzing actual performance data.
• Setting targets in later years.
Responding to inquiries or audits
For more information:
TIPS publications are available online at [insert website].
Acknowledgements:
Our thanks to those whose experience and insights helped shape this publication, including Gerry Britan and
Subhi Mehdi of USAID’s Office of Management Policy, Budget and Performance (MPBP). This publication was
updated by Jill Tirnauer of Management Systems International.
Comments can be directed to:
Gerald Britan, Ph.D.
Tel: (202) 712-1158
[email protected]
Contracted under RAN-M-00-04-00049-A-FY0S-84
Integrated Managing for Results II
7
NUMBER 9
2011 Printing
PERFORMANCE MONITORING & EVALUATION
TIPS
CONDUCTING CUSTOMER SERVICE ASSESSMENTS
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to
performance monitoring and evaluation. This publication is a supplemental reference to the
Automated Directive Service (ADS) Chapter 203.
WHAT IS A CUSTOMER
SERVICE ASSESSMENT?
Under USAID’s new operations system, Agency operating units are required to
routinely and systematically
assess customer needs for,
perceptions of, and reactions
to USAID programs.
A customer service assessment is a management tool for understanding USAID’s programs
from the customer’s perspective. Most often
these assessments seek feedback from customers about a program’s service delivery performance. The Agency seeks views from both
ultimate customers (the end-users, or beneficiaries, of USAID activities—usually disadvantaged
groups) and intermediate customers (persons
or organizations using USAID resources, services, or products to serve the needs of the
ultimate customers).
This TIPS gives practical advice about customer service
assessments— for example,
when they should be conducted, what methods may
be used, and what information can be usefully included.
Customer service assessments may also be used
to elicit opinions from customers or potential
customers about USAID’s strategic plans, development objectives, or other planning issues.
1
For example, the operating unit may seek their
views on development needs and priorities to
help identify new, relevant activities.
WHO DOES CUSTOMER
SERVICE ASSESSMENTS?
WHY CONDUCT
CUSTOMER SERVICE
ASSESSMENTS?
USAID guidance specifies that all operating
units should develop a customer service plan.
The plan should include information about customers’ needs, preferences, and reactions as an
element in a unit’s planning, achieving, perfor-
USAID’s reengineered operating system calls
for regularly conducting customer service assessments for all program activities. Experience indicates that effective customer feedback
on service delivery improves performance,
achieves better results, and creates a more participatory working environment for programs,
and thus increases sustainability.
Box 1.The Customer Service Plan
The customer service plan presents the
operating unit’s vision for including customers and partners to achieve its objectives.
It explains how customer feedback will be
incorporated to determine customer needs
and perceptions
of services provided, and how this feedback
will be regularly incorporated into the unit’s
operations. The customer service plan is a
management tool for the operating unit and
does not require USAID/W approval.
Specifically, the plan
These assessments provide USAID staff with
the information they need for making constructive changes in the design and execution of development programs. This information may also
be shared with partners and customers as an
element in a collaborative, ongoing relationship.
In addition, customer service assessments provide input for reporting on results, allocating
resources, and presenting the operating unit’s
development programs to external audiences.
• Identifies the ultimate and intermediate customers for service delivery and
segments customer groups for different
programs, products, or services
Customer service assessments are relevant not
only to program-funded activities directed to
customers external to USAID. They can also
be very useful in assessing services provided to
internal USAID customers.
• Describes and regularly schedules appropriate means for assessing service
delivery, performance, and customer
satisfaction
Moreover, customer service assessments are
federally mandated. The Government Performance and Results Act of 1993 and Executive
Order 12862 of 1993 direct federal agencies to
reorient their programs toward achievement
of measurable results that reflect customers’
needs and to systematically assess those needs.
Agencies must report annually to the Administration on customer service performance.
• Establishes service principles and specifies measurable service performance
standards indicates staff responsibilities
for managing customer service activities—including assessments
• Specifies the resources required for customer service activities and assessments.
2
formance in delivering the program’s products
and services.
mance monitoring and evaluation functions (see
Box 1). Depending on the scope of its program
operations, an operating unit may find it needs
to plan several customer service assessments.
The various assessments might be tailored to
different strategic objectives, program activities
and services, or customer groups (differentiated, for example, by gender, ethnicity, or income).
Responsibility for designing and managing these
assessments typically is assigned to the relevant
development objective.
Unless the service or product delivery is satisfactory (i.e., timely, relevant, accessible, good
quality) from the perspective of the customers,
it is unlikely that the program will achieve its
substantive development results, which, after all,
ultimately depend on customers’ participation
and use of the service or product. For example,
a family planning program is unlikely to achieve
reduced fertility rates unless customers are satisfied with the contraceptive products
it offers and the delivery mechanism it uses to
provide them. If not sufficiently satisfied, customers will simply not use them.
HOW DO CUSTOMER
SERVICE ASSESSMENTS
COMPLEMENT PERFORMANCE MONITORING
AND EVALUATION?
Customer service assessments thus complement broader performance monitoring and
evaluation systems by monitoring a specific
type of result: service delivery performance
Performance monitoring and evaluation broad- from the customer’s perspective. By providing
ly addresses the results or outcomes of a pro- managers with information on whether cusgram.These results reflect objectives chosen by tomers are satisfied with and using a program’s
the operating unit (in consultation with part- products and services, these assessments are
ners and customer representatives) and may especially useful for giving early indications of
encompass several types of results.
whether longer term substantive development
results are likely to be met.
Often they are medium- to longer-term developmental changes or impacts. Examples: reduc- Both customer service assessments and perfortions in fertility rates, increases in income, im- mance monitoring and evaluation use the same
provements in agricultural yields, reductions in array of standard social science investigation
forest land destroyed.
techniques—surveys, rapid and participatory
appraisal, document reviews, and the like. In
some cases, the same survey or rapid appraisal
may even be used to gather both types of information. For example, a survey of customers of
an irrigation program might ask questions about
service delivery aspects (e.g., access, timeliness,
quality, use of irrigation water) and questions
concerning longer term development results
(e.g., yields, income).
Another type of result often included in performance monitoring and evaluation involves customer perceptions and responses to goods or
services delivered by a program— for example,
the percentage of women satisfied with the maternity care they receive, or the proportion of
farmers who have tried a new seed variety and
intend to use it again. Customer service assessments look at this type of result—customer
satisfaction, perceptions, preferences, and related opinions about the operating unit’s per3
planning the assessment should 1) identify the
purpose and intended uses of the information,
2) clarify the program products or services being assessed, 3) identify the customer groups
involved, and 4) define the issues the study will
address. Moreover, the scope of work typically discusses data collection methods, analysis
techniques, reporting and dissemination plans,
and a budget and time schedule.
STEPS IN CONDUCTING A
CUSTOMER SERVICE
ASSESSMENT
Step 1. Decide when the assessment
should be done.
Customer service assessments should be conducted whenever the operating unit requires
customer information for its management purposes. The general timing and frequency of customer service assessments is typically outlined
in the unit’s customer service plan.
Specific issues to be assessed will vary with the
development objective, program activities under way, socioeconomic conditions, and other
factors. However, customer service assessments generally aim at understanding
• Customer views regarding the importance
Customer service assessments are likely to
of various USAID-provided services (e.g.,
be most effective if they are planned to coortraining, information, commodities, technidinate with critical points in cycles associated
cal assistance) to their own needs and priwith the program being assessed (crop cycles,
orities
local school year cycles, host country fiscal year
cycles, etc.) as well as with the Agency’s own • Customer judgments, based on measurable
annual reporting and funding cycles.
service standards, on how well USAID is
performing service delivery
Customer service assessments will be most
valuable as management and reporting tools if • Customer comparisons of USAID service
they are carried out some months in advance of
delivery with that of other providers.
the operating unit’s annual planning and reporting process. For example, if a unit’s results re- Open-ended inquiry is especially well suited for
view and resources request (R4) report is to be addressing the first issue.The other two may be
completed by February, the customer service measured and analyzed quantitatively or qualiassessment might be conducted in November. tatively by consulting with ultimate or intermediate customers with respect to a number of
However, the precise scheduling and execution service delivery attributes or criteria important
of assessments is a task appropriate for those
responsible for results in a program sector—
members of the strategic objective or results
Box 2.
package team.
Illustrative Criteria For Assessing
Service Delivery
Step 2. Design the assessment.
Convenience. Ease of working with the
operating unit, simple processes, minimal
red tape, easy physical access to contacts
Depending on the scale of the effort, an operating unit may wish to develop a scope of work for
a customer service assessment. At a minimum,
4
Responsiveness. Follow up promptly, meet
changing needs, solve problems, answer questions, return calls
and program activity.
Reliability. On-time delivery that is thorough, accurate, complete
With its objective clearly in mind, and the information to be collected carefully specified, the
operating unit may decide in-house resources,
external assistance consultants, or a combination of the two, to conduct the assessment.
Step 3. Conduct the assessment.
Quality of products and services. Perform as intended; flexible in meeting local
needs; professionally qualified personnel
Contact personnel. Professional, knowledgable, understand local culture, language
skills
Select from a broad range of methods. A customer service assessment is not just a survey. It may
use a broad repertory of inquiry tools designed
to elicit information about the needs, preferences, or reactions of customers regarding a
USAID activity, product or service. Methods
may include the following:
to customer satisfaction (see Box 2).
• Formal customer surveys
In more formal surveys, for example, customers
may be asked to rate services and products on,
say, a 1-to-5 scale indicating their level of satisfaction with specific service characteristics or
attributes they consider important (e.g., quality,
reliability, responsiveness). In addition to rating
the actual services, customers may be asked
what they would consider “excellent” service,
referring to the same service attributes and using the same 5-point scale. Analysis of the gap
between what customers expect as an ideal
standard and what they perceive they actually
receive indicates the areas of service delivery
needing improvement.
• rapid appraisal methods (e.g., focus groups,
town meetings, interviews with key informants)
Breadth of choice. Sufficient choices to
meet customer needs and preferences
• Participatory appraisal techniques, in which
customers plan analyze, self-monitor, evaluate or set priorities for activities
• Document reviews, including systematic
use of social science research conducted by
others.
Use systematic research methods. A hastily prepared and executed effort does not provide
quality customer service assessment informaIn more qualitative approaches, such as focus tion. Sound social science methods are essengroups, customers discuss these issues among tial.
themselves while researchers listen carefully to
their perspectives. Operating units and teams Practice triangulation. To the extent resources
should design their customer assessments to and time permit, it is preferable to gather incollect customer feedback on service delivery formation from several sources and methods,
issues and attributes they believe are most im- rather than relying on just one. Such triangulaportant to achieving sustainable results toward tion will build confidence in findings and proa clearly defined strategic objective. These is- vide adequate depth of information for good
sues will vary with the nature of the objective decision-making and program management. In
5
particular, quantitative surveys and qualitative
studies often complement each other. Whereas
a quantitative survey can produce statistical
measurements of customer satisfaction (e.g.,
with quality, timeliness, or other aspects of a
program operation) that can be generalized
to a whole population, qualitative studies can
provide an in-depth understanding and insight
into customer perceptions and expectations on
these issues.
and encourage closer rapport with customers and partners. Moreover, they encourage a
more collaborative, participatory, and effective
approach to achievement of objectives.
Conduct assessments routinely. Customer service
assessments are designed to be consciously
iterative. In other words, they are undertaken
periodically to enable the operating unit to
build a foundation of findings over time to inform management of changing customer needs
and perceptions. Maintaining an outreach orientation will help the program adapt to changing
circumstances as reflected in customer views.
H. S. Plunkett and Elizabeth Baltimore, Customer
Focus Cookbook, USAID/M/ROR, August 1996.
Selected Further Reading
Resource Manual for Customer Surveys. Statistical
Policy Office, Office of Management and Budget. October 1993.
Zeithaml, Valarie A; A. Parasuraman; and Leonard L.Berry. Delivering Quality Service. New York:
Free Press
Step 4. Broadly disseminate and use
assessment findings to improve performance.
Customer service assessments gain value when
broadly disseminated within the operating unit,
to other operating units active in similar program sectors, to partners, and more widely
within USAID. Sharing this information is also
important to maintaining open, transparent relations with customers themselves.
Assessment findings provide operating unit
managers with insight on what is important to
customers and how well the unit is delivering
its programs. They also can help identify operations that need quality improvement, provide
early detection of problems, and direct attention to areas where remedial action may be
taken to improve delivery of services.
Customer assessments form the basis for review of and recommitment to service principles. They enable measurement of service
delivery performance against service standards
6
NUMBER 10
2011 Printing
PERFORMANCE MONITORING & EVALUATION
TIPS
CONDUCTING FOCUS GROUP INTERVIEWS
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to
peroformance monitoring and evaluation. This publication is a supplemental reference to the
Automated Directive Service (ADS) Chapter 203.
WHAT IS A FOCUS GROUP
INTERVIEW?
USAID’s guidelines encourage use of rapid, lowcost methods to collect
information on the
performance of
development assistance
activities.
A focus group interview is an inexpensive, rapid
appraisal technique that can provide managers with a wealth of qualitative information on
performance of development activities, services, and products, or other issues. A facilitator
guides 7 to 11 people in a discussion of their
experiences, feelings, and preferences about a
topic. The facilitator raises issues identified in
a discussion guide and uses probing techniques
to solicit views, ideas, and other information.
Sessions typically last one to two hours.
Focus group interviews,
the subject of this TIPS,
is one such method.
ADVANTAGES AND
LIMITATIONS
1
cannot be explained recommendations and
This technique has several advantages. It is low
suggestions are needed from customers,
cost and provides speedy results. Its flexible forpartners, experts, or other stakeholders
mat allows the facilitator to explore unanticipated issues and encourages interaction among
participants. In a group setting participants pro- For example, focus groups were used to unvide checks and balances, thus minimizing false cover problems in a Nepal family planning program where facilities were underutilized, and
or extreme views.
to obtain suggestions for improvements from
Focus groups have some limitations, however. customers.The focus groups revealed that rural
The flexible format makes it susceptible to fa- women considered family planning important.
cilitator bias, which can undermine the validity However, they did not use the clinics because of
and reliability of findings. Discussions can be caste system barriers and the demeaning mansidetracked or dominated by a few vocal individ- ner of clinic staff. Focus group participants suguals. Focus group interviews generate relevant gested appointing staff of the same social status
qualitative information, but no quantitative data to ensure that rural women were treated with
from which generalizations can be made for a respect. They also suggested that rural women
whole population. Moreover, the information disseminate information to their neighbors
can be difficult to analyze; comments should be about the health clinic.
interpreted in the context of the group setting.
Before deciding whether to use focus group interviews as a source of information, the study
WHEN ARE FOCUS GROUP purpose needs to be clarified. This requires
identifying who will use the information, deterINTERVIEWS USEFUL?
mining what information is needed, and understanding why the information is needed. Once
this is done, an appropriate methodology can
Focus group interviews can be useful in all phasbe selected. (See Tips 5 Using Rapid Appraisal
es of development activities— planning, impleMethods for additional information on selecting
mentation, monitoring, and evaluation. They can
appraisal techniques.)
be used to solicit views, insights, and recommendations of program staff, customers, stakeholders, technical experts, or other groups.
STEPS IN CONDUCTING
FOCUS GROUP
INTERVIEWS
They are especially appropriate when:
• program activities are being planned and it
is important for managers to understand
customers’ and other stakeholders’ attitudes, preferences or needs
Follow this step-by-step advice to help ensure
high-quality results.
• specific services or outreach approaches Step 1. Select the team
have to take into account customers’ prefConducting a focus group interview requires a
erences
small team, with at least a facilitator to guide
• major program implementation problems the discussion and a rapporteur to record it.
The facilitator should be a native speaker who
2
can put people at ease. The team should have Step 3. Decide on timing and location
substantive knowledge of the topic under discussion.
Discussions last one to two hours and should
be conducted in a convenient location with
Skills and experience in conducting focus some degree of privacy. Focus groups in a small
groups are also important. If the interviews village arouse curiosity and can result in uninare to be conducted by members of a broader vited participants. Open places are not good
evaluation team without previous experience spots for discussions.
in focus group techniques, training is suggested.
This training can take the form of role playing, Step 4. Prepare the discussion guide
formalized instruction on topic sequencing and
probing for generating and managing group dis- The discussion guide is an outline, prepared in
cussions, as well as pre-testing discussion guides advance, that covers the topics and issues to be
in pilot groups.
discussed. It should contain few items, allowing
some time and flexibility to pursue unanticipatStep 2. Select the participants
ed but relevant issues.
First, identify the types of groups and institutions that should be represented (such as program managers, customers, partners, technical experts, government officials) in the focus
groups. This will be determined by the informtion needs of the study. Often separate focus
groups are held for each type of group. Second,
identify the most suitable people in each group.
One of the best approaches is to consult key
informants who know about local conditions.
It is prudent to consult several informants to
minimize the biases of individual preferences.
Excerpt from a Discussion
Guide on Curative
Health Services
(20-30 minutes)
Q. Who treats/cures your children
when they get sick? Why?
Note: Look for opinions about
Each focus group should be 7 to 11 people to
allow the smooth flow of conversation.
• outcomes and results
• provider-user relations
• costs (consultations, transportation, medicine)
• waiting time
• physical aspects (privacy, cleanliness)
• availability of drugs, lab services
• access (distance, availability of
transportation)
• follow-up at home
Participants should be homogenous, from similar socioeconomic and cultural backgrounds.
They should share common traits related to the
discussion topic. For example, in a discussion
on contraceptive use, older and younger women should participate in separate focus groups.
Younger women may be reluctant to discuss
sexual behavior among their elders, especially if
it deviates from tradition. Ideally, people should
not know each other. Anonymity lowers inhibition and prevents formation of cliques.
3
The guide provides the framework for the facilitator to explore, probe, and ask questions.
Initiating each topic with a carefully crafted
question will help keep the discussion focused.
Using a guide also increases the comprehensiveness of the data and makes data collection
more efficient. Its flexibility, however can mean
that different focus groups are asked different
questions, reducing the credibility of the findings. An excerpt from a discussion guide used
in Bolivia to assess child survival services provides an illustration. (See box on page 3)
• What do you think about corruption in the
criminal justice system?
• How do you feel about the three parties
running in upcoming national elections?
Use probing techniques. When participants give
incomplete or irrelevant answers, the facilitator
can probe for fuller, clearer responses. A few
suggested techniques:
Repeat the question—repetition gives more time
to think
Step 5. Conduct the interview
Adopt sophisticated naivete” posture—convey
limited understanding of the issue and ask for
specific details
Establish rapport. Often participants do not
know what to expect from focus group discussions. It is helpful for the facilitator to outline
the purpose and format of the discussion at the
beginning of the session, and set the group at
ease. Participants should be told that the discussion is informal, everyone is expected to
participate, and divergent views are welcome.
Pause for the answer—a thoughtful nod or expectant look can convey that you want a fuller
answer
Repeat the reply—hearing it again sometimes
Phrase questions carefully. Certain types of ques- stimulates conversation. Ask when, what,
tions impede group discussions. For example, where, which, and how questions—they proyes-or-no questions are one dimensional and voke more detailed information
do not stimulate discussion. “Why” questions
put people on the defensive and cause them to Use neutral comments— Anything else?” Why do
take “politically correct” sides on controversial you feel this way?”
issues.
Control the discussion. In most groups a few indiOpen-ended questions are more useful be- viduals dominate the discussion. To balance out
cause they allow participants to tell their story participation:
in their own words and add details that can re• Address questions to individuals who are
sult in unanticipated findings. For example:
reluctant to talk
• What do you think about the criminal justice system?
• Give nonverbal cues (look in another direction or stop taking notes when an individual
• How do you feel about the upcoming natalks for an extended period)
tional elections?
• Intervene, politely summarize the point,
If the discussion is too broad the facilitator can
then refocus the discussion
narrow responses by asking such questions as:
4
• Take advantage of a pause and say, “Thank
you for that interesting idea, perhaps we can
discuss it in a separate session. Meanwhile
with your consent, I would like to move on
to another item.”
trends andpatterns, strongly held or frequently
aired opinions.
Read each transcript. Highlight sections that correspond to the discussion guide questions and
mark comments that could be used in the final
report.
Minimize group pressure. When an idea is being
adopted without any general discussion or disagreement, more than likely group pressure is
occurring. To minimize group pressure the facilitator can probe for alternate views. For example, the facilitator can raise another issue, or
say, “We had an interesting discussion but let’s
explore other alter natives.”
Analyze each question separately. After reviewing
all the responses to a question or topic, write a
summary statement that describes the discussion. In analyzing the results, the team should
consider:
• Words. Weigh the meaning of words participants used. Can a variety of words and
phrases categorize similar responses?
Step 6. Record the discussion
A rapporteur should perform this function.
Tape recordings in conjunction with written • Framework. Consider the circumstances in
notes are useful. Notes should be extensive
which a comment was made (context of
and reflect the content of the discussion as well
previous discussions, tone and intensity of
as nonverbal behavior (facial expressions, hand
the comment).
movements).
• Internal agreement. Figure out whether shifts
Shortly after each group interview, the team
in opinions during the discussion were
should summarize the information, the team’s
caused by group pressure.
impressions, and implications of the information for the study.
• Precision of responses. Decide which responses were based on personal experience and
Discussion should be reported in participants’
give them greater weight than those based
language, retaining their phrases and grammation vague impersonal impressions.
cal use. Summarizing or paraphrasing responses
can be misleading. For instance, a verbatim reply • The big picture. Pinpoint major ideas. Allo“Yes, indeed! I am positive,” loses its intensity
cate time to step back and reflect on major
when recorded as “Yes.”
findings.
Step 7. Analyze results
• Purpose of the report. Consider the objectives of the study and the information
needed for decisionmaking. The type and
After each session, the team should assemble
scope of reporting will guide the analytical
the interview notes (transcripts of each focus
process. For example, focus group reports
group interview), the summaries, and any other
typically are: (1) brief oral reports that highrelevant data to analyze trends and patterns.
light key findings; (2) descriptive reports
The following method can be used.
that summarize the discussion; and (3) analytical reports that provide trends, patterns,
Read summaries all at one time. Note potential
5
or findings and include selected comments.
Focus Group Interviews of Navarongo Community
Health and Family Planning Project in Ghana
The Ghanaian Ministry of Health launched a small pilot project in three villages
in 1994 to assess community reaction to family planning and elicit community
advice on program design and management. A new model of service deliverywas introduced: community health nurses were retrained as community health
officers living in the communities and providing village-based clinical services.
Focus group discussions were used to identify constraints to introducing family planning services and clarify ways to design operations that villagers value.
Discussions revealed that many women want more control over their ability
to reproduce, but believe their preferences are irrelevant to decisions made
in the male dominated lineage system. This indicated that outreach programs
aimed primarily at women are insufficient. Social groups must be included to
legitimize and support individuals’ family-planning decisions. Focus group discussions also revealed women’s concerns about the confidentiality of information and services. These findings preclude development of a conventional communitybased distribution program, since villagers clearly prefer outside service
delivery workers to those who are community members.
Selected Further Reading
Krishna Kumar, Conducting Group Interviews in
Developing Countries, A.I.D. Program Design and
Evaluation Methodology Report No. 8, 1987
(PN-AAL-088)
Richard A. Krueger, Focus Groups: A Practical
Guide for Applied Research, Sage Publications,
1988
6
2009, NUMBER 12
2ND EDITION
PERFORMANCE MONITORING & EVALUATION
TIPS
DATA QUALITY STANDARDS
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to
performance monitoring and evaluation. This publication is a supplemental reference to the
Automated Directive System (ADS) Chapter 203.
WHY IS DATA
QUALITY
IMPORTANT?
Results-focused development
programming
requires
managers to design and
implement programs based
on evidence. Since data play a
central role in establishing
effective
performance
management systems, it is
essential to ensure good data
quality
(see
Figure
1).
Without this, decision makers
do not know whether to have
confidence in the data, or
worse, could make decisions
based on misleading data.
Attention to data quality
assists in:
Figure 1. Data Quality Plays a Central Role in Developing
Effective Performance Management Systems
Cycle:
Plan: Identify or Refine Key Program Objectives
Design: Develop or Refine the Performance Management Plan
Analyze Data
Use Data: Use Findings from Data Analysis to Improve Program Effectiveness
Data
Quality
 Ensuring
that
limited
development resources are
used as effectively as
possible
1
 Ensuring
that
Agency
program
and
budget
decisions in Washington
and the field are as well
informed
possible
as
practically
 Meeting the requirements
of
the
Government
Performance and Results
Act (GPRA)
 Reporting the impact of
USAID programs to external
stakeholders,
including
senior management, OMB,
the Congress, and the
public with confidence
DATA QUALITY
STANDARDS
Data quality is one element of
a
larger
interrelated
performance
management
system. Data quality flows
from a well designed and
logical strategic plan where
Assistance Objectives (AOs)
and Intermediate Results (IRs)
are clearly identified.
If a
result is poorly defined, it is
difficult to identify quality
indicators,
and
further,
without quality indicators, the
resulting data will often have
data quality problems.
One key challenge is to
determine what level of data
quality is acceptable (or “good
enough”) for management
purposes. It is important to
understand that we rarely
require the same degree of
rigor as needed in research or
for laboratory experiments.
Standards for data quality
must be keyed to our
intended use of the data. That
is, the level of accuracy,
currency,
precision,
and
reliability
of
performance
The Five Data Quality
Standards
1. Validity
2. Reliability
3. Precision
4. Integrity
5. Timeliness
information
should
be
consistent
with
the
requirements
of
good
management.
Determining
appropriate
or
adequate
thresholds of indicator and
data quality is not an exact
science.
This task is made
even more difficult by the
complicated and often datapoor development settings in
which USAID operates.
As
with
performance
indicators, we sometimes have
to consider trade-offs, or
make informed judgments,
when applying the standards
for data quality. This is
especially true if, as is often
the case, USAID relies on
others to provide data for
indicators. For example, if our
only existing source of data
for a critical economic growth
indicator is the Ministry of
Finance, and we know that the
Ministry’s
data
collection
methods are less than perfect,
we may have to weigh the
alternatives of relying on lessthan-ideal data, having no
data at all, or conducting a
potentially
costly
USAIDfunded
primary
data
collection effort. In this case,
2
a decision must be made as to
whether the Ministry’s data
would allow the Assistance
Objective team to have
confidence when assessing
program
performance
or
whether they are so flawed as
to be useless, or perhaps
misleading, in reporting and
managing for results.
The
main point is that managers
should not let the ideal drive
out the good.
1. VALIDITY
Validity refers to the extent to
which a measure actually
represents what we intend to
measure.1
Though simple in principle,
validity can be difficult to
assess in practice, particularly
when
measuring
social
phenomena.
For example,
how can we measure political
power or sustainability? Is the
poverty gap a good measure
of the extent of a country’s
poverty? However, even valid
indicators have little value, if
the data collected do not
correctly measure the variable
or characteristic encompassed
by the indicator. It is quite
possible, in other words, to
identify valid indicators but to
then
collect
inaccurate,
unrepresentative,
or
incomplete data.
In such
cases, the quality of the
indicator is moot. It would be
equally undesirable to collect
1
This criterion is closely related
to “directness” criteria for
indicators.
good data
indicator.
for
an
invalid
There are a number of ways to
organize or present concepts
related to data validity. In the
USAID context, we focus on
three key dimensions of
validity that are most often
relevant
to
development
programming, including: face
validity,
attribution,
and
measurement error.
FACE VALIDITY
Face validity means that an
outsider or an expert in the
field would agree that the
data is a true measure of the
result. For data to have high
face validity, the data must be
true representations of the
indicator, and the indicator
must be a valid measure of
the result. For example:
Result:
Increased
household income in a
target district
Indicator:
Value
of
median household income
in the target district
In this case, the indicator has a
high degree of face validity
when compared to the result.
That is, an external observer is
likely to agree that the data
measure
the
intended
objective. On the other hand,
consider
the
following
example:
Result:
Increased
household income in a
target district
Indicator:
Number of
houses in the target
community with tin roofs
This example does not appear
to have a high degree of face
validity as a measure of
increased income, because it
is not immediately clear how
tin roofs are related to
increased income.
The
indicator above is a proxy
indicator
for
increased
income.
Proxy indicators
measure results indirectly, and
their validity hinges on the
assumptions made to relate
the indicator to the result. If
we assume that 1) household
income data are too costly to
obtain and 2) research shows
that when the poor have
increased income, they are
likely to spend it on tin roofs,
then this indicator could be an
appropriate
proxy
for
increased income.
ATTRIBUTION
Attribution focuses on the
extent to which a change in
the data is related to USAID
interventions. The concept of
attribution is discussed in
detail as a criterion for
indicator
selection,
but
reemerges when assessing
validity.
Attribution means
that changes in the data can
be plausibly associated with
USAID interventions.
For
example, an indicator that
measures changes at the
national level is not usually
appropriate for a program
targeting a few areas or a
particular segment of the
3
population.
following:
Consider the
Result:
revenues
in
municipalities.
Increased
targeted
Indicator: Number of
municipalities where tax
revenues have increased
by 5%.
In this case, assume that
increased
revenues
are
measured
among
all
municipalities
nationwide,
while the program only
focuses on a targeted group
of municipalities. This means
that the data would not be a
valid measure of performance
because the overall result is
not reasonably attributable to
program activities.
MEASUREMENT ERROR
Measurement error results
primarily from the poor
design or management of
data collection processes.
Examples include leading
questions,
unrepresentative
sampling,
or
inadequate
training of data collectors.
Even if data have high face
validity, they still might be an
inaccurate measure of our
result due to bias or error in
the measurement process.
Judgments about acceptable
measurement error should
reflect technical assessments
about what level of reductions
in measurement error are
possible and practical. This
can be assessed on the basis
of cost as well as management
judgments about what level of
accuracy
decisions.
is
needed
for
Some degree of measurement
error is inevitable, particularly
when dealing with social and
economic changes, but the
level of measurement error
associated
with
all
performance data collected or
used by operating units
should not be so large as to 1)
call into question either the
direction or degree of change
reflected by the data or 2)
overwhelm the amount of
anticipated change in an
indicator
(making
it
impossible for managers to
determine whether progress.
reflected in the data is a result
of actual change or of
measurement error). The two
main sources of measurement
error are sampling and nonsampling error.
Sampling Error (or
representativeness)
Data
are
said
to
be
representative
if
they
accurately
reflect
the
population they are intended
to
describe.
The
representativeness of data is a
function of the process used
to select a sample of the
population from which data
will be collected.
It is often not possible, or
even desirable, to collect data
from
every
individual,
household, or community
involved in a program due to
resource
or
practical
constraints. In these cases,
data are collected from a
sample to infer the status of
the population as a whole. If
we are interested in describing
the characteristics of a
country’s primary schools, for
example, we would not need
to examine every school in the
country. Depending on our
focus, a sample of a hundred
schools might be enough.
However, when the sample
used to collect data are not
representative
of
the
population as a whole,
significant
bias
can
be
introduced into the data. For
example, if we only use data
from 100 schools in the capital
area of the country, our data
will
not
likely
be
representative of all primary
schools in the country.
Drawing a sample that will
allow managers to confidently
generalize data/findings to
the population requires that
two basic criteria are met: 1)
that all units of a population
(e.g., households, schools,
enterprises) have an equal
chance of being selected for
the sample and 2) that the
sample is of adequate size.
The sample size necessary to
ensure that resulting data are
representative to any specified
degree can vary substantially,
depending on the unit of
analysis, the size of the
population, the variance of the
characteristics being tracked,
and
the
number
of
characteristics that we need to
analyze. Moreover, during
data collection it is rarely
possible to obtain data for
every member of an initially
4
chosen sample. Rather, there
are established techniques for
determining acceptable levels
of non-response or for
substituting new respondents.
If a sample is necessary, it is
important for managers to
consider the sample size and
method relative to the data
needs.
While data validity
should always be a concern,
there may be situations where
accuracy is a particular
priority. In these cases, it may
be useful to consult a
sampling expert to ensure the
data are representative.
Non-Sampling Error
Non-sampling error includes
poor design of the data
collection instrument, poorly
trained
or
partisan
enumerators, or the use of
questions (often related to
sensitive subjects) that elicit
incomplete
or
untruthful
answers from respondents.
Consider the earlier example:
Result: Increased
household income in a
target district
Indicator: Value of
median household
income in the target
district
While these data appear to
have high face validity, there is
the potential for significant
measurement error through
reporting bias. If households
are asked about their income,
they might be tempted to
income
to
under-report
demonstrate the need for
additional assistance (or overreport
to
demonstrate
success). A similar type of
reporting bias may occur
when data is collected in
groups or with observers, as
respondents may modify their
responses to match group or
observer norms. This can be a
particular source of bias when
collecting data on vulnerable
groups. Likewise, survey or
interview
questions
and
sequencing
should
be
developed in a way that
minimizes the potential for
the leading of respondents to
predetermined responses. In
order to minimize nonsampling measurement error,
managers should carefully
plan and vet the data
collection process with a
careful eye towards potential
sources of bias.
Minimizing Measurement
Error
Keep in mind that USAID is
primarily
concerned
with
learning,
with
reasonable
confidence, that anticipated
improvements have occurred,
not with reducing error below
some arbitrary level. 2 Since it
is impossible to completely
eliminate measurement error,
and reducing error tends to
become
increasingly
expensive or difficult, it is
important to consider what an
2
For additional information, refer
to Common Problems/Issues with
Using Secondary Data in the CDIE
Resource Book on Strategic
Planning and Performance
Monitoring, April 1997.
acceptable level of error
would be.
Unfortunately,
there is no simple standard
that can be applied across all
of the data collected for
USAID’s varied programs and
results.
As
performance
management plans (PMPs) are
developed, teams should:
 Identify the existing or
potential sources of error
for each indicator and
document this in the PMP.
 Assess how this error
compares
with
the
magnitude of expected
change. If the anticipated
change is less than the
measurement error, then
the data are not valid.
 Decide whether alternative
data sources (or indicators)
need to be explored as
better alternatives or to
complement the data to
improve data validity.
2. RELIABILITY
Data should reflect stable and
consistent
data
collection
processes
and
analysis
methods over time.
Reliability is important so that
changes in data can be
recognized as true changes
rather than reflections of poor
or changed data collection
methods. For example, if we
use
a
thermometer
to
measure a child’s temperature
repeatedly and the results
vary from 95 to 105 degrees,
even though we know the
child’s temperature hasn’t
changed, the thermometer is
5
not a reliable instrument for
measuring fever.
In other
words, if a data collection
process is unreliable due to
changes in the data collection
instrument,
different
implementation across data
collectors, or poor question
choice, it will be difficult for
managers to determine if
changes in data over the life
of the project reflect true
changes or random error in
the data collection process.
Consider
the
following
examples:
Indicator: Percent
increase in income
among target
beneficiaries.
The first year, the project
reports
increased
total
income, including income as a
result of off-farm resources.
The second year a new
manager is responsible for
data collection, and only farm
based income is reported.
The third year, questions arise
as to how “farm based
income” is defined. In this
case, the reliability of the data
comes into question because
managers are not sure
whether changes in the data
are due to real change or
changes in definitions. The
following is another example:
Indicator: Increased
volume of agricultural
commodities sold by
farmers.
A scale is used to measure
volume
of
agricultural
commodities sold in the
What’s the Difference
Between Validity and
Reliability?
Validity refers to the
extent to which a
measure actually
represents what we
intend to measure.
Reliability refers to the
stability of the
measurement process.
That is, assuming there is
no real change in the
variable being measured,
would the same
measurement process
provide the same result if
the process were
repeated over and over?
market. The scale is jostled
around in the back of the
truck. As a result, it is no
longer properly calibrated at
each stop. Because of this,
the scale yields unreliable
data, and it is difficult for
managers
to
determine
whether changes in the data
truly reflect changes in
volume sold.
3. PRECISION
Precise data have a sufficient
level of detail to present a fair
picture of performance and
enable management decisionmaking.
The level of precision or detail
reflected in the data should be
smaller (or finer) than the
margin of error, or the tool of
measurement is considered
too imprecise. For some
indicators, for which the
magnitude
of
expected
change is large, even relatively
large measurement errors may
be perfectly tolerable; for
other
indicators,
small
amounts of change will be
important and even moderate
levels of measurement error
will be unacceptable.
Example: The number of
politically
active
nongovernmental organizations
(NGOs) is 900. Preliminary
data shows that after a few
years this had grown to
30,000 NGOs. In this case, a
10 percent measurement error
(+/- 3,000 NGOs) would be
essentially
irrelevant.
Similarly, it is not important to
know precisely whether there
are 29,999 or 30,001 NGOs. A
less precise level of detail is
still sufficient to be confident
in the magnitude of change.
Consider
an
alternative
scenario. If the second data
point is 1,000, a 10 percent
measurement error (+/- 100)
would
be
completely
unacceptable
because
it
would represent all of the
apparent change in the data.
4. INTEGRITY
Integrity focuses on whether
there is improper manipulation
of data.
Data that are collected,
analyzed and reported should
have established mechanisms
in
place
to
reduce
manipulation.
There are
generally two types of issues
that affect data integrity. The
first is transcription error. The
second, and somewhat more
complex issue, is whether
there is any incentive on the
6
part of the data source to
manipulate the data for
political or personal reasons.
Transcription Error
Transcription error refers to
simple data entry errors made
when transcribing data from
one document (electronic or
paper) or database to another.
Transcription
error
is
avoidable,
and
Missions
should seek to eliminate any
such error when producing
internal or external reports
and other documents. When
the data presented in a
document produced by an
operating unit are different
from the data (for the same
indicator and time frame)
presented in the original
source simply because of data
entry or copying mistakes, a
transcription
error
has
occurred.
Such differences
(unless due to rounding) can
be easily avoided by careful
cross-checking of data against
the original source. Rounding
may result in a slight
difference from the source
data but may be readily
justified when the underlying
data do not support such
specificity, or when the use of
the data does not benefit
materially from the originally
reported level of detail. (For
example, when making cost or
budget
projections,
we
typically
round
numbers.
When we make payments to
vendors, we do not round the
amount
paid
in
the
accounting ledger. Different
purposes can accept different
levels of specificity.)
Technology can help to
reduce transcription error.
Systems can be designed so
that the data source can enter
data directly into a database—
reducing the need to send in a
paper report that is then
entered into the system.
However, this requires access
to computers and reliable
internet services. Additionally,
databases can be developed
with internal consistency or
range checks to minimize
transcription errors.
The use of preliminary or
partial data should not be
confused with transcription
error. There are times, where
it makes sense to use partial
data (clearly identified as
preliminary or partial) to
inform management decisions
or to report on performance
because these are the best
data currently available. When
preliminary or partial data are
updated by the original
source, USAID should quickly
follow suit, and note that it
has done so. Any discrepancy
between preliminary data
included in a dated USAID
document and data that were
subsequently updated in an
original source does not
constitute transcription error.
Manipulation
A somewhat more complex
issue is whether data is
manipulated.
Manipulation
should be considered 1) if
there may be incentive on the
part of those that report data
to skew the data to benefit
the project or program and
managers suspect that this
may be a problem, 2) if
managers
believe
that
numbers
appear
to
be
unusually favorable, or 3) if
the data are of high value and
managers want to ensure the
integrity of the data.
There are a number of ways in
which managers can address
manipulation.
First, simply
understand the data collection
process. A well organized and
structured process is less likely
to be subject to manipulation
because each step in the
process is clearly documented
and handled in a standard
way. Second, be aware of
potential issues. If managers
have reason to believe that
data are manipulated, then
they should further explore
the issues. Managers can do
this by periodically spot
checking or verifying the data.
This establishes a principle
that the quality of the data is
important and helps to
determine
whether
manipulation is indeed a
problem.
If there is
substantial concern about this
issue,
managers
might
conduct a Data Quality
Assessment (DQA) for the AO,
IR, or specific data in question.
Example: A project assists
the Ministry of Water to
reduce
water
loss
for
agricultural use. The Ministry
reports key statistics on water
loss to the project. These
statistics are critical for the
Ministry, the project and
USAID to understand program
performance. Because of the
7
importance of the data, a
study is commissioned to
examine data quality and
more specifically whether
there is any tendency for the
data to be inflated. The study
finds that there is a very slight
tendency to inflate the data,
but it is within an acceptable
range.
5. TIMELINESS
Data should be available and
up to date enough to meet
management needs.
There are two key aspects of
timeliness. First, data must be
available frequently enough
to influence management
decision
making.
For
performance indicators for
which annual data collection is
not practical, operating units
will collect data regularly, but
at longer time intervals.
Second, data should be
current or, in other words,
sufficiently up to date to be
useful in decision-making. As
a general guideline, data
should lag no more than three
years.
Certainly, decisionmaking should be informed
by the most current data that
are
practically
available.
Frequently,
though,
data
obtained from a secondary
source, and at times even
USAID-funded primary data
collection,
will
reflect
substantial time lags between
initial data collection and final
analysis and publication. Many
of these time lags are
unavoidable,
even
if
considerable
additional
resources
were
to
be
expended.
Sometimes
preliminary estimates may be
obtainable, but they should be
clearly flagged as such and
replaced as soon as possible
as the final data become
available from the source.
The
following
example
demonstrates issues related to
timeliness:
Result: Primary school
attrition in a targeted
region reduced.
Indicator: Rate of
student attrition at
targeted schools.
In August 2009, the Ministry
of Education published full
enrollment analysis for the
2007 school year.
In this case, currency is a
problem because there is a 2
year time lag for these data.
While it is optimal to collect
and report data based on the
U.S. Government fiscal year,
there are often a number of
practical challenges in doing
so. We recognize that data
may come from preceding
calendar or fiscal years.
Moreover, data often measure
results for the specific point in
time that the data were
collected, not from September
to September, or December to
December.
Often the realities of the
recipient country context will
dictate the appropriate timing
of the data collection effort,
rather than the U.S. fiscal year.
For example, if agricultural
yields are at their peak in July,
then data collection efforts to
measure yields should be
conducted in July of each
year. Moreover, to the extent
that
USAID
relies
on
secondary data sources and
partners for data collection,
we may not be able to dictate
exact timing
ASSESSING DATA
QUALITY
Approaches and steps for how
to assess data quality are
discussed in more detail in
TIPS 18: Conducting Data
Quality Assessments. USAID
policy requires managers to
understand the strengths and
weaknesses of the data they
use on an on-going basis. In
addition, a Data Quality
Assessment (DQA) must be
conducted at least once every
3 years for those data
reported to Washington (ADS
203.3.5.2).
For more information:
TIPS publications are available online at [insert website]
Acknowledgements:
Our thanks to those whose experience and insights helped shape this publication including Gerry Britan
and Subhi Mehdi of USAID’s Office of Management Policy, Budget and Performance (MPBP). This
publication was updated by Michelle Adams-Matson of Management Systems International (MSI).
Comments regarding this publication can be directed to:
Gerald Britan, Ph.D.
Tel: (202) 712-1158
[email protected]
Contracted under RAN-M-00-04-00049-A-FY0S-84
Integrated Managing for Results II
8
NUMBER 13
2ND EDITION, 2010 DRAFT
PERFORMANCE MONITORING & EVALUATION
TIPS
BUILDING A RESULTS FRAMEWORK
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to
performance monitoring and evaluation. This publication is a supplemental reference to the
Automated Directive System (ADS) Chapter 203.
WHAT IS A RESULTS
FRAMEWORK?
The Results Framework (RF) is a
graphic representation of a
strategy to achieve a specific
objective that is grounded in
cause-and-effect logic.
The RF
includes the Assistance Objective
(AO) and Intermediate Results
(IRs), whether funded by USAID
or partners, necessary to achieve
the objective (see Figure 1 for an
example). The RF also includes
the critical assumptions that must
hold true for the strategy to
remain valid.
The
Results
represents
Framework
a development hypothesis or a
theory about how intended
change will occur. The RF shows
how the achievement of lower
level objectives (IRs) leads to the
achievement of the next higher
order of objectives, ultimately
resulting in the AO.
In short, a person looking at a
Results Framework should be
able to understand the basic
theory for how key program
objectives will be achieved. The
Results
Framework
is
an
important tool because it helps
managers identify and focus on
key objectives within a complex
development environment.
1
A RESULTS FRAMEWORK
INCLUDES:

An Assistance Objective (AO)

Intermediate Results (IR)

Hypothesized cause and
effect linkages

Critical Assumptions
WHY IS THE RESULTS
FRAMEWORK
IMPORTANT?
The development of a Results
Framework
represents
an
important first step in forming
the actual strategy. It facilitates
analytic thinking and helps
What’s the Difference
Between a Results Framework
and the Foreign Assistance
Framework (FAF)?
In one word, accountability. The
results framework identifies an
objective that a Mission or Office
will be held accountable for
achieving in a specific country or
program environment. The
Foreign Assistance Framework
outlines broad goals and
objectives (e.g. Peace and
Security) or, in other words,
programming categories.
Achievement of Mission or
Office AOs should contribute to
those broader FAF objectives.
program managers gain clarity
around
key
objectives.
Ultimately, it sets the foundation
not only for the strategy, but also
for numerous other management
and
planning
functions
downstream, including project
design, monitoring, evaluation,
and program management. To
summarize,
the
Results
Framework:
 Provides an opportunity to
build consensus and ownership
around shared objectives not
only among AO team members
but also, more broadly, with
host-country representatives,
partners, and stakeholders.
 Facilitates agreement with
other
actors
(such
as
USAID/Washington, other USG
entities, the host country, and
other donors) on the expected
results and resources necessary
to achieve those results. The
AO is the focal point of the
agreement
between
USAID/Washington and the
Mission. It is also the basis for
Assistance
Agreements
(formerly
called
Strategic
Objective
Assistance
Agreements).
 Functions as an effective
communication tool because it
succinctly captures the key
elements of a program’s intent
and content.
 Establishes the foundation to
design
monitoring
and
evaluation
systems.
Information from performance
monitoring and evaluation
systems should also inform the
development of new RFs.
 Identifies the objectives that
drive project design.
In order to be an effective tool, a
Results Framework should be
current. RFs should be revised
when 1) results are not achieved
or completed sooner than
expected, 2) critical assumptions
are no longer valid, 3) the
underlying development theory
must be modified, or 4) critical
problems with policy, operations,
or resources were not adequately
recognized.
KEY CONCEPTS
THE RESULTS FRAMEWORK
IS PART OF A BROADER
STRATEGY
While the Results Framework is
one of the core elements of a
strategy, it alone does not
constitute a complete strategy.
Typically it is complimented by
narrative that further describes
the thinking behind the RF, the
relationships
between
the
objectives, and the identification
of synergies. As a team develops
the RF, broader strategic issues
2
should be considered, including
the following:
 What has led the team to
propose
the
Results
Framework?
 What is strategic about what is
being proposed (that is, does it
reflect
a
comparative
advantage or a specific niche)?
 What are the main strategic
issues?
 What is different in the new
strategy when compared to the
old?
 What synergies emerge? How
are
cross-cutting
issues
addressed?
How can these
issues be tackled in project
level
planning
and
implementation?
THE UNDERPINNING OF THE
RESULTS FRAMEWORK
A good Results Framework is not
only based on logic. It draws on
analysis, standard theories in a
technical
sector,
and
the
expertise
of
on-the-ground
managers.
Supporting Analysis
Before developing a Results
Framework, the team should
determine what analysis exists
and what analysis must yet be
completed
to
construct
a
development hypothesis with a
reasonable level of confidence.
Evaluations
constitute
an
important source of analysis,
identify important lessons from
past programs, and may explore
the validity of causal linkages that
can be used to influence future
programming.
Analysis of past
performance monitoring data is
also an important source of
information.
FIGURE 2. SETTING THE CONTEXT
FOR PARTICIPATION
External Forces
(Host Country
Strategy)
Standard Sector Theories
Sectors, particularly those that
USAID has worked in for some
time, often identify a set of
common elements that constitute
theories for how to accomplish
certain
objectives.
These
common elements form a basic
―template‖ of sorts to consider in
developing an RF. For example,
democracy
and
governance
experts often refer to addressing
supply and demand.
Supply
represents
the
ability
of
government to play its role
effectively or provide effective
services. Demand represents the
ability of civil society to demand
or
advocate
for
change.
Education
generally
requires
improved quality in teaching and
curriculum,
community
engagement,
and
adequate
facilities. Health often requires
improved quality of services, as
well as access to -- and greater
awareness of – those services.
An understanding of these
common strategic elements is
useful because they lay out a
standard set of components that
a team must consider in
developing a good RF. Although,
not all of these elements will
apply to all countries in the same
way, they form a starting point to
inform the team’s thinking. As
the team makes decisions about
what (or what not) to address,
this becomes a part of the logic
The
―Fit‖
USAID Mission/
Vision
that is presented in the narrative.
Technical experts can assist teams
in understanding standard sector
theories. In addition, a number
of USAID publications outline
broader sector strategies or
provide guidance on how to
develop strategies in particular
technical areas1.
On-the-Ground Knowledge
and Experience
Program managers are an
important source of knowledge
on the unique program or incountry factors that should be
considered in the development of
the Results Framework. They are
best able to examine different
types of information, including
1
Examples include: Hansen,
Gary. 1996. Constituencies for
Reform: Strategic Approaches for
Donor-Supported Civic Advocacy
Groups or USAID. 2008. Securing
the Future: A Strategy for
Economic Growth.
3
Internal
Capacity
analyses and standard sector
theories, and tailor a strategy for
a specific country or program
environment.
PARTICIPATION AND
OWNERSHIP
Development
of
a
Results
Framework presents an important
opportunity for USAID to engage
its own teams, the host country,
civil society, other donors, and
other
partners
in
defining
program objectives. Experience
has shown that a Results
Framework built out of a
participatory process results in a
more effective strategy.
Recent donor commitments to
the Paris Declaration and the
Accra Agenda for Action reinforce
these points. USAID has agreed
to increase ownership, align
systems
with
country-led
strategies, use partner systems,
harmonize aid efforts, manage for
development
results,
and
establish mutual accountability.
Common questions include,
―how
do
we
manage
participation?‖ or ―how do we
avoid raising expectations that
we cannot meet?‖
One
approach for setting the context
for effective participation is to
simply set expectations with
participants before engaging in
strategic discussions. In essence,
USAID is looking for the
―strategic fit‖ (see Figure 2). That
is, USAID seeks the intersection
between what the host country
wants, what USAID is capable of
delivering, and the vision for the
program.
WHOLE-OF- GOVERNMENT
APPROACHES
Efforts are underway to institute
planning processes that take into
account the U.S. Government’s
overall approach in a particular
country.
A
whole-ofapproach
may
government
identify larger goals or objectives
to which many USG entities
contribute.
Essentially, those
objectives would be at a higher
level or above the level of
accountability of any one USG
agency alone. USAID Assistance
Objectives
should
clearly
contribute to those larger goals,
but also reflect what the USAID
Mission can be held accountable
for within a specified timeframe
and within budget parameters.
The
whole-of-government
approach may be reflected at a
lower level in the Results
Framework as well.
The RF
provides flexibility to include the
objectives of other
GUIDELINES FOR CONSTRUCTING AOs AND IRs
AOs and IRs should be:

Results Statements. AOs and IRs should express an outcome. In other words,
the results of actions, not the actions or processes themselves. For example,
the statement ―increased economic growth in targets sectors‖ is a result, while
the statement ―increased promotion of market-oriented policies‖ is more
process oriented.

Clear and Measurable. AOs and IRs should be stated clearly and precisely, and
in a way that can be objectively measured. For example, the statement
―increased ability of entrepreneurs to respond to an improved policy, legal,
and regulatory environment‖ is both ambiguous and subjective. How one
defines or measures ―ability to respond‖ to a changing policy environment is
unclear and open to different interpretations. A more precise and measurable
results statement in this case is ―increased level of investment.‖ It is true that
USAID often seeks results that are not easily quantified. In these cases, it is
critical to define what exactly is meant by key terms. For example, what is
meant by ―improved business environment‖? As this is discussed, appropriate
measures begin to emerge.

Unidimensional. AOs or IRs ideally consist of one clear overarching objective.
The Results Framework is intended to represent a discrete hypothesis with
cause-and-effect linkages. When too many dimensions are included, that
function is lost because lower level results do not really ―add up‖ to higher
level results. Unidimensional objectives permit a more straightforward
assessment of performance. For example, the statement ―healthier, better
educated, higher-income families‖ is an unacceptable multidimensional result
because it includes diverse components that may not be well-defined and
may be difficult to manage and measure. There are limited exceptions. It may
be appropriate for a result to contain more than one dimension when the
result is 1) achievable by a common set of mutually-reinforcing Intermediate
Results or 2) implemented in an integrated manner (ADS 201.3.8).
actors (whether other USG
entities, donors, the host country,
or other partners) where the
achievement of those objectives
are essential for USAID to achieve
its AO.
For example, if a
program achieves a specific
objective that contributes to
USAID’s AO, it should be
reflected as an IR.
This can
facilitate greater coordination of
efforts.
THE LINKAGE TO PROJECTS
The RF should form the
foundation for project planning.
4
Project teams may continue to
flesh out the Results Framework
in further detail or may use the
Logical Framework2. Either way,
all projects and activities should
be designed to accomplish the
AO and some combination of one
or more IRs.
2
The Logical Framework (or
logframe for short) is a project
design tool that complements the
Results Framework. It is also
based on cause-and-effect
linkages. For further information
reference ADS 201.3.11.8.
THE PROCESS FOR
DEVELOPING A
RESULTS
FRAMEWORK
SETTING UP THE PROCESS
Missions may use a variety of
approaches to develop their
respective results frameworks. In
setting up the process, consider
the following three questions.
When
should
the
results
frameworks be developed? It is
often helpful to think about a
point in time at which the team
will have enough analysis and
information
to
confidently
construct a results framework.
Who is going to participate
(and at what points in the
process)?
It is important to
develop a schedule and plan out
the process for engaging partners
and stakeholders. There are a
number of options (or a
combination) that might be
considered:
 Invite
key
partners
or
stakeholders
to
results
framework
development
sessions. If this is done, it may
be useful to incorporate some
training
on
the
results
framework methodology in
advance. Figure 3 outlines the
basic building blocks and
defines terms used in strategic
planning
across
different
organizations.
 The AO team may develop a
preliminary results framework
and hold sessions with key
counterparts to present the
draft strategy and obtain
feedback.
 Conduct a strategy workshop
for AO teams to present out
RFs and discuss strategic issues.
Although these options require
some time and effort, the results
framework will be more complete
and representative.
What process and approach
will be used to develop the
results frameworks?
We
strongly recommend that the AO
team hold group sessions to
construct the results framework.
It is often helpful to have one
person
(preferably
with
experience in strategic planning
and facilitation) to lead these
sessions.
This person should
focus on drawing out the ideas of
the group and translating them
into the results framework.
STEP 1. IDENTIFY THE
ASSISTANCE OBJECTIVE
The Assistance Objective (AO) is
the center point for any results
framework and is defined as:
The most ambitious result
(intended measurable change)
that a USAID Mission/Office,
along with its partners, can
materially affect, and for which
it is willing to be held
accountable (ADS 201.3.8).
Defining an AO at an appropriate
level of impact is one of the most
critical and difficult tasks a team
faces.
The AO forms the
5
―It is critical to stress the importance
of not rushing to finalize a results
framework. It is necessary to take
time for the process to mature and to
be truly participative.‖
—USAID staff member in Africa
standard by which the Mission or
Office is willing to be judged in
terms of its performance.
The
concept of ―managing for results‖
(a USAID value also reflected in
the Paris Declaration) is premised
on this idea.
The task can be challenging,
because an AO should reflect a
balance of two conflicting
and
considerations—ambition
accountability. On the one hand,
every team wants to deliver
significant impact for a given
investment. On the other hand,
there are a number of factors
outside the control of the team.
In fact, as one moves up the
Results Framework toward the
AO, USAID is more dependent on
other development partners to
achieve the result.
Identifying an appropriate level
of ambition for an AO depends
on a number of factors and will
be different for each country
context. For example, in one
country it may be appropriate for
the AO to be ―increased use of
family planning methods‖ while
in another, ―decreased total
fertility‖ (a higher level objective)
would be more suitable. Where
to set the objective is influenced
by the following factors:
 Programming
history.
There
are
different
expectations for more
mature programs, where
higher level impacts and
greater sustainability are
expected.
 The magnitude of the
development problem.
 The timeframe
strategy.
for
Figure 3. Results Framework Logic
So What?
Necessary
and
Sufficient
the
 The
range
of
resources
available or expected.
The AO should represent the
team’s best assessment of what
can realistically be achieved. In
other words, the AO team should
be able to make a plausible case
that the appropriate analysis has
been done and the likelihood of
success is great enough to
warrant investing resources in the
AO.
STEP 2. IDENTIFY
INTERMEDIATE RESULTS
After agreeing on the AO, the
team must identify the set of
―lower level‖ Intermediate Results
necessary to achieve the AO. An
Intermediate Result is defined as:
An important result that is
seen as an essential step to
achieving a final result or
outcome.
IRs
are
How?
measurable results that may
capture a number of
discrete and more specific
results (ADS 201.3.8.4).
As the team moves down from
the AO to IRs, it is useful to ask
―how‖ can the AO be achieved?
By answering this question, the
team begins to formulate the IRs
(see Figure 3). The team should
assess relevant country and
sector conditions and draw on
development experience in other
countries to better understand
the changes that must occur if
the AO is to be attained.
The
Results
Framework
methodology
is
sufficiently
flexible to allow the AO team to
include Intermediate Results that
are supported by other actors
when they are relevant and
critical to achieving the AO. For
example, if another donor is
6
building schools that are
essential for USAID to
accomplish an education AO
(e.g.
increased
primary
school completion), then
that should be reflected as
an IR because it is a
necessary ingredient for
success.
Initially, the AO team might
identify a large number of
possible results relevant to
the AO.
However, it is
important to eventually settle on
the critical set of Intermediate
Results. There is no set number
for how many IRs (or levels of IRs)
are appropriate. The number of
Intermediate Results will vary
with the scope and complexity of
the AO. Eventually, the team
should arrive at a final set of IRs
that
members
believe
are
reasonable. It is customary for
USAID Missions to submit a
Results Framework with one or
two
levels
of
IRs
to
USAID/Washington for review.
The key point is that there should
be enough information to
adequately
convey
the
development hypothesis.
So What is Causal Logic Anyway?
Causal logic is based on the concept of cause-and-effect. That is, the accomplishment of lower-level
objectives ―cause‖ the next higher-level objective (or the effect) to occur. In the following example, the
hypothesis is that if IR 1, 2, and 3 occur, it will lead to the AO.
AO: Increased
Completion of
Primary School
IR 1: Improved
Quality of
Teaching
STEP 3. CLARIFY THE
RESULTS FRAMEWORK
LOGIC
Through
the
process
of
identifying Intermediate Results,
the team begins to construct the
cause-and-effect logic that is
central to the Results Framework.
Once the team has identified the
Intermediate Results that support
an objective, it must review and
confirm this logic.
The accomplishment of lower
level results, taken as a group,
should result in the achievement
of the next higher objective. As
the team moves up the Results
Framework, they should ask, ―so
what?‖ If we accomplish these
lower
level
objectives,
is
something
of
significance
achieved at the next higher level?
The
higher-order
result
establishes the ―lens‖ through
which lower-level results are
viewed. For example, if one IR is
―Increased Opportunities for Outof-School Youth to Acquire Life
Skills,‖ then, by definition, all
lower level IRs would focus on
IR 2: Improved
Curriculum
IR 3: Increased
Parental
Commitment to
Education
the target population established
(out-of-school youth).
As the team looks across the
Results Framework, it should ask
whether the Intermediate Results
are necessary and sufficient to
achieve the AO.
Results Framework logic is not
always linear.
There may be
relationships across results or
even with other AOs. This can
sometimes be demonstrated on
the graphic (e.g., through the use
of arrows or dotted boxes with
some explanation) or simply in
the narrative. In some cases,
teams find a number of causal
connections in an RF. However,
teams have to find a balance
between the two extremes- on
the one hand, where logic is too
simple and linear and, on the
other, a situation where all
objectives are related to all
others.
STEP 4. IDENTIFY CRITICAL
ASSUMPTIONS
The next step is to identify the set
of critical assumptions that are
relevant to the achievement of
7
the AO. A critical assumption is
defined as:
―….a general condition under
which
the
development
hypothesis will hold true.
Critical
assumptions
are
outside
the
control
or
influence of USAID and its
partners (in other words, they
are not results), but they
reflect conditions that are
likely to affect the achievement
of results in the Results
Framework.
Critical
assumptions may also be
expressed
as
risks
or
vulnerabilities…‖
(ADS
201.3.8.3)
Identifying critical assumptions,
assessing associated risks, and
determining how they should be
addressed is a part of the
strategic
planning
process.
Assessing risk is a matter of
balancing the likelihood that the
critical assumption will hold true
with the ability of the team to
address the issue. For example,
consider the critical assumption
―adequate rainfall.‖
If this
assumption has held true for the
What is NOT Causal Logic?
Categorical Logic. Lower level results are simply sub-categories rather than cause and effect, as
demonstrated in the example below.
AO: Increased
Completion of
Primary School
IR 1: Improved
Pre-Primary
School
IR 2: Improved
Primary
Education
IR 3: Improved
Secondary
Education
Definitional Logic. Lower-level results are a restatement (or further definition) of a higher-level objective.
The use of definitional logic results in a problem later when identifying performance indicators because it is
difficult to differentiate indicators at each level.
IR: Strengthened
Institution
IR: Institutional
Capacity to Deliver
Goods & Services
target region only two of the past
six years, the risk associated with
this assumption is so great that it
poses a risk to the strategy.
the years when a drought may
occur.
In cases like this, the AO team
should attempt to identify ways
to actively address the problem.
For example, the team might
include efforts to improve water
storage or irrigation methods, or
increase use of drought-resistant
seeds or farming techniques.
This would then become an IR (a
specific
objective
to
be
accomplished by the program)
rather than a critical assumption.
Another option for the team is to
develop contingency plans for
As a final step, the AO team
should step back from the Results
Framework and review it as a
whole. The RF should be
straightforward
and
understandable. Check that the
results contained in the RF are
measurable and feasible with
anticipated USAID and partner
resource levels. This is also a
good point at which to identify
synergies between objectives and
across AOs.
STEP 5. COMPLETE THE
RESULTS FRAMEWORK
8
STEP 6. IDENTIFY
PRELIMINARY
PERFORMANCE MEASURES
Agency policies (ADS 201.3.8.6)
require that the AO team present
proposed indicators for the AO
with baseline data and targets.
The AO, along with indicators and
targets, represents the specific
results that will be achieved vis-avis the investment. To the extent
possible, indicators for IRs with
baseline and targets should be
included as well.
1.
Figure 1. Illustrative Results Framework
AO:
Increased
Production by
Farmers in the
Upper River Zone
IR:
Farmers’ Access to
Commercial
Capital Increased
IR: Farmers’
Capacity to
Develop Bank
Loan Applications
Increased
(4 years)
IR: Banks’ Loan
Policies Become
More Favorable
for the Rural
Sector
(3 years)
2.
3.
4.
Critical Assumptions
Market prices for farmers’ products remain stable
or increase.
Prices of agricultural inputs remain stable or
decrease.
Roads needed to get produce to market are
maintained.
Rainfall and other critical weather conditions
remain stable.
IR:
Farmers’ Transport
Costs Decreased
IR: Additional
Local Wholesale
Market Facilities
Constructed (with
the World Bank)
IR:
Farmers’
Knowledge About
Effective
Production
Methods
Increased
IR: Village
Associations
Capacity to
Negotiate
Contracts
Increased (4 years)
IR: New
Technologies
Available
(World Bank)
(
Key
USAID
Responsible
Partner(s)(4
Responsible
9
USAID +
Partner(s)
Responsible
IR: Farmers’
Exposure to OnFarm Experiences
of Peers Increased
Figure 3. The Fundamental Building Blocks for Planning
ASSISTANCE OBJECTIVE (AO)
AO
The highest level objective for which USAID is
willing to be held accountable. AOs may also
be referred to as outcomes, impacts, or results.
Increased Primary School Completion
INTERMEDIATE RESULTS (IRs)
IR
Interim events, occurrences, or conditions that
are essential for achieving the AO. IRs may
also be referred to as outcomes or results.
Teaching Skills Improved
OUTPUT
OUTPUT
Products or services produced as a result of
internal activity.
Number of teachers trained
INPUT
INPUT
Resources used to produce an output.
Funding or person days of training
10
Figure 4. Sample Results Framework and Crosswalk of FAF Program Hierarchy and a
Results Framework
F Program
Hierarchy for
Budgeting and
Reporting
Illustrative Results Framework for
Program Planning
Assistance Objective: Economic Competitiveness of
Private Enterprises Improved
IR 1: Enabling Environment for
Enterprises Improved
IR 1.1 Licensing
and registration
requirements for
enterprises
streamlined
The Illustrative Results Framework
links to the FAF Program
Hierarchy as follows:
• Objective 4 Economic Growth
• Program Areas 4.6 (Private Sector
•
•
•
•
•
•
•
Competitiveness) and 4.7
(Economic Opportunity
Program Elements 4.6.1, 4.6.2, 4.7
Sub-Elements 4.6.12 and 4.7.2.1
Sub-Element 4.6.1.3
Sub-Element 4.7.2.2
Sub-Element 4.6.2.1
Sub-Element 4.7.3
Sub-Element 4.6.2.4
IR 1.3
Regulatory
environment for
micro and small
enterprises
improved
IR 1.2
Commercial laws
that support
market-oriented
transactions
promoted
IR 2: Private Sector
Capacity Strengthened
IR 2.1
Competitiveness
of targeted
enterprises
improved
IR 2.2
Productivity of
microenterprises in
targeted
geographic
regions
increased
Critical Assumptions:
•
•
Key political leaders, including the President and the
Minister of Trade and Labor, will continue to support
policy reforms that advance private enterprise-led
growth.
Government will sign the Libonia Free Trade
Agreement, which will open up opportunities for
enterprises targeted under IR 2.1.
Note: The arrows demonstrate the linkage of AO1, IR 1, and IR 1.1 to the FAF. As an example, IR1 links to the program element 4.6.1
“Business Enabling Environment”. IR 1.1 links to 4.7.2.1 “Reduce Barriers to Registering Micro and Small Business”.
11
IR 2.3
Information
Exchange
Improved
For more information:
TIPS publications are available online at [insert website].
Acknowledgements:
Our thanks to those whose experience and insights helped shape this publication including Gerry Britan
and Subhi Mehdi of USAID’s Office of Management Policy, Budget and Performance (MPBP). This
publication was updated by Michelle Adams-Matson, of Management Systems International.
Comments can be directed to:
Gerald Britan, Ph.D.
Tel: (202) 712-1158
[email protected]
Contracted under RAN-M-00-04-00049-A-FY0S-84
Integrated Managing for Results II
12
NUMBER 15
2011 Printing
PERFORMANCE MONITORING & EVALUATION
TIPS
MEASURING INSTITUTIONAL CAPACITY
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to
peroformance monitoring and evaluation. This publication is a supplemental reference to the
Automated Directive Service (ADS) Chapter 203.
What are the strengths and limitations of each
approach with regard to internal bias, quantification, or comparability over time or across
organizations?
INTRODUCTION
This PME Tips gives USAID managers information on measuring institutional capacity,* including some tools that measure the capacity of an
entire organization as well as others that look
at individual components or functions of an organization. The discussion concentrates on the
internal capacities of individual organizations,
rather than on the entire institutional context
in which organizations function. This Tips is not
about how to actually strengthen an institution, nor is it about how to assess the eventual
impact of an organization’s work. Rather, it is
limited to a specific topic: how to measure an
institution’s capacities.
How will the data be collected and how participatory can and should the measurement process be?
Measuring institutional capacity might be one
important aspect of a broader program in institutional strengthening; it may help managers
make strategic, operational, or funding decisions;
or it may help explain institutional strengthening activities and related performance.
It addresses the following questions:
Whatever the reason for assessing institutional
capacity, this Tips presents managers with several tools for identifying institutional strengths
and weaknesses.
Which measurement approaches are most
useful for particular types of capacity building?
The paper will define and discuss capacity assessment in general and present several ap1
able to accomplish their mission and provide
for their own needs in the long run.
­
USAID operating units build capacity with a broad
spectrum of partner and customer organizations.
These include but are not limited to:
proaches for measuring institutional capacity.
We assess the measurement features of each
approach to help USAID managers select the
tool that best fits their diverse management
and reporting needs. The paper is organized as
follows:
1. Background: Institutional Capacity • American private voluntary organizations
Building and USAID
(PVOs)
2. How to Measure Institutional Capacity
• Local and international nongovernmental
organizations (NGOs) and other civil society organizations (CSOs)
3. Measurement Issues
4. Institutional Assessment Tools
• Community-based membership cooperatives, such as a water users group
5. Measuring Individual Organizational
Components
• Networks and associations of organizations
6. Developing Indicators
• Political parties
7. Practical Tips for a Busy USAID Manager
• Government entities (ministries, departments, agencies, subunits, policy analysis
units, health clinics, schools)
BACKGROUND:
INSTITUTIONAL CAPACITY
BUILDING AND USAID
• Private sector organizations (financial institutions, companies, small businesses and
other forprofit organizations)
• Regional institutions
USAID operating units must work closely with
partner and customer organizations to meet
program objectives across all Agency goal areas, among them Peace and Security, Governing
Justly and Democratically, Economic Growth,
Investing in People, and Humanitarian Assistance. In the course of planning, implementing,
and measuring their programs, USAID managers often find that a partner or customer organization’s lack of capacity stands in the way
of achieving results. Increasing the capacity of
partner and customer organizations helps them
carry out their mandate effectively and function
more efficiently. Strong organizations are more
The Agency uses a variety of techniques to
build organizational capacity. The most common involve providing technical assistance, advisory services, and long-term consultants to
organizations, to help them build the skills and
experience necessary to contribute successfully to sustainable development. Other techniques include providing direct inputs, such as
financial, human, and technological resources.
Finally, USAID helps establish mentoring relationships; provides opportunities for formal
study in-country, in the United States or in
third countries; and it sets up internships or
2
apprenticeships with other organizations. The
goal of strengthening an institution is usually to
improve the organization’s overall performance
and viability by improving administrative and
management functions, increasing the effectiveness of service provision, enhancing the organization’s structure and culture, and furthering its sustainability. Institutional strengthening
programs may address one or more of these
components.
the capacity of an organization to help make
decisions about awarding grants or holding
grantees accountable for results. In this case,
the assessment is more of an external oversight/audit of an organization hired to carry out
Agency programs. Or, the manager may have a
programmatic commitment to strengthen the
abilities of customer and partner organizations.
Different tools and methods are available for
both situations. This paper deals primarily with
programs that fit the latter description.
In most cases, USAID managers are concerned
with institutional strengthening because they
are interested in the eventual program-level results (and the sustainability of these results) that
these stronger organizations can help achieve.
While recognizing the need to address eventual results, this Tips looks primarily at ways to
measure institutional capacity. Understanding
and measuring institutional capacity are critical
and often more complex than measuring the
services and products an organization delivers.
Within USAID, the former Office of Private and
Voluntary Cooperation (PVC) took the lead on
building the capacity of nongovernmental organization (NGO) and private voluntary organization (PVO) partners. PVC has defined development objectives and intermediate results aimed
specifically at improving the internal capacity
of U.S. PVOs. PVC has studied different approaches to institutional capacity building and
has begun to develop a comprehensive capacity assessment tool called discussion-oriented
Measuring organizational capacity is important organizational self-assessment, described in exbecause it both guides USAID interventions ample 1 in this paper. In addition to DOSA, PVC
and allows managers to demonstrate and re- has developed several indicators for measuring
port on progress. The data that emerge from institutional capacity development.
measuring institutional capacity are commonly
used in a number of valuable ways. These data PVC specifically targets NGOs and PVOs
establish baselines and provide the basis for and is particularly concerned with enhancsetting targets for improvements. They help ex- ing partnerships. USAID missions, by contrast,
plain where or why something is going wrong; work with a broader range of organizations
they identify changes to specific program in- on activities aimed at increasing institutional
terventions and activities that address areas of capacity. Such programs usually view instipoor performance; they inform managers of the tutional capacity as a means to achieve highimpact of an intervention or the effectiveness er level program results, rather than as an
of an intervention strategy; and they identify end in itself.
lessons learned.They are also useful for reporting to Washington and to partners.
HOW TO MEASURE
INSTITUTIONAL CAPACITY
It is important to note the difference between
assessing capacity for contracting and grantmaking decisions versus for a “capacity build- An organization can be thought of as a system
ing” relationship with partner/customer organi- of related components that work together to
zations. A USAID manager may want to assess achieve an agreed-upon mission. The follow3
ing list of organizational components is not
all-inclusive, nor does it apply universally to
all organizations. Rather, the components are
representative of most organizations involved
in development work and will vary according
to the type of organization and the context in
which it functions.
• External relations
Administrative and Support Functions
• Other
Resources
• Human
• Financial
• Administrative procedures and management systems
MANAGEMENT ISSUES
• Financial management (budgeting, accounting, fundraising, sustainability)
This TIPS presents capacity-assessment tools
and other measurement approaches that, while
similar in some ways, vary in both their emphasis and their method for evaluating an organization’s capacity. Some use scoring systems and
others don’t; some use questionnaires while
others employ focus groups; some use external evaluators , and others use selfassessments;
some emphasize problem solving, while others concentrate on appreciating organzational
strengths. Some tools can be used to measure
the same standard across many organizations,
while others are organization specific. Many of
the tools are designed so that the measurement
process is just as important as, if not more important than, the resulting information. They
may involve group discussions, workshops, or
exercises, and may explicitly attempt to be participatory. Such tools try to create a learning
opportunity for the organization’s members, so
that the assessment itself becomes an integral
part of the capacity-building effort.
• Human resource management (staff recruitment, placement, support)
• Management of other resources (information, equipment, infrastructure)
Technical/Program Functions
• Service delivery system
• Program planning
• Program monitoring and evaluation
• Use and management of technical knowledge and skills
Structure and Culture
• Organizational identity and culture
• Vision and purpose
Because of each user’s different needs, it would
be difficult to use this TIPS as a screen to predetermine the best capacity-assessment tool for
each situation. Rather, managers are encouraged to adopt the approaches most appropriate
to their program and to adapt the tools best
suited for local needs. To assist managers in
identifying the most useful tools and approach-
• Leadership capacity and style
• Organizational values
• Governance approach
4
es, we consider the following issues for each of
the tools presented:
methods are hands-on and highly participatory, involving a wide range of customers,
partners, and stakeholders, while others
are more exclusive, relying on the opinion
of one or two specialists. In most cases, it
is best to use more than one data collection method.
• Type of organization measured. Many
of the instruments developed to measure
institutional capacity are designed specifically for measuring NGOs and PVOs. Most
of these can be adapted easily for use with
other types of organizations, including government entities.
• Objectivity. By their nature, measures of
institutional capacity are subjective. They
rely heavily on individual perception, judgment, and interpretation. Some tools are
better than others at limiting this subjectivity. For instance, they balance perceptions with more empirical observations, or
they clearly define the capacity area being
measured and the criteria against which
it is being judged. Nevertheless, users of
these tools should be aware of the limitations to the findings.
• Comparability across organizations.
To measure multiple organizations, to compare them with each other, or to aggregate
the results of activities aimed at strengthening more than one organization, the tool
used should measure the same capacity
areas for all the organizations and use the
same scoring criteria and measurement
processes. Note, however, that a standard
tool, applied to diverse organizations, is
less able to respond to specific organizational or environmental circumstances. This
is less of a problem if a group of organizations, using the same standard tool, has
designed its diagnostic instrument together
(see the following discussion of PROSE).
• Quantification. Using numbers to represent capacity can be helpful when they
are recognized as relative and not absolute
measures. Many tools for measuring institutional capacity rely on ordinal scales.
Ordinal scales are scales in which values
can be ranked from high to low or more
to less in relation to each other. They are
useful in ordering by rank along a continuum, but they can also be misleading.
Despite the use of scoring criteria and
guidelines, one person’s “3” may be someone else’s “4.” In addition, ordinal scales
do not indicate how far apart one score is
from another. (For example, is the distance
between “agree” and “strongly agree” the
same as the distance between “disagree”
and “strongly disagree”?) Qualitative
descriptions of an organization’s capacity
level are a good complement to ordinal
scales.
• Comparability over time. In many cases, the value of measuring institutional capacity lies in the ability to track changes in
one organization over time. That requires
consistency in method and approach. A
measurement instrument, once selected
and adapted to the needs of a particular
organization, must be applied the same way
each time it is used. Otherwise, any shifts
that are noted may reflect a change in the
measurement technique rather than an
actual change in the organization.
• Data collection. Data can be collected
in a variety of ways: questionnaires, focus
groups, interviews, document searches, and
observation, to name only some. Some
• Internal versus external assessments.
Some tools require the use of external
5
facilitators or assessors; others offer a
process that the organization itself can
follow. Both methods can produce useful
data, and neither is automatically better
than the other. Internal assessments can
facilitate increased management use and
better understanding of an assessment’s
findings, since the members of the organization themselves are carrying out the
assessment. By contrast, the risk of bias
and subjectivity is higher in internal assessments. External assessments may be more
objective. They are less likely to introduce
internal bias and can make use of external
expertise. The downside is that external
assessors may be less likely to u cover
what is really going on inside an organization.
same format.
• Practicality. The best measurement
systems are designed to be as simple as
possible-- not too time consuming, not unreasonably costly, yet able to provide managers with good information often enough
to meet their management needs. Managers should take practicality into account
when selecting a measurement tool. They
should consider the level of effort and
resources required to develop the instrument and collect and analyze the data, and
think about how often and at what point
during the management cycle the data will
be available to managers.
Background
• Background of the methodology/tool
• Process (how the methodology/tool is used
in the field)
• Product (the types of outputs expected)
• Assessment (a discussion of the uses and
relative strengths of each methodology/
tool)
• An example of what the methodology/tool
looks like
PARTICIPATORY, RESULTS-ORIENTED
SELF-EVALUATION
The participatory, results-oriented self-evaluation (PROSE) method was developed by Evan
Bloom of Pact and Beryl Levinger of the Education Development Center. It has the dual
purpose of both assessing and enhancing organizational capacities. The PROSE method produces an assessment tool customized to the
organizations being measured. It is designed to
compare capacities across a set of peer organizations, called a cohort group, which allows
for benchmarking and networking among the
organizations. PROSE tools measure and profile
organizational capacities and assess, over time,
how strengthening activities affect organizational capacity. In addition, through a facilitated
workshop, PROSE tools are designed to allow
organizations to build staff capacity; create consensus around future organizational capacitybuilding activities; and select, implement, and
track organizational change and development
strategies.
INSTITUTIONAL
ASSESMENT TOOLS
This section describes capacity measurement
tools that USAID and other development organizations use. You can find complete references
and Web sites in the resources section at the One example of an instrument developed using
end of the paper. For each tool, we follow the the PROSE method is the discussion-oriented
6
organizational self-assessment. DOSA was developed in 1997 for the Office of Private and
Voluntary Cooperation and was designed specifically for a cohort of USAID PVO grantees.
Participatory, Results-Oriented
Self-Evaluation
Type of Organization Measured
Process
NGOs/PVOs; adaptable to other types of organizations
Developers of the PROSE method recommend
that organizations participate in DOSA or develop a customized DOSA-like tool to better fit
their organization’s specific circumstances. The
general PROSE process for developing such a
tool is as follows: After a cohort group of organizations is defined, the organizations meet in
a workshop setting to design the assessment
tool. With the help of a facilitator, they begin by
pointing to the critical organizational capacities
they want to measure and enhance. The cohort
group then develops two sets of questions: discussion questions and individual questionnaire
items. The discussion questions are designed to
get the group thinking about key issues. Further,
these structured discussion questions minimize
bias by pointing assessment team members toward a common set of events, policies, or conditions. The questionnaire items then capture
group members’ assessments of those issues
on an ordinal scale. During the workshop, both
sets of questions are revised until the cohort
group is satisfied. Near the end of the process,
tools or standards from similar organizations
can be introduced to check the cohort group’s
work against an external example. If the tool
is expected to compare several organizations
within the same cohort group, the tool must be
implemented by facilitators trained to administer it effectively and consistently across the
organizations.
Features
• Cross-organizational comparisons can be
made
• Measures change in one organization or a
cohort of organizations over time
• Measures well-defined capacity areas
against well-defined criteria
• Assessment based primarily upon perceived capacities
• Produces numeric score on capacity areas
• Assessment should be done with the help
of an outside facilitator or trained insider
• Data collected through group discussion
and individual questionnaires given to a
cross-section of the organization’s staff
team meets for four to six hours and should
represent a cross-functional, crosshierarchical
sample from the organization. Participants respond anonymously to a questionnaire, selecting the best response to statements about the
organization’s practices (1=strongly disagree,
2=disagree, 3=neutral, 4=agree, 5=strongly
Once the instrument is designed, it is applied
agree) in six capacity areas:
to each of the organizations in the cohort. In
the case of DOSA, the facilitator leads a team
• External Relations
of the organization’s members through a series
(constituency development, fund-raising
of group discussions interspersed with individand communications)
ual responses to 100 questionnaire items. The
7
Example 1. Excerpt From DOSA, a PROSE Tool
The DOSA questionnaire can be found in annex 1a
The following is a brief example drawn from the Human Resource Management section of the
DOSA questionnaire:
Discussion Questions
a. When was our most recent staff training?
b. How often over the last 12 months have we held staff training events?
Questionnaire items for individual response
Strongly
Disagress
Neutral
Agree
Strongly
Disagree
Agree
1. We routinely offer
1
2
3
4
5
staff training.
Discussion Questions
a. What are three primary, ongoing functions (e.g., monitoring and evaluation, proposal writing, resource mobilization) that we carry out to achieve our mission?
b. To what extent does staff, as a group, have the requisite skills to carry out these functions?
c. To what extent is the number of employees carrying out these functions commensurate
with work demands?
Questionnaire items for individual response
Strongly
Disagress
Neutral
Disagree
2. We have the ap1
2
3
propriate staff skills
to achieve our mission
3. We have the ap1
2
3
propriate staff numbers to achieve our
mission
Agree
4
4
Strongly
Agree
5
5
*The annexes for this paper are available separately and can be obtained through the USAID
Development Experience Clearinghouse at http://dec.usaid.gov/index.cfm
• Financial Resource Management
(budgeting, forecasting, and cash management)
• Human Resource Management
(staff training, supervision, and personnel
practices)
8
• Organizational Learning
(teamwork and information sharing)
check on the perceived capacities reported by
individual organizational members. It also helps
identify capacity areas that all members agree
• Strategic Management
need immediate attention.
(planning, governance, mission, and partnering) Because the cohort organizations develop the
specifics of the instrument together and share
• Service Delivery
a common understanding and application of the
(field-based program practices and sustainabil- approach, PROSE is relatively good at comparity issues)
ing organizations with each other or rolling up
results to report on a group of organizations
Although the analysis is statistically complex, together. However, the discussions could influquestionnaires can be scored and graphics pro- ence the scoring if facilitators are not consisduced using instructions provided on the DOSA tent in their administration of the tool.
Web site. In the case of DOSA, the DOSA team
in Washington processes the results and posts INSTITUTIONAL DEVELOPMENT FRAMEthem on the Internet. The assessment tool can WORK
be readministered annually to monitor organizational changes.
Background
The institutional development framework (IDF)
is a tool kit developed by Mark Renzi of Management Systems International. It has been used
in USAID/Namibia’s Living in a Finite Environment project as well as several other USAID
programs. Designed specifically to help nonprofit organizations improve efficiency and become more effective, the IDF is best suited for
the assessment of a single organization, rather
than a cohort group (as opposed to PROSE).
The kit contains three tools (Institutional Development Framework, Institutional Development Profile, and Institutional Development
Calculation Sheet), which help an organization
determine where it stands on a variety of organizational components, identify priority areas
of improvement, set targets, and measure progress over time. While it can be adapted for any
organization, the IDF was originally formulated
for environmental NGOs.
Product
PROSE instruments produce two types of
scores and accompanying graphics. The first is a
capacity score, which indicates how an organization perceives its strengths and weaknesses in
each of the capacity and subcapacity areas. The
second is a consensus score, which shows the
degree to which the assessment team members
agree on their evaluation of the organization’s
capacity.
Assessment
Unless the existing DOSA questions are used,
developing a PROSE instrument from scratch
can be time consuming and generally requires
facilitators to guide the process of developing and using the instrument. PROSE, like most
other such instruments, is based on perceived
capacities and does not currently include a
method for measuring externally observable
performance in various capacity areas (although
this is under consideration). It is unique among
the instruments in this paper in its use of a
consensus score. The consensus score acts as a
Process
An organization can use the IDF tools either
with or without the help of a facilitator. The
IDF identifies five organizational capacity areas,
9
Institutional Development Framework
Type of Organization Measured
NGOs/PVOs; adaptable to other types of organizations
Features
• Can be used, with limitations, to compare across organizations
• Measures change in the same organization over time
• Measures well-defined capacity areas against well-defined criteria
• Assessment based primarily upon perceived capacities
• Produces numeric score on capacity areas
• Produces qualitative description of an organization’s capacity in terms of developmental stages
• Assessment can be done internally or with help of an outside facilitator
•
• Data collected through group discussion with as many staff as feasible
called resource characteristics. Each capacity (public relations, ability to work with local
area is further broken down into key compo- communities, ability to work with government
nents, including:
bodies, ability to work with other NGOs)
Each key component within a capacity area is
• Oversight/Vision
rated at one of four stages along an organiza(board, mission, autonomy)
tional development continuum (1= start up, 2=
development, 3= expansion/consolidation, and
• Management Resources
4= sustainability). IDF offers criteria describing
(leadership style, participatory managment,
each stage of development for each of the key
management systems, planning, community
components (see example 2 below).
participation, monitoring, evaluation)
Different processes can be used depending on
• Human Resources
the organization’s size and the desired out(staff skills, staff development, organizational
come. Small organizations usually involve as
diversity)
many staff as possible; larger organizations may
work in small groups or use a few key infor• Financial Resources
mants. Members of the organization can modify
(financial management, financial vulnerability,
the Institutional Development Framework to fit
financial solvency)
their organization. Nonapplicable areas can be
ignored and new areas can be added, although
• External Resources
the creator of the tool warns against complete10
ly rewriting the criteria. Through discussion, the
participating members then use the criteria to
determine where along the development continuum their organization is situated for each
component. The resulting graphic, the Institutional Development Profile (IDP), uses bars or
“x”s to show where the organization ranks on
each key component.Through a facilitated meeting or group discussion, organization members
then determine which areas of organizational
capacity are most important to the organization
and which need priority attention for improvement. Using the IDP, they can visually mark their
targets for the future.
The IDF also provides numeric ratings. Each key
component can be rated on a scale of 1 to 4,
and all components can be averaged together
to provide a summary score for each capacity area. This allows numeric targets to be set
and monitored. The Institutional Development
Calculation Sheet is a simple table that permits
the organization to track progress over time by
recording the score of each component along
the development continuum.
Example 2. Excerpt From the IDF Tool
The following is an excerpt from the Financial Management section of the Institutional Development Framework. The entire framework appears in annex 2.
Resource
Key
Characteristic Component
Financial
Management
Budget as
Management
Tools
Cash
Controls
Financial
Security
Criteria for Each Progressive Stage
(the Development Continuum)
Start Up
Development
Expansion and
Consolidation
1
2
3
Total expendiBudgets are
Budgets are
not used as
developed for ture is usually
within 20% of
management
project actools.
tivities, but are budget, but
often over- or actual activity
underspent
often diverge
by more than from budget
20%.
predictions.
Improved
No clear
Financial
procedures ex- controls exist financial control
ist for handling but lack a sys- systems exist.
payables and
tematic office
receivables.
procedure.
Financing
comes from
only one
source.
Financing
comes from
multiple
sources, but
90% or more
from one
source.
11
No single
source of funding provides
more than 60%
of funding.
Sustainability
4
Budgets are
integral part of
project management
and are adjusted as project
implementation
warrants.
Excellent cash
controls for
payables and
receivables and
established
budget procedures.
No single source
provides more
than 40% of
funding.
Product
Process
The IDF produces a graphic that shows the
component parts of an organization and the organization’s ratings for each component at different points in time. It also provides a numeric
score/rating of capacity in each key component
and capacity area.
The OCAT is intended to be a participatory
self-assessment but may be modified to be an
external evaluation. An assessment team, composed of organizational members (representing different functions of the organization) plus
some external helpers, modifies the OCAT assessment sheet to meet its needs (annex 3).The
Assessment
assessment sheet consists of a series of statements under seven capacity areas (with subThe IDF is an example of a tool that not only elements). The assessment team then identifies
helps assess and measure an organization’s ca- sources of information, assigns tasks, and uses a
pacity but also sets priorities for future change variety of techniques (individual interviews, foand improvements. Compared with some of the cus groups, among others) to collect the inforother tools, IDF is relatively good at tracking mation they will later record on the assessment
one organization’s change over time because of sheet. The assessment team assigns a score to
the consistent criteria used for each progres- each capacity area statement (1=needs urgent
sive stage of development. It is probably not attention and improvement; 2=needs attention;
as well suited for making cross-organizational 3=needs improvement; 4=needs improvement
comparisons, because it allows for adjustment in limited aspects; but not major or urgent;
to fit the needs of each individual organization. 5=room for some improvement; 6=no need for
immediate improvement).The assessment team
ORGANIZATIONAL CAPACITY ASSESMENT would have to develop precise criteria for what
TOOL
rates as a “1” or a “2,” etc.
Background
The capacity areas and sub-elements are:
Pact developed the organizational capacity assessment tool (OCAT) in response to a need
to examine the impact of NGO capacity-building activities. Like the Institutional Development Framework, OCAT is better suited for
measuring one organization over time. The
OCAT differs substantially from the IDF in
its data collection technique. It is designed to
identify an organization’s relative strengths and
weaknesses and provides the baseline information needed to develop strengthening interventions. It can also be used to monitor progress.
The OCAT is well known; other development
organizations have widely adapted it. Designed
to be modified for each measurement situation,
the OCAT can also be standardized and used
across organizations.
• Governance
(board, mission/goal, constituency, leadership,
legal status)
• Management Practices
(organizational structure, information
management, administration procedures,
personnel, planning, program development,
program reporting)
• Human Resources
(human resources development, staff roles,
work organization, diversity issues, supervisory
practices, salary and benefits)
12
• Financial Resources
(accounting, budgeting, financial/inventory
Example 3. Excerpt From an Adaptation of the OCAT
USAID/Madagascar developed a capacity assessment tool based on the OCAT, but tailored it to
its own need to measure 21 partner institutions implementing reproductive health programs,
including the Ministry of Health. The mission tried to measure different types of organizations
and compare them by creating a standardized instrument to use with all the organizations.
Combining the OCAT results with additional information from facilitated discussions, the mission was able to summarize how different types of organizations perceived different aspects of
their capacity and recommend future strengthening programs.
Some of the difficulties that USAID/Madagascar encountered when using the tool included
having to translate questions from French to Malagasy, possibly losing some of their meaning;
finding that some respondents were unable to answer some questions because they had no
experience with the part of the organization to which the questions referred; discovering that
some respondents had difficulty separating the subject area of the questionnaire (family planning) from their work in other health areas; and having difficulty scheduling meetings because
of the organizations’ heavy workload. Moreover, the mission noted that the instrument is based
on perceptions and is self-scored, with the resulting potential for bias.a
Below is an excerpt from the “communications/extension to customers” component of the
OCAT used by USAID/Madagascar. The entire questionnaire is in annex 4.
Classification Scale
0 Nonexistent or out of order
1 Requires urgent attention and upgrading
2 Requires overall attention and upgrading
3 Requires upgrading in certain areas, but neither major nor urgent
4 Operating, but could benefit from certain improvements
5 Operating well in all regards
Communications/Extension to Customers
a. The institution has in each clinic a staff trained and competent in
counseling all customers.
1 2 3 4 5
b. The institution is able to identify and develop key messages for exten- 1 2 3 4 5
sion among potential customers, and it can produce or obtain materials
for communicating such messages.
c. A well-organized community extension is practiced by the clinic’s
staff or other workers affiliated with the institution, whether they are
salaried or volunteers. A system exists for supervising extension workers and monitoring their effectiveness.
13
1 2 3 4 5
controls, financial reporting)
• Service Delivery
(sectoral expertise, constituency, impact
assessment)
The IDF and the OCAT are similar in several
ways, but the processes differ. The OCAT uses
an assessment team that conducts research before completing the assessment sheet. For the
IDF, organization members meet and fill out the
sheet (determine their capacities) without the
intermediate data collection step (the OCAT,
by design, relies on evidence to supplement
perceptions when conducting an assessment,
and the IDF does not). The OCAT’s data-gathering step allows for systematic cross-checking
of perceived capacities with actual or observable “facts.” It is more inductive, building up to
the capacity description, while the IDF attempts
to characterize the organization along the development continuum from the beginning. The
OCAT categorizes an organization’s capacity
areas into one of four developmental stages.
Unlike the IDF, which uses the stages as the criteria by which members rate their organization,
the OCAT uses them as descriptors once the
rating has been done.
• External Relations
(constituency relations, inter-NGO collaboration, public relations, local resources, media)
• Sustainability
(program/benefit sustainability, organizational
sustainability, financial sustainability, resource
base sustainability)
After gathering data, the assessment team
meets to reach a consensus on the rating of
each element. With the help of an OCAT rating sheet, averages can be calculated for each
capacity area. These numeric scores indicate
the relative need for improvement in each area.
They also correspond to a more qualitative description of the organization’s developmental
stage. Each capacity area can be characterized
as nascent, emerging, expanding, or mature. DYNAMIC PARTICIPATORY
OCAT provides a table (similar to the IDF), INSTITUTIONAL DIAGNOSIS
“NGO Organizational Development—Stages
and Characteristics” that describes organiza- Background
tional capacities at each stage of development.
The dynamic participatory institutional diagnoProduct
sis (DPID) was developed by the Senegal PVO/
NGO support project in conjunction with the
The OCAT provides numeric ratings for each New TransCentury Foundation and Yirawah Incapacity area. In addition, it gives organizations ternational. It is a rapid and intensive facilitated
a description of their capacity areas in terms assessment of the overall strengths and weakof progressive stages of organizational develop- nesses of an organization. This methodology
ment. This information can be presented graph- explores member perceptions of an organizaically as well as in narrative form.
tion and the organization’s relationship with its
environment. DPID is highly participatory; an
Assessment
organization assesses itself in the absence of
external benchmarks or objectives to take full
The OCAT identifies areas of organization- advantage of its specific context, such as culture
al strength and weakness and tracks related and attitudes.
changes from one measurement period to the
next.
Process
14
Example 4. An Application of DPID
Since the DPID is such an individualized and flexible tool, every application will be different.
The DPID does not lend itself easily to an example as do the other tools in this Tips. Below
is an anecdote about one West African organization’s use of the DPID as reported by the
Senegal DPIPVO/NGO support project.
A Federation of Farmers’ Cooperatives with about 15,000 members in the Sahel was looking for a unique and efficient approach to redress some of the organization’s problems. The
federation suffered from internal strife and a tarnished reputation, impeding its ability to raise
funds. Through DPID, the federation conducted a critical in-depth analysis of its operational
and management systems, resulting in the adoption of “10 emergency measures” addressing
leadership weaknesses, management systems, and operational procedures. Subsequently, the
organization underwent internal restructuring, including an overhaul of financial and administrative systems. One specific result of the DPID analysis was that federation members gained
more influence over the operations of the federation.
An outside facilitator conducts the DPID over
5 to 10 days. It takes place during a series of
working sessions in which the facilitator leads
an organization’s members through several
stages: discussion of the services; operations
and results of the organization; exploration of
the issues affecting the organization; and summarization of the “state of the organization.”
During the discussions, members analyze the
following features of the organization:
They examine each element with reference to
institutional behavior, human behavior, management, administration, know-how, philosophy
and values, and sensitive points.
• Identity
Assessment
• Mission
Unlike the previously described tools, the DPID
does not use ranking, scoring, or questionnaires,
nor does it assess the organization along a continuum of developmental stages. Assessment
is based purely on group reflection. The DPID
requires a facilitator experienced in leading a
group through this type of analysis.
• Means and Resources
• Environment
• Management
• Internal Operations
• Service Provided and Results
Product
A written description of the state of the organization can result from the working sessions.The
analysis is qualitative without numeric scoring.
The DPID is open ended but somewhat systematic in covering a predefined set of organizational functions. Because of its flexibility, the
DPID is organization specific and should not
15
ORGANIZATIONAL CAPACITY
INDICATOR
Dynamic Participatory
Institutional Diagnosis
Background
From 1994 through 1997, the Christian Reformed World Relief Committee (CRWRC)
conducted research on organizational capacitybuilding with the Weatherhead School of Management at Case Western Reserve University
and more than 100 local NGOs around the
world. The results of this research led them to
replace their earlier system, the Skill Rating System, with an approach to capacity building and
assessment based on “appreciative inquiry.” Appreciative inquiry is a methodology that emphasizes an organization’s strengths and potential
more than its problems. It highlights those qualities that give life to an organization and sustain its ongoing capacity. Rather than providing
a standardized tool, the organizational capacity
indicator assumes that capacity monitoring is
unique to each organization and in the organization’s own self-interest. The organizational
capacity indicator (OCI) builds ownership because each organization creates its own capacity
assessment tool. Capacity areas are self-defined
and vary from organization to organization.
Type of Organization Measured
NGOs/PVOs; adaptable to other types of organizations
Features
• Difficult to compare across organizations
• Difficult to compare the same organization over time
• Capacity areas and criteria for measurement are loosely defined
• Assessment based primarily upon perceived capacities
• Produces qualitative description of an
organization’s capacity
• Assessment done with the help of an
outside facilitator
• Data collected through group discussion
with the organization’s staff
Process
Although organizations create their own tool
under the OCI, they all follow a similar probe used to compare organiza tions. Nor is it a
cess in doing so. As they involve all partners
rigorous means of monitoring an organization’s
and stakeholders as much as possible, the parchange over time. Since the DPID does not use
ticipants “appreciate” the organization’s history
external standards to assess institutional caand culture.Together they explore peak experipacities, it should not be used to track accountences, best practices, and future hopes for the
ability. Collecting information from the DPID,
organization. Next, the participants identify the
as well as using it, should offer organizations a
forces and factors that have made the organizaprocess to assess their needs, improve commution’s positive experiences possible. These benications, and solve problems around a range of
come the capacity areas that the organization
organizational issues at a given moment.
tries to monitor and improve.
Next, the participants develop a list of “provoca16
ences related to each capacity component. The
organization should monitor itself by this process twice a year. The results of the assessment
should be used to encourage future development, plans, and aspirations.
tive propositions” for each capacity area. These
propositions, visions of what each capacity area
should ideally look like in the future, contribute
to the overall objective: that each organization
will be able to measure itself against its own
vision for the future, not some external standard. Each capacity area is defined by the most
ambitious vision of what the organization can
become in that area. Specific indicators or behaviors are then identified to show the capacity
area in practice. Next, the organization designs
a process for assessing itself and sharing experi-
Product
Each time a different organization uses the
methodology, a different product specific to
that organization is developed. Thus, each tool
will contain a unique set of capacity areas, an
Example 5. Excerpt From an OCI Tool
The following is an excerpt of one section from the capacity assessment tool developed by
CRWRC’s partners in Asia, using the OCI method. (The entire tool can be found in annex 5.)
It offers a menu of capacity areas and indicators from which an organization can choose and
then modify for its own use. It identifies nine capacity areas, and under each area is a “provocative proposition” or vision of where the organization wants to be in that area. It provides an
extensive list of indicators for each capacity area, and it describes the process for developing
and using the tool. Staff and partners meet regularly to determine their capacity on the chosen indicators. Capacity level can be indicated pictorially, for example by the stages of growth
of a tree or degrees of happy faces.
Capacity Area
A clear vision, mission, strategy, and set of shared values
Proposition
Our vision expresses our purpose for existing: our dreams, aspirations, and concerns for the
poor. Our mission expresses how we reach our vision. Our strategy expresses the approach
we use to accomplish our goals. The shared values that we hold create a common understanding and inspire us to work together to achieve our goal.
Selected Indicators
• Every person can state the mission and vision in his or her own words
• There is a yearly or a six-month plan, checked monthly
• Operations/activities are within the vision, mission, and goal of the organization
• Staff know why they do what they’re doing
• Every staff member has a clear workplan for meeting the strategy
• Regular meetings review and affirm the strategy
17
• Possible to measure well-defined capacity areas across well-defined criteria
organization capacity areas like the DPID does.
The OCI is the only tool presented in this paper in which the capacity areas are entirely self
defined. It is also unique in its emphasis on the
positive, rather than on problems. Further, the
OCI is more rigorous than the DPID, in that it
asks each organization to set goals and develop
indicators as part of the assessment process. It
also calls for a scoring system to be developed,
like the more formal tools (PROSE, IDF, OCAT).
Because indicators and targets are developed
for each capacity area, the tool allows for relatively consistent measurement over time. OCI
is not designed to compare organizations with
each other or to aggregate the capacity measures of a number of organizations; however, it
has proven useful in allowing organizations to
learn from each other and in helping outsiders
assess and understand partner organizations.
• Assessment based primarily upon perceived capacities
THE YES/NO CHECKLIST OR
“SCORECARD”
• Produces numeric or pictorial score on
capacity areas
Background
Organizational Capacity
Indicator
Type of Organization Measured
NGOs/PVOs; adaptable to other types of
organizations
Features
• Difficult to comparably measure across
organizations
• Measures change in the same organization over time
A scorecard/checklist is a list of characteristics
or events against which a yes/no score is assigned. These individual scores are aggregated
and presented as an index. Checklists can effectively track processes, outputs, or more general
characteristics of an organization. In addition,
they may be used to measure processes or outputs of an organization correlated to specific
areas of capacity development.
• Assessment done internally
• Data collected through group discussion
with organization’s staff
evaluation process, and scoring methods. In
general, the product comprises a written description of where the organization wants to be
in each capacity area, a list of indicators that can Scorecards/checklists can be used either to
be used to track progress toward the targeted measure a single capacity component of an
organization or several rolled together. Scorelevel in a capacity area, and a scoring system.
cards/checklists are designed to produce a
quantitative score that can be used by itself or
Assessment
as a target (though a scorecard/checklist withLike the DPID, the OCI is highly participatory out an aggregate score is also helpful).
and values internal standards and perceptions.
Both tools explicitly reject the use of external
standards. However, the OCI does not desi nate
18
Process
The Yes/No Checklist
“Scorecard”
To construct a scorecard, follow these general
steps: First, clarify what the overall phenomena
to be measured are and identify the components that, when combined, cover the phenomenon fairly well. Next, develop a set of characteristics or indicators that together capture the
relevant phenomena. If desired, and if evidence
and analysis show that certain characteristics
are truly more influential in achieving the overall result being addressed, define a weight to be
assigned to each characteristic/indicator. Then
rate the organization(s) on each characteristic
using a well defined data collection approach.
The approach could range from interviewing
organization members to reviewing organization documents, or it could consist of a combination of methods. Finally, if desired and appropriate, sum the score for the organization(s).
Type of Organization Measured
All types of organizations
Features
• Cross-organizational comparisons can
be made
• Measures change in the same organization over time
• Measures well-defined capacity areas
against well-defined criteria
Product
• Possible to balance perceptions with
empirical observations
• Produces numeric score on capacity
areas
A scorecard/checklist results in a scored listing
of important characteristics of an organization
and can also be aggregated to get a summary
score.
• Assessment can be done by an external
evaluator or internally
• Data collected through interviews, observation, documents, involving a limited
number of staff
Assessment
A scorecard/checklist should be used when the
characteristics to be scored are unambiguous.
There is no room for “somewhat” or “yes, but . .
.” with the scorecard technique.The wording of
each characteristic should be clear and terms
should be well defined. Because scorecards/
checklists are usually based on observable facts,
processes, and documents, they are more objective than most of the tools outlined in this Tips.
This, in turn, makes them particularly useful for
cross-organizational comparisons, or tracking
organizations over time; that is, they achieve
better measurement consistency and comparability. Yet concentrating on observable facts can
be limiting, if such facts are not complemented
19
with descriptive and perceptionbased information. Though a person outside the organization
frequently completes the scorecard/checklist,
self-assessment is also possible. Unlike other
tools that require facilitators to conduct or
interpret them, individuals who are not highly
trained can also use scorecards. Further, since
scorecards are usually tightly defined and specific, they are often a cheaper measurement
tool.
Example 6. A Scorecard
USAID/Mozambique developed the following scorecard to measure various aspects of institutional capacity in partner civil society organizations. The following example measures democratic governance.
Increased Democratic Governance Within Civil Society Organizations
Characteristics
Score Multiplied Weight
By
1. Leaders (board member or equivalent) of the
X
3
CSO electedby secret ballot. No=0 pts.Yes=1 pt.
2. General assembly meetings are adequately
announced at least two weeks in advance to all
members (1 pt.) and held at least twice a year (1
pt.). Otherwise=0 pt.
3. Annual budget presented for member approval. No=0 pts.Yes=1 pt.
4. Elected leaders separate from paid employees.
No=0 pts.Yes=1 pt.
5. Board meetings open to ordinary members
(nonboard members). No=0 pts.Yes=1 pt.
X
2
X
2
X
2
X
1
Weighted
Score
Total
external relations and internal governance.
development professionals also
MEASURING INDIVIDUAL Organizational
use other tools to measure specific capacity
ORGANIZATIONAL
areas. Some drawbacks of these tools are that
they require specialized technical expertise and
COMPONENTS
they can be costly to use on a regular basis.
Other tools may require some initial training
In some cases, USAID is not trying to strength- but can be much more easily institutionalized.
en the whole organization, but rather specific Below we have identified some tools for meaparts of it that need special intervention. In many suring selected organizational components.
cases, the best way of measuring more specific (You will find complete reference information
organizational changes is to use portions of the for these tools in the resources section of this
instruments described. For instance, the IDF Tips.)
has a comparatively well-developed section
on management resources (leadership style, STRUCTURE AND CULTURE
participatory management, planning, monitoring and evaluation, and management systems). The Preferred Organizational Structure instruSimilarly, the OCAT has some good sections on ment is designed to assess many aspects of or20
ganizational structure, such as formality of rules,
communication lines, and decision-making. This
tool requires organizational development skills,
both to conduct the assessment and to interpret the results.
HUMAN RESOURCES AND THEIR
MANAGEMENT
First, the development of indicators should be
driven by the informational needs of managers,
from both USAID and the given relevant organizations-- to inform strategic and operational
decisions and to assist in reporting and communicating to partners and other stakeholders.
At times, there is a tendency to identify or design a data collection instrument without giving
too much thought to exactly what information
will be needed for management and reporting.
In these situations, indicators tend to be developed on the basis of the data that have been
collected, rather than on what managers need.
More to the point, the development of indicators should follow a thorough assessment of
informational needs and precede the identification of a data collection instrument. Managers should first determine their informational
needs; from these needs, they should articulate
and define indicators; and only then, with this
information in hand, they would identify or
develop an instrument to collect the required
data. This means that, in most cases, indicators
should not be derived, post facto, from a data
collection tool. Rather, the data collection tool
should be designed with the given indicators in
mind. Second, indicators should be developed
for management decisions at all levels (input indicators, output indicators, process indicators,
and outcome/impact indicators). With USAID’s
increased emphasis on results, managers sometimes may concentrate primarily on strategic
indicators (for development objectives and
intermediate results). While an emphasis on
results is appropriate, particularly for USAID
managers, tracking operational-level information for the organizations supported through
a given Agency program is critical if managers
are to understand if, to what degree, and how
the organizations are increasing their capacities. The instruments outlined in this paper can
provide data for indicators defined at various
management levels.
Many personnel assessments exist, including the
Job Description Index and the Job Diagnostic
Survey, both of which measure different aspects
of job satisfaction, skills, and task significance.
However, skilled human resource practitioners
must administer them. Other assessments, such
as the Alexander Team Effectiveness Critique,
have been used to examine the state and functioning of work teams and can easily be applied
in the field.
SERVICE DELIVERY
Often, a customer survey is one of the best
ways to measure the efficiency and effectiveness of a service delivery system. A specific
customer survey would need to be designed
relative to each situation. Example 7 shows a
sample customer service assessment.
DEVELOPING INDICATORS
Indicators permit managers to track and understand activity/program performance at both
the operational (inputs, outputs, processes)
and strategic (development objectives and intermediate results) levels. To managers familiar
with the development and use of indicators, it
may seem straightforward to derive indicators
from the instruments presented in the preceding pages. However, several critical points will
ensure that the indicators developed within the
context of these instruments are useful to managers.
Finally, indicators should meet the criteria out21
Example 7. A Customer Service Assessment
1. In the past 12 months, have you ever contacted a municipal office to complain about something such as poor
city services or a rude city official, or any other reason?
________No ________Yes
If YES:
1a. How many different problems or complaints did you contact the municipality about in the last 12 months?
________One ________Two ________Three to five ________More than five
1b. Please describe briefly the nature of the complaint starting with the one you feel was most important.
1._______________________________________________
2._______________________________________________
3._______________________________________________
2. Which department or officials did you contact initially regarding these complaints?
____Mayor’s office
____Council member
____Police
____Sanitation
____Public works
____Roads
____Housing
____Health
____Other________________________________________
2a. Were you generally satisfied with the city’s response? (IF DISSATISFIED, ASK: What were the major
reasons for your dissatisfaction?)
_____Response not yet completed
_____Satisfied
_____Dissatisfied, never responded or corrected condition
_____Dissatisfied, poor quality or incorrect response was provided
_____Dissatisfied, took too long to complete response, had to keep pressuring for results, red tape, etc.
_____Dissatisfied, personnel were discourteous, negative, etc.
_____Dissatisfied, other_____________________________
3. Overall, are you satisfied with the usefulness, courtesy and effectiveness of the municipal department or
official that you contacted?
_____Definitely yes
_____Generally yes
_____Generally no (explain)__________________________
_____Definitely no (explain)__________________________
Survey adapted from Hatry, Blair, and others, 1992.
22
lined in USAID’s Automated Directives System
and related pieces of Agency guidance such as
CDIE’s Performance Monitoring and Evaluation Tips #6, “Selecting Performance Indicators,” and Tips #12, “Guidelines for Indicator
and Data Quality.” That is, indicators should be
direct, objective, practical, and adequate. Once
an indicator has been decided upon, it is important to document the relevant technical details:
a precise definition of the indicator; a detailed
description of the data source; and a thorough
explanation of the data collection method. (Refer to Tips #7, “Preparing a Performance Monitoring Plan.”)
data). If a result refers to multiple organizations,
it might be useful to frame an indicator in terms
of the number or percent of the organizations
that meet or exceed a given threshold score or
development stage, on the basis of an aggregate
index or the score of a single element for each
organization. The key is to ensure that the indicator reflects the result and to then identify the
most appropriate and useful measurement instrument.
RESULTS-LEVEL INDICATORS
Example 8 includes real indicators used by USAID missions in 1998 to report on strategic objectives and intermediate results in institutional
capacity strengthening.
USAID managers spend substantial time and
energy developing indicators for development
objectives and intermediate results related to
institutional capacity. The range of the Agency’s
institutional strengthening programs is broad,
as is the range of the indicators that track the
programs’ results. Some results reflect multiple
organizations and others relate to a single organization. Additionally, of those results that relate to multiple organizations, some may refer
to organizations from only one sector while
others may capture organizations from a number of sectors. Results related to institutional
strengthening also vary relative to the level of
change they indicate-- such as an increase in institutional capacity versus the eventual impact
generated by such an i crease-- and with regard to whether they reflect strengthening of
the whole organization(s) or just one or several
elements. It is relatively easy to develop indicators for all types of results and to use the instruments outlined in this Tips to collect the necessary data. For example, when a result refers
to strengthening a single organization, across
all elements, an aggregate index or “score” of
institutional strength may be an appropriate indicator (an instrument based on the IDF or the
scorecard model might be used to collect such
PRACTICAL TIPS FOR A
BUSY USAID MANAGER
This TIPS introduces critical issues related to
measuring institutional capacity. It presents a
number of approaches that managers of development programs and activities currently use
in the field. In this section we summarize the
preceding discussion by offering several quick
tips that USAID managers should find useful as
they design, modify, and implement their own
approaches for measuring institutional capacity.
1. Carefully review the informational needs of
the relevant managers and the characteristics of the organization to be measured to
facilitate development of indicators. Identify
your information needs and develop indicators
before you choose an instrument.
23
2. To assist you in selecting an appropriate
measurement tool, ask yourself the following
questions as they pertain to your institutional
capacity measurement situation. Equipped
with the answers to these questions, you
Example 8.
Selected Institutional Capacity Indicators From USAID Missions
Indicator
• Number of institutions meeting at least
80% of their targeted improvements
To Measure
Institutions strengthened (entire organization)
• Amount of funds raised from non-USAID
Institutions more financially sustainable
sources
• Number of organizations where USAID
contribution is less than 25% of revenues
• Number of organizations where at least
five funding sources contribute at least 10%
each
• Percent of suspected polio cases investiga- Organization’s service delivery systems
tee within 48 hours
strengthened
• Number of governmental units displaying
Local government management capacities
improved practices, such as open and trans- improved
parent financial systems, set organizational
procedures, accountability, participatory
decision-making, by-laws and elections
can scan the “features list” that describes
every tool in this paper to identify which
measurement approaches to explore further.
• Is the objective to measure the entire organization? Or is it to measure
specific elements of the organization?
If the latter, what are the specific capacity areas of functions to be measured?
• How will the information be used?
To measure change in an organization over time? To compare organizations with each other?
To inform procurement decisions?
To hold an organization accountable
for achieving results or implementing
reforms?
• What type of organizations are you
measuring? Are there any particular
measurement issues pertaining to
this type of organization that must
be considered?
• How participatory do you want the
measurement process to be?
• What is the purpose of the intervention? To strengthen an organization?
24
• Will organization members themselves or outsiders conduct the
assessment?
• What product do you want the measurement tool to generate?
and involve some form of ordinal scaling/
scoring. When reviewing data, managers
should therefore zero in on the direction and
general degree of change. Do not be overly
concerned about small changes; avoid false
precision.
• Do you want the measurement process to be an institution-strengthening exercise in itself?i. Do you need
an instrument that measures one
organization? Several organizations
againstindividual criteria? Or several organizations against standard
criteria?
5. Cost matters-- and so does the frequency
and timing of data collection. Data need to
be available frequently enough, and at the
right point in the program cycle, to inform
operational and strategic management decisions. Additionally, the management benefits
of data should exceed the costs associated
with their collection.
3. If you are concerned about data reliability, apply measurement instruments consistently over
time and across organizations to ensure data
reliability. You can adapt and adjust tools as
needed, but once you develop the instru- 6. The process of measuring institutional capacity
can contribute substantially to increasing an orment, use it consistently.
ganization’s strength. A number of measurement approaches are explicitly designed as
4. When interpreting and drawing conclusions
learning opportunities for organizations;
from collected data, remember the limits of the
that is, to identify problems and suggest rerelevant measurement tool. Most methods for
lated solutions, to improve communication,
measuring institutional capacity are subjecor to facilitate a consensus around future
tive, as they are based on the perceptions
priorities
of those participating in the assessment,
This TIPS was prepared for CDIE by Alan Lessik and
Victoria Michener of Management Systems International.
Bibliography
RESOURCES
Booth, W.; and R. Morin. 1996. Assessing Organizational Capacity Through Participatory
Monitoring and Evaluation Handbook. Prepared for the Pact Ethiopian NGO Sector
Enhancement Initiative. Washington: USAID.
Center for Democracy and Governance. 1998. Handbook of Democracy and Governance
25
Program Indicators.Washington: U.S. Agency for International Development.
Christian Reformed World Relief Committee. 1997. Partnering to Build and Measure
Organizational Capacity. Grand Rapids, Mich.
Cooper, S.; and R. O’Connor. 1993. “Standards for Organizational Consultation: Assessment and Evaluation Instruments.” Journal of Counseling and Development 71: 651-9.
Counterpart International. N.d. “CAP Monitoring and Evaluation Questionnaire.”
—N.d. “Manual for the Workshop on Development of a Training and Technical Assistance Plan (TTAP).”
—N.d. “Institutional Assessment Indicators.”
Drucker, P.; and C. Roseum. 1993. How to Assess Your Nonprofit Organization with Peter Drucker’s Five
Important Questions: User Guide for Boards, Staff,Volunteers and Facilitators. Jossey--Bass .
Eade, D. 1997. Capacity-Building: An Approach to People-Centred Development. Oxford: Oxfam.
Fowler, A.; L. Goold; and R. James. 1995. Participatory Self Assessment of NGO Capacity. INTRAC
Occasional Papers Series No. 10. Oxford.
Hatry, H.; L. Blair; D. Fisk; J. Grenier; J. Hall; and P. Schaenman. 1992. How Effective Are Your Community
Services? Procedures for Measuring Their Quality. Washington: The Urban
Institute.
International Working Group on Capacity Building of Southern NGOs. 1998. “Southern NGO Capacity Building: Issues and Priorities.” New Delhi: Society for Participatory Research in Asia.
International Working Group on Capacity Building for NGOs. 1998. “Strengthening Southern NGOs: The Donor Perspective.” Washington: USAID and The World Bank.
Kelleher, D. and K. McLaren with R. Bisson. 1996. “Grabbing the Tiger by the Tail: NGOs
Learning forOrganizational Change.” Canadian Council for International Cooperation.
Lent, D. October 1996. “What is Institutional Capacity?” On Track: The Reengineering Digest. 2 (7): 3. Washington: U.S. Agency for International Development.
Levinger, B. and E. Bloom. 1997. Introduction to DOSA: An Outline Presentation.
http://www.edc.org/int/capdev/dosafile/dosintr.htm.
Lusthaus, C., G. Anderson, and E. Murphy. 1995. “Institutional Assessment: A Framework for
Strengthening Organizational Capacity for IDRC’s Research Partners.” IDRC.
26
Mentz, J.C.N. 1997. “Personal and Institutional Factors in Capacity Building and Institutional
Development.” European Centre for Development Policy Management Working Paper
No. 14.
Morgan, P.; and A. Qualman. 1996. “Institutional and Capacity Development, Results-Based
Management and Organisational Performance.” Canadian International Development
Agency.
New TransCentury Foundation. 1996. Practical Approaches to PVO/NGO Capacity Building:
Lessons from the Field (five monographs). Washington: U.S.Agency for International
Development.
Pact. N.d. “What is Prose?”
—1998. “Pact Organizational Capacity Assessment Training of Trainers.” 7-8 January.
Renzi, M. 1996. “An Integrated Tool Kit for Institutional Development.”Public Administration
and Development 16: 469-83.
—N.d. “The Institutional Framework: Frequently Asked Questions.” Unpublished paper.
Management Systems International.
Sahley, C. 1995. “Strengthening the Capacity of NGOs: Cases of Small Enterprise Development
Agencies in Africa.” INTRAC NGO Management and Policy Series. Oxford.
Save the Children. N.d. Institutional Strengthening Indicators: Self Assessment for NGOs. UNDP. 1997.
Capacity Assessment and Development. Technical Advisory Paper No. 3, Management
Development and Governance Division. New York.
Bureau for Policy and Program Coordination. 1995. USAID-U.S. PVO Partnership. Policy
Guidance. Washington: U.S. Agency for International Development.
Office of Private and Voluntary Cooperation. 1998. USAID Support for NGO Capacity-Building:
Approaches, Examples, Mechanisms. Washington: U.S. Agency for International
Development.
—1998. Results Review Fiscal Year 1997. Washington: U.S. Agency for International
Development.
NPI Learning Team. 1997. New Partnerships Initiative: A Strategic Approach to Development
Partnering. Washington: U.S. Agency for International Development. 23
USAID/Brazil. 1998. Fiscal Year 2000 Results Review and Resource Request.
USAID/Guatemala. 1998. Fiscal Year 2000 Results Review and Resource Request.
27
USAID/Indonesia. 1998. Fiscal Year 2000 Results Review and Resource Request.
USAID/Madagascar. 1998. Fiscal Year 2000 Results Review and Resource Request.
—1997. Institutional Capacity Needs Assessment.
USAID/Mexico. 1998. The FY 1999--FY 2003 Country Strategy for USAID in Mexico.
USAID/Mozambique. 1998. Fiscal Year 2000 Results Review and Resource Request.
USAID/West Bank--Gaza. 1998. Fiscal Year 2000 Results Review and Resource Request.
Whorton, J.; and D. Morgan. 1975. Measuring Community Performance: A Handbook of
Indicators, University of Oklahoma.
World Bank. 1996. Partnership for Capacity Building in Africa: Strategy and Program of Action.
Washington.
World Learning. 1998. Institutional Analysis Instrument: An NGO Development Tool.
Sources of Information on Institutional Capacity Measurement Tools
Discussion-Oriented Organizational Self-Assessment:
http://www.edc.org/int/capdev/dosafile/dosintr.htm.
Institutional Development Framework: Management Systems International. Washington.
Organizational Capacity Assessment Tool: http://www.pactworld.org/ocat.html Pact.Washington.
Dynamic Participatory Institutional Diagnostic: New TransCentury Foundation. Arlington,Va.
Organizational Capacity Indicator: Christian Reformed World Relief Committee. Grand
Rapids, Mich.
Smith, P.; L. Kendall; and C. Hulin. 1969. The Measurement of Satisfaction in Work and
Retirement. Rand McNally.
Hackman, J.R.; and G.R. Oldham. 1975. “Job Diagnostic Survey: Development of the Job
Diagnostic Survey”
Journal of Applied Psychology 60: 159-70.
Goodstein, L.D.; and J.W. Pfeiffer, eds. 1985. Alexander Team Effectiveness Critique:The 1995
Annual: Developing Human Resources. Pfeiffer & Co.
28
Bourgeois, L.J.; D.W. McAllister; and T.R. Mitchell. 1978. “Preferred Organizational Structure:
The Effects of Different Organizational Environments Upon Decisions About
Organizational Structure.” Academy of Management Journal 21: 508-14.
Kraut, A. 1996. Customer and Employee Surveys: Organizational Surveys:Tools for Assessment and
Change. Jossey-Bass Publishers. 24
29
NUMBER 16
1ST EDITION 2010
PERFORMANCE MONITORING & EVALUATION
TIPS
CONDUCTING MIXED-METHOD EVALUATIONS
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to
performance monitoring and evaluation. This publication is a supplemental reference to the
Automated Directive System (ADS) Chapter 203.
INTRODUCTION
This TIPS provides guidance on
using a mixed-methods approach
for
evaluation
research.
Frequently, evaluation statements
of work specify that a mix of
methods be used to answer
evaluation questions. This TIPS
includes the rationale for using a
mixed-method evaluation design,
guidance for selecting among
methods (with an example from
an evaluation of a training
program) and examples of
techniques for analyzing data
collected with several different
methods
(including
―parallel
analysis‖).
MIXED-METHOD
EVALUATIONS
DEFINED
A mixed-method evaluation is
one that uses two or more
techniques or methods to collect
the data needed to answer one or
more evaluation questions. Some
of the different data collection
methods that might be combined
in
an
evaluation
include
structured
observations,
key
informant interviews, pre- and
post-test surveys, and reviews of
government statistics. This could
involve the collection and use of
both quantitative and qualitative
data to analyze and identify
findings
and
to
develop
conclusions in response to the
evaluation questions.
1
RATIONALE FOR
USING A MIXEDMETHOD
EVALUATION DESIGN
There are several possible cases
in which it would be highly
beneficial to employ mixedmethods in an evaluation design:
 When a mix of different
methods is used to collect data
from different sources to
provide independent estimates
of key indicators—and those
estimates complement one
another—it
increases
the
validity of conclusions related
to an evaluation question. This
is referred to as triangulation.
(See TIPS 5: Rapid Appraisal,
and Bamberger, Rugh and
Key Steps in Developing a Mixed-Method Evaluation Design and Analysis
Strategy
1. In order to determine the methods that will be employed, carefully review the purpose of the evaluation and the
primary evaluation questions. Then select the methods that will be the most useful and cost-effective to answer
each question in the time period allotted for the evaluation. Sometimes it is apparent that there is one method
that can be used to answer most, but not all, aspects of the evaluation question.
2. Select complementary methods to cover different aspects of the evaluation question (for example, the how and
why issues) that the first method selected cannot alone answer, and/or to enrich and strengthen data analysis
and interpretation of findings.
3. In situations when the strength of findings and conclusions for a key question is absolutely essential, employ a
triangulation strategy. What additional data sources and methods can be used to obtain information to answer
the same question in order to increase the validity of findings from the first method selected?
4. Re-examine the purpose of the evaluation and the methods initially selected to ensure that all aspects of the
primary evaluation questions are covered thoroughly. This is the basis of the evaluation design. Develop data
collection instruments accordingly.
5. Design a data analysis strategy to analyze the data that will be generated from the selection of methods chosen
for the evaluation.
6. Ensure that the evaluation team composition includes members that are well-versed and experienced in applying
each type of data collection method and subsequent analysis.
7. Ensure that there is sufficient time in the evaluation statement of work for evaluators to fully analyze data
analysis
canof conducting
also  Mixed-method
evaluations
generated[2006]
from eachfor
method
employedricher
and to realize
the and
benefits
a mixed method evaluation.
Mabry
further
explanation and descriptions of
triangulation strategies used in
evaluations.)
provide a better understanding
of the context in which a
program operates.
 When reliance on one method
There are a number of additional
benefits derived from using a mix
of methods in any given
evaluation.
alone may not be sufficient to
answer all aspects of each
evaluation question.
 When the data collected from
one method can help interpret
findings from the analysis of
data collected from another
method.
For
example,
qualitative data from in-depth
interviews or focus groups can
help
interpret
statistical
patterns from quantitative data
collected through a randomsample survey. This yields a
 Using
mixed-methods can
more readily yield examples of
unanticipated
changes
or
responses.
 Mixed-method
evaluations
have the potential of surfacing
other key issues and providing
a deeper understanding of
program context that should
be considered when analyzing
data and developing findings
and conclusions.
2
often yield a wider range of
points of view that might
otherwise be missed.
DETERMINING
WHICH METHODS TO
USE
In a mixed-method evaluation,
the evaluator may use a
combination of methods, such as
a survey using comparison
groups in a quasi-experimental or
experimental design, a review of
key documents, a reanalysis of
government statistics, in-depth
interviews with key informants,
focus groups, and structured
observations. The selection of
methods, or mix, depends on the
nature of the evaluation purpose
and the key questions to be
addressed.
SELECTION OF DATA
COLLECTION
METHODS – AN
EXAMPLE
The selection of which methods
to use in an evaluation is
driven by the key evaluation
questions to be addressed.
Frequently,
one
primary
evaluation method is apparent.
For
example,
suppose
an
organization wants to know
about the effectiveness of a pilot
training program conducted for
100 individuals to set up their
own small businesses after the
completion of the training.
The evaluator should ask what
methods are most useful and
cost-effective to assess the
question of the effectiveness of
that training program within the
given time frame allotted for the
evaluation. The answer to this
question must be based on the
stated outcome expected from
the training program. In this
example, let us say that the
organization’s expectations were
that, within one year, 70 percent
of the 100 individuals that were
trained will have used their new
skills and knowledge to start a
small business.
What is the best method to
determine whether this outcome
has been achieved? The most
costeffective
means
of
answering this question is to
survey 100 percent of the
individuals who graduated from
the training program using a
close-ended questionnaire.
It
follows that a survey instrument
should be designed to determine
if these individuals have actually
succeeded in starting up a new
business.
While this sounds relatively
straightforward, organizations are
often interested in related issues.
If less than 70 percent of the
individuals started a new business
one year after completion of the
training,
the
organization
generally wants to know why
some
graduates
from
the
program were successful while
others were not. Did the training
these individuals received actually
help them start up a small
business? Were there topics that
should have been covered to
more thoroughly prepare them
for the realities of setting up a
business?
Were there other
topics that should have been
addressed?
In summary, this
organization wants to learn not
only whether at least 70 percent
of the individuals trained have
started up a business, but also
how effectively the training
equipped them to do so. It also
wants to know both the strengths
and the shortcomings of the
training so that it can improve
future training programs.
The organization may also want
to know if there were factors
outside the actual intervention
that had a bearing on the
training’s success or failure. For
example, did some individuals
find employment instead? Was
3
access to finance a problem? Did
they conduct an adequate market
analysis? Did some individuals
start with prior business skills?
Are there factors in the local
economy, such as local business
regulations, that either promote
or discourage small business
start-ups? There are numerous
factors
which
could
have
influenced this outcome.
The selection of additional
methods to be employed is,
again, based on the nature of
each aspect of the issue or set
of related questions that the
organization wants to probe.
To continue with this example,
the evaluator might expand the
number of survey questions to
address issues related to the
effectiveness of the training and
external factors such as access to
finance.
These
additional
questions can be designed to
yield additional quantitative data
and to probe for information
such as the level of satisfaction
with the training program, the
usefulness
of
the
training
program
in
establishing
a
business, whether the training
graduate
received
a
small
business start-up loan, if the size
of the loan the graduate received
was sufficient, and whether
graduates are still in the process
of starting up their businesses or
instead have found employment.
Intake data from the training
program on characteristics of
each trainee can also be
examined to see if there are any
particular characteristics, such as
sex or ethnic background, that
can be correlated with the survey
findings.
It is important to draw on
additional methods to help
explain the statistical findings
from the survey, probe the
strengths and shortcomings of
the training program, further
understand issues related to
access to finance, and identify
external factors affecting success
in starting a business.
In this
case, the evaluation design could
focus on a sub-set of the 100
individuals to obtain additional
qualitative information.
A
selected group of 25 people
could be asked to answer an
additional series of open-ended
questions during the same
interview session, expanding it
from 30 minutes to 60 minutes.
Whereas asking 100 people
open-ended questions would be
better than just 25 people, costs
prohibit interviewing the entire
group.
Using
the
same
example,
suppose the organization has
learned
through
informal
feedback that access to finance is
likely a key factor in determining
success in business start-up in
addition to the training program
itself.
Depending on the
evaluation
findings,
the
organization may want to design
a finance program that increases
access to loans for small business
start-ups. To determine the
validity of this assumption, the
evaluation design relies on a
triangulation approach to assess
whether and how access to
finance for business start-ups
provides further explanations
regarding success or failure
outcomes. The design includes a
plan to collect data from two
other sources using a separate
data collection method for each
source. The first data source
includes the quantitative data
from the survey of the 100
training
graduates.
The
evaluation designers determine
that the second data source will
be the managers of local banks
and credit unions that survey
respondents reported having
approached for start-up loans.
In-depth interviews will be
conducted
to
record
and
understand policies for lending to
entrepreneurs trying to establish
small businesses, the application
of those policies, and other
business practices with respect to
prospective clients.
The third
data source is comprised of bank
loan statistics for entrepreneurs
who have applied to start up
small businesses. Now there are
three independent data sources
using different data collection
methods to assess whether
access to finance is an additional
key factor in determining small
business start-up success.
In this example, the total mix of
methods the evaluator would use
includes the following: the survey
of all 100 training graduates, data
from open-ended questions from
a subset of graduates selected for
longer interviews, analysis of
training intake data on trainee
characteristics,
in-depth
interviews with managers of
lending institutions, and an
examination of loan data. The
use of mixed-methods was
necessary because the client
organization in this case not only
wanted to know how effective the
pilot training course was based
4
on its own measure of program
success, but also whether access
to finance contributed to either
success or failure in starting up a
new business. The analysis of the
data will be used to strengthen
the training design and content
employed in the pilot training
course, and as previously stated,
perhaps to design a microfinance
program.
The last step in the process of
designing
a
mixed-method
evaluation is to determine how
the data derived from using
mixed-methods will be analyzed
to produce findings and to
determine the key conclusions.
ANALYZING DATA
FROM A MIXEDMETHOD
EVALUATION –
DESIGNING A DATA
ANALYSIS STRATEGY
It is important to design the data
analysis strategy before the
actual data collection begins.
Having done so, the evaluator
can begin thinking about trends
in findings from different sets of
data to see if findings converge
or diverge.
Analyzing data
collected from a mixture of
methods is admittedly more
complicated than analyzing the
data derived from one method.
This entails a process in which
quantitative and qualitative data
analysis strategies are eventually
connected to determine and
understand key findings. Several
different techniques can be used
to analyze data from mixedmethods approaches, including
parallel
analysis,
conversion
analysis,
sequential
analysis,
multilevel analysis, and data
synthesis. The choice of analytical
techniques should be matched
with the purpose of the
evaluation using mixed-methods.
Table 1 briefly describes the
different analysis techniques and
the situations in which each
method is best applied.
In
complex
evaluations
with
multiple issues to address, skilled
evaluators may use more than
one of these techniques to
analyze the data.
EXAMPLE OF
APPLICATION
Here we present an example of
parallel
mixed-data
analysis,
because it is the most widely
used analytical technique in
mixed-method evaluations. This
is followed by examples of how
to resolve situations where
divergent findings arise from the
analysis of data collected through
a triangulation process.
PARALLEL
MIXED-DATA
ANALYSIS
Parallel mixed-data analysis is
comprised of two major steps:
Step 1: This involves two or
more analytical processes. The
data collected from each method
employed must be analyzed
separately.
For example, a
statistical analysis of quantitative
data derived from a survey, a set
of height/weight measures, or a
set of government statistics is
conducted. Then, a separate and
independent
analysis
is
conducted of qualitative data
derived from, for example, indepth interviews, case studies,
focus groups, or structured
observations
to
determine
emergent
themes,
broad
patterns, and contextual factors.
The main point is that the
analysis of data collected from
each
method
must
be
conducted independently.
Step 2: Once the analysis of the
data generated by each data
collection method is completed,
the evaluator focuses on how the
analysis and findings from each
data set can inform, explain,
and/or strengthen findings from
the other data set. There are two
possible
primary
analytical
methods for doing this – and
sometimes both methods are
used in the same evaluation.
Again, the method used depends
on the purpose of the evaluation.
 In cases where more than one
method is used specifically to
strengthen
and
validate
findings for the same question
through a triangulation design,
the evaluator compares the
findings from the independent
analysis on each data set to
determine if there is a
convergence of findings. This
method is used when it is
critical to produce defensible
conclusions that can be used to
inform
major
program
decisions (e.g., end or extend a
program).
 To interpret or explain findings
from
quantitative
5
analysis,
evaluators use findings from
the analysis of qualitative data.
This method can provide a
richer analysis and set of
explanations affecting program
outcomes that enhance the
utility of the evaluation for
program
managers.
Conversely,
patterns
and
associations arising from the
analysis of quantitative data
can inform additional patterns
to look for in analyzing
qualitative data. The analysis
of qualitative data can also
enhance the understanding of
important program context
data. This method is often used
when program managers want
to know not only whether or
not a program is achieving its
intended results, but also, why
or why not.
WHEN FINDINGS
CONVERGE
DO
NOT
In cases where mixed-method
evaluations employ triangulation,
it is not unusual that findings
from the separate analysis of
each
data
set
do
not
automatically converge. If this
occurs, the evaluator must try to
resolve the conflict among
divergent findings. This is not a
disaster. Often this kind of
situation
can
present
an
opportunity to generate more
nuanced
explanations
and
important additional findings that
are of great value.
One method evaluators use when
findings from different methods
diverge is to carefully re-examine
the raw qualitative data through
a second and more in-depth
content analysis. This is done to
determine if there were any
factors or issues that were missed
when these data were first being
organized for analysis.
The
results of this third layer of
analysis can produce a deeper
understanding of the data, and
can then be used to generate
new interpretations.
In some
cases, other factors external to
the program might be discovered
through contextual analysis of
economic, social or political
conditions or an analysis of
operations and interventions
across program sites.
Another approach is to reanalyze
all the disaggregated data in
each data set separately, by
characteristics of the respondents
as appropriate to the study, such
as age, gender, educational
background, economic strata,
etc., and/or by geography/locale
of respondents.
The results of this analysis may
yield other information that can
help to resolve the divergence of
findings. In this case, the
evaluator should attempt to rank
order these factors in terms of
frequency of occurrence. This
further analysis will provide
additional explanations for the
variances in findings. While most
professionals build this type of
disaggregation into the analysis
of the data during the design
phase of the evaluation, it is
worth reexamining patterns from
disaggregated data.
Evaluators should also check for
data quality issues, such as the
validity of secondary data sources
or possible errors in survey data
from incomplete recording or
incorrect coding of responses.
(See TIPS 12: Data Quality
Standards.) If the evaluators are
still at the program site, it is
possible to resolve data quality
issues with limited follow-up data
collection by, for example,
conducting in-depth interviews
with key informants (if time and
budget permit).
In cases where an overall
summative program conclusion is
required, another analytical tool
that is used to resolve divergent
findings is the data synthesis
method.
(See Table 2.) This
method rates the strength of
findings generated from the
analysis of each data set based
on the intensity of the impact
(e.g., on a scale from very high
positive to very high negative)
and the quality and validity of the
data. An overall rating is assigned
for each data set, but different
weights can then be assigned to
different data sets if the evaluator
knows that certain data sources
or methods for collecting data
are stronger than others.
Ultimately, an index is created
based on the average of those
ratings to synthesize an overall
program effect on the outcome.
See McConney, Rudd and Ayres
(2002) to learn more about this
method.
REPORTING ON
MIXED-METHOD
EVALUATIONS
Mixed-method
evaluations
generate a great deal of data,
6
and, to profit from the use of
those methods, evaluators must
use and analyze all of the data
sets. Through the use of mixedmethod evaluations, findings and
conclusions can be enriched and
strengthened. Yet there is a
tendency to underuse, or even
not to use, all the data collected
for the evaluation. Evaluators can
rely too heavily on one particular
data source if it generates easily
digestible and understandable
information for a program
manager. For example, in many
cases data generated from
qualitative
methods
are
insufficiently analyzed. In some
cases only findings from one
source are reported.
One
way
to
prevent
underutilization of findings is to
write a statement of work that
provides the evaluator sufficient
time to analyze the data sets
from each method employed,
and hence to develop valid
findings, explanations, and strong
conclusions that a program
manager
can
use
with
confidence.
Additionally,
statements of work for evaluation
should require evidence of, and
reporting on, the analysis of data
sets from each method that was
used to collect data, or
methodological justification for
having discarded any data sets.
REFERENCES
Bamberger, Michael, Jim Rugh and Linda Mabry. Real World Evaluation: Working Under Budget,
Time, Data and Political Constraints, Chapter 13, ―Mixed-Method Evaluation,‖ pp. 303-322, Sage
Publications Inc., Thousand Oaks, CA, 2006.
Greene, Jennifer C. and Valerie J. Caracelli. ―Defining and Describing the Paradigm Issue in Mixedmethods Evaluation,” in Advances in Mixed-Method Evaluation: The Challenges and Benefits of
Integrating Diverse Paradigms, Green and Caracelli eds. New Directions for Evaluation. Josey-Bass
Publishers, No. 74, Summer 1997, pp 5-17.
Mark, Melvin M., Irwin Feller and Scott B. Button. ―Integrating Qualitative Methods in a
Predominantly Quantitative Evaluation: A Case Study and Some Reflections,‖ in Advances in
Mixed-Method Evaluation: The Challenges and Benefits of Integrating Diverse Paradigms, Green
and Caracelli eds. New Directions for Evaluation. Josey-Bass Publishers, No. 74, Summer 1997, pp
47-59.
McConney, Andrew, Andy Rudd, and Robert Ayres. ―Getting to the Bottom Line: A Method for
Synthesizing Findings Within Mixed-method Program Evaluations,‖ in American Journal of
Evaluation, Vol. 3, No. 2, 2002, pp. 121-140.
Teddlie, Charles and Abbas Tashakkori, Foundations of Mixed-methods Research: Integrating
Quantitative and Qualitative Approaches in the Behavioral Science, Sage Publications, Inc., Los
Angeles, 2009.
7
TABLE 1 – METHODS FOR ANALYZING MIXED-METHODS DATA1
Analytical
Method
Brief Description
Best for…
Parallel
Two or more data sets collected using a mix of
Triangulation designs to look for
methods (quantitative and qualitative) are analyzed
convergence of findings when the strength
independently. The findings are then combined or
of the findings and conclusions is critical,
integrated.
or to use analysis of qualitative data to
yield deeper explanations of findings from
quantitative data analysis.
Conversion
Two types of data are generated from one data source
Extending the findings of one data set, say,
beginning with the form (quantitative or qualitative) of
quantitative, to generate additional
the original data source that was collected. Then the
findings and/or to compare and potentially
data are converted into either numerical or narrative
strengthen the findings generated from a
data. A common example is the transformation of
complimentary set of, say, qualitative data.
qualitative narrative data into numerical data for
statistical analysis (e.g., on the simplest level,
frequency counts of certain responses).
Sequential
A chronological analysis of two or more data sets
Testing hypotheses generated from the
(quantitative and qualitative) where the results of the
analysis of the first data set.
analysis from the first data set are used to inform the
analysis of the second data set. The type of analysis
conducted on the second data set is dependent on the
outcome of the first data set.
Multilevel
Qualitative and quantitative techniques are used at
Evaluations where organizational units for
different levels of aggregation within a study from at
study are nested (e.g., patient, nurse,
least two data sources to answer interrelated evaluation
doctor, hospital, hospital administrator in
questions. One type of analysis (qualitative) is used at
an evaluation to understand the quality of
one level (e.g., patient) and another type of analysis
patient treatment).
(quantitative) is used in at least one other level (e.g.,
nurse).
Data
Synthesis
A multi-step analytical process in which: 1) a rating of
Providing a bottom-line measure in cases
program effectiveness using the analysis of each data
where the evaluation purpose is to provide
set is conducted (e.g., large positive effect, small
a summative program-wise conclusion
positive effect, no discernable effect, small negative
when findings from mixed-method
effect, large negative effect; 2) quality of evidence
evaluations using a triangulation strategy
assessments are conducted for each data set using
do not converge and appear to be
“criteria of worth” to rate the quality and validity of each
irresolvable, yet a defensible conclusion is
data set gathered; 3) using the ratings collected under
needed to make a firm program decision.
the first two steps, develop an aggregated equation for
Note: there may still be some divergence in
each outcome under consideration to assess the overall
the evaluation findings from mixed data
strength and validity of each finding; and 4) average
sets that the evaluator can still attempt to
outcome-wise effectiveness estimates to produce one
resolve and/or explore to further enrich the
overall program-wise effectiveness index.
analysis and findings.
1
See Teddlie and Tashakkori (2009) and Mark, Feller and Button (1997) for examples and further explanations of parallel data analysis.
See Teddlie and Tashakkori (2009) on conversion, sequential, multilevel, and fully integrated mixed-methods data analysis; and
McConney, Rudd, and Ayers (2002), for a further explanation of data synthesis analysis.
8
For more information:
TIPS publications are available online at [insert website].
Acknowledgements:
Our thanks to those whose experience and insights helped shape this publication including USAID’s
Office of Management Policy, Budget and Performance (MPBP). This publication was written by Dr.
Patricia Vondal of Management Systems International.
Comments regarding this publication can be directed to:
Gerald Britan, Ph.D.
Tel: (202) 712-1158
[email protected]
Contracted under RAN-M-00-04-00049-A-FY0S-84
Integrated Managing for Results II
9
NUMBER 17
1ST EDITION, 2010
PERFORMANCE MONITORING & EVALUATION
TIPS
CONSTRUCTING AN EVALUATION
REPORT
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to performance
monitoring and evaluation. This publication is a supplemental reference to the Automated Directive
System (ADS) Chapter 203.
INTRODUCTION
This TIPS has three purposes. First,
it provides guidance for evaluators
on the structure, content, and style
of evaluation reports. Second, it
offers
USAID
officials,
who
commission evaluations, ideas on
how to define the main deliverable.
Third, it provides USAID officials
with guidance on reviewing and
approving evaluation reports.
The main theme is a simple one: how
to make an evaluation report useful to
its readers. Readers typically include
a
variety
of
development
stakeholders and professionals; yet,
the most important are the
policymakers and managers who
need credible information for
program or project decision-making.
Part of the primary purpose of an
evaluation usually entails informing
this audience.
To be useful, an evaluation report
should address the evaluation
questions and issues with accurate
and data-driven findings, justifiable
conclusions,
and
practical
recommendations. It should reflect
the use of sound evaluation
methodology and data collection,
and report the limitations of each.
Finally, an evaluation should be
written with a structure and style
that promote learning and action.
Five common problems emerge in
relation to evaluation reports.
These problems are as follows:
• An unclear description of the
program strategy and the specific
results it is designed to achieve.
• Inadequate description of the
evaluation’s purpose, intended
uses, and the specific evaluation
questions to be addressed.
• Imprecise analysis and reporting
of quantitative and qualitative data
collected during the evaluation.
1
• A lack of clear distinctions
between findings and conclusions.
• Conclusions
that
are
not
grounded in the facts and
recommendations that do not
flow logically from conclusions.
This guidance offers tips that apply
to an evaluation report for any type
of evaluation — be it formative,
summative (or impact), a rapid
appraisal evaluation, or one using
more rigorous methods.
Evaluation reports should be readily
understood and should identify key
points clearly, distinctly, and
succinctly. (ADS 203.3.6.6)
A PROPOSED
REPORT OUTLINE
Table 1 presents a suggested outline
and approximate page lengths for a
typical evaluation report.
The
evaluation team can, of course,
modify this outline as needed. As
indicated in the table, however,
some elements are essential parts of
any report.
This outline can also help USAID
managers define the key deliverable
in an Evaluation Statement of Work
(SOW) (see TIPS 3: Preparing an
Evaluation SOW).
We will focus particular attention
on the section of the report that
covers findings, conclusions, and
recommendations. This section
represents the core element of the
evaluation report.
BEFORE THE
WRITING BEGINS
Before the report writing begins, the
evaluation team must complete two
critical tasks: 1) establish clear and
defensible findings, conclusions, and
recommendations
that
clearly
address the evaluation questions;
and 2) decide how to organize the
report in a way that conveys these
elements most effectively.
FINDINGS,
CONCLUSIONS, AND
RECOMMENDATIONS
One of the most important tasks in
constructing an evaluation report is
to organize the report into three
main elements: findings, conclusions,
and recommendations (see Figure
1). This structure brings rigor to
the evaluation and ensures that each
element can ultimately be traced
back to the basic facts. It is this
structure that sets evaluation apart
from other types of analysis.
Once the research stage of an
evaluation is complete, the team has
typically collected a great deal of
data in order to answer the
evaluation questions. Depending on
the methods used, these data can
include observations, responses to
survey questions, opinions and facts
from key informants, secondary data
from a ministry, and so on. The
team’s first task is to turn these raw
data into findings.
Suppose, for example, that USAID
has charged an evaluation team with
answering the following evaluation
question (among others):
“How adequate are the prenatal
services provided by the Ministry of
Health’s
rural
clinics
in
Northeastern District?”
To answer this question, their
research in the district included site
visits to a random sample of rural
clinics,
discussions
with
knowledgeable health professionals,
and a survey of women who have
used clinic prenatal services during
the past year. The team analyzed
the raw, qualitative data and
identified the following findings:
• Of the 20 randomly-sampled rural
clinics visited, four clinics met all
six established standards of care,
FIGURE 1.
ORGANIZING KEY ELEMENTS
OF THE EVALUATION
REPORT
Recommendations
Proposed actions for management
Conclusions
Interpretations and judgments
based on the findings
Findings
Empirical facts collected during the
evaluation
while the other 16 (80 percent)
failed to meet at least two
standards. The most commonly
unmet standard (13 clinics) was
“maintenance of minimum staffpatient ratios.”
• In 14 of the 16 clinics failing to
meet two or more standards, not
one of the directors was able to
state the minimum staff-patient
ratios for nurse practitioners,
nurses, and prenatal educators.
TYPICAL PROBLEMS WITH FINDINGS
Findings that:
1.
Are not organized to address the evaluation questions — the reader must
figure out where they fit.
2.
Lack precision and/or context —the reader cannot interpret their relative
strength.
Incorrect: “Some respondents said ’x,’ a few said ’y,’ and others said ’z.’”
Correct: “Twelve of the 20 respondents (60 percent) said ’x,’ five (25
percent) said ’y,’ and three (15 percent) said ’z.’ ”
3.
Mix findings and conclusions.
Incorrect: “The fact that 82 percent of the target group was aware of the
media campaign indicates its effectiveness.”
Correct: Finding: “Eighty-two percent of the target group was aware of the
media campaign.” Conclusion: “The media campaign was effective.”
2
TYPICAL PROBLEMS WITH
CONCLUSIONS
Conclusions that:
1.
2.
3.
4.
Restate findings.
Incorrect: “The project met its
performance targets with respect
to outputs and results.”
Correct: “The project’s strategy
was successful.”
Are vaguely stated.
Incorrect: “The project could
have been more responsive to its
target group.”
Correct: “The project failed to
address the different needs of
targeted women and men.”
Are based on only one of several
findings and data sources.
Include respondents’ conclusions,
which are really findings.
Incorrect: “All four focus groups
of project beneficiaries judged the
project to be effective.”
Correct: “Based on our focus
group data and quantifiable data on
key results indicators, we conclude
that the project was effective.”
• Of 36 women who had used their
rural clinics’ prenatal services
during the past year, 27 (76
percent) stated that they were
“very
dissatisfied”
or
“dissatisfied,” on a scale of 1-5
from “very dissatisfied” to “very
satisfied.” The most frequently
cited reason for dissatisfaction
was “long waits for service” (cited
by 64 percent of the 27
dissatisfied women).
• Six of the seven key informants
who offered an opinion on the
adequacy of prenatal services for
the rural poor in the district
noted that an insufficient number
of prenatal care staff was a “major
problem” in rural clinics.
These findings are the empirical facts
collected by the evaluation team.
Evaluation findings are analogous to
the evidence presented in a court of
law or a patient’s symptoms
identified during a visit to the
doctor. Once the evaluation team
has correctly laid out all the findings
against each evaluation question,
only then should conclusions be
drawn for each question. This is
where many teams tend to confuse
findings and conclusions both in
their analysis and in the final report.
Conclusions represent the team’s
judgments based on the findings.
These are analogous to a court
jury’s decision to acquit or convict
based on the evidence presented or
a doctor’s diagnosis based on the
symptoms. The team must keep
findings and conclusions distinctly
separate
from
each
other.
However, there must also be a clear
and logical relationship between
findings and conclusions.
In our example of the prenatal
services evaluation, examples of
reasonable conclusions might be as
follows:
• In general, the levels of prenatal
care staff in Northeastern
District’s
rural
clinics
are
insufficient.
• The Ministry of Health’s periodic
informational bulletins to clinic
directors regarding the standards
of prenatal care are not sufficient
to ensure that standards are
understood and implemented.
However, sometimes the team’s
findings from different data sources
are not so clear-cut in one direction
as this one. In those cases, the team
must weigh the relative credibility of
the data sources and the quality of
the data, and make a judgment call.
The team might state that a
definitive conclusion cannot be
made, or it might draw a more
3
guarded conclusion such as the
following:
“The preponderance of the
evidence suggests that prenatal
care is weak.”
The team should never omit
contradictory findings from its
analysis and report in order to have
more
definitive
conclusions.
Remember,
conclusions
are
interpretations and judgments made
TYPICAL PROBLEMS WITH
RECOMMENDATIONS
Recommendations that:
1. Are unclear about the action to be
taken.
Incorrect: “Something needs to be
done to improve extension
services.”
Correct: “To improve extension
services, the Ministry of Agriculture
should implement a comprehensive
introductory training program for all
new extension workers and annual
refresher training programs for all
extension workers. “
2. Fail to specify who should take
action.
Incorrect: “Sidewalk ramps for the
disabled should be installed.”
Correct: “Through matching grant
funds from the Ministry of Social
Affairs, municipal governments
should install sidewalk ramps for the
disabled.”
3. Are not supported by any findings
and conclusions
4. Are not realistic with respect to
time and/or costs.
Incorrect: The Ministry of Social
Affairs should ensure that all
municipal sidewalks have ramps for
the disabled within two years.
Correct: The Ministry of Social
Affairs should implement a gradually
expanding program to ensure that all
municipal sidewalks have ramps for
the disabled within 15 years.
on the basis of the findings.
Sometimes we see reports that
include conclusions derived from
preconceived notions or opinions
developed
through
experience
gained outside the evaluation,
especially by members of the team
who have substantive expertise on a
particular topic. We do not
recommend this, because it can
distort the evaluation. That is, the
role of the evaluator is to present
the findings, conclusions, and
recommendations in a logical order.
Opinions outside this framework
are then, by definition, not
substantiated by the facts at hand. If
any of these opinions are directly
relevant to the evaluation questions
and come from conclusions drawn
from prior research or secondary
sources, then the data upon which
they are based should be presented
among the evaluation’s findings.
FIGURE 3
OPTIONS FOR REPORTING
FINDINGS, CONCLUSIONS,
AND RECOMMENDATIONS
OPTION 1
FINDINGS
Evaluation Question 1
Evaluation Question 2
CONCLUSIONS
Evaluation Question 1
Evaluation Question 2
OPTION 2
EVALUATION
QUESTION 1
Findings
Conclusions
Recommendations
EVALUATION
QUESTION 2
Findings
RECOMMENDATIONS Conclusions
Evaluation Question 1
Evaluation Question 2 Recommendations
OPTION 3
Mix the two approaches. Identify which
evaluation questions are distinct and which
are interrelated. For distinct questions, use
option 1 and for the latter, use option 2.
FIGURE 2
Tracking the linkages is one way to help ensure a credible report, with
information that will be useful.
Evaluation Question #1:
FINDINGS
CONCLUSIONS
RECOMMENDATIONS
XXXXXX
YYYYYY
ZZZZZZ
XXXXXX
ZZZZZZ
XXXXXX
YYYYYY
Once conclusions are complete, the
team is ready to make its
recommendations.
Too often
recommendations do not flow from
the team’s conclusions or, worse,
they are not related to the original
evaluation purpose and evaluation
questions. They may be good ideas,
but they do not belong in this
section of the report.
As an
alternative, they could be included in
an annex with a note that they are
derived
from
coincidental
observations made by the team or
from team members’ experiences
elsewhere.
Using our example related to rural
health clinics, a few possible
recommendations could emerge as
follows:
• The
Ministry
of
Health’s
Northeastern
District
office
should develop and implement an
annual prenatal standards-of-care
training program for all its rural
clinic directors. The program
would cover….
• The Northeaster District office
should
conduct
a
formal
assessment of prenatal care
staffing levels in all its rural clinics.
• Based on the assessment, the
4
ZZZZZZ
Northeastern
District
office
should establish and implement a
five-year plan for hiring and
placing needed prenatal care staff
in its rural clinics on a mostneedy-first basis.
Although
the
basic
recommendations should be derived
from conclusions and findings, this is
where the team can include ideas
and options for implementing
recommendations that may be based
on their substantive expertise and
best
practices
drawn
from
experience outside the evaluation
itself. Usefulness is paramount.
When developing recommendations,
consider practicality. Circumstances
or resources may limit the extent to
which a recommendation can be
implemented. If practicality is an
issue — as is often the case — the
evaluation team may need to ramp
down recommendations, present
them in terms of incremental steps,
or suggest other options. In order
to be useful, it is essential that
recommendations be actionable or,
in other words, feasible in light of
the human, technical, and financial
resources available.
Weak connections between findings,
conclusions, and recommendations
can undermine the user’s confidence
in evaluation results. As a result, we
encourage teams—or, better yet, a
colleague who has not been
involved—to review the logic before
beginning to write the report. For
each evaluation question, present all
the findings, conclusions, and
recommendations in a format similar
to the one outlined in Figure 2.
Starting with the conclusions in the
center, track each one back to the
findings that support it, and decide
whether the findings truly warrant
the conclusion being made. If not,
revise the conclusion as needed.
Then track each recommendation to
the conclusion(s) from which it
flows, and revise if necessary.
CHOOSE THE BEST
APPROACH FOR
STRUCTURING THE
REPORT
Depending on the nature of the
evaluation questions and the
findings,
conclusions,
and
recommendations, the team has a
few options for structuring this part
of the report (see Figure 3). The
objective is to present the report in
a way that makes it as easy as
possible for the reader to digest all
of the information. Options are
discussed below.
Option 1- Distinct Questions
If all the evaluation questions are
distinct from one another and the
relevant findings, conclusions, and
recommendations do not cut across
questions, then one option is to
organize the report around each
evaluation question. That is, each
question will include a section
including its relevant findings,
conclusions, and recommendations.
Option 2- Interrelated
Questions
If, however, the questions are
closely interrelated and there are
findings,
conclusions,
and/or
recommendations that apply to
more than one question, then it may
be preferable to put all the findings
for all the evaluation questions in
one section, all the conclusions in
another,
and
all
the
recommendations in a third.
Option 3- Mixed
If the situation is mixed—where a
few but not all the questions are
closely interrelated—then use a
mixed
approach.
Group
the
interrelated questions and their
findings,
conclusions,
and
recommendations into one subsection, and treat the stand-alone
questions and their respective
findings,
conclusions,
and
recommendations in separate subsections.
The important point is that the team
should be sure to keep findings,
conclusions, and recommendations
separate and distinctly labeled as such.
Finally, some evaluators think it
more useful to present the
conclusions first, and then follow
with the findings supporting them.
This helps the reader see the
“bottom line” first and then make a
judgment as to whether the
conclusions are warranted by the
findings.
OTHER KEY
SECTIONS OF THE
REPORT
THE EXECUTIVE
SUMMARY
The Executive Summary should
stand alone as an abbreviated
version of the entire report. Often
it is the only thing that busy
managers read. The Executive
Summary should be a “mirror
image” of the full report—it should
contain no new information that is
not in the main report.
This
principle also applies to making the
Executive Summary and the full
report equivalent with respect to
presenting positive and negative
evaluation results.
Although all sections of the full
report are summarized in the
Executive Summary, less emphasis is
given to an overview of the project
and the description of the evaluation
purpose and methodology than is
given to the findings, conclusions,
and recommendations. Decisionmakers
are
generally
more
interested in the latter.
The Executive Summary should be
written after the main report has
been drafted. Many people believe
that a good Executive Summary
should not exceed two pages, but
there is no formal rule in USAID on
this. Finally, an Executive Summary
should be written in a way that will
entice interested stakeholders to go
on to read the full report.
DESCRIPTION OF THE
PROJECT
Many evaluation reports give only
cursory
attention
to
the
development
problem
(or
opportunity) that motivated the
project in the first place, or to the
5
FIGURE 4. SUMMARY OF EVALUATION DESIGN AND METHODS (an illustration)
Evaluation
Question
Type of Analysis
Conducted
Data Sources and
Methods Used
Type and Size of
Sample
1. How
adequate are
the prenatal
services
provided by
the Ministry
of Health’s
(MOH) rural
clinics in
Northeastern
District?
Comparison of rural
clinics’ prenatal
service delivery to
national standards
MOH manual of rural
clinic standards of care
Structured observations
and staff interviews at
rural clinics
Twenty clinics,
randomly sampled
from 68 total in
Northeastern District
Three of the originally
sampled clinics were closed
when the team visited. To
replace each, the team
visited the closest open
clinic. As a result, the sample
was not totally random.
Description, based on
a content analysis of
expert opinions
Key informant interviews
with health care experts in
the district and the MOH
Ten experts identified
by project & MOH
staff
Only seven of the 10
experts had an opinion
about prenatal care in the
district.
Description and
comparison of ratings
among women in the
district and two other
similar rural districts
In-person survey of
recipients of prenatal
services at clinics in the
district and two other
districts
Random samples of 40
women listed in clinic
records as having
received prenatal
services during the
past year from each of
the three districts’
clinics
Of the total 120 women
sampled, the team was able
to conduct interviews with
only 36 in the district, and
24 and 28 in the other two
districts. The levels of
confidence for generalizing
to the populations of service
recipients were __, __, and
__, respectively.
“theory of change” that underpins
USAID’s intervention. The “theory
of change” includes what the project
intends to do and the results which
the activities are intended to
produce. TIPS 13: Building a Results
Framework is a particularly useful
reference and provides additional
detail on logic models.
If the team cannot find a description
of these hypotheses or any model of
the project’s cause-and-effect logic
such as a Results Framework or a
Logical Framework, this should be
noted. The evaluation team will
then have to summarize the project
strategy in terms of the “if-then”
propositions that show how the
project designers envisioned the
interventions as leading to desired
results.
In describing the project, the
evaluation team should be clear
about what USAID tried to improve,
eliminate, or otherwise change for
the better. What was the “gap”
between conditions at the start of
the project and the more desirable
conditions that USAID wanted to
establish with the project? The team
should indicate whether the project
design documents and/or the recall
of interviewed project designers
offered a clear picture
of the specific economic and social
factors that contributed to the
problem — with baseline data, if
available. Sometimes photographs
and
maps
of
before-project
conditions, such as the physical
characteristics and locations of rural
prenatal clinics in our example, can
be used to illustrate the main
problem(s).
It is equally important to include
basic information about when the
project was undertaken, its cost, its
intended beneficiaries, and where it
was implemented (e.g., country-wide
or only in specific districts). It can
be particularly useful to include a
6
Limitations
map that shows the project’s target
areas.
A good description also identifies
the organizations that implement the
project, the kind of mechanism used
(e.g., contract, grant, or cooperative
agreement), and whether and how
the project has been modified during
implementation.
Finally,
the
description
should
include
information about context, such as
conflict or drought, and other
government or donor activities
focused on achieving the same or
parallel results.
THE EVALUATION
PURPOSE AND
METHODOLOGY
The credibility of an evaluation
team’s findings, conclusions, and
recommendations rests heavily on
the quality of the research design, as
well as on data collection methods
and analysis used. The reader needs
to understand what the team did
and why in order to make informed
judgments
about
credibility.
Presentation of the evaluation design
and methods is often best done
through a short
summary in the text of the report
and a more detailed methods annex
that
includes
the
evaluation
instruments. Figure 4 provides a
sample summary of the design and
methodology that can be included in
the body of the evaluation report.
From a broad point of view, what
research design did the team use to
answer each evaluation question?
Did the team use description (e.g.,
to document what happened),
comparisons (e.g., of baseline data
or targets to actual data, of actual
practice to standards, among target
sub-populations or locations), or
cause-effect research (e.g., to
determine whether the project
made a difference)? To do causeeffect analysis, for example, did the
team use one or more quasiexperimental approaches, such as
time-series analysis or use of nonproject comparison groups (see
TIPS 11: The Role of Evaluation)?
More specifically, what data collection
methods did the team use to get the
evidence needed for each evaluation
question? Did the team use key
informant interviews, focus groups,
surveys,
on-site
observation
methods, analyses of secondary data,
and other methods? How many
people did they interview or survey,
how many sites did they visit, and
how did they select their samples?
and developing the findings and
conclusions that follow in the
report. The reader needs to know
these limitations in order to make
informed judgments about the
evaluation’s
credibility
and
usefulness.
Most evaluations suffer from one or
more constraints that affect the
comprehensiveness and validity of
findings and conclusions. These may
include overall limitations on time
and
resources,
unanticipated
problems in reaching all the key
informants and survey respondents,
unexpected problems with the
quality of secondary data from the
host-country government, and the
like. In the methodology section, the
team
should
address
these
limitations and their implications for
answering the evaluation questions
When writing its report, the
evaluation team must always
remember the composition of its
audience. The team is writing for
policymakers,
managers,
and
takeholders, not for fellow social
science
researchers
or
for
publication in a professional journal.
To that end, the style of writing
should make it as easy as possible
for the intended audience to
understand and digest what the
team is presenting. For further
suggestions on writing an evaluation
in reader-friendly style, see Table 2.
7
READER-FRIENDLY
STYLE
TABLE 1. SUGGESTED OUTLINE FOR AN EVALUATION REPORT1
Element
Approximate
Number of
Pages
Description and Tips for the Evaluation Team
Title Page
1 (but no page
number)
Essential. Should include the words “U.S. Agency for International
Development” with the acronym “USAID,” the USAID logo, and the
project/contract number under which the evaluation was conducted. See
USAID Branding and Marking Guidelines (http://www.usaid.gov/branding/)
for logo and other specifics. Give the title of the evaluation; the name of
the USAID office receiving the evaluation; the name(s), title(s), and
organizational affiliation(s) of the author(s); and the date of the report.
Contents
As needed, and
start with
Roman
numeral ii.
Essential. Should list all the sections that follow, including Annexes. For
multi-page chapters, include chapter headings and first- and second-level
headings. List (with page numbers) all figures, tables, boxes, and other
titled graphics.
Foreword
1
Optional. An introductory note written by someone other than the
author(s), if needed. For example, it might mention that this evaluation is
one in a series of evaluations or special studies being sponsored by USAID.
Acknowledgements
1
Optional. The authors thank the various people who provided support
during the evaluation.
Preface
1
Optional. Introductory or incidental notes by the authors, but not material
essential to understanding the text. Acknowledgements could be included
here if desired.
Executive Summary
2-3; 5 at most
Essential, unless the report is so brief that a summary is not needed. (See
discussion on p. 5)
Glossary
1
Optional. Is useful if the report uses technical or project-specific
terminology that would be unfamiliar to some readers.
Acronyms and
Abbreviations
1
Essential, if they are used in the report. Include only those acronyms that
are actually used. See Table 3 for more advice on using acronyms.
I. Introduction
5-10 pages,
starting with
Arabic numeral
1.
Optional. The two sections listed under Introduction here could be
separate, stand-alone chapters. If so, a separate Introduction may not be
needed.
Description of the
Project
The Evaluation Purpose
and Methodology
II. Findings, Conclusions,
and Recommendations
Essential. Describe the context in which the USAID project took place—
e.g., relevant history, demography, political situation, etc. Describe the
specific development problem that prompted USAID to implement the
project, the theory underlying the project, and details of project
implementation to date. (See more tips on p. 6.)
Essential. Describe who commissioned the evaluation, why they
commissioned it, what information they want, and how they intend to use
the information (and refer to the Annex that includes the Statement of
Work). Provide the specific evaluation questions, and briefly describe the
evaluation design and the analytical and data collection methods used to
answer them. Describe the evaluation team (i.e., names, qualifications, and
roles), what the team did (e.g., reviewed relevant documents, analyzed
secondary data, interviewed key informants, conducted a survey, conducted
site visits), and when and where they did it. Describe the major limitations
encountered in data collection and analysis that have implications for
reviewing the results of the evaluation. Finally, refer to the Annex that
provides a fuller description of all of the above, including a list of
documents/data sets reviewed, a list of individuals interviewed, copies of
the data collection instruments used, and descriptions of sampling
procedures (if any) and data analysis procedures. (See more tips on p. 6.)
20-30 pages
Essential. However, in some cases, the evaluation user does not want
recommendations, only findings and conclusions. This material may be
8
TABLE 1. SUGGESTED OUTLINE FOR AN EVALUATION REPORT1
Element
Approximate
Number of
Pages
Description and Tips for the Evaluation Team
organized in different ways and divided into several chapters. (A detailed
discussion of developing defensible findings, conclusions, and
recommendations and structural options for reporting them is on p 2 and
p. 5)
III. Summary of
Recommendations
1-2 pages
Essential or optional, depending on how findings, conclusions and
recommendations are presented in the section above. (See a discussion of
options on p. 4.) If all the recommendations related to all the evaluation
questions are grouped in one section of the report, this summary is not
needed. However, if findings, conclusions, and recommendations are
reported together in separate sections for each evaluation question, then a
summary of all recommendations, organized under each of the evaluation
questions, is essential.
IV. Lessons Learned
As needed
Required if the SOW calls for it; otherwise optional. Lessons learned
and/or best practices gleaned from the evaluation provide other users, both
within USAID and outside, with ideas for the design and implementation of
related or similar projects in the future.
Some are
essential and
some are
optional as
noted.
Essential. Lets the reader see exactly what USAID initially expected in the
evaluation.
Annexes
Statement of Work
Evaluation Design and
Methodology
Essential. Provides a more complete description of the evaluation
questions, design, and methods used. Also includes copies of data
collection instruments (e.g., interview guides, survey instruments, etc.) and
describes the sampling and analysis procedures that were used.
List of Persons
Interviewed
Essential. However, specific names of individuals might be withheld in order
to protect their safety.
List of Documents
Reviewed
Essential. Includes written and electronic documents reviewed, background
literature, secondary data sources, citations of websites consulted.
Dissenting Views
If needed. Include if a team member or a major stakeholder does not agree
with one or more findings, conclusions, or recommendations.
Recommendation
Action Checklist
Optional. As a service to the user organization, this chart can help with
follow-up to the evaluation. It includes a list of all recommendations
organized by evaluation question, a column for decisions to accept or reject
each recommendation, a column for the decision maker’s initials, a column
for the reason a recommendation is being rejected, and, for each accepted
recommendation, columns for the actions to be taken, by when, and by
whom.
1
The guidance and suggestions in this table were drawn from the writers’ experience and from the “CDIE Publications
Style Guide: Guidelines for Project Managers, Authors, & Editors,” compiled by Brian Furness and John Engels, December
2001. The guide, which includes many tips on writing style, editing, referencing citations, and using Word and Excel is
available online at http://kambing.ui.ac.id/bebas/v01/DEC-USAID/Other/publications-style-guide.pdf. Other useful
guidance: ADS 320 (http://www.usaid.gov/policy/ads/300/320.pdf ; http://www.usaid.gov/branding; and
http://www.usaid.gov/branding/Graphic Standards Manual.pdf.
9
TABLE 2. THE QUICK REFERENCE GUIDE FOR A READER-FRIENDLY TECHNICAL STYLE
Writing Style—
Keep It Simple
and Correct!
Avoid meaningless precision. Decide how much precision is really necessary. Instead of “62.45
percent,” might “62.5 percent” or “62 percent” be sufficient? The same goes for averages and other
calculations.
Use technical terms and jargon only when necessary. Make sure to define them for the unfamiliar
readers.
Don’t overuse footnotes. Use them only to provide additional information which, if included in the
text, would be distracting and cause a loss of the train of thought.
Use Tables,
Charts and Other
Graphics to
Enhance
Understanding
Avoid long, “data-dump”paragraphs filled with numbers and percentages. Use tables, line graphs, bar
charts, pie charts, and other visual displays of data, and summarize the main points in the text. In
addition to increasing understanding, these displays provide visual relief from long narrative tracts.
Be creative—but not too creative. Choose and design tables and charts carefully with the reader in
mind.
Make every visual display of data a self-contained item. It should have a meaningful title and headings
for every column; a graph should have labels on each axis; a pie or bar chart should have labels for
every element.
Choose shades and colors carefully. Expect that consumers will reproduce the report in black and
white and make copies of copies. Make sure that the reader can distinguish clearly among colors or
shades among multiple bars and pie-chart segments. Consider using textured fillings (such as hatch
marks or dots) rather than colors or shades.
Provide “n’s” in all displays which involve data drawn from samples or populations. For example,
the total number of cases or survey respondents should be under the title of a table (n = 100). If a
table column includes types of responses from some, but not all, survey respondents to a specific
question, say, 92 respondents, the column head should include the total number who responded to
the question (n = 92).
Refer to every visual display of data in the text. Present it after mentioning it in the text and as soon
after as practical, without interrupting paragraphs.
Number tables and figures separately, and number each consecutively in the body of the report.
Consult the CDIE style guide for more detailed recommendations on tables and graphics.
Punctuate the
Text with Other
Interesting
Features
Put representative quotations gleaned during data collection in text boxes. Maintain balance
between negative and positive comments to reflect the content of the report. Identify the sources
of all quotes. If confidentiality must be maintained, identify sources in general terms, such as “a clinic
care giver” or “a key informant.”
Provide little “stories” or cases that illustrate findings. For example, a brief anecdotal story in a text
box about how a woman used a clinic’s services to ensure a healthy pregnancy can enliven, and
humanize, the quantitative findings.
Use photos and maps where appropriate. For example, a map of a district with all the rural clinics
providing prenatal care and the concentrations of rural residents can effectively demonstrate
adequate or inadequate access to care.
Don’t overdo it. Strike a reader-friendly balance between the main content and illustrative material.
In using illustrative material, select content that supports main points, not distracts from them.
Finally…
Remember that the reader’s need to understand, not the writer’s need to impress, is paramount.
Be consistent with the chosen format and style throughout the report.
Sources: “CDIE Publications Style Guide: Guidelines for Project Managers, Authors, & Editors,” compiled by Brian
Furness and John Engels, December 2001 (http://kambing.ui.ac.id/bebas/v01/DEC-USAID/Other/publications-styleguide.pdf); USAID’s Graphics Standards Manual (http://www.usaid.gov/branding/USAID_Graphic_Standards_Manual.pdf);
and the authors extensive experience with good and difficult-to-read evaluation reports.
10
For more information:
TIPS publications are available online at [insert website].
Acknowledgements:
Our thanks to those whose experience and insights helped shape this publication including Gerry Britan and
Subhi Mehdi of USAID’s Office of Management Policy, Budget and Performance (MPBP). This publication was
written by Larry Beyna of Management Systems International (MSI).
Comments regarding this publication can be directed to:
Gerald Britan, Ph.D.
Tel: (202) 712-1158
[email protected]
Contracted under RAN-M-00-04-00049-A-FY0S-84
Integrated Managing for Results II
11
NUMBER 18
1ST EDITION, 2010
PERFORMANCE MONITORING & EVALUATION
TIPS
CONDUCTING DATA QUALITY ASSESSMENTS
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to
performance monitoring and evaluation. This publication is a supplemental reference to the
Automated Directive System (ADS) Chapter 203.
THE PURPOSE OF
THE DATA QUALITY
ASSESSMENT
Data quality assessments (DQAs)
help managers to understand how
confident they should be in the
data used to manage a program and
report on its success. USAID’s
ADS notes that the purpose of the
Data Quality Assessment is to:
“…ensure that the USAID
Mission/Office
and
Assistance Objective (AO)
Team are aware of the
strengths and weaknesses of
the data, as determined by
applying the five data quality
standards …and are aware
of the extent to which the
data integrity can be trusted
to influence management
decisions.” (ADS 203.3.5.2)
This purpose is important to keep
in mind when considering how to
do a data quality assessment. A
data quality assessment is of little
use unless front line managers
comprehend key data quality issues
and are able to improve the
performance management system.
THE DATA QUALITY
STANDARDS
Five key data quality standards are
used to assess quality. These are:
• Validity
• Reliability
• Precision
• Integrity
• Timeliness
A more detailed discussion of each
standard is included in TIPS 12:
Data Quality Standards.
WHAT IS REQUIRED?
USAID POLICY
While managers are required to
understand data quality on an
ongoing basis, a data quality
assessment must also be conducted
at least once every three years for
those
data
reported
to
Washington. As a matter of good
management, program managers
may decide to conduct DQAs
more frequently or for a broader
range of data where potential
issues emerge.
The ADS does not prescribe a
specific way to conduct a DQA. A
variety of approaches can be used.
Documentation may be as simple
1
as a memo to the files, or it could
take the form of a formal report.
The most appropriate approach
will
reflect
a
number
of
considerations,
such
as
management need, the type of data
collected, the data source, the
importance of the data, or
suspected data quality issues. The
key is to document the findings,
whether formal or informal.
A DQA focuses on applying the
data
quality
standards
and
examining
the
systems
and
approaches for collecting data to
determine whether they are likely
to produce high quality data over
time. In other words, if the data
quality standards are met and the
data collection methodology is well
designed, then it is likely that good
quality data will result.
This “systematic approach” is
valuable because it assesses a
broader set of issues that are likely
to ensure data quality over time (as
opposed to whether one specific
number is accurate or not). For
example, it is possible to report a
number correctly, but that number
may not be valid1 as the following
example demonstrates.
Example:
A program works
across a range of municipalities
(both urban and rural).
It is
reported that local governments
have increased revenues by 5%.
These data may be correct.
However, if only major urban areas
have been included, these data are
not valid.
That is, they do not
measure the intended result.
VERIFICATION OF DATA
Verification of data means that the
reviewer follows a specific datum
to its source, confirming that it has
supporting documentation and is
accurate—as is often done in
audits.
The DQA may not
necessarily verify that all individual
numbers reported are accurate.
The ADS notes that when assessing
data from partners, the DQA
should focus on “the apparent
accuracy and consistency of the
data.” As an example, Missions
often report data on the number of
individuals trained. Rather than
verifying each number reported,
the DQA might examine each
project’s system for collecting and
maintaining those data. If there is a
good system in place, we know
that it is highly likely that the data
produced will be of high quality.
“…data used for
management purposes
have different standards
than data used for
research.
Having said this, it is certainly
advisable to periodically verify
actual data as part of the larger
performance management system.
Project managers may:
Choose a few indicators to verify
periodically throughout the course
of the year.
Occasionally spot check data (for
example, when visiting the field).
HOW GOOD DO DATA
HAVE TO BE?
Refer to TIPS 12: Data Quality
Standards for a full discussion of all
the data quality standards.
1
In development, there are rarely
perfect data. Moreover, data used
for management purposes have
different standards than data used
2
for research. There is often a
direct trade-off between cost and
quality.
Each manager is
responsible for ensuring the highest
quality data possible given the
resources and the management
context. In some cases, simpler,
lower-cost approaches may be
most appropriate. In other cases,
where indicators measure progress
in major areas of investment,
higher data quality is expected.
OPTIONS AND
APPROACHES FOR
CONDUCTING DQAS
A data quality assessment is both a
process for reviewing data to
understand
strengths
and
weaknesses
as
well
as
documentation. A DQA can be
done in a variety of ways ranging
from the more informal to the
formal (see Figure 1).
In our
experience, a combination of
informal, on-going and systematic
assessments work best, in most
cases, to ensure good data quality.
INFORMAL OPTIONS
Informal approaches can be ongoing or driven by specific issues as
they emerge.
These approaches
depend more on the front line
manager’s in-depth knowledge of
the program.
Findings are
documented by the manager in
memos
or
notes
in
the
Performance Management Plan
(PMP).
Example:
An
implementer
reports
that
civil
society
organizations (CSOs) have initiated
50 advocacy campaigns.
This
number seems unusually high. The
project
manager
calls
the
Implementer to understand why
the number is so high in
FIGURE 1. OPTIONS FOR CONDUCTING DATA QUALITY ASSESSMENTS- THE CONTINUUM
Informal Options
• Conducted internally by the
AO team
• Ongoing (driven by
emerging and specific
issues)
• More dependent on the
AO team and individual
manager’s expertise &
knowledge of the program
• Conducted by the program
manager
• Product: Documented in
memos, notes in the PMP
comparison to previously reported
numbers and explores whether a
consistent
methodology
for
collecting the data has been used
(i.e., whether the standard of
reliability has been met).
The
project manager documents his or
her findings in a memo and
maintains that information in the
files.
Informal approaches should be
incorporated into Mission systems
as a normal part of performance
management. The advantages and
disadvantages of this approach are
as follows:
Advantages
• Managers
incorporate
data
quality as a part of on-going
work processes.
• Issues can be addressed and
corrected quickly.
• Managers establish a principle
that data quality is important.
Disadvantages
• It is not systematic and may not
be complete. That is, because
informal
assessments
are
normally driven by more
Semi-Formal Partnership
• Draws on both
management expertise and
M&E expertise
• Periodic & systematic
• Facilitated and coordinated
by the M&E expert, but AO
team members are active
participants
• Product: Data Quality
Assessment Report
immediate
management
concerns, the manager may miss
larger issues that are not readily
apparent (for example, whether
the data are attributable to
USAID programs).
• There is no comprehensive
document that addresses the
DQA requirement.
• Managers may not have enough
expertise to identify more
complicated data quality issues,
audit
vulnerabilities,
and
formulate solutions.
SEMI-FORMAL /
PARTNERSHIP OPTIONS
Semi-formal or partnership options
are characterized by a more
periodic and systematic review of
data quality. These DQAs should
ideally be led and conducted by
USAID staff. One approach is to
partner
a
monitoring
and
evaluation (M&E) expert with the
Mission’s AO team to conduct the
assessment jointly.
The M&E
expert can organize the process,
develop
standard
approaches,
facilitate
sessions,
assist
in
identifying potential data quality
issues and solutions, and may
3
Formal Options
• Driven by broader
programmatic needs, as
warranted
• More dependent on
external technical expertise
and/or specific types of data
expertise
• Product: Either a Data
Quality Assessment report
or addressed as a part of
another report
document the outcomes of the
assessment. This option draws on
the experience of AO team
members as well as the broader
knowledge and skills of the M&E
expert.
Engaging front line
mangers in the DQA process has
the additional advantage of making
them more aware of the strengths
and weaknesses of the data—one
of the stated purposes of the DQA.
The advantages and disadvantages
of this approach are summarized
below:
Advantages
• Produces a systematic and
comprehensive report with
specific recommendations for
improvement.
• Engages AO team members in
the data quality assessment.
• Draws on the complementary
skills of front line managers and
M&E experts.
• Assessing data quality is a matter
of understanding trade-offs and
context in terms of deciding
what data is “good enough” for a
program. An M&E expert can be
useful in guiding AO team
members through this process in
order to ensure that audit
vulnerabilities are adequately
addressed.
• Does not require
external team.
a
large
These types of data quality
assessments require a high degree
of rigor and specific, in-depth
technical expertise.
Advantages
and disadvantages are as follow:
Disadvantages
Advantages
• The Mission may use an internal
M&E expert or hire someone
from the outside.
However,
hiring an outside expert will
require additional resources, and
external contracting requires
some time.
• Produces a systematic and
comprehensive assessment, with
specific recommendations.
• Examines data quality issues with
rigor and based on specific, indepth technical expertise.
• Because of the additional time
and planning required, this
approach is less useful for
addressing immediate problems.
• Fulfills two important purposes,
in that it can be designed to
improve data collection systems
both within USAID and for the
beneficiary.
FORMAL OPTIONS
Disadvantages
At the other end of the continuum,
there may be a few select situations
where Missions need a more
rigorous and formal data quality
assessment.
• Often conducted by an external
team of experts, entailing more
time and cost than other
options.
Example:
A Mission invests
substantial funding into a highprofile program that is designed to
increase the efficiency of water use.
Critical performance data comes
from the Ministry of Water, and is
used both for performance
management and reporting to key
stakeholders,
including
the
Congress. The Mission is unsure as
to the quality of those data. Given
the high level interest and level of
resources invested in the program,
a data quality assessment is
conducted by a team including
technical experts to review data
and
identify
specific
recommendations
for
improvement. Recommendations
will be incorporated into the
technical assistance provided to the
Ministry to improve their own
capacity to track these data over
time.
• Generally
involvement
managers.
less
direct
by front line
• Often examines data through a
very technical lens.
It is
important to ensure that
broader management issues are
adequately addressed.
THE PROCESS
The Mission will also have to
determine
whether
outside
assistance is required.
Some
Missions have internal M&E staff
with the appropriate skills to
facilitate this process.
Other
Missions may wish to hire an
outside M&E expert(s) with
experience in conducting DQAs.
AO team members should also be
part of the team.
DATA SOURCES
Primary Data: Collected directly by
USAID.
Secondary Data: Collected from and
other sources, such as implementing
partners, host country governments,
other donors, etc.
STEP 2. DEVELOP AN
OVERALL APPROACH
AND SCHEDULE
The team leader must convey the
objectives, process, and schedule
for conducting the DQA to team
members. This option is premised
on the idea that the M&E expert(s)
work closely in partnership with
AO
team
members
and
implementing partners to jointly
assess data quality. This requires
active participation and encourages
managers to fully explore and
understand the strengths and
weaknesses of the data.
For purposes of this TIPS, we will
outline a set of illustrative steps for
the middle (or semi-formal/
partnership) option.
In reality,
these steps are often iterative.
STEP 3. IDENTIFY THE
INDICATORS TO BE
INCLUDED IN THE
REVIEW
STEP 1. IDENTIFY THE
DQA TEAM
It is helpful to compile a list of all
indicators that will be included in
the DQA. This normally includes:
Identify one person to lead the
DQA process for the Mission. This
person is often the Program
Officer or an M&E expert. The
leader is responsible for setting up
the
overall
process
and
coordinating with the AO teams.
4
• All indicators reported to
USAID/Washington (required).
• Any indicators with suspected
data quality issues.
• Indicators for program areas
that are of high importance.
This list can also function as a
central guide as to how each
indicator is assessed and to
summarize where follow-on action
is needed.
STEP 4. CATEGORIZE
INDICATORS
With the introduction of standard
indicators,
the
number
of
indicators that Missions report to
USAID/Washington has increased
substantially. This means that it is
important to develop practical and
streamlined
approaches
for
conducting DQAs. One way to do
this is to separate indicators into
two categories, as follows:
Outcome Level Indicators
Outcome level indicators measure
AOs or Intermediate Results (IRs).
Figure 2 provides examples of
indicators at each level.
The
standards for good data quality are
applied to results level data in
order to assess data quality.
The data quality assessment
worksheet (see Table 1) has been
developed as a tool to assess each
indicator against each of these
standards.
Output Indicators
Many of the data quality standards
are not applicable to output
indicators in the same way as
outcome level indicators.
For
example, the number of individuals
trained by a project is an output
indicator. Whether data are valid,
timely, or precise is almost never
an issue for this type of an
indicator. However, it is important
to ensure that there are good data
collection and data maintenance
systems in place. Hence, a simpler
and more streamlined approach
can be used to focus on the most
relevant issues. Table 2 outlines a
sample matrix for assessing output
indicators. This matrix:
• Identifies the indicator.
• Clearly outlines
collection method.
the
data
• Identifies key data quality issues.
• Notes whether further action is
necessary.
• Provides specific information on
who was consulted and when.
STEP 5. HOLD WORKING
SESSIONS TO REVIEW
INDICATORS
Hold working sessions with AO
team members.
Implementing
partners may be included at this
5
point as well. In order to use time
efficiently, the team may decide to
focus these sessions on resultslevel indicators. These working
sessions can be used to:
• Explain the purpose and process
for conducting the DQA.
• Review data quality standards for
each results-level indicator,
including the data collection
systems and processes.
• Identify issues or concerns that
require further review.
STEP 6. HOLD SESSIONS
WITH IMPLEMENTING
PARTNERS TO REVIEW
INDICATORS
If the implementing partner was
included in the previous working
session, results-level indicators will
already have been discussed. This
session may then focus on
reviewing the remaining outputlevel indicators with implementers
who often maintain the systems to
collect the data for these types of
indicators. Focus on reviewing the
systems and processes to collect
and maintain data. This session
provides a good opportunity to
identify solutions or recommenddations for improvement.
STEP 7. PREPARE THE
DQA DOCUMENT
As information is gathered, the
team should record findings on the
worksheets provided.
It is
particularly important to include
recommendations for action at the
conclusion of each worksheet.
Once this is completed, it is often
useful to include an introduction
to:
• Outline the overall approach and
methodology used.
• Highlight key data quality issues
that are important for senior
management.
• Summarize
recommendations
for
improving
performance
management systems.
AO
team
members
and
participating implementers should
have an opportunity to review the
first draft. Any comments or issues
can then be incorporated and the
DQA finalized.
STEP 8. FOLLOW UP ON
ACTIONS
Finally, it is important to ensure
that there is a process to follow-up
on recommendations.
Some
recommendations
may
be
addressed internally by the team
handling management needs or
audit vulnerabilities. For example,
the AO team may need to work
with a Ministry to ensure that data
can be disaggregated in a way that
correlates precisely to the target
group. Other issues may need to
be addressed during the Mission’s
portfolio reviews.
CONSIDER THE
SOURCE – PRIMARY
VS. SECONDARY
DATA
PRIMARY DATA
USAID is able to exercise a higher
degree of control over primary
data that it collects itself than over
secondary data collected by others.
As a result, specific standards
should be incorporated into the
data collection process. Primary
data collection requires that:
• Written procedures are in place
for data collection.
• Data are collected from year to
year using a consistent collection
process.
• Data
are
collected
using
methods
to
address
and
minimize sampling and nonsampling errors.
• Data are collected by qualified
personnel and these personnel
are properly supervised.
• Duplicate data are detected.
• Safeguards are in place to
prevent unauthorized changes to
the data.
• Source
documents
are
maintained and readily available.
• If the data collection process is
contracted
out,
these
requirements
should
be
incorporated directly into the
statement of work.
SECONDARY DATA
Secondary data are collected from
other sources, such as host
country
governments,
implementing partners, or from
other organizations. The range of
control that USAID has over
secondary data varies.
For
example, if USAID uses data from a
survey commissioned by another
donor, then there is little control
over
the
data
collection
methodology. On the other hand,
USAID does have more influence
over
data
derived
from
implementing partners. In some
cases,
specific
data
quality
requirements may be included in
the contract. In addition, project
performance management plans
6
(PMPs) are often reviewed or
approved by USAID. Some ways in
which to address data quality are
summarized below.
Data from Implementing
Partners
• Spot check data.
• Incorporate specific data quality
requirements as part of the
SOW, RFP, or RFA.
• Review data quality collection
and maintenance procedures.
Data from Other Secondary
Sources
Data from other secondary
sources includes data from host
countries, government, and other
donors.
• Understand the methodology.
Documentation often includes a
description of the methodology
used to collect data.
It is
important to understand this
section so that limitations (and
what the data can and cannot
say) are clearly understood by
decision makers.
• Request a briefing on the
methodology, including data
collection
and
analysis
procedures, potential limitations
of the data, and plans for
improvement (if possible).
• If data are derived from host
country organizations, then it
may be appropriate to discuss
how assistance can be provided
to strengthen the quality of the
data. For example, projects may
include technical assistance to
improve management and/or
M&E systems.
TABLE 1. THE DQA WORKSHEET FOR OUTCOME LEVEL INDICATORS
Directions: Use the following worksheet to complete an assessment of data for outcome level indicators against the
five data quality standards outlined in the ADS. A comprehensive discussion of each criterion is included in TIPS 12
Data Quality Standards.
Data Quality Assessment Worksheet
Assistance Objective (AO) or Intermediate Result (IR):
Indicator:
Reviewer(s):
Date Reviewed:
Data Source:
Is the Indicator Reported to USAID/W?
Criterion
Definition
1. Validity
Do the data clearly and adequately
represent the intended result? Some issues
to consider are:
Face Validity. Would an outsider or an
expert in the field agree that the indicator is
a valid and logical measure for the stated
result?
Attribution. Does the indicator measure
the contribution of the project?
Measurement Error. Are there any
measurement errors that could affect the
data? Both sampling and non-sampling error
should be reviewed.
2. Integrity
Do the data collected, analyzed and
reported have established mechanisms in
place to reduce manipulation or simple
errors in transcription?
3. Precision
Are data sufficiently precise to present a fair
picture of performance and enable
management decision-making at the
appropriate levels?
4. Reliability
Do data reflect stable and consistent data
collection processes and analysis methods
over time?
5. Timeliness
Are data timely enough to influence
management decision-making (i.e., in terms
of frequency and currency)?
Yes or No
Explanation
Note: This criterion requires the reviewer to
understand what mechanisms are in place to
reduce the possibility of manipulation or
transcription error.
Note: This criterion requires the reviewer to ensure
that the indicator definition is operationally precise
(i.e. it clearly defines the exact data to be collected)
and to verify that the data is, in fact, collected
according to that standard definition consistently
over time.
A Summary of Key Issues and Recommendations:
7
Table 2. SAMPLE DQA FOR OUTPUT INDICATORS: THE MATRIX APPROACH
Document
Source
Data
Source
Data Collection Method/
Key Data Quality Issue
Further Action
Additional Comments/
Notes
AO or IR
Indicators
1.
Number of investment
measures made consistent
with international investment
agreements as a result of
USG assistance
Quarterly
Report
Project
A
A consultant works directly with the
committee in charge of simplifying procedures
and updates the number of measures
regularly on the website
(www.mdspdres.com). The implementer has
stated that data submitted includes
projections for the upcoming fiscal year
rather than actual results.
Yes. Ensure that
only actual results
within specified
timeframes are
used for
reporting.
Meeting with COTR 6/20/10
and 7/6/10.
2.
Number of public and private
sector standards-setting
bodies that have adopted
internationally accepted
guidelines for standards
setting as a result of USG
assistance
Semi-Annual
Report
Project
A
No issues. Project works only with one body
(the Industrial Standards-Setting Service) and
maintains supporting documentation.
No.
Meeting with COTR and COP
on 6/20/10.
3.
Number of legal, regulatory,
or institutional actions taken
to improve implementation
or compliance with
international trade and
investment agreements due
to support from USGassisted organizations
Quarterly
Report
Project
A
Project has reported “number of Regional
Investment Centers”. This is not the same as
counting “actions”, so this must be corrected.
Yes. Ensure that
the correct
definition is
applied.
Meeting with COTR, COP,
and Finance Manager and M&E
specialist on 6/20/10. The
indicator was clarified and the
data collection process will be
adjusted accordingly.
4.
Number of Trade and
Investment Environment
diagnostics conducted
Quarterly
Report
Projects
A and B
No issues. A study on the investment
promotion policy was carried out by the
project. When the report is presented and
validated the project considers it
“conducted”.
No.
Meeting with CTO and COPs
on 6/25/10.
8
For more information:
TIPS publications are available online at [insert website].
Acknowledgements:
Our thanks to those whose experience and insights helped shape this publication including Gerry Britan and
Subhi Mehdi of USAID’s Office of Management Policy, Budget and Performance (MPBP). This publication was
written by Michelle Adams-Matson, of Management Systems International.
Comments can be directed to:
Gerald Britan, Ph.D.
Tel: (202) 712-1158
[email protected]
Contracted under RAN-M-00-04-00049-A-FY0S-84
Integrated Managing for Results II
9
NUMBER 19
1ST EDITION, 2010 DRAFT
PERFORMANCE MONITORING & EVALUATION
TIPS
RIGOROUS IMPACT EVALUATION
ABOUT TIPS
These TIPS provide practical advice and suggestions to USAID managers on issues related to
performance monitoring and evaluation. This publication is a supplemental reference to the
Automated Directive System (ADS) Chapter 203.
WHAT IS RIGOROUS
IMPACT
EVALUATION?
Rigorous impact evaluations are
useful for determining the effects
of
USAID
programs
on
outcomes.
This
type
of
evaluation allows managers to
test development hypotheses by
comparing changes in one or
more specific outcomes to
changes that occur in the
absence
of
the
program.
Evaluators
term
this
the
counterfactual. Rigorous impact
evaluations
typically
use
comparison groups, composed of
individuals or communities that
do not participate in the
program. The comparison group
FIGURE 1. DEFINITIONS OF IMPACT EVALUATION
• An evaluation that looks at the impact of an intervention on final welfare
outcomes, rather than only at project outputs, or a process evaluation which
focuses on implementation.
• An evaluation carried out some time (five to ten years) after the
intervention has been completed, to allow time for impact to appear.
• An evaluation considering all interventions within a given sector or
geographical area.
• An evaluation concerned with establishing the counterfactual, i.e., the
difference the project made (how indicators behaved with the project
compared to how they would have been without it).
is examined in relation to the
treatment group to determine
the effects of the USAID program
or project.
Impact evaluations may be
defined in a number of ways (see
Figure 1). For purposes of this
TIPS, rigorous impact evaluation
1
is defined by the evaluation
design (quasi-experimental and
experimental) rather than the
topic being evaluated. These
methods can be used to attribute
change at any program or project
outcome level, including
Intermediate Results (IR), sub-IRs,
and Assistance Objectives (AO).
Decisions about whether a
rigorous impact evaluation would
be appropriate and what type of
rigorous impact evaluation to
conduct are best made during
the program or project design
phase, since many types of
rigorous impact
evaluation can only be utilized if
comparison groups are
established and baseline data is
collected before a program or
project intervention begins.
WHY ARE RIGOROUS
IMPACT
EVALUATIONS
IMPORTANT?
A rigorous impact evaluation
enables managers to determine
the extent to which a USAID
program or project actually
caused observed changes.
A Performance Management Plan
(PMP) should contain all of the
tools necessary to track key
objectives (see also TIPS 7
Preparing
a
Performance
Management Plan).
However,
comparing
data
from
performance indicators against
baseline values demonstrates
only
whether
change
has
occurred,
with
very
little
information about what actually
caused the observed change.
USAID program managers can
only say that the program is
correlated with changes in
outcome, but cannot confidently
attribute that change to the
program.
FIGURE 2. A WORD ABOUT WORDS
Many of the terms used in rigorous evaluations hint at the origin of these
methods: medical and laboratory experimental research. The activities of a
program or project are often called the intervention or the independent
variable, and the outcome variables of interest are known as dependent
variables. The target population is the group of all individuals (if the unit of
analysis or unit is the individual) who share certain characteristics sought by
the program, whether or not those individuals actually participate in the
program. Those from the target population who actually participate are
known as the treatment group, and the group used to measure what would
have happened to the treatment group had they not participated in the
program (the counterfactual) is known as a control group if they are selected
randomly, as in an experimental evaluation, or, more generally, as a
comparison group if they are selected by other means, as in a quasiexperimental evaluation.
There are normally a number of
factors, outside of the program,
that might influence an outcome.
These are called confounding
factors. Examples of confounding
factors include programs run by
other donors, natural events (e.g.,
rainfall, drought, earthquake,
etc.), government policy changes,
or even maturation (the natural
changes that happen in an
individual or community over
time). Because of the potential
contribution
of
these
confounding factors, the program
manager cannot claim with full
certainty that the program
caused the observed changes or
results.
In some cases, the intervention
causes all observed change. That
is, the group receiving USAID
assistance will have improved
significantly while a similar, nonparticipating group will have
stayed roughly the same.
In
other situations, the target group
may have already been improving
and the program helped to
accelerate that positive change.
Rigorous
evaluations
are
2
designed to identify the effects of
the program of interest even in
these cases, where both the
target
group
and
nonparticipating groups may have
both changed, only at different
rates. By identifying the effects
caused by a program, rigorous
evaluations
help
USAID,
implementing partners and key
stakeholders
learn
which
program or approaches are most
effective, which is critical for
effective
development
programming.
WHEN SHOULD
THESE METHODS BE
USED?
Rigorous impact evaluations can
yield very strong evidence of
program effects. Nevertheless,
this method is not appropriate
for all situations.
Rigorous
impact evaluations often involve
extra costs for data collection and
always require careful planning
during program implementation.
To determine whether a rigorous
impact evaluation is appropriate,
potential cost should be weighed
against the need for and
usefulness of the information.
Rigorous
impact
evaluations
answer
evaluation
questions
concerning the causal effects of a
program.
However, other
evaluation designs may be more
appropriate for answering other
types of evaluation questions.
For example, the analysis of ‘why’
and ‘how’ observed changes,
particularly unintended changes,
were produced may be more
effectively answered using other
evaluation methods, including
participatory evaluations or rapid
appraisals. Similarly, there are
situations
when
rigorous
evaluations, which often use
comparison groups, will not be
advisable, or even possible. For
example, assistance focusing on
political parties can be difficult to
evaluate using rigorous methods,
as this type of assistance is
typically offered to all parties,
making the identification of a
comparison group difficult or
impossible. Other methods may
be more appropriate and yield
conclusions
with
sufficient
credibility
for
programmatic
decision-making.
rigorous
impact
While
evaluations are sometimes used
to examine the effects of only
one
program
or
project
approach,
rigorous
impact
evaluations are also extremely
useful for answering questions
about the effectiveness of
alternative
approaches
for
achieving a given result, e.g.,
which of several approaches for
improving farm productivity, or
for delivering legal services, are
most effective.
Missions should consider using
rigorous evaluations strategically
to answer specific questions
about the effectiveness of key
approaches.
When multiple
rigorous evaluations are carried
out across Missions on a similar
topic or approach, the results can
be used to identify approaches
that can be generalized to other
settings, leading to significant
advances
in
programmatic
knowledge. Rigorous methods
are often useful when:
 Multiple
approaches
to
achieving desired results have
been suggested, and it is
unclear which approach is the
most effective or efficient;
 An approach is likely to be
replicated if successful, and
clear evidence of program
effects are desired before
scaling up;
 A program uses a large amount
of resources or affects a large
number of people; and
 In general, little is known about
the effects of an important
program or approach, as is
often the case with new or
innovative approaches.
PLANNING
Rigorous methods require strong
performance
management
systems to be built around a
clear, logical results framework
(see TIPS 13 Building a Results
Framework). The development
hypothesis should clearly define
the logic of the program, with
3
particular emphasis on the
intervention
(independent
variable) and the principal
anticipated results (dependent
variables), and provides the basis
for the questions that will be
addressed by the rigorous
evaluation.
Rigorous evaluation builds upon
the indicators defined for each
level of result, from inputs to
outcomes, and requires high data
quality.
Because
quasiexperimental and experimental
designs typically answer very
specific evaluation questions and
are generally analyzed using
quantitative methods, they can
be paired with other evaluation
tools and methods to provide
context, triangulate evaluation
conclusions, and examine how
and why effects were produced
(or not) by a program. This is
termed mixed method evaluation
(see TIPS 16, Mixed Method
Evaluations).
Unlike
most
evaluations
conducted by USAID, rigorous
impact evaluations are usually
only possible, and are always
most effective, when planned
before project implementation
begins.
Evaluators need time
prior to implementation to
identify appropriate indicators,
identify a comparison group, and
set baseline values. If rigorous
evaluations are not planned prior
to implementation, the number
of potential evaluation design
options is reduced, often leaving
alternatives that are either more
complicated or less rigorous. As
a result, Missions should consider
the feasibility of and need for a
Observed Change
Outcome of Interest
FIGURE 3. CONFOUNDING EFFECTS
Program
Effect
Confounding
Effect
Baseline
Follow-up
= Target Group
= Comparison Group
rigorous evaluation prior to and
during project design.
DESIGN
Although
there
are
many
variations, rigorous evaluations
are divided into two categories:
quasi-experimental
and
experimental. Both categories of
rigorous evaluations rely on the
same basic concept - using the
counterfactual to estimate the
changes caused by the program.
The counterfactual answers the
question, “What would have
happened to program participants
if they had not participated in the
program?” The comparison of
the
counterfactual
to
the
observed change in the group
receiving USAID assistance is the
true measurement of a program’s
effects.
While
before
and
after
measurements of a single group
using a baseline allow the
measurement of a single group
both with and without program
participation, this design does
not control for all the other
confounding factors that might
influence the participating group
during program implementation.
Well constructed, comparison
groups provide a clear picture of
the effects of program or project
interventions on the target group
by
differentiating
program/project effects from the
effects of multiple other factors in
the environment that affect both
the target and comparison
groups. This means that in
situations where economic or
other factors affecting
both
groups make everyone better
off, it will still be possible to see
the additional or incremental
improvement caused by the
program or project, as Figure 3
illustrates.
QUASI-EXPERIMENTAL
EVALUATIONS
To estimate program effects,
quasi-experimental designs rely
on measurements of a nonrandomly selected comparison
group. The most common means
for selecting a comparison group
is
matching,
wherein
the
4
evaluator ‘hand-picks’ a group of
similar units based on observable
characteristics that are thought to
influence the outcome.
For
example, the evaluation of
an
agriculture program aimed at
increasing crop yield might seek
to
compare
participating
communities
against
other
communities with similar weather
patterns,
soil
types,
and
traditional crops, as communities
sharing
these
critical
characteristics would be most
likely to behave similarly to the
treatment group in the absence
of the program.
However, program participants
are often selected based on
certain characteristics, whether it
is level of need, motivation,
location, social or political factors,
or some other factor.
While
evaluators can often identify and
match many of these variables, it
is impossible to match all factors
that might create differences
between the treatment and
comparison groups, particularly
characteristics that are more
difficult to measure or are
unobservable, such as motivation
or social cohesion. For example,
if a program is targeted at
WHAT IS EXPERIMENTAL AND
QUASI-EXPERIMENTAL
EVALUATION?
Experimental design is based on a
the selection of the comparison and
treatment group through random
sampling.
Quasi-experimental design is
based on a comparison group that
is chosen by the evaluator (that is,
not based on random sampling).
FIGURE 4.
QUASI-EXPERIMENTAL EVALUATION OF THE KENYA NATIONAL CIVIC EDUCATION PROGRAM
PHASE II (NCEP II)
NCEP II, funded by USAID in collaboration with other donors, reached an estimated 10 million individuals through
workshops, drama events, cultural gatherings and mass media campaigns aimed at changing individuals’ awareness,
competence and engagement in issues related to democracy, human rights, governance, constitutionalism, and
nation-building. To determine the program’s impacts on these outcomes of interest, NCEP as evaluated using a
quasi-experimental design with a matched comparison group.
Evaluators matched participants to a comparison group of non-participating individuals who shared geographic and
demographic characteristics (such as age, gender, education, and involvement with CSOs). This comparison group
was compared to the treatment group along the outcomes of interest to identify program effects. The evaluators
found that the program had significant long term effects, particularly on ‘civic competence and involvement’ and
‘identity and ethnic group relations, but had only negligible impact on ‘Democratic Values, Rights, and
Responsibilities’. The design also allowed the evaluators to assess the conditions under which the program was
most successful. They found confirmation of prior assertions of the critical role in creating lasting impact of multiple
exposures to civic education programs through multiple participatory methods.
- ‘The Impact of the Second National Kenya Civic Education Programme (NECP II-URAIA) on Democratic Attitudes,
Values, and Behavior’, Steven E. Finkel and Jeremy Horowitz, MSI
communities that are likely
succeed, then the target group
might be expected to improve
relative to a comparison group
that was not chosen based on the
same factors. Failing to account
for this in the selection of the
comparison group would lead to
a biased estimate of program
impact. Selection bias is the
difference
between
the
comparison group and the
treatment group caused by the
inability to completely match on
all characteristics, and the
uncertainty
or
error
this
generates in the measurement of
program effects.
Other
common
quasiexperimental designs, in addition
to matching, are described below.
Non-Equivalent Group Design.
This is the most common quasiexperimental design in which a
comparison group is hand-picked
to match the treatment group as
closely as possible. Since handpicking the comparison group
cannot completely match all
characteristics with the treatment
group, the groups are considered
to be ‘non-equivalent’.
significantly different except in
terms of eligibility for the
program. Because of this, the
group just above the cut-off
serves as a comparison group for
those just below (or vice versa) in
a regression discontinuity design.
Regression
Discontinuity.
Programs often have eligibility
criteria based on a cut-off score
or value of a targeting variable.
Examples
include
programs
accepting only households with
income
below
2,000
USD,
organizations registered for at
least two years, or applicants
scoring above a 65 on a pre-test.
In each of these cases, it is likely
that individuals or organizations
just above and just below the
cut-off value would demonstrate
only marginal or incremental
differences in the absence of
USAID assistance, as families
earning 2,001 USD compared to
1,999 USD are unlikely to be
Propensity Score Matching. This
method is based on the same
rationale as regular matching: a
comparison group is selected
based on shared observable
characteristics with the treatment
group.
However, rather than
‘hand-picking’ matches based on
a small number of variables,
propensity score matching uses a
statistical process to combine
information
from
all
data
collected
on
the
target
population to create the most
accurate matches possible based
on observable characteristics.
5
Interrupted Time Series.1 Some
programs will encounter
situations where a comparison
group is not possible, often
because the intervention affects
everyone at once, as is typically
the case with policy change. In
these cases, data on the outcome
of interest are recorded at
numerous intervals before and
after the program or activity take
places. The data form a timeseries or trend, which the
evaluator analyzes for significant
changes around the time of the
intervention. Large spikes or
drops immediately after the
intervention signal changes
caused by the program. This
method is slightly different from
the other rigorous methods as it
does not use a comparison group
to rule out potentially
confounding factors, leading to
increased uncertainty in
evaluation conclusions.
Interrupted time series are most
effective when data are collected
regularly both before and after
the intervention, leading to a
long time series, and alternative
causes are monitored.
EXPERIMENTAL EVALUATION
In an experimental evaluation, the
treatment
and
comparison
groups are selected from the
target population by a random
process. For example, from a
target
population
of
50
communities that meet the
1
Interrupted time series is normally
viewed as a type of impact evaluation.
It is typically considered quasiexperiemental although it does not use a
comparison group.
eligibility (or targeting) criteria of
a program, the evaluator uses a
coin flip, lottery, computer
program, or some other random
process to determine the 25
communities that will participate
in the program (treatment group)
and the 25 communities that will
not (control group, as the
comparison group is called when
it is selected randomly). Because
they use random selection
processes,
experimental
evaluations are often called
randomized
evaluations
or
randomized
controlled
trials
(RCTs).
Random selection from a target
population into treatment and
control groups is the most
effective tool for eliminating
selection bias because it removes
the possibility of any individual
characteristic
influencing
selection. Because units are not
assigned to treatment or control
groups
based
on
specific
characteristics, but rather are
divided
randomly,
all
characteristics that might lead to
selection bias, such as motivation,
poverty level, or proximity, will be
roughly equally divided between
the treatment and control
groups.
If an evaluator uses
random assignment to determine
treatment and control groups,
she might, by chance, get two or
three
very
motivated
communities in a row assigned to
the treatment group, but if the
program is working in more than
a handful of communities, the
number
of
motivated
communities will likely balance
6
out between treatment
control in the end.
and
Because
random
selection
completely eliminates selection
bias, experimental evaluations are
often easier to analyze and
provide more credible evidence
than quasi experimental designs.
Random assignment can be done
with any type of unit, whether the
unit is the individual, groups of
individuals (e.g., communities or
districts),
organizations,
or
facilities (e.g., health center or
school) and usually follows one of
the designs discussed below.
Simple Random Assignment.
When the number of program
participants has been decided
and additional eligible individuals
are identified, simple random
assignment through a coin flip or
lottery can be used to select the
treatment group and control
groups.
Programs
often
encounter
‘excess
demand’
naturally (for example in training
programs, participation in study
tours, or where resources limit
the
number
of
partner
organizations),
and
simple
random assignment can be an
easy and fair way to determine
participation while maximizing
the
potential
for
credible
evaluation conclusions.
Phased-In Selection. In some
programs, the delivery of the
intervention does not begin
everywhere at the same time. For
capacity or logistical reasons,
some units receive the program
intervention earlier than others.
This type of schedule creates a
natural opportunity for using an
FIGURE 5.
EXPERIMENTAL EVALUATION OF THE IMPACTS OF EXPANDING CREDIT ACCESS IN
SOUTH AFRICA
While commercial loans are a central component of most microfinance strategies, there is much less consensus on
whether consumer loans are also for economic development. Microfinance in the form loans for household
consumption or investment has been criticized as unproductive, usurious, and a contributor to debt cycles or traps.
In an evaluation partially funded by USAID, researchers used an experimental evaluation designed to test the impacts
of access to consumer loans on household consumption, investment, education, health, wealth, and well-being.
From a group of 787 applicants who were just below the credit score needed for loan acceptance, the researchers
randomly selected 325 (treatment group) that would be approved for a loan. The treatment group was surveyed,
along with the remaining 462 who were randomly denied (control group), eight months after their loan application to
estimate the effects of receiving access to consumer credit. The evaluators found that the treatment group was more
likely to retain wage employment, less likely to experience severe hunger in their households, and less likely to be
impoverished than the control group providing strong evidence of the benefits of expanding access to consumer
loans.
-‘Expanding Credit Access: Estimating the Impacts’, Dean Karlan and Jonathan Zinman,
http://www.povertyactionlab.org/projects/print.php?pid=62
experimental design. Consider a
project where the delivery of a
radio-based
civic
education
program was scheduled to
operate in 100 communities
during year one, another 100
during year two, and a final 100
during year three. The year of
participation can be randomly
assigned. Communities selected
to participate in year one would
be designated as the first
treatment group (T1). For that
year, all the other communities
that would participate in Years
Two and Three form the initial
control group. In the second
year, the next 100 communities
would become the second
treatment group (T2), while the
final 100 communities would
continue to serve as the control
group. Random assignment to
the year of participation ensures
that
all
communities
will
participate in the program but
also maximizes evaluation rigor
by reducing selection bias, which
could be significant if only the
most motivated communities
participate in Year One.
Blocked
(or
Stratified)
Assignment. When it is known in
advance that the units to which a
program intervention could be
delivered differ in one or more
ways that might influence the
program outcome, (e.g., age, size
of the community in which they
are located, ethnicity, etc.),
evaluators may wish to take extra
steps to ensure that such
conditions are evenly distributed
between
an
evaluation’s
treatment and control groups. In
a simple block (stratified) design,
an evaluation might separate
men and women, and then use
randomized assignment within
each block to construct the
evaluation’s
treatment
and
control groups, thus ensuring a
specified number or percentage
7
of men and women in each
group.
Multiple Treatments. It is
possible that multiple approaches
will be proposed or implemented
for the achievement of a given
result. If a program is interested
in
testing
the
relative
effectiveness of three different
strategies or approaches, eligible
units can be randomly divided
into three groups. Each group
participates in one approach, and
the results can be compared to
determine which approach is
most effective. Variations on this
design can include additional
groups to test combined or
holistic approaches and a control
group to test the overall
effectiveness of each approach.
COMMON
QUESTIONS AND
CHALLENGES
While
rigorous
evaluations
require significant attention to
detail in advance, they need not
be impossibly complex. Many of
the most common questions and
challenges can be anticipated and
minimized.
COST
Rigorous evaluations will almost
always cost more than standard
evaluations that do not require
comparison groups.
However,
the
additional
cost
can
sometimes
be
quite
low
depending on the type and
availability of data to be
collected.
Moreover, findings
from rigorous evaluations may
lead to future cost-savings,
through improved programming
and more efficient use of
resources over the longer term.
Nevertheless, program managers
must anticipate these additional
costs, including the additional
planning requirements, in terms
of staffing and budget needs.
ETHICS
The use of comparison groups is
sometimes criticized for denying
treatment
to
potential
beneficiaries.
However, every
program has finite resources and
must select a limited number of
program participants. Random
selection of program participants
is often viewed, even by those
beneficiaries
who
are
not
selected, as being the fairest and
most transparent method for
determining participation.
A second, more powerful, ethical
question emerges when a
program
seeks
to
target
participants that are thought to
be most in need of the program.
In
some
cases,
rigorous
evaluations require a relaxing of
targeting
requirements
(as
discussed in Figure 6) in order to
identify enough similar units to
constitute a comparison group,
meaning that perhaps some of
those identified as the ‘neediest’
might be assigned to the
comparison group. However, it is
often the case that the criteria
used to target groups do not
provide a degree of precision
required to confidently rankorder
potential
participants.
Moreover, rigorous evaluations
can help identify which groups
benefit most, thereby improving
targeting for future programs.
SPILLOVER
Programs are often designed to
incorporate ‘multiplier effects’
whereby program effects in one
community naturally spread to
others nearby.
While these
effects help to broaden the
impact of a program, they can
result in bias in conclusions when
the effects on the treatment
group spillover to the comparison
group. When comparison groups
also benefit from a program, then
they no longer measure only the
confounding effects, but also a
portion of the program effect.
This leads to underestimation of
program impact since they
8
FIGURE 6. TARGETING IN
RIGOROUS EVALUATIONS
Programs often have specific
eligibility requirements without
which a potential participant could
not feasibly participate.
Other
programs target certain groups
because of perceived need or
likelihood of success. Targeting is
still
possible
with
rigorous
evaluations, whether experimental
or quasi-experimental, but must be
approached in a slightly different
manner. If a program intends to
work in 25 communities, rather than
defining
one
group
of
25
communities that meet the criteria
and participate in the program, it
might be necessary to identify a
group of 50 communities that meet
the eligibility or targeting criteria
and will be split into the treatment
and comparison group.
This
reduces the potential for selection
bias while still permitting the
program to target certain groups.
In situations where no additional
communities meet the eligibility
criteria and the criteria cannot be
relaxed, phase-in or multiple
treatment approaches, as discussed
below, might be appropriate.
appear better off than they would
have been in the absence of the
program.
In some cases,
spillovers can be mapped and
measured but, most often, they
must be controlled in advance by
selecting treatment and control
groups or units that are unlikely
to significantly interact with one
another.
A special case of
spillover occurs in substitution
bias wherein governments or
other donors target only the
comparison group to fill in gaps
of service. This is best avoided by
ensuring coordination between
the
program
and
development actors.
other
SAMPLE SIZE
During the analysis phase,
rigorous evaluations typically use
statistical tests to determine
whether any observed differences
between
treatment
and
comparison groups represent
actual differences (that would
then, in a well designed
evaluation, be attributed to the
program)
or
whether
the
difference could have occurred
due to chance alone. The ability
to make this distinction depends
principally on the size of the
change and the total number of
units in the treatment and
comparison groups, or sample
size. The more units, or higher
the sample size, the easier it is to
attribute change to the program
rather than to random variations.
During
the
design
phase,
rigorous
impact
evaluations
typically calculate the number of
units (or sample size) required to
confidently identify changes of
the size anticipated by the
program. An adequate sample
size helps prevent declaring a
successful project ineffectual
(false negative) or declaring an
ineffectual project successful
(false positive). Although sample
9
size calculations should be done
before each program, as a rule of
thumb,
rigorous
impact
evaluations are rarely undertaken
with less than 50 units of analysis.
RESOURCES
This TIPS is intended to provide
an introduction to rigorous
impact evaluations.
Additional
resources are provided on the
next page for further reference.
Further Reference
Initiatives and Case Studies:
-
-
-
-
Office of Management and Budget (OMB):
o http://www.whitehouse.gov/OMB/part/2004_program_eval.pdf
o http://www.whitehouse.gov/omb/assets/memoranda_2010/m10-01.pdf
U.S. Government Accountability Office (GAO):
o http://www.gao.gov/new.items/d1030.pdf
USAID:
o Evaluating Democracy and Governance Effectiveness (EDGE):
http://www.usaid.gov/our_work/democracy_and_governance/technical_areas/dg_office/eval
uation.html
o Measure Evaluation:
http://www.cpc.unc.edu/measure/approaches/evaluation/evaluation.html
o The Private Sector Development (PSD) Impact Evaluation Initiative:
www.microlinks.org/psdimpact
Millennium Challenge Corporation (MCC) Impact Evaluations:
http://www.mcc.gov/mcc/panda/activities/impactevaluation/index.shtml
World Bank:
o The Spanish Trust Fund for Impact Evaluation:
http://web.worldbank.org/WBSITE/EXTERNAL/EXTABOUTUS/ORGANIZATION/EXTHDNETW
ORK/EXTHDOFFICE/0,,contentMDK:22383030~menuPK:6508083~pagePK:64168445~piPK:6
4168309~theSitePK:5485727,00.html
o The Network of Networks on Impact Evaluation: http://www.worldbank.org/ieg/nonie/
o The Development Impact Evaluation Initiative:
http://web.worldbank.org/WBSITE/EXTERNAL/EXTDEC/EXTDEVIMPEVAINI/0,,menuPK:39982
81~pagePK:64168427~piPK:64168435~theSitePK:3998212,00.html
Others:
o Center for Global Development’s ‘Evaluation Gap Working Group’:
http://www.cgdev.org/section/initiatives/_active/evalgap
o International Initiative for Impact Evaluation: http://www.3ieimpact.org/
Additional Information:
-
-
Sample Size and Power Calculations:
o http://www.statsoft.com/textbook/stpowan.html
o http://www.mdrc.org/publications/437/full.pdf
World Bank: ‘Evaluating the Impact of Development Projects on Poverty: A Handbook for
Practitioners’:
o http://web.worldbank.org/WBSITE/EXTERNAL/TOPICS/EXTPOVERTY/EXTISPMA/0,,contentM
DK:20194198~pagePK:148956~piPK:216618~theSitePK:384329,00.html
Poverty Action Lab’s ‘Evaluating Social Programs’ Course: http://www.povertyactionlab.org/course/
10
For more information:
TIPS publications are available online at [insert website]
Acknowledgements:
Our thanks to those whose experience and insights helped shape this publication including USAID’s
Office of Management Policy, Budget and Performance (MPBP). This publication was written by Michael
Duthie of Management Systems International.
Comments regarding this publication can be directed to:
Gerald Britan, Ph.D.
Tel: (202) 712-1158
[email protected]
Contracted under RAN-M-00-04-00049-A-FY0S-84
Integrated Managing for Results II
11