Evaluating Social Program1
Evaluating Social Program1
Evaluating Social Program1
• Two questions
First: Why assess needs and resources?
- a deeper understanding of the community (needs ,
resources, assets, culture, social sturctures, history,
etc) allows understanding how to address the
community’s needs and use its resources;
- Assessments encourage community members to
consider the community's assets and how to use
them, as well as the community's needs and how to
address them.
WHY DEVELOP A PLAN FOR ASSESSING LOCAL NEEDS AND RESOURCES?
• SO
It may be important to address the community's priorities
first, in order to establish trust and show respect, even if
you don't believe that those priorities are in fact the most
important issues.
Building relationships and credibility may be more important
at the beginning of a long association than immediately
tackling what seems to be the most pressing need.
Among other things, community members' priorities may be
the right ones: they may see underlying factors that you
don't yet understand.
WHO SHOULD BE INVOLVED IN DEVELOPING A PLAN FOR ASSESSING LOCAL NEEDS AND
RESOURCES?
SWOT stands
for: Strength, Weakness, Opportunity, Threat.
A SWOT analysis guides you to identify your
organization’s strengths and weaknesses (S-
W), as well as broader opportunities and
threats (O-T). Developing a fuller awareness of
the situation helps with both strategic
planning and decision-making.
WHAT IS A SWOT ANALYSIS AND WHY SHOULD YOU USE ONE?
HOW TO EVALUATE
Outline
Causal Inference and Counterfactuals
Randomized Selection Methods
Diff erence-in-Diff erences
Matching
Combining Methods
Causal Inference
Assessing the impact of a program on a series
of out-comes is equivalent to assessing the
causal effect of the program on those outcomes
Most policy questions involve cause-and-eff ect
relationships: Does teacher training improve
students’ test scores? Do conditional cash
transfer programs cause better health
outcomes in children? Do vocational training
programs increase trainees’ incomes?
Causal Inference
To establish causality between a program and
an out-come, we use impact evaluation
methods to rule out the possibility that any
factors other than the program of interest
explain the observed impact.
What is the impact or causal effect of a
program P on an outcome of interest Y? is
given by the basic impact evaluation formula:
α = (Y | P = 1) − (Y | P = 0).
Causal Inference
(α): the causal impact (α) of a program (P) ;
(Y) is the outcome.
Example : if P denotes a vocational training program
and Y denotes income, then the causal impact of the
vocational training program (α) is the difference
between a person’s income (Y) after participating in
the vocational training program (in other words, when
P = 1) and the same person’s income (Y) at the same
point in time if he or she had not participated in the
program (in other words, when P = 0).
Causal Inference
We would like to measure income at the same point in
time for the same unit of observation (a person, in this
case), but in two different states of the world.
By comparing the same individual with herself at the
same moment, we would have managed to eliminate
any outside factors that might also have explained the
difference in outcomes. We could then be confident
that the relation-ship between the vocational training
program and income is causal.
Causal Inference
So the impact (α) of a program as the difference
in outcomes (Y) for the same individual with
and without participation in a program. Yet we
know that measuring the same person in two
different states at the same time is impossible.
the counterfactual is what the outcome (Y)
would have been in the absence of a program
(P)
Counterfactual
The question is how to estimate the
counterfactual?
Answer : use comparison groups or “control
groups”
How to estimate counterfactual
Solving the counterfactual problem requires the
evaluator to identify a “perfect clone” for each program
participant
Key goal of an impact evaluation is to identify a group of
program participants (the treatment group) and a group
of nonparticipants (the comparison group) that are
statistically identical in the absence of the program. If the
two groups are identical, excepting only that one group
participates in the program and the other does not, then
we can be sure that any difference in outcomes must be
due to the program.
How to estimate counterfactual
• For example, the average age in the treatment group
should be the same as the average age in the
comparison group. Second, the treatment and
comparison groups should react to the program in
the same way. For example, the incomes of units in
the treatment group should be as likely to benefit
from training as the incomes of the comparison
group. Third, the treatment and comparison groups
cannot be differentially exposed to other
interventions during the evaluation period.
How to estimate counterfactual
An invalid comparison group is one that differs
from the treatment group in some way other
than the absence of the treatment. Those
additional differences can cause our impact
estimate to be invalid or, in statistical terms,
biased: it will not estimate the true impact of the
program. Rather, it will estimate the effect of the
program mixed with the effect of those other
differences
How to estimate counterfactual
How to construct valid comparison groups that
will allow you to estimate the counterfactual?
First method : before-and-after, or pre-post,
comparisons that compare the outcomes of
program participants prior to and subsequent to
the introduction of a program;
Second method: with-and-without comparisons
between units that choose to enroll and
units that choose not to enroll.
Comparing Before and After
Components:
Inputs: Resources at the disposal of the project,
including staff and budget;
Activities: Actions taken or work performed to
convert inputs into outputs;
Outputs: The tangible goods and services that the
project activities produce (They are directly under the
control of the implementing agency.)
Results chain
Indicators
Hypothese for the impact evaluation
Hypothese 11
Hypothese Hypothese 2 Hypothese 3 Hypothese 4 Hypothese 5
Indicators
Needs assessment
Data to collect
Research Imput
questions Indicators
Hypothese
Output Indicators
Hypothese
Outcomes Indicators
Hypothese
Final outcomes Indicators
Example: Theory of ChangeTop-Down
Example
Case study
Category Assumption
Outputs (New curriculum approved Teachers willing to be trained
Teachers trained Youth trained in business youth can attend training
skills
Outcome (Curriculum widely - Curriculum accepted by local school
implemented authorities
Better under- standing of business - Better quality of teaching lead to the fact
Improved soft skills that youth can attend school regularly
Improved employability
Increased interest for higher education
Higher-Level outcomes (Lower youth New skills are demanded and rewarded
unemployment Higher house- hold by labor market
income)
ModuleV
So?
As a program managers and evaluators, we
first: establish our questions and learning
objectives
Second: select the most appropriate evaluation
tool to provide the necessary information.
What Is the Purpose of the Evaluation?
Evaluations address three types of questions:
First: Descriptive questions: describe processes, conditions,
organizational relationships, and stakeholder views.
Second: Normative questions: compare what is
taking place to what should be taking place. They
compare the current situation with the specific
objectives and targets that have been defined
(Has our project been implemented as intended?
Is it performing as expected?).
What Is the Purpose of the Evaluation?
Descriptive Cause-and-
Normative
(only) effect
Process
evaluation
Cost-effectiveness/Cost-
Linking Evaluation Questions to Evaluation
Design
No Evaluation: if a program manager requires only
descriptive information about the intervention, for
example, because the project is in a very early stage
and the objective is to obtain some general
information about how the program is being
implemented.
Performance evaluations: assess how well program
objectives have been formulated as well as the
program’s progress in achieving these objectives.
Linking Evaluation Questions to Evaluation
Design
Such evaluations can be carried out across all stages of
implementation, but they are particularly common
for mid-term reviews (when their focus is on learning
for program management) or at program completion
(when their focus is on account- ability and lessons
learned for future interventions).
Performance evaluations may incorporate more
extensive data collection, such as a before-and-after
comparison of participant outcomes or additional
qualitative tools.
Linking Evaluation Questions to Evaluation
Design
Process Evaluation: allows to fully understanding
how
a program works and seek to assess how well a
program is being implemented. it determines
whether there are gaps between planned and
realized activities and outputs and try to understand
the reasons for gaps. Building on descriptive
information such as what activities are being offered
and who is participating in the program (or who is
not).
Linking Evaluation Questions to Evaluation
Design
Process evaluations tend to rely on a mix of
quantitative and qualitative tools, including key
informant interviews, user satisfaction surveys, direct
observation, and focus groups.
Impact evaluations are the only type of evaluation to
specifically answer cause-and-effect questions in a
quantifiable manner. Such questions require us to
determine not only whether the desired outcomes
occurred but also whether those outcomes occurred
because the program was implemented.
Linking Evaluation Questions to Evaluation
Design
To estimate the causal effect of a program on outcomes
of interest, any method chosen must estimate the so-
called counterfactual, that is, what would have
happened to program participants in the absence of
the program. To do this, impact evaluations require
finding a comparison group; This is what makes
impact evaluations different from all other
evaluations. As a result, they tend to require more
time and quantitative skills, and they typically cost
more than other evaluation types.
Linking Evaluation Questions to Evaluation
Design
Cost-Effectiveness and Cost-Benefit Analyses: assess
monetary and nonmonetary program costs and
compare them with alternative uses of the same
resources and the benefits produced by the
intervention (Baker 2000). Cost-effectiveness analysis
(CEA) measures the cost per output or outcome (e.g.,
$300 per youth trained, $500per job created) and
compares this cost to similar interventions of our
own and other organizations.
•
Linking Evaluation Questions to Evaluation
Design
It thus answers the question about how much output or
outcome we get per dollar spent (descriptive) and whether
there is a gap with our expectations (normative). Cost-benefit
analysis (CBA), in turn, weighs the total expected costs against
the total expected benefits (outcomes) of an intervention,
where both costs and benefits are typically expressed in
monetary terms. For instance, if our program were to help 500
youth find and keep jobs or set up sustainable small
businesses, we would estimate the aggregate benefits in terms
of higher incomes, better health, lower crime, etc., and
compare these benefits to the overall costs of the
intervention.
Does Our Operational Context Fit the Desired
Type of Evaluation?
After formulating the right questions and identifying a
potential type of evaluation, we need to assess the
operational context of the intervention to
understand what evaluation can be implemented
within the given constraints.
Timing
Questions about what kind of information is needed
are closely related to the question of when the
results of the evaluation need to be available.
Knowing when they need to be available determines
when the information needs to be collected.
Does Our Operational Context Fit the Desired
Type of Evaluation?
At What Stage of the Program Is the Information Needed?
How Long Does the Evaluation Take? Information needs vary
depending on the program lifecycle. For example, a program that
has just been planned may require a cost-effectiveness analysis to
help determine whether or not to implement the program. How
long an evaluation takes partly depends on the methods used for
collecting and analyzing data, which differ according to the type
of evaluation, and on the breadth and depth desired for the
particular study, which differ within each type of evaluation
At what point in a program’s lifecycle different evaluation strategies are best conducted?
-Resolve
operational issues, -Run fully
-Establish people Expand the program if
operational program Phase 2 succeeds
interest in with limited
program, Incorporate lessons
geographical scope from impact
-Verify that - Conduct an impact evaluation
participants are evaluation to analyze
satisfied with the Continue monitoring
the effect of the and performance
progam program evaluations
Types of Programs That Usually Justify an Impact
Evaluation
Strategically relevant and influential;
Innovative or untested;
Replicable.
Overview of main evaluation types
Cost-
Performance Process Evaluation Impact Effectiveness
Evaluation Evaluation and Cost-
Benefit
•
Analyses
•
What are Do programs Are adequate resources How have Are program
have clear and systems participants’ costs justified
the main objectives? (management, well-being compared
questions Is the program information, etc.) in changed as a with similar
answered design place? result of the interventions?
appropriate to Is the program being intervention? Are aggregate
by this type achieve the implemented according Are there any program costs
of objectives? to design? unintended justified in
evaluation? To what extent What are the actual consequences terms benefits
have program steps and activities , positive or achieved?
objectives been involved in delivering a negative, on
achieved? product or service? program
Do priorities need What do beneficiaries or participants?
to be changed? other stakeholders know
or think about the
program?
Evaluation Evaluation and Cost-Bene
Analyses
131
Impact evaluations
132
Impact evaluations
134
what is impact?
135
What is counterfactual?
136
Illustration of a program impact
Outcome
Outcome
Without intervention
Without intervention
137
Time Time
What is counterfactual?
138
How to estimate impact?
139
How to estimate impact?
Eligible Eligible
Random
assignement
Eligible non
Treatment Comparison Treatment
participants
Comparison
141
Key Points
146
How It Works
Total population
Select sample
Step 2:
Treatment Comparison
147
How It Works
149
How It Works
Random assignment
Comparison
Treatment (1,000) (1.000)
152
How It Works
153
How It Works
154
How It Works
155
How It Works
156
How It Works
157
How It Works
159
How It Works
160
When Can I Use a Lottery Design?
161
Advantages of a Lottery Design?
162
Advantages of a Lottery Design?
163
Disadvantages of a Lottery Design?
164
Method 2: Randomized Phase-In Design
165
How It Works
Random assignment
Year 1
Year 2
Year 3
Treatment Comparison
167
How It Works
169
When Can I Use A Phase-in Design?
170
Advantages of Randomized Phase-In Design
171
Disadvantages of Randomized Phase-In Design
173
Method 3: Discontinuity Design
174
How It Works
175
How It Works
176
How It Works
177
Sample discontinuity chart
178
How It Works
179
When Can I Use a Discontinuity Design?
181
Disadvantages of a Discontinuity Design
183
Overview of impact evaluation techniques
184
Overview of impact evaluation techniques
Methodology Description Comparison Required Data When to Use?
Group Assumptions Needed
185
Key Points
187
Module VI
Cost-Effectiveness Analysis
Why cost-effectiveness analysis?
191
What is cost-effectiveness analysis (CEA)
192
• It is a quantitative method to assess wether
the effect of intervention is justified by its
costs
193
cost-effectiveness analysis (CEA)
and cost-benefit analysis (CBA)
• Both CEA and CBA can be used before the
intervention or during or after the program.
However, only retrospective analysis will provide
practitioners with the full information of actual
costs and benefits to determine the overall success
of the intervention. In fact, an impact evaluation is
a necessary condition for having a reliable estimate
of the program’s direct and indirect benefits (in this
course, we only emphasize on direct benefits).
194
Weighing costs and benefits
COSTS BENEFITS
195
How to choose between cost-effectiveness
analysis(CEA) and cost-benefit analysis (CBA)
Trying to put a dollar value on many intangible
benefits may be difficult and subjective, and it can
represent a big challenge, especially for CBA. Hence,
CBA is usually considered most useful when there are
multiple types of benefits and consensus about how
to quantify them in monetary terms.
196
How to Calculate Net Benefits for Participants
197
How to Calculate Net Benefits for Participants
198
How to Calculate Net Benefits for Participants
200
How to Calculate Net Benefits for Participants
Second, the program must have equal or greater return than running
realized the same or greater impact per person for less money? This
options. The program that has the greatest impact but costs the
201
How to Calculate Net Benefits for Participants
202
How to Calculate Net Benefits for Participants
203
How to Calculate Net Benefits for
Participants
Discounting is simply a way of taking account of
time by attaching lower weights to cash flows
(or benefits) which occur in the more distant
future than to those which occur relatively
soon.
The cost – effectiveness ratio will then be the
discounted costs divided by the discounted
physical outputs.
204
How to Calculate Net Benefits for Participants
207
Bibligraphy
• Bamberger, M., Rugh, J. et, L. (2012). RealWorld Evaluation : Working Under Budget,
Time, Data, and Political Constraints (2ème éd.). Sage Publications. Chen, H.,
Donaldson, S., Mark, M. (éds.) (2011). Advancing Validity in Outcome Évaluation :
Theory and Practice. New Directions for Evaluation. No130. Jossey-Bass et American
Evaluation Association. Funnell, S. et Rogers, P (2011). Purposeful Program Theory :
Effective Use of Theories of Change and Logic Models. Jossey-Bass Publications.
• Jones, H. (2011). « A guide to monitoring and evaluating policy influence. » Londres :
Overseas Development Institute. http://www.odi.org.uk/ resources/docs/6453.pdf
• Mayne, J. (2001). « Addressing attribution through contribution analysis : Using
performance measures sensibly », Canadian Journal of Program Evaluation, 16(1), 1-
24.
• Mayne, J. (2011). « Contribution Analysis : Advancing Cause and Effect. » In
Evaluating the Complex : Attribution, Contribution, and Beyond, édité par Robert
Schwartz, Mita Mara, et Kim Forss. Londres : Transaction Publishing
Bibligraphy
• OCDE/CAD. (2011). Evaluating Impacts : An Overview. http://www.oecd.org/document/17/0,3
746,en_21571361_34047972_46179985_1_1_1_1,00. html
• OCDE/CAD. (2002). Glossaire des principaux termes relatifs à l’évaluation et la gestion axée sur
les résultats. http://www.oecd.org/dataoecd/29/21/2754804.pdf Malhotra, A. (2010).
« Connecting Routine M&E with Impact Evaluation, » International Center for Research on
Women. Présentation à InterAction/3ie Impact Evaluation Design Clinic, Washington, D.C., avril
2010. http:// www.interaction.org/sites/default/files/18/ Anju_Malhotra_-
_Connecting_Routine_M&E_ with_Impact_Evaluation.pdf
• Paul, S. (2002). Holding the State to Account : Citizen Monitoring in Action. Bangalore : Public
Affairs Centre.
• Pawson, R. et Tilley, N. (1997). Realistic Evaluation. Sage Publications.
• Scriven, M. (1991). Evaluation Thesaurus (4ème éd). Sage Publications.
• Woolcock, M. (2011). « Guest Post on The Importance of Time and Trajectories in
Understanding Project Effectiveness. » Development Impact. Blogs.worldbank.org.
http://blogs.worldbank.org/impactevaluations/ guest-post-michael-woolcock-on-the-
importanceof-time-and-trajectories-in-understanding-project-effe.