Transcript Document
IDEV 624 – Monitoring and Evaluation Evaluating Program Impact Elke de Buhr, PhD Payson Center for International Development Tulane University Process vs. Outcome/Impact Monitoring Process Monitoring Outcome Impact Monitoring Evaluation LFM USAID Results Framework A Public Health Questions Approach to HIV/AIDS M&E Are we doing them on a large enough scale? Determining Collective OUTCOMES Effectiveness & IMPACTS OUTCOMES Are we doing them right? Monitoring & Evaluating National OUTPUTS Programs ACTIVITIES Are collective efforts being implemented on a large enough scale to impact the epidemic? (coverage; impact)? Surveys & Surveillance Are interventions working/making a difference? Outcome Evaluation Studies Are we implementing the program as planned? Outputs Monitoring What are we doing? Are we doing it right? Process Monitoring & Evaluation, Quality Assessments Are we doing What interventions and resources are needed? INPUTS the right Needs, Resource, Response Analysis & Input Monitoring things? Understanding What interventions can work (efficacy & effectiveness)? Potential Efficacy & Effectiveness Studies, Formative & Summative Evaluation, Research Synthesis Responses What are the contributing factors? Determinants Research Problem Identification What is the problem? Situation Analysis & Surveillance 7/21/2015 (UNAIDS 2008) Strategic Planning for M&E: Setting Realistic Expectations All Most Input/ Output Monitoring Process Evaluation Number of Projects Some Outcome Monitoring / Evaluation Few* Impact Monitoring / Evaluation Levels of Monitoring & Evaluation Effort *Disease impact monitoring is synonymous with disease surveillance and should be part of all national-level efforts, but cannot be easily linked to specific projects 7/21/2015 4 Monitoring Strategy • Process Activities • Outcome/Impact Goals and Objectives Impact Evaluation Impact Evaluation • Impact evaluations are undertaken to find out whether a program has accomplished its intended effects • Directed at the net effects of an intervention, impact evaluations produce "an estimate of the impact of the intervention uncontaminated by the influence of other processes and events that also may affect the behavior or conditions at which the social program being evaluated is directed” (Rossi/Freeman 1989: 229) • Ideally, impact assessments establish causality by means of a randomized experiment Outcome vs. Impact • Outcome level: Status of an outcome at some point of time • Outcome change: Difference between outcome levels at different points in time • Impact/program effect: Proportion of an outcome change that can be attributed uniquely to a program as opposed to the influence of some other factor (Rossi/Lipsey/Freeman 2004) Outcome vs. Impact (cont.) • Impact/program effect: the value added or net gain that would not have occurred without the program and the only part of the outcome for which the program can honestly take credit – Most demanding evaluation task – Time-consuming and costly (Rossi/Lipsey/Freeman 2004: 207) Outline of an Impact Evaluation 1. Unit of analysis 2. Research question/hypothesis 3. Evaluation design 4. Sampling method 5. Impact indicators 6. Data analysis plan 1. Unit of Analysis Unit of Analysis • Unit of analysis: The units on which outcome measures are taken in an impact assessment and, correspondingly, the units on which data are available for analysis • The unit of analysis in impact assessments is determined by 1. 2. the nature of the intervention and the targets to which the intervention is directed • Can be individuals, households, neighborhoods, organizations, geographic areas, etc. (Rossi/Lipsey/Freeman 2004) What are your program’s units of analysis? 2. Research Question/Hypothesis Hypothesis • Hypothesis: Formal statement that predicts relationship between one or more factors and the problem under study • Support or reject the null hypothesis • Null = no relationship • Test: – Compare same variable over time – Comparison between two or more groups Can you formulate a null hypothesis for your program? 3. Evaluation Design Evaluation Designs • Evaluation strategies: – Comparisons over time – Comparison between groups • Research designs: – Pre-test/Post-test designs – Time series – Quasi-experiments – Randomized experiments Comparisons Over Time Time O1 X Pretest/Posttest design O2 Time O1 O2 O3 X O4 O5 O6 O3 X O4 Time O1 X O2 X Longitudinal designs / Time series Effect of Intervention? (Fisher, A A and J R Foreit Designing HIV/AIDS Intervention Studies: An Operations Handbook Population Council: May 2002, p.56) Effect of Intervention? (Fisher and Foreit, p.57) Effect of Intervention? (Fisher and Foreit, p. 57) Effect of Intervention? (Fisher and Foreit, p. 58) Comparisons Between Groups Time Experimental group O1 Comparison group O3 X O2 O4 Quasiexperimental design Time Experimental group Control group R O1 O3 X O2 O4 Experimental design Randomized Experiments • “Flagships of impact assessment” (Rossi/Lipsey/Freeman 2004: 262) • When conducted well, provide the most credible conclusions about program effects • Isolate the effects of the intervention being evaluated by ensuring that intervention and control group are statistically equivalent except for the intervention received • In practice, it is sufficient if groups, as aggregates, are comparable with regard to any characteristic relevant to the outcome Randomization • Randomization: Assignment of potential targets to intervention and control groups on the basis of chance so that every unit in a target population has the same probability as any other to be selected for either group • Approximations of randomization: Acceptable if the groups that are being compared do not differ on any characteristic relevant to the intervention or the expected outcomes ( Quasi-experiments) (Rossi/Lipsey/Freeman 2004) Feasible? • Randomized experiments are not feasible for all impact assessments • Results may be ambiguous if – program in early stages of implementation – interventions change in ways experiments cannot easily capture • In addition, the method may – be perceived as unfair or unethical (requires withholding services from parts of the target population) – be too resource intensive (technical expertise, time, costs, etc.) – cause disruption in program procedures for delivering services, create artificial situation Quasi-Experimental Designs • Often used when it is not feasible to randomly assign targets to intervention and control groups • Types of quasi-experimental designs: matched controls, statistical controls, reflexive controls, etc. • Threats to validity: Selection bias, secular trends, interfering events, maturation Threats to Validity Threats to Internal Validity • INTERNAL VALIDITY: Any changes that are observed in the dependent variable are due to the effect of the independent variable. They are not due to some other independent variables (extraneous variables, alternative explanations, rival hypotheses). The extraneous variables need to be controlled for in order to be sure that any results are due to the treatment and thus the study is internally valid. • Threat of History: Study participants may have had outside learning experiences and enhanced their knowledge on a topic and thus score better when they are assessed after an intervention independent from the impact of the intervention. (No control group) Threat of Maturation: Study participants may have matured in their ability to understand concepts and developed learning skills over time and thus score better when they are assessed after an intervention independent from the impact of the intervention. (No control group) Threat of Mortality: Study participants may drop out and do not participate in all measures. Those that drop out are likely to differ from those that continue to participate. (No pretest) Treat of Testing: Study participants might do better on the posttest compared to the pretest simply because they take the same test a second time. Threat of Instrumentation: The posttest may have been revised or otherwise modified compared to the pretest and the two test are not comparable anymore. John Henry Effect: Control group may try extra hard after not becoming part of the “chosen” group (compensatory rivalry). Resentful Demoralization of Control Group: Opposite of John Henry Effect. Control group may be demoralized and perform below normal after not becoming part of the “chosen” group. Compensatory Equalization: Control group may feel disadvantaged for not being part of the “chosen” group and receive extra resources to keep everybody happy. This can cloud the effect if the intervention. Statistical Regression: Threat to validity in cases in which the researcher uses extreme groups as study participants that have been selected based on test scores. Due to the role that chance plays in test scores, the scores of students that score at the bottom of the normal curve are likely to go up, the scores of those that score at the top will go down if they are assessed a second time. Differential Selection: Experimental and control group differ in its characteristics. This may influence the results. Selection-Maturation Interaction: Combines the threats to validity described as differential selection and maturation. If experimental and control group differ in important respects, as for example age, differences in achievement might be due to this maturational characteristic rather than the treatment. Experimental Treatment Diffusion: Close proximity of treatment and control group might result in treatment diffusion. This clouds the effect of the intervention. • • • • • • • • • • • Threats to Validity Matrix History Maturation Mortality Testing Instrumentation One-Shot Case Study YES YES YES - - One-Group PretestPosttest Design YES YES CONT. YES Time Series Design YES CONT. CONT. Pretest-Posttest Control Group Design CONT. CONT. Posttest-Only Control Group Design CONT. Single-Factor Multiple Treatment Designs John Henry Effect Compensatory Equalization Differential Selection - - - MAYBE - - - YES MAYBE - - - CONT. CONT. CONT. MAYBE MAYBE CONT. CONT. YES - - MAYBE MAYBE CONT. CONT. CONT. CONT. CONT. CONT. MAYBE MAYBE CONT. Solomon 4 – Group Design CONT. CONT. CONT. CONT. CONT. MAYBE MAYBE CONT. Factorial Design CONT. CONT. CONT. CONT. CONT. MAYBE MAYBE CONT. Static-Group Comparison Design CONT. CONT. YES - - MAYBE MAYBE YES Nonequivalent Control Group Design CONT. CONT. CONT. CONT. CONT. MAYBE MAYBE CONT. Research Designs - Variations A. Simple Designs B. Cross-Sectional Studies C. Longitudinal Studies D. Experimental Designs A. Simple Designs • One-Shot Case Study XO • One-Group Pretest-Posttest Design OXO • Time Series Design OOOOXOOOO R = Random assignment of subjects to conditions X = Experimental treatment O = Observation of the dependent variable (pretest, posttest, interim measure, etc.) B. Cross-Sectional Studies Group 1 Comparison of groups. One point in time. Variations: Case-control study Group 2 Group 3 Case-Control Study Group 1 (with characteristic) Event(s) Group 2 (without characteristic) Comparison of groups. One point in time. Major limitations: Cannot be sure that population has not changed since event(s). C. Longitudinal Studies Population Population Population Comparison of population over time. Repeated measurements. Variations: Panel study, Cohort study Panel Study Group 1 Group 1 Group 1 Measures change over time. Repeated data collection from same individuals. Major limitations: High drop-out rates pose threat to internal validity. Cohort Study Cohort (1) Cohort (2) Cohort (3) Measures change over time. Repeated data collection from same cohort but different individuals. Major limitations: Measures total change but fluctuations within cohort are not assessed. D. Experimental Designs Group 1 Experiment Group 1 Pre-Test Post-Test Group 2 Group 2 Compares group(s) exposed to treatment with group not exposed to treatment. Measures at two points of time. Variations: True experimental design, Quasi-experimental design True Experimental Design Group 1 Target population Groups assigned randomly. Experiment Group 1 Pre-Test Post-Test Group 2 Group 2 Compares group(s) exposed to treatment with group not exposed to treatment. Measures at two points of time. Research subjects are assigned randomly to treatment and control group. Major limitations: Not feasible for all research & ethical problems. True Experimental Designs • True experimental designs use control groups and random assignment of participants Variations: • Pretest-Posttest Control Group Design • Posttest-Only Control Group Design • Single-Factor Multiple Treatment Designs • Solomon 4 – Group Design • Factorial Design Pretest-Posttest Control Group Design ROXO RO O • The randomly assigned experimental group receives the treatment and the control group receives no treatment or an alternative treatment Posttest-Only Control Group Design RXO R O • Like previous but without pretest. Single-Factor Multiple Treatment Designs R O X1 O R O X2 O RO O • Extension of Pretest-Posttest Control Group Design • Sample is assigned randomly to one of several conditions Solomon 4 – Group Design ROXO RO O R XO R O • Developed by researchers that worried about the effect of pretesting on the validity of the results. Factorial Design Two Independent Variables A B AxB Three Independent Variables A B C AxB AxC BxC AxBxC • Allows to include more than one independent variable. • Test for the effects of different kinds of variables that might be expected to influence outcomes (gender, age, etc.). Quasi-Experimental Design Group 1 Target population Groups not assigned randomly. Experiment Group 1 Pre-Test Post-Test Group 2 Group 2 Compares group(s) exposed to treatment with group not exposed to treatment. Measures at two points of time. Random assignment not possible. Major limitations: Not a true experiment. Threats to validity. ( Selection bias) Quasi-Experimental Designs • Quasi-experimental designs lack the random assignment of experimental designs. Variations: • Static-Group Comparison Design XO ------------O • Nonequivalent Control Group Design OXO ------------O O Choosing an Evaluation Design Impact Evaluation Strategy • Comparison – Same group (over time) – Different groups • Design balances accuracy and reliability with cost and feasibility What is a “good enough” research design? Research Design Flow-Chart Research Design Observational Study Cross-Sectional Experimental Study Longitudinal Methods Survey Research Participant Observation Single Group True Experiment Quasi-Experiment Methods Clinical Experiment Natural Experiment Comparison Group Flow Chart (Methodologist Toolchest, Version 3.0) 4. Sampling Methods Sample Selection • Sample size • Sampling frame • Sample selection = sampling – Probability sampling – Nonprobability sampling Sampling Methods • Census vs. Sampling – Census measures all units in a population – Sampling identifies and measures a subset of individuals within the population • Probability vs. Non-Probability Sampling – Probability sampling results in a sample that is representative of the target population – A non-probability sample is not representative of any population Probability Sampling • Sample representative of the target population, large sample size – Simple random/systematic sampling . – Stratified random/systematic sampling – Cluster sampling – Experimental and quasi-experimental designs Advantages • Findings representative of the population • Advanced statistical analysis Disadvantages • Costly and time consuming (depending on target population) • Significant training needs 5. Impact Indicators Concepts, Variables and Indicators Example 1 Example 2 Example 3 Concepts Size Health Variables Area Economic well-being Income per capita Indicators Square kilometers (Phuong Pham, Introduction to Quantitative Analysis) Life Expectancy Purchasing Average Power Parity years of life (PPP) GNP if born in ($) per capita 1970 Indicator Criteria 1. Measurable (able to be recorded and analyzed in quantitative or qualitative terms) 2. Precise (defined the same way by all people) 3. Consistent (not changing over time so that it always measures the same thing) 4. Sensitive (changing proportionally in response to actual changes in the condition or item being measured) Categorical vs. Continuous Variables • Continuous variables – A variable that can be measured (weight, height, age, etc.) • Categorical variables – A variable that cannot be measured but can be categorized (ethnic group, age group, educational level, socio-economic class, etc.) 6. Data Analysis Plan Data Analysis • Type of variable – Categorical – Continuous • Type of data analysis – Descriptive analysis – Hypothesis testing Descriptive Analysis vs. Hypothesis Testing • Descriptive data analysis – Organizing and summarizing data • Statistical inference – Procedure by which we reach a conclusion about a population on the basis of the information contained in a sample that has been drawn from that population (Phuong Pham, Introduction to Quantitative Analysis) Exercise Exercise • Outline an Outcome and/or Impact Evaluation for your program • Include a description of: 1. 2. 3. 4. 5. 6. Unit of analysis Research question/hypothesis Evaluation design Sampling method Impact indicators Data analysis plan