1 / 30

Using Randomized Evaluations to Improve Policy

Nandini Krishnan. Using Randomized Evaluations to Improve Policy. Objective. To Identify the causal effect of an intervention separate the impact of the program from other factors Need to find out what would have happened without the program

yair
Télécharger la présentation

Using Randomized Evaluations to Improve Policy

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Nandini Krishnan Using Randomized Evaluations to Improve Policy

  2. Objective • To Identify the causal effect of an intervention • separate the impact of the program from other factors • Need to find out what would have happened without the program • Cannot observe the same person with and without the program at the same point of time >> Create a valid counterfactual

  3. Correlation is not causation 1) ? Fertilizer Use High Yield OR ? 2) ? High Yield Knowledge of Technologies Fertilizer Use

  4. Illustration: Fertilizer Voucher Program (Before-After) (+6) BIASED measure of the program impact 4

  5. Motivation • Hard to distinguish causation from correlation from statistical analysis of existing data • However complex the statistics can only see that X moves with Y • Hard to correct for unobserved characteristics, like motivation/ability • May be very important- also affect outcomes of interest • Selection bias a major issue for impact evaluation • Projects started at specific times and places for particular reasons • Participants may be selected or self-select into programs • First farmers to adopt a new technology are likely to be very different from the average farmer, looking at their yields will give you a misleading impression of the benefits of a new technology

  6. (+4) Impact of the program Illustration: Fertilizer Voucher Program (Valid Counterfactual) (+2) Impact of other (external) factors 6

  7. Experimental Design • All those in the study have the same chance of being in the treatment or comparison group • By design, treatment and comparison have the same characteristics (observed and unobserved), on average • Only difference is treatment • With large sample, all characteristics average out • Unbiased impact estimates

  8. Options for Randomization • Lottery (0nly some receive) • Lottery to receive information about a new agricultural technology • Random phase-in (everyone gets it eventually) • Train some farmers groups each year • Variation in treatment • Some get information on new seed, others get credit, others get demonstration plot on their land etc • Encouragement design • One farmers support center per district • Some farmers get travel voucher to attend the center

  9. Lottery among the qualified Must receive the program Randomize who gets the program Not suitable for the program

  10. Opportunities • Budget constraint prevents full coverage • Random assignment (lottery) is fair and transparent • Limited implementation capacity • Phase-in gives all the same chance to go first • No evidence on which alternative is best • Random assignment to alternatives with equal ex ante chance of success

  11. Opportunities for Randomization • Take up of existing program is not complete • Provide information or incentive for some to sign up- Randomize encouragement • Pilot a new program • Good opportunity to test design before scaling up • Operational changes to ongoing programs • Good opportunity to test changes before scaling them up

  12. Different levels you can randomize at • Individual • Farm • Farmers’ Association • Irrigation block • Village level • Women’s association • Youth groups • School level

  13. Group or individual randomization? • If a program impacts a whole group-- usually randomize whole community to treatment or comparison • Easier to get big enough sample if randomize individuals Individual randomization Group randomization

  14. Unit of Randomization • Randomizing at higher level sometimes necessary: • Political constraints on differential treatment within community • Practical constraints—confusing to implement different versions • Spillover effects may require higher level randomization • Randomizing at group level requires many groups because of within community correlation

  15. Elements of an experimental design

  16. External and Internal Validity (1) • External validity • The evaluation sample is representative of the total population • The results in the sample represent the results in the population We can apply the lessons to the whole population • Internal validity • The intervention and comparison groups are truly comparable •  estimated effect of the intervention/program on the evaluated population reflects the real impact on thatpopulation

  17. External and Internal Validity (2) • An evaluation can have internal validity without external validity • Example: A randomized evaluation of encouraging women to stand for elections in urban areas may not tell us much about impact of a similar program in rural areas • An evaluation without internal validity, can’t have external validity • If you don’t know whether a program works in one place, then you have learnt nothing about whether it works elsewhere.

  18. Random Sample- Randomization Randomization Internal& external validity National Population Samples National Population

  19. Stratification Randomization Internal validity Population Population stratum Samples of Population Stratum

  20. Representative but biased: useless National Population Randomization Biased assignmentUSELESS! 20

  21. Efficacy & Effectiveness • Efficacy • Proof of concept • Smaller scale • Pilot in ideal conditions • Effectiveness • At scale • Prevailing implementation arrangements -- “real life” • Higher or lower impact? • Higher or lower costs?

  22. Advantages of “experiments” • Clear and precise causal impact • Relative to other methods • Much easier to analyze- Difference in averages • Cheaper (smaller sample sizes) • Easier to explain • More convincing to policymakers • Methodologically uncontroversial

  23. Randomly assigning wheat plots to receive regular irrigation water Wheat plants do NOT • Raise ethical or practical concerns about randomization • Fail to comply with Treatment • Find a better Treatment • Move away—so lost to measurement • Refuse to answer questionnaires • Human beings can be a little more challenging!

  24. What if there are constraints on randomization? • Some interventions can’t be assigned randomly • Partial take up or demand-driven interventions: Randomly promote the program to some • Participants make their own choices about adoption • Perhaps there is contamination- for instance, if some in the control group take-up treatment 24

  25. Random Promotion(Encouragement Design) • Those who get promotion are more likely to enroll • But who got promotion was determined randomly, so not correlated with other observables/non-observables • Compare average outcomes of two groups: promoted/not promoted • Effect of offering the encouragement (Intent-To-Treat) • Effect of the intervention on the complier population (Local Average Treatment Effect) • LATE= ITT/proportion of those who took it up

  26. Randomization

  27. Random encouragement

  28. Common pitfalls to avoid • Calculating sample size incorrectly • Randomizing one district to treatment and one district to control and calculating sample size on number of people you interview • Collecting data in treatment and control differently • Counting those assigned to treatment who do not take up program as control—don’t undo your randomization!! 28

  29. When is it really not possible? • The treatment already assigned and announced and no possibility for expansion of treatment • The program is over (retrospective) • Universal take up already • Program is national and non excludable • Freedom of the press, exchange rate policy (sometimes some components can be randomized) • Sample size is too small to make it worth it

  30. Thank You

More Related