APPAM Institutional Member Forum - Leverage Points for Stronger, More Informative Impact Evaluations of Social Programs
August 18, 2020
Virtually, through Zoom
Researchers conducting rigorous evaluations of social programs continue to push the methodological frontier—by trying to get “inside the black box” to discover what drives impacts, by working to generalize results from sample to nation, and by seeking to validate quasi-experimental approaches using experimental benchmarks. As these investigations proceed, some fundamentals of strong social experiments are under-attended.This forum sponsored by Westat and APPAM will address three of these, presenting ideas of universal relevance to random assignment evaluations illustrated through employment assistance studies.
One fundamental move forward is to test only policies/programs with strong potential for favorable impacts. The forum’s opening session presents three strategies for following this path—strategies for obtaining qualitative and quantitative assessments of an employment intervention’s potential for generating earnings gains before implementing a large-scale experimental test. Each of these strategies—which can be applied beyond workforce policy—is underutilized by evaluators and evaluation planners.
Another fundamental need in social experiments is to observe the right policy contrast between the intervention scenario and the counterfactual world (the treatment and control group contrast in a randomized experiment). The right contrast depends on a study’s motivating purpose: What policy or funding decision will be made using the impact estimate? What long-term conclusion on an employment strategy’s overall effectiveness will be drawn from the evidence? The forum’s second session frames these issues and presents new methods for analyzing and interpreting data from workforce assistance experiments that get the right answers to the right intervention/counterfactual contrasts.
None of these strategies will work unless well-chosen and well-constructed social experiments are put into the field effectively. The final forum session, a roundtable, invites all attendees to share the latest best practices for assuring site readiness and intervention fidelity when conducting employment-focused random assignment evaluations. It is experience from the field and practitioners’ synergies that will push progress in these aspects of social experimentation.
Finally, the forum concludes with reflections on the messages of the day, addressing the following questions: Has the forum provided good advice for guiding future impact evaluations—in employment policy and in general? What are the next topics that should be addressed to strengthen implementation and scientific tools for more informative social policy impact evaluations?
Registration will open mid July.
Schedule at a Glance*
|8:00 AM – 9:00 AM
||Registration & coffee
|9:00 AM – 9:20 AM
|9:20 AM – 10:35 AM
Assessing an Intervention’s Potential Benefits Prior to Launch of a Full-Scale Experiment
Framer: Laurie May
- Annalisa Mastri, Mathematica
- Laurie May, Westat
- John Martinez, MDRC
|10:35 AM – 10:45 AM
|10:45 AM – 12:00 PM
Defining and Implementing the Most Policy-Relevant Counterfactuals
Framer: Laura Peck
- Laura Peck, Abt Associates
- Dana Rotz, Mathematica
- Stephen Bell, Westat
|12:00 PM – 12:20 PM
|12:20 PM – 1:30 PM
Assuring Site Readiness and Intervention Fidelity
Framer: Patty Trope
|1:35 PM – 2:00 PM
Reflection, questions, comments
- Naomi Goldstein, OPRE/ACF/DHHS
*subject to change due to the new, virtual, format.