ArticleTargetedEvaluationsCanHelpPolicymakersSetPriorities.pdf

1 pewtrusts.org/resultsfirst

Targeted Evaluations Can Help
Policymakers Set Priorities
A policymaker’s guide to building evaluation capacity

An issue brief from the Pew-MacArthur Results First Initiative March 2018

Overview
State and local leaders allocate millions of dollars each year to programs designed to deliver services to their
constituents, but often the results go unmeasured. Without data on the effect, if any, that programs are having on
participants and communities, policymakers are unable to discern which ones work, which don’t, and how best to
distribute limited public resources.

Governments have a range of evaluation tools to gauge the effectiveness and efficiency of public programs. One type
they can employ is called an impact evaluation, which is a targeted study of how a particular program or intervention
affects specific outcomes. These evaluations allow governments to isolate the effect of a program or initiative on a
group, similar to the way that medical researchers test the effectiveness of a treatment or drug on a population. When
conducted correctly, impact evaluations can help policymakers decide when to scale up programs that work, when to
scale back or eliminate those that don’t work, how to reallocate resources to better use, and when to improve those that
show potential.

But states and counties face real challenges in conducting impact evaluations, and typically don’t have the resources,
time, or ability to evaluate every program. Recent innovations in techniques, increased access to data, and expanded
partnerships between governments and researchers have made it easier for governments to rigorously assess the
effectiveness of their programs, and they can remain judicious with their limited capacity and target evaluations carefully.

Program
Assessment

Budget
Development

Implementation
Oversight

Outcome
Monitoring

Targeted
Evaluation

This brief is one in a series about the five key components of evidence-based policymaking as identified in “Evidence-Based
Policymaking: A Guide for Effective Government,” a 2014 report by the Pew-MacArthur Results First Initiative. The other components
are program assessment, budget development, implementation oversight, and outcome monitoring.

http://www.pewtrusts.org/en/projects/pew-macarthur-results-first-initiative

2 pewtrusts.org/resultsfirst

This brief provides options for government leaders to build their capacity to conduct these targeted studies by:

1. Hiring or training staff to facilitate impact evaluations.

2. Building partnerships with external research entities to leverage expertise.

3. Making better use of existing administrative data systems to reduce costs.

4. Aligning data policies and funding to support impact evaluation.

The brief also includes examples of how state and local governments have used impact evaluations to assess
their programmatic investments, and details when such an assessment is worthwhile.

What Is an Impact Evaluation and How Is It Different From Outcome Monitoring?

Impact evaluations use a randomized controlled trial (RCT) design or quasi-experimental
design (QED) to rigorously assess effectiveness. Both RCTs and QEDs use treatment (program
participants) and comparison (nonparticipants) groups to evaluate outcomes against what
would have occurred without the program. When done well, these evaluations provide
policymakers with evidence of a program’s effectiveness, helping inform their programmatic,
policy, and funding decisions.

Impact evaluations provide different information and answer different questions than outcome
monitoring systems that many state and local governments operate. Those systems track
performance data that can be used to illustrate trends in the program’s outcomes and compare
that performance to prior years or other benchmarks, but this information cannot show what is
driving those results.

For example, consider a state that implements a summer program to help students who are
reading below grade level. Performance data show that participants scored significantly higher
on English language arts exams their senior year of high school than students who were similarly
struggling to read at grade level but did not participate in the program. Is the summer reading
program the reason the test scores went up? More importantly for state policymakers, is this
a program the state should expand in to improve student learning? Outcome monitoring
can’t answer those questions. But impact evaluations can isolate the effects of the program and
control for other factors that could influence student test scores, thus revealing more details
about possible linkages.

http://www.pewtrusts.org/en/projects/pew-macarthur-results-first-initiative

3 pewtrusts.org/resultsfirst

How impact evaluation can support more effective
government programs
Government leaders have a lot to gain from impact evaluations. They can use information from these studies,
alongside performance data, to decide when to:

Scale up what works. Policymakers in Chicago relied on evidence from impact evaluations when deciding
whether to scale up an intervention aimed at reducing violent crime. They turned to the University of Chicago
Urban Labs,1 which helps governments and nonprofit organizations test programs across public safety,
education, health, poverty, energy, and environment areas, to conduct multiple impact evaluations of the city’s
Becoming a Man2 program, which targets at-risk youth. Results from a 2016 impact evaluation showed that
people who participated in the program had up to 50 percent fewer arrests for violent crime and increased
their on-time graduation rates by up to 19 percent compared with a similar group of individuals who did not
participate.3 Because of the program’s success, the city expanded the program later that year to serve an
additional 1,300 youths.4

Improve programs that show promise. The New York City Mayor’s Office for Economic Opportunity (NYC
Opportunity) regularly conducts impact evaluations to test local anti-poverty programs and uses the findings
to make needed improvements. For instance, the office’s Justice Corps5 initiative has undergone multiple
modifications since its inception in 2008, driven by findings from internal performance monitoring and
evaluations, including an RCT study.6 The findings showed positive impacts on participants’ employment rates
and wages but no effect on educational attainment or recidivism. This led to several programmatic adjustments,
such as an expansion of youth development training for staff and implementation of a new risk-needs assessment
and case management toolkit, which further evaluation showed improved outcomes.7

Scale back or replace programs that don’t work. The Iowa Department of Corrections decided to scale back
a community-based domestic violence intervention after impact evaluations in other states did not show the
desired results. The department redirected resources to an alternative intervention8 and contracted with a local
university to test its effect on reducing the number of persons who reoffend. The impact evaluation demonstrated
positive effects, including a significantly lower recidivism rate than another domestic violence program operating
in the state.9 This evaluation showed Iowa policymakers that the funds redirected from the old program to the
new one ($4.85 million dedicated since fiscal year 2011) were not wasted.10

http://www.pewtrusts.org/en/projects/pew-macarthur-results-first-initiative

4 pewtrusts.org/resultsfirst

Is an Impact Evaluation Always the Right Tool?

Before state and local governments spend limited funds and staff time on an impact study, it is
important that they understand the questions being asked—and that they match the questions
with the appropriate type of evaluation.

The table below shows four types of evaluations commonly used to answer policymakers’
questions, the basic questions each study best addresses, and how each can help policymakers
make informed choices. For instance, an outcome study could show policymakers whether
participants in a college access program reported increased understanding of college readiness
resources available to them after completing the program. An impact study, on the other hand,
would show policymakers whether the program, as opposed to other factors, contributed to the
increased understanding of college readiness resources.

Sources: Centers for Disease Control and Prevention, “Types of Evaluation,” accessed July 28, 2017, https://www.
cdc.gov/std/Program/pupestd/Types%20of%20Evaluation.pdf; The University of Minnesota; “Different Types of
Evaluation,” accessed July 28, 2017, https://cyfar.org/different-types-evaluation#Formative; Permanency Innovations
Initiative, “Family Search and Engagement (FSE) Program Manual” (2016), https://www.acf.hhs.gov/sites/default/files/
cb/pii_fse_program_manual.pdf; Kathryn E. Newcomer, Harry P. Hatry and Joseph S. Wholey, eds., Handbook of Practical
Program Evaluation, 4th ed., (New Jersey: John Wiley & Sons, Inc., 2015), 27–29; Michael Q. Patton, Utilization-Focused
Evaluation, 4th ed., (Thousand Oaks, CA: Sage Publications, 2008)

Evaluation type What the evaluation can answer What the evaluation does for policymakers

Process

• Has the program been implemented as intended?

• What types of implementation issues have
emerged and how can they be addressed?

• What new ideas are emerging that can be tried out
and tested?

• Demonstrates whether
programs are being delivered
with fidelity to the program
model

Outcome

• Did participants report desired changes after
completing the program?

• For whom did the program work best?

• Shows whether desired
intermediate outcomes
are being achieved

Impact

• Is the program effective at achieving desired
outcomes?

• How did program participants do in achieving
desired outcomes compared with nonparticipants?

• Provides rigorous evidence
on a program’s impact

Cost-benefit Analysis

• Do the program’s benefits outweigh its costs?

• Is the program more cost-beneficial than other
programs addressing the same problem?

• Compares the expected return
on investment of multiple
programs that address the
same problem

http://www.pewtrusts.org/en/projects/pew-macarthur-results-first-initiative

https://www.cdc.gov/std/Program/pupestd/Types%20of%20Evaluation.pdf

https://www.cdc.gov/std/Program/pupestd/Types%20of%20Evaluation.pdf

https://www.acf.hhs.gov/sites/default/files/cb/pii_fse_program_manual.pdf

https://www.acf.hhs.gov/sites/default/files/cb/pii_fse_program_manual.pdf

5 pewtrusts.org/resultsfirst

How to build impact evaluation capacity
To leverage the benefits of impact evaluations, governments need to build their capacity—expertise, data, and
funding—to conduct them effectively. Building this capacity can be challenging, even to officials who understand
the importance of these studies. But government officials can choose an approach that best fits their available
resources. These include: developing internal staff; building partnerships with external research entities; making
better use of administrative data systems that enable researchers to use existing data to conduct impact
evaluations; and aligning data policies and funding to support evaluations.

Hire or train staff to facilitate impact evaluations
In to effectively and more frequently complete impact evaluations, governments can either hire research
staff with the requisite skills to conduct these studies, or hire new or train existing staff to manage impact
evaluation contracts. Governments can develop staff capacity in individual agencies and offices, as well as in
centralized offices such as the office of a mayor, county commissioner, governor, or legislature.

Some states have hired staff with the requisite skills to conduct impact studies. For example, the Washington
Department of Social and Health Services’ Research and Data Analysis (RDA) office has about 70 full-time
employees who perform a range of analytical and data management tasks, including impact evaluations. About
70 percent of these staff are funded through time-limited grants and projects, and 30 percent are supported by
legislative appropriation. RDA has evaluated innovative pilot programs as well as long-standing projects that had
not previously been studied. These evaluations have significantly affected policy and program decisions in the
state, such as a 2009 study of a chronic care management practice that led to a scaling up of two of the state’s
health projects.11

While Results First researchers found few offices in the 50 states, other than RDA, that regularly conduct
impact evaluations, they did identify several offices that conduct other types of evaluations. For example, some
legislative audit and research offices, as well as some agency research divisions, perform outcome evaluations
that aim to measure the results of state programs or policies but are unable to control for external factors likely to
influence those outcomes. Policymakers could work with these offices to identify opportunities to conduct impact
evaluations, particularly on high dollar programs, or programs being considered for expansion.

Even where governments contract out impact studies to universities or external evaluation firms—a viable option
for jurisdictions with limited resources—Results First researchers found that maintaining a small level of staff
knowledgeable in evaluation has substantial benefits. Such staff can, for instance, help select external evaluators,
manage evaluation contracts, collaborate on choosing study designs, and assist with data access. These staff
may also have a deeper understanding of the data and issues relevant to a program than external evaluators do,
and can help facilitate communication and knowledge transfer between external evaluators and program staff.12
Michael Martinez-Schiferl, research and evaluation supervisor at the Colorado Department of Human Services,
noted, “Program knowledge is very important to the research aspect. Having internal research staff embedded
with program staff promotes collaboration on research and provides opportunities for research staff to develop
some program expertise, as there are so many nuances about the program that they couldn’t understand from an
outside perspective.”13

NYC Opportunity contracts out its evaluations while maintaining staff to oversee the work. The staff members
manage the contracts of the independent evaluation firms, oversee impact studies of the anti-poverty programs
and key mayoral initiatives, and work in partnership with city agencies to use the evaluation results to inform
decisions to expand, improve, or discontinue programs.14 NYC Opportunity has a dedicated fund from the city

http://www.pewtrusts.org/en/projects/pew-macarthur-results-first-initiative

6 pewtrusts.org/resultsfirst

to support its staff and external evaluations, but also seeks funding from federal grants and philanthropy to
supplement this work.15

In the past 10 years, the organization has launched more than 70 programs, most of which have undergone an
evaluation and some of which have become a national model for success.16 One example is the City University of
New York’s Accelerated Study in Associate Programs, which provides extensive financial, academic, and personal
support to working adults who are pursuing an associate’s degree. The program’s first impact evaluation, done in
partnership with a research organization, showed promising early results on academic outcomes, including lower
drop-out rates and higher total credits accumulated, and a subsequent study demonstrated increased graduation
rates.17 Following these successful evaluations, the program is expanding from 4,000 students in fiscal 2014 to
25,000 students in fiscal 2019, and has been replicated in other parts of the country.18

Build partnerships with external research entities to leverage expertise
Governments can help fill evaluation capacity gaps by creating long-term partnerships with external research
entities, such as local universities, to perform an evaluation in its entirety or provide technical assistance or training.

For example, Brian Clapier, former associate commissioner at the City of New York’s Administration for Children’s
Services (ACS), attributed some of his office’s most successful evaluations to partnerships with universities.19
“Based on my experience the research-to-policy gap is a real challenge. One key strategy is to bring in the
research partners (often from universities), and have these partners conduct the evaluations. Once the evaluation
has occurred, strategically placed public agency staff can bridge the findings of the evaluation to program staff
responsible for the policy development.”

This approach proved successful when, in 2012, ACS contracted with Chapin Hall at the University of Chicago
to implement and test a pilot child welfare program to promote healthy families and child well-being. The study
found that two of the program’s interventions—KEEP (Keeping Foster and Kin Parents Supported and Trained)
and Parenting Through Change—were 11 percent more likely than a comparison group to achieve permanency.20
Based on the results, ACS continued funding both programs.

Results First researchers identified several government-university partnerships—some that perform impact
evaluations for programs spanning a range of policy areas and others that evaluate the same program over a
period of time. For instance, state and county agencies in Wisconsin frequently partner with the University of
Wisconsin Population Health Institute to perform evaluations on a range of human services programs, including
behavioral health, child welfare, juvenile justice, and health.21 The New York State Office of Children and Family
Services, on the other hand, has partnered with the Center for Human Services Research at the University of
Albany since 1995 to perform multiple evaluations—including impact studies—of one child welfare program,
Healthy Families New York.22

Several jurisdictions have joined forces with policy labs—typically housed in universities—to help them design
and test the effectiveness of government programs. For example, the Colorado Evaluation and Action Lab is a
new government-research partnership that will help state officials to evaluate public programs and policies.23

Government leaders also partner with individual researchers, rather than a research organization, whose interests
and skills align with a particular policy or program they want to evaluate. South Carolina’s Department of Health
and Human Services, with the help of the Abdul Latif Jameel Poverty Action Lab (J-PAL), a research center at the
Massachusetts Institute of Technology, paired up with health-focused researchers at Northwestern University to
develop a randomized evaluation of assignment to Medicaid managed care plans.24

http://www.pewtrusts.org/en/projects/pew-macarthur-results-first-initiative

7 pewtrusts.org/resultsfirst

Policymakers should pair these research partners with trained government staff in the offices that operate (or
oversee providers who operate) the programs being evaluated.

Make better use of existing administrative data systems to reduce impact
evaluation costs
Most governments maintain reporting systems that collect administrative data, such as criminal arrest or
education records, which could be used to help conduct impact evaluations at a lower cost than collecting
this information from scratch.25 For example, New Mexico’s Department of Corrections conducted a quasi-
experimental design (QED) study of a substance use dis program using administrative data from three
state correctional offices. With a small evaluation budget, the department was able to successfully answer
policymakers’ questions about whether a program affected recidivism or substance use dis relapse rates.26

Some states, such as Washington and South Carolina, have built sophisticated data warehouses that link data across
multiple agencies and can be used for performing evaluations. For example, the South Carolina Revenue and Fiscal
Affairs Office (RFA) operates an integrated data warehouse that receives copies of agency databases used for program
administration and research.27 While the originating agencies maintain control over the use of the data, the RFA
provides guidance on sharing and usage agreements to help researchers access the information to evaluate programs.28
A new Pay for Success project focused on expanding an evidence-based family support and coaching program in the
state will use data from the warehouse to evaluate the program’s impact and calculate its return on investment.29

Washington state’s RDA also uses administrative data for most of its impact evaluations, which enables
the agency to conduct more frequent evaluations on a wide range of programs. “We can knock out a quasi-
experimental evaluation, assuming there’s no new data to collect, in a relatively short time and at a fraction of
the cost of doing an external evaluation,” said health economist David Mancuso.30

Both federal and private entities are creating funding opportunities to support state and local governments in
using administrative data to support low-cost RCTs. (See Appendix A for a list of funding sources.)

Align data policies and funding to support evaluation
Two key obstacles to conducting impact evaluations are accessing the data necessary for the study and finding
the resources to fund it. Yet state and county officials have found creative ways to mitigate these seemingly
formidable challenges.

To generate new evidence on what works, researchers need access to government data, but service providers and
government agencies may be hesitant to share data due to privacy issues or concerns over how the data might be used.
Government leaders can alleviate these sensitivities and facilitate information access by developing sharing and usage
agreements that outline the purpose of the data sharing, how it will be shared with the public, and privacy protections.31

For example, the Colorado Department of Education established a data-sharing and usage agreement with
Mathematica Policy Research to allow Mathematica to access the department’s administrative data to conduct
an impact evaluation of a new charter school program’s effects on education achievement.32 The agreement
outlines the types of data to be shared, as well as the responsibility of the requestor to use the data only for the
purposes outlined in the agreement (in other words, the impact study), to secure and later destroy the shared
data, and to share analyses with the department prior to publication. Because of the shared data, Mathematica
could perform an impact evaluation that showed state officials that taxpayer investments in the program had
positive impacts on reading and math skills at the elementary, middle school, and high school grade levels.33

http://www.pewtrusts.org/en/projects/pew-macarthur-results-first-initiative

8 pewtrusts.org/resultsfirst

States and localities can finance impact evaluations by setting aside internal funding for the studies, allowing
governments to select programs in most need of an evaluation rather than being subject to the priorities of
external funders. Results First researchers identified several ways governments are setting aside funding for
rigorous evaluations, including through legislation, grants, and budget allocations.

For example, California passed a law in 2014 that sets aside funds to award contracts to recipients who agree to
partner with an independent evaluator to assess the effectiveness of programs funded through the contracts.34
Three counties received $1.25 million to $2 million in 2016 to implement and evaluate selected social services
programs.35 Though the law does not require recipients to conduct an impact evaluation, Alameda County
is performing an RCT of a life coaching and mentoring services program aimed at reducing recidivism and
increasing employment.36

Some state and local governments dedicate funds to support staff who oversee or conduct impact evaluations.
Washington state’s RDA receives approximately 30 percent of its funding from a legislative appropriation that
includes support for research staff who manage evaluations.37 RDA supplements this funding with federal grants,
including Medicaid and a Substance Abuse and Mental Health Block Grant.

Even when state and local governments build impact evaluation staff or set aside funds to support these studies,
additional federal and private funding can help fill remaining capacity and funding gaps.

The federal government has provided several competitive and formula grant opportunities. For example,
the Institute of Education Sciences released a request for applications in 2017 for low-cost RCTs or QEDs of
education interventions.38 The U.S. Department of Health and Human Services has also provided grants that
included funding to evaluate child welfare and teen pregnancy interventions.39 While these opportunities provide
substantial support for impact evaluations, they should not be a substitute for using existing government
resources to support this work; many are one-time grants that limit support to one study, and some target
programs or policies that might not be an area of need in a particular jurisdiction.

Even jurisdictions that have never completed an impact evaluation have opportunities to start building this
capacity through external sources. For instance, in 2016 J-PAL launched the State and Local Innovation
Initiative to help jurisdictions perform randomized studies of social programs,40 with eight jurisdictions chosen
to participate in the first two rounds.41 In addition to funding, each will receive technical support and custom
trainings to expand the internal capacity to create, use, and share rigorous evidence. (See Appendix A for more
information on funding opportunities.)

How to select programs for an impact evaluation
While state and local governments have demonstrated the value of impact studies to assess programmatic
investments, it is not practical (or even necessary) to rigorously evaluate every program. Decision-makers can
identify and prioritize which programs to study by considering four key questions:

1. Does the program have an evidence base? To identify programs that could benefit from an impact evaluation,
governments can inventory the programs they fund and determine which ones have evidence supporting their
effectiveness. National research clearinghouses—which review and aggregate impact evaluations in to rate
programs by their level of evidence of effectiveness—can help determine if local programs have existing evidence.
Governments can prioritize evaluations for programs that do not have strong evidence of their effectiveness.

2. Has the program been properly implemented? Poorly implemented programs are less likely to achieve the
outcomes that leaders and residents expect, which would impair the results of an impact assessment.

http://www.pewtrusts.org/en/projects/pew-macarthur-results-first-initiative

9 pewtrusts.org/resultsfirst

3. Are the right data available for an impact study? To conduct an impact study, evaluators need access to
the right kinds of data. If the data are owned by other parties (e.g., another agency or program provider) or
do not exist, governments should consider the feasibility of getting the data, which could entail developing
data-sharing agreements or spending additional funds and time to collect new data.

4. Does the program serve a significant number of people and/or is it a large budget item? Programs that have
a higher number of clients and/or are costlier typically have a larger impact on a government’s budget than
those that are less prescribed or less costly, and may be more attractive options for an impact evaluation.

Decision-makers may find that some of their untested programs are not good candidates for an impact evaluation.
In that case, they can take other steps to ensure these programs are generating positive results, such as tracking
outcomes of participants and reviewing implementation to identify any issues with operation and delivery. Decision-
makers can review these programs again at a later time to determine if they have become evaluation-ready.

Conclusion
Policymakers care about funding what works, and impact evaluations are an important tool that can be used to
inform better, data-driven decisions. Impact evaluations provide critical information on program effectiveness,
which policymakers can consider when making decisions about when to scale up, scale back, or make
adjustments to a particular program or initiative.

By building their jurisdiction’s capacity to evaluate untested programs, policymakers can hold themselves
accountable to the public, and ensure that the state’s public dollars are directed to those programs that yield
positive results. While challenges still …

Place your order
(550 words)

Approximate price: $22

Calculate the price of your order

550 words
We'll send you the first draft for approval by September 11, 2018 at 10:52 AM
Total price:
$26
The price is based on these factors:
Academic level
Number of pages
Urgency
Basic features
  • Free title page and bibliography
  • Unlimited revisions
  • Plagiarism-free guarantee
  • Money-back guarantee
  • 24/7 support
On-demand options
  • Writer’s samples
  • Part-by-part delivery
  • Overnight delivery
  • Copies of used sources
  • Expert Proofreading
Paper format
  • 275 words per page
  • 12 pt Arial/Times New Roman
  • Double line spacing
  • Any citation style (APA, MLA, Chicago/Turabian, Harvard)

Our guarantees

Delivering a high-quality product at a reasonable price is not enough anymore.
That’s why we have developed 5 beneficial guarantees that will make your experience with our service enjoyable, easy, and safe.

Money-back guarantee

You have to be 100% sure of the quality of your product to give a money-back guarantee. This describes us perfectly. Make sure that this guarantee is totally transparent.

Read more

Zero-plagiarism guarantee

Each paper is composed from scratch, according to your instructions. It is then checked by our plagiarism-detection software. There is no gap where plagiarism could squeeze in.

Read more

Free-revision policy

Thanks to our free revisions, there is no way for you to be unsatisfied. We will work on your paper until you are completely happy with the result.

Read more

Privacy policy

Your email is safe, as we store it according to international data protection rules. Your bank details are secure, as we use only reliable payment systems.

Read more

Fair-cooperation guarantee

By sending us your money, you buy the service we provide. Check out our terms and conditions if you prefer business talks to be laid out in official language.

Read more

Order your essay today and save 30% with the discount code HAPPY