Clean Air Act of 1970

1
MEANS-TESTED EMPLOYMENT AND
TRAINING PROGRAMS
Burt S. Barnow
George Washington University
Jeffrey Smith
University of Michigan
For Presentation At The Organization For Economic Cooperation
And Development, Paris, France
February 2, 2016
Topics
2





Types of programs covered
Rationales for government support for E&T programs
History of US E&T programs from the Great Depression to present
Funding patterns over time and current programs
Evaluation issues






Use of RCTs
Major alternatives to RCTs
Other issues: data, general equilibrium, follow-up
Findings from evaluations of WIA, Job Corps, other programs
Program operation issues
Conclusions
Types of Activities Covered
3

Skill development increase vocational skills through classroom or on-the-job training;

Job development programs consist of public employment programs where jobs are
specifically created for the participants;

Employability development programs improve personal attitudes and attributes
needed for employment (soft skills)

Work experience programs provide employment experiences intended to help workers
gain the same attitudes and attributes as employability development programs.

Labor exchange programs that help match job seekers with job openings;

Counseling and assessment and labor market information (LMI) help workers
learn more about their abilities and aptitudes, and provide information about the current and
future labor market.
Note: Although it is sometimes assumed that E&T means training, Barnow and Trutko (2007)
found < 50% of WIA exiters received training
Source: Butler and Hobbie (1976) augmented
Focus on “Means Tested” Programs
4

Many US programs not means tested
 Unemployment
insurance
 Vocational education
 Employment service (labor exchange services)
 Registered apprenticeship programs
 WIA/WIOA not means tested for dislocated workers and
only for adults if insufficient funds in local areas
 Also exclude place-based programs, in-school youth
programs, state/local funded programs
Why Have Government Support?
5






E&T programs not public goods or natural monopoly, so
could rely on private sector
Musgrave “merit goods” argument
Imperfect access to capital for poor
Compensation for government actions or unforeseen
events (displaced workers)
Imperfect information
Some rationales call for means testing, but others do not
History of E&T Programs:
Programs Established in the Great Depression
6



Under Hoover, Reconstruction Finance Administration spent $300M
on work relief, employing up to 2M people
Many work relief programs under FDR—largest was Civil Works
Administration with 4.3M workers
In 1933 Wagner-Peyser Act established the Employment Service
 Not means tested, so not covered in depth here
 ES often used to enforce ALMP provisions, e.g., UI work test
 Real budget has been reduced for years for ES, leading to emphasis
today on self-service and staff-assisted labor exchange
Programs in 1960s
7


New Deal programs (except ES) stopped in 1943—just ES
lived on
Training programs emerged in 1960s
Area Redevelopment Act (1961) was 1st program
 Manpower Development and Training Act (1962)

Originally passed to deal with “automation” (technical change)
 No automation, so focus on the poor with classroom training and OJT
(2/3 CT and 1/3 OJT)
 Served 1.9M participants 1963-1972
 Note: Ashenfelter did pioneering evaluations of training for MDTA

Programs from 1973-1998
8

Comprehensive Employment and Training Act (CETA)





Ran from 1973-1983 and established system of local agencies running
programs
Included public service employment, which grew to be the largest
component of CETA
Concern about “fiscal substitution” in PSE programs led to restrictions
on people and work, making program ultimately unpopular with all
parties (although estimated to have large impact on earnings)
Concern about “creaming” was large, leading to special programs for
Native Americans & farmworkers
Non-experimental evaluations of CETA, all using the same data, had
huge range of estimates, setting the stage for RCT evaluations (see
Barnow 1987)
Programs from 1973-1998
9


Large youth initiative in 1977 proved poor youth do want
to work, but little else
Two changes to CETA have endured:
 Private
Sector Initiative Program discovered employers
 Government economists in DOL developed first
performance measures and adjustment models

CETA replaced by Job Training Partnership Act in 1982
Job Training Partnership Act 1982-1998
Retained Basic CETA Structure
10








Programs for economically disadvantaged youth and adults continued to be locally
administered;
States assumed a much greater role in monitoring performance of local programs;
Private sector was given the opportunity to play a major role in guiding and/or operating the
local programs;
System was to be performance driven, with local programs held accountable and rewarded or
sanctioned based on their performance;
Program added for dislocated workers
Amendments in 1992 restricted who could be served and how served—65% of participants
had to be “hard to serve”
First major DOL program evaluated with RCT showed modest impacts for men and women,
but no impacts for out-of-school youth
No evaluation of programs for dislocated workers or in-school youth
Workforce Investment Act of 1998
11






Continued devolution of authority to states
Called for services through One-Stop Career Centers (now called American
Job Centers)
One-Stops were to have universal access—idea was to avoid stigma
Over a dozen mandatory partners in One-Stops who were required to pay
for infrastructure (in theory)
To avoid rushing people into training who did not need it (and to save $),
participants (customers now) were to go through sequence of services:
core, intensive, and then training
Training was reserved for the poor if not enough funds available for all
Workforce Investment Act of 1998
12





Customers were to have choice of training programs through individual
training accounts, which were like vouchers
To help assure high-quality training vendors, states were to establish an
eligible training provider list by program, with standards for getting on list and
staying on list
Performance measurement structure similar to JTPA, but DOL dropped
statistical adjustments in favor of “negotiated” standards
Summer youth employment program abolished and year-round youth
employment program established
Youth programs were now required to spend at least 30% of funds on outof-school youth
Changes in Workforce Innovation and
Opportunity Act (WIOA) of 2014
13





Data on training providers’ outcomes must be made available
Allows states to transfer unlimited amounts of their grant between
the adult and dislocated worker programs
Adds “basic skills deficient” as a priority category, along with low
income, for Adult services
Requires that 75 percent of Youth funds be used for out-of-school
youth, a large increase over the 30 percent required under WIA
Combines core and intensive service categories into “career services”
and abolishes requirement that customers pass through core and
intensive services before receiving training
Changes in Workforce Innovation and
Opportunity Act (WIOA) of 2014
14




Permits direct contracts with higher education institutions
(class-size contracts rather than just ITAs)
Strengthens the requirements for partners in American Job
Centers: ES required to be in AJCs, and TANF a mandatory
partner
Reduces required employer contributions for customized and
sectoral training programs
Includes specific performance measures for WIOA and other
E&T programs, with employer satisfaction and longer followup than WIA
15
Employment and Training
Expenditure Patterns Over Time




Funding has generally declined in real terms since the 1980s
The share of GDP devoted to E&T programs (except ES) has shrunk from
.094% in 1985 to .048% in 2012
The Recovery Act greatly increased activity temporarily during Great
Recession, but funding ended while unemployment still high
Funding affected in part by evaluations
 Youth funding greatly reduced after National JTPA Study showed youth
programs ineffective
 Job Corps funding increased when initial results showed program
effective (but not reduced when results not sustained)
 Dislocated worker funding has increased over time despite lack of
evidence on program effectiveness
1965
1966
1967
1968
1969
1970
1971
1972
1973
1974
1975
1976
1977
1978
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
2012
2012 Dollars (thousands)
Funding for DOL Employment & Training
Programs, 1965-2012
16
60000,000
50000,000
40000,000
30000,000
20000,000
10000,000
,0
Year
Funding as Percentage of GDP
DOL Employment & Training Programs, 1965-2012
17
0.90%
0.80%
0.70%
Percentage of GDP
0.60%
0.50%
0.40%
0.30%
0.20%
0.10%
0.00%
1965
1970
1975
1980
1985
1990
Year
1995
2000
2005
2010
Current DOL Employment
and Training Programs
18




Currently 14 DOL programs with at least $30M annual
funding
Two largest programs are Job Corps, residential program
for poor youth, and WIA Dislocated Worker program
Programs mostly targeted by economic status, age,
reason for lack of employment
Wagner-Peyser Employment Service is major
exception—open to all
Current Funding for Major DOL Employment and
Training Programs
19
Job Corps
DOL / Employment Training
Administration
$1,684
WIA Dislocated Workers
DOL / Employment Training
Administration
$1,219b
WIA Youth Activities
DOL / Employment Training
Administration
$818
WIA Adult Program
DOL / Employment Training
Administration
$764
Wagner-Peyser Funded Employment DOL / Employment Training
Service
Administration
$664*
Senior Community Service
Employment Program
DOL / Employment Training
Administration
$433
DOL / Employment Training
Administration
$306c
Trade Adjustment Assistance (TAA)
Disabled Veterans Outreach Program
(DVOP) and Local Veterans’
Employment Representative Program DOL / Veterans' Employment and
(LVER)
Training Service
H-1B Job Training Grants
DOL / Employment Training
Administration
$175
$166**
Major Programs Outside DOL
20

Pell Grants support higher education for low-income
students
 Total
support for Pell Grants $33.7B
 Support for E&T from Pell estimated to be $8.7B, more than
funding from all 3 WIA funds + ES

Other non-DOL programs also large
 TANF
welfare program spends $1.5B on E&T
 Adult education spends $564M
 SNAP (Food Stamps) E&T has budget of $416M
Current Funding for Employment and Training
Programs Outside DOL
21
Pell Grants
Ed / Office of Vocational and
Adult Education
Temporary Assistance for Needy HHS / Administration for
Families (TANF) Grants
Children & Families
$8,181
$1,517d
Adult Education - Grants to
States
Ed / Office of Vocational and
Adult Education
$564e
SNAP Employment & Training
USDA / Food and Nutrition
Service
$416f
22
Are There Too Many Employment and
Training Programs?





In 1994 GAO claimed 154 E&T programs, but many
were not programs (e.g., incentive payments)
In 2011, GAO counted 47 and we count 20 with at least
$30M
Many programs are pilots or have special target groups
Biggest issues are ES/WIOA and TANF/WIOA
Duplication has some advantages, but overall hard to
argue there are not too many
Evaluation Issues for E&T Programs
23
Basic equation is
Yi  DY
i i1  (1  Di )Y0i
where:
Yi is the outcome of interest
Di is treatment status
Y01 is the outcome without treatment and
Y11 is the outcome with treatment
 Problem is we do not observe outcome with and without
the treatment for same person

What Do We Want to Estimate?
24



Most commonly we want average treatment effect on the
treated:
E(Y1  Y0 | D  1)
Sometimes we want average treatment effect for entire
population, which could differ if impact varies by
selection:
E (Y1  Y0 )
Sometimes interested in quantile treatment effects, e.g.,
impact at median or other point
25
What Do We Assume on Treatment
Effect?


Key issue is whether to assume common treatment effect
Older literature assumed impact identical for all, but theory and evidence
suggest otherwise





Many programs involve selection decision, which might depend on impact
Treatment received not identical, so impacts likely to vary
Issue of whether impacts vary by economic conditions or characteristics
ultimately an empirical one—why assume it away?
At minimum, most studies look at impacts by sex, often by race/ethnicity
Paper reviews Heckman Robb (1985) and Heckman, LaLonde, and Smith
(1999) models to show importance of selection in estimating impact, e.g.,
ATET>ATE>ATNT
Random Assignment not a Cure-All
26




Heckman and Smith (1995) discuss randomization bias as
potential issue
Randomization while keeping enrollment constant requires
increasing number of applicants to program
If selection depends on expected impact and impact varies by
number selected, random assignment will give impact for
wrong program size
Ideally, sites should also be randomly selected—sometimes
this works (Job Corps, WIA), but not always (JTPA)
Random Assignment not a Cure-All
27


Another problem is that randomization can lead to substitution bias,
where control group receives similar treatments—in JTPA
evaluation large share of control group received training
Not all in experiment comply with assignment




No-show rate can be high: estimate effect on treatment on treated, but is
this what we want?
Can use IV (Bloom1984) HST (1998) if impact on no shows = 0
Crossovers from C to T bigger problem but can be dealt with (Orr
1998)
See Greenberg and Barnow (2014) and Barnow (2011) for
examples of things that can go wrong
Non-Experimental Approaches
28

Include rich mix of covariates and assume that selection is
based on observable variables


Propensity score matching


Widely used, with mixed results when tested with RCT data
Regression discontinuity designs


Widely used but often with no proof valid assumptions
Rare in E&T context
Difference-in difference models

Often combined with other approaches, good control on nonvarying unobservables
Data and Measurement Issues
29



Data on service receipt often not measured well,
especially for control group
Administrative data and survey data have different
strengths and weaknesses, and they can lead to
contrasting impact findings (Barnow and Greenberg
2015)
Length of follow-up is very important for CBA; see Job
Corps analysis (Schochet et al. 2006) and JTPA longterm follow-up (GAO 1996)
Some Issues for Cost-Benefit Analysis
30




When performed, usually compare average costs and
benefits—more work on marginal BCA needed
Assumptions after observed follow-up are key—Job
Corps is good example
Limited analyses of outcomes other than earnings, e.g.,
crime, fertility, health
Valuing “leisure” time of participants (to themselves and
society) difficult and rarely done
General Equilibrium Effects
31





Displacement, where participants take jobs that would have
gone to control group members, can make social gains less
If T group enters different labor market, wages to C group
could increase, making social gains more
Large programs could change relative prices
Scale effects can be captured if scale varied across labor
markets; see Crepon et al. (2013)
Possible to estimate general equilibrium models, like
Davidson & Woodbury (1993), Lise et al. (2004), and
Heckman et al. (2004), all of which found large GE effects
Findings from Major WIA Evaluations
32

We present findings from 3 major studies, all using exact
matching and PSM
 Studies
differ in states, time period, variables controlled for,
and method to some extent

Heinrich et al. (2013) (training v. no training)
 For
adult women, ~$800/quarter Q4-Q16
 For adult men, ~$500/quarter in later quarters
 For dislocated workers, no patterns of gains
Findings from Major WIA Evaluations
33

Andersson et al. (2013)
 Adults
M/F pooled gain $300-$400 quarter in later quarters
 Dislocated workers lose ~125/quarter in one state and gain
~$300/quarter in other state

Hollenbeck (2009) Indiana pooled M/F
gain $549 in 3rd quarter after exit and $463 in 7th
quarter after exit
 Dislocated workers gain $410 in 3rd quarter after exit and
$310 in 7th quarter
 Adults
Summary of WIA Findings
34



Researchers generally find modest, positive earnings gains for
adults from training, that appear to persist for several years
Findings for dislocated workers much less consistent, often
zero or negative, perhaps because populations differ or
perhaps because hard to distinguish temporary from
permanent shocks
Results from WIA RCT due later this year will help sort this
out, particularly inconsistencies on dislocated workers
Evaluation of Job Corps
35


Job Corps is long-term residential program for poor
youth and has larger budget than any other DOL E&T
program
Exemplary RCT evaluation strategy includes
 Most
Job Corps sites included good for external validity
 Small control group at each site to reduce bias
 Use of administrative and survey data
 Analysis of outcomes including crime
36
Evaluation of Job Corps:
Major Findings






Job Corps increases education and training for T group by
about 1 academic year
Job Corps increased literacy skills
For first 2 years after random assignment, participants earned
more, ~12% more in years 3 and 4 after random assignment
In years 5-10, no difference in earnings for T and C groups
Job Corps reduced crime ~5 percentage points
Overall, B<C except for 20-24 year olds
Other E&T Programs of Interest
37





Trade Adjustment Assistance evaluated by quasi-experimental
methods had no impact
Many studies of welfare to work programs, often showing
modest impacts
Evaluation of three sectoral programs by Public-Private
Ventures using RCTs found large impacts for 2 years after
random assignment
Evaluations of dislocated worker programs mixed, but little
evidence training valuable
Little credible evidence that youth programs effective
Program Operation Issues
38




Research on vouchers mixed, generally indicating vouchers
popular but do not improve impacts much, if at all
Performance measurement studies show typical measures not
correlated with impacts and often have perverse incentives
More studies of participation would be useful for
understanding programs and evaluations
There is current interest in career pathway programs, training
to obtain industry-sponsored credentials, and sectoral
training programs, with evaluations underway
Summary and Suggestions
39





Programs for poor adults pass CBA test, but do not make
participants self-sufficient: can we do better?
Results for dislocated workers sparse and mixed: will WIA
evaluation change things?
Youth programs disappointing: can we build on Job Corps
findings to do better?
Getting good cost data very difficult and makes good CBA
challenging
Although RCTs have key role in evaluations, use of nonexperimental designs important for looking at marginal
program changes