Ch. 3 rev

1
CJE3444
Crime Prevention
Chapter 3
Evaluation and Crime Prevention
Dr. Elizabeth C. Buchholz
Types of Evaluation
 Evaluation of crime prevention refers to investigating the impact of a prevention technique or
intervention on the level of subsequent crime, fear, or other intended outcome.
 Investigating the impact
 Impact Evaluation
 Focus on what changes occur after the introduction of the policy, intervention, or
program.
 For instance: Do neighborhood watch problems reduce fear and actual crime rates once
introduced?
 Do police patrols reduce drug sales in an area?
Impact (outcome) Evaluation
cont’d
 In crime prevention, pose interesting problems
 Crime prevention initiatives rarely rely on a single intervention or approach
 The target of the initiatives is a neighborhood or other geographic area
 Neighborhoods cannot be isolated
 Many interventions are not uniformly applied across an area or adopted by all
residents
 Competing issues of crime displacement and diffusion of benefits
Impact (outcome) Evaluation
 Problems:
 The target of the initiatives (unit of analysis for evaluation) is a neighborhood or
other geographic area
 Neighborhoods cannot be isolated
 Many interventions are not uniformly applied across an area or adopted by
all residents
 Issues of crime displacement and diffusion of benefits
Impact Evaluation
 Problems (cont’d):
 Crime prevention initiatives rarely rely on a single intervention approach.
Programs use a “menu” of different activities at the same time
 Watch scheme
 Property identification
 Neighborhood cleanup
 Periodic meeting
 Newsletter
Obstacles with Impact Evaluation
 Which of the program initiatives are most effective?
 What is the unit of analysis?
 Is crime displaced (moved to another area) or is it prevented?
09.03.13
2
Process Evaluation
 Considers the implementation of a program or initiative and involves determining the
procedures used to implement a specific program.
 More of a qualitative approach to evaluation.
 Offer a detailed descriptive account of the program and its implementation
 Information is pivotal in answering questions about the context of an intervention and what
actually took place in the initiative
 Unfortunately, many evaluations look only at the process
Process Evaluation
 Mission/goals of the program
 Level and quality of the program staff
 Funding and other resources of the program
 Obstacles faced in implementing and sustaining the initiative
 Degree to which the project was carried out as planned
 Level of support for the program
 Degree to which the clients complied with the operation
Problems with Process Evaluation
 The impact of the program is often difficult to comprehend
 Success is often determined by “false” results:
 Number of meetings
 Number of members involved
 Length of program
 Financial support
 Are not true measures of success
Cost Benefit Evaluations
 Seeks to assess whether the costs of an intervention are justified by the benefits or outcomes
that accrue from it.
 A cost benefit analysis should involve an impact and process evaluation to determine the
program’s worth.
 You cannot determine if costs are justified if you do not measure whether or not the program
is able to bring about the expected change.
Problems with Cost-Benefit Evaluations
 Largest problem involves setting monetary values on factors that are not easily enumerated
(fear, trauma, emotional loss, etc.)
 Another problem is making certain that all of the costs involved in the program (and related
to the program operations) are counted
Theories and Measurement
 Programs are often created in a theoretical vacuum
 Those implementing and evaluating the intervention do not pay attention to the
theoretical assumptions underlying the prevention program.
 Evaluations undertaken in a theoretical vacuum may still provide answers regarding whether
the program had the intended impact, just not why
 Basic questions on the assumed causal mechanism are often ignored
 Residents and planners want results but often fail to ask the questions of why or how:
09.03.13
3
 Why would an educational program reduce aggressive behavior?
 This results in “Basic” evaluations of yes it was a success or no it was not
 These evaluations fail to tell us “why”
Theories and Measurement
 Programs that are not evaluated:
 Fail to tell us why the program is or is not successful
 Can provide only limited insight to whether the program can be implemented to other
places or at other times
 Many investigations might not be necessary if the underlying theory for the
intervention was examined
• Implementing curfew laws when most crimes involving youths occur during
the afternoon
Why the resistance?
1. “Outcome Myopia”
 Programs and the evaluators are only interested in whether the program works
and not how or why it works.
 Does not take into account the possibility that other factors are at work
 Does not tell why a program does not work
2. Many program administrators assume that a program works
 They believe it’s only “common sense” that it works
 Blind belief in programs
 Have the ear of politicians who can provide legislative and funding
3. Many programs are the results of grassroots efforts
 Not interested in evaluations, as long as they’re happy with it
Theories and Evaluation
 Evaluation of crime prevention strategies is key
 Most programs fail to have adequate evaluation strategies
 Examples
 Juvenile curfew laws
 Weed and Seed programs
Measurement Problems
 Measuring key outcome variables when the intervention is geographically based
• Is area being measured the same as the area the police intervention was
implemented?
• Many crime prevention programs are based on neighborhoods or other small
geographic areas that do not coincide with specific police reporting areas
 Conundrum in crime prevention whereas programs often try to simultaneously reduce the
level of crime while increasing the reporting of crime to the police
 Possible solutions
• Using victim survey data
THEORY AND MEASUREMENT
 Measurement Issues cont’d
Victim data is not always available, and the collection of that data can be both timeconsuming and costly
Operationalizing key variables such as fear is not a straightforward enterprise
09.03.13
4
Finding ways to uncover the competing influences in the project that mask the outcomes is
difficult
THEORY AND MEASUREMENT
 Follow-Up Periods
How long after the implementation of the program or intervention will changes in crime (or
other outcome) appear?
Is there a possibility that over time any initial changes will diminish or disappear?
There is no rule on the appropriate follow-up time
Evaluation should look to the underlying theory for guidance
Ideal would be one where follow-up data is gathered at different intervals
METHOD FOR EVALUATION
Experimental Design
 The preferred approach by many evaluators due to a number of strengths:
 A true experimental design is also known as a randomized control trial
o the gold standard in evaluation
 Relies on the random assignment of cases into experimental and control groups, increases the
likelihood that the two groups being compared are equivalent.
 There is enough control over the evaluation to make certain that the experimental group
receives the treatment or intervention, while the control group does not.
Experimental Design
 A strength of Experimental Design is controlling for internal validity
 Internal validity deals with factors that could cause the results other than the measures that
were implemented
 Having a control group limits this threat to the program
Experimental Design cont’d
 The Maryland Scale of Scientific Methods (see Table 3.2) outlines the measures for
evaluation
 Developed for a review of literature on what works in prevention for the U.S.
Congress
 How closely a study adhered to the standards of a true experimental design
 Suggests that policy makers should only consider research that meets the gold
standard and that research funds should only be expended when an experimental
design is used
 Not many programs meet the experimental design goal
Other Threats to Evaluation
 Generalizability
 Are they applicable in other places, settings, and times?
 External Validity
 Wide range of potential problems inherent in trying to replicate the findings of any program
evaluation
 Difficult, if not impossible, to randomly assign communities to experimental and control
groups
 Matching cannot guarantee that the areas are comparable
 Mo way to isolate the experimental and control communities from all other influences
09.03.13
5
 Experimental Design evaluations often guide the project instead of the project being guiding
by the needs of the city
 Experimental Design (cont.)
Problems (cont.)
Underlying problem is that experimental designs fail to consider the context within which a
program or intervention operates
Too easy to jump to a conclusion that something does or does not work
Negative findings may be the result of factors such as poor program implementation,
misspecification of the appropriate target or causal mechanism underlying the problem, or
resistance by the target
Realistic Evaluation
 Rather than relying exclusively on experimental approaches, evaluation needs to observe the
phenomenon in its entirety.
 Two key ideas: mechanism and context
 Mechanism: understanding “what it is about a program which makes it work”
 By what process does an intervention impact an outcome measure such as crime or
fear of crime?
 Context: “the relationship between causal mechanisms and their effects is not fixed, but
contingent”
Summary
 Evaluation or analysis of crime prevention programs are often seen as meaningless or too cost
and are not completed
 A good evaluation is important for the growth and success of the program
 The remaining chapters will focus on different programs/designs and evaluation of program
success will be discussed for each
09.03.13