Understanding and Mitigating Bias in Forensic Evaluation 1 In press, International Journal of Forensic Mental Health Understanding and Mitigating Bias in Forensic Evaluation: Lessons from Forensic Science Patricia A. Zapf1 and Itiel E. Dror2 1 John Jay College of Criminal Justice, The City University of New York 2 University College London Understanding and Mitigating Bias in Forensic Evaluation 2 Abstract Research and commentary have emerged in the last decade surrounding cognitive bias in forensic examinations, both with respect to various domains within forensic science as well as with respect to forensic psychology. Indeed, in 2009 the National Research Council (NRC) issued a 352-page report entitled, Strengthening Forensic Science in the United States: A Path Forward that delineated several weaknesses within the various forensic science domains and proposed a series of reforms to improve the issue of reliability within the forensic sciences. Since the NRC report various commentators have written about the impact of cognitive biases in the forensic sciences and have proposed solutions to mitigate the impact of these biases. The purpose of this paper is to examine and consider the various influences that can bias observations and inferences in forensic evaluation and to apply what we know from forensic science to propose possible solutions to these problems. We use Sir Francis Bacon’s doctrine of idols—which underpins modern scientific method—to expand Dror’s (2015) five-level taxonomy of the various stages at which bias can originate within forensic science to create a seven-level taxonomy. We describe the ways in which biases can arise and impact work in forensic evaluation at these seven levels, highlighting potential solutions and various means of mitigating the impact of these biases, and conclude with a proposal for using scientific principles to improve forensic evaluation. Key Words: bias; cognitive bias; cognitive factors; forensic evaluation; forensic psychology Understanding and Mitigating Bias in Forensic Evaluation 3 Understanding and Mitigating the Impact of Bias in Forensic Evaluation: Lessons from Forensic Science Research and commentary have emerged in the last decade surrounding cognitive bias in forensic examinations, both with respect to various domains within forensic science (e.g., fingerprinting, see Dror & Rosenthal, 2008; DNA, see Dror & Hampikian, 2011) as well as with respect to forensic psychology (i.e., forensic evaluations in civil and criminal domains; see Murrie, Boccaccini, Guarnera, & Rufino, 2013; Neal & Brodsky, 2016). Indeed, in 2009 the National Research Council (NRC) issued a 352-page report entitled, Strengthening Forensic Science in the United States: A Path Forward that delineated several weaknesses within the various forensic science domains and proposed a series of reforms to improve the issue of reliability within the forensic sciences. Prominent among these weaknesses was the issue of cognitive factors, which impact an examiner’s understanding, analysis, and interpretation of data. Cognitive factors are relevant to all aspects of human perception and decision-making and thus are implicated in many domains of forensic science as well as in forensic evaluation.1 Since the NRC Report, several researchers and commentators have written about the impact of cognitive biases in the forensic sciences (e.g., Dror, 2015; Dror, Thompson, Meissner, Kornfield, Krane, Saks & Risinger, 2015) and have proposed solutions to mitigate the impact of these biases (Found & Ganas, 2013; Krane et al, 2008, Mattijssen et al., 2015). The overarching goal in all of this, of course, is to make the various disciplines within the field of forensics as scientific as possible and to limit the impact of 1 We use the all encompassing terms forensic evaluation and forensic evaluators to refer to the panoply of forensic mental health assessments conducted by mental health professionals working at the intersection of the mental health and legal systems. Understanding and Mitigating Bias in Forensic Evaluation 4 factors that bias examiners and might impede impartial and objective observations and conclusions. While we acknowledge differences between the workflow and roles of various forensic science practitioners and forensic mental health evaluators, we also believe that there are overarching similarities in the tasks required between the forensic science and forensic mental health evaluation domains. Across these two domains, examiners and evaluators are tasked with collecting and considering various relevant pieces of data in arriving at a conclusion or opinion and, across both of these domains, irrelevant information can change the way an examiner/evaluator interprets the relevant data. Bias mechanism, such as bias cascade and bias snowball, can impact examiners in forensic science as well as in forensic psychology (Dror, Morgan, Rando, & Nakhaeizadeh, 2017) We also acknowledge that completely unbiased and impartial examinations (forensic science) and evaluations (forensic mental health) are aspirational in nature and that, because examiners/evaluators are human, it is impossible to eliminate all bias. Nevertheless, the intent here is to shed light on the issue of bias, the various stages at which it can exert influence, and possible ways in which to think about minimizing its impact on decisions and outcomes in forensic evaluation. The purpose of this paper is to examine and consider the various influences that can bias observations and inferences in forensic evaluation and to apply what we know from forensic science to propose possible solutions to these problems. We use Sir Francis Bacon’s doctrine of idols—which underpins modern scientific method—to expand a fivelevel taxonomy (Dror, 2015) of the various stages at which bias can originate within forensic science to create a seven-level taxonomy applicable to forensic evaluation. We Understanding and Mitigating Bias in Forensic Evaluation 5 describe the ways in which biases can arise and impact work in forensic evaluation at these various levels, highlighting potential solutions and various means of attempting to mitigate the impact of these biases, and conclude with a proposal for next steps on the path forward with the hope that increased awareness of and exposure to these issues will continue to stimulate further research and discussion in this area. Sir Francis Bacon, who laid the foundations for modern science, believed that scientific knowledge could only arise if we avoid factors that distort and prevent objectivity. Nearly 400 years ago, Bacon developed the doctrine of “idols,” in which he set out the various obstacles that he believed stood in the way of truth and science—false idols that prevent us from making accurate observations and achieving understanding by distorting the truth and, therefore, stand in the way of science. These obstacles were categorized into idola tribus (idols of the tribe), idola spectus (idols of the den or cave), idola fori (idols of the market), and idola theatric (idols of the theatre). The implications of these obstacles for forensic science have been described in Dror (2009). For example, Bacon makes the case that experiences, education, training, and other personal traits (the idola spectus) that derive from nurture, can cause people to misperceive and misinterpret nature differently. That is, because of individual differences in their upbringing, experiences, and professional affiliations, people develop personal allegiances, ideologies, theories, and beliefs, and these may “corrupt the light of nature” (Bacon, 1620/1902, p. 21). Bacon’s doctrine of idols distinguishes between idols that are a result of our physical nature (e.g., human cognitive architecture) and the ways in which we were nurtured (e.g., experiences), and those that result from our social nature and the fact that Understanding and Mitigating Bias in Forensic Evaluation 6 we are social animals who interact with others in communities and work together. The first two idols—those of the tribe and the den—result from our physical nature and upbringing respectively, whereas the others—those of the market and theatre—result from our social nature and our interactions with others. Although we acknowledge an imperfect alignment between the idols as Bacon conceptualized them and how they might be conceptualized and interpreted today, and indeed including how we might have done so in this article, this conceptualization nonetheless provides a useful framework within which to consider challenges and obstacles in various domains. Here we elaborate on Dror’s (2009) use of this framework for forensic science by integrating it into a discussion of the challenges and obstacles facing forensic evaluation. In parallel and in addition to Bacon’s four idols, Dror and his colleagues have discussed various levels at which cognitive factors might interfere with objective observations and inferences and contribute to bias within the forensic sciences (see Dror, 2015; Stoel, Berger, Kerkhoff, Mattijssen, & Dror, 2015). These various levels encompass both case-specific components as well as more general, base-rate and organizational factors. When considering Bacon’s doctrine of idols in light of the various factors that might introduce bias into forensic decision-making, it appears that Bacon’s idols can be integrated into these levels to further expand and enrich the taxonomy. Here we present a seven-level taxonomy that integrates Bacon’s doctrine of idols with the previous work of Dror and colleagues on the various sources of bias that might be introduced, and apply these to forensic evaluation. Understanding and Mitigating Bias in Forensic Evaluation 7 Forensic Evaluation as akin to Scientific Investigation Forensic evaluation requires the collection and examination of various pieces of data to arrive at an opinion regarding a particular legal issue at hand. The most common type of forensic evaluation conducted is the examination of a defendant’s competency to stand trial, with more than 60,000 evaluations of this type being ordered each year in the United States (see Melton, Petrila, Poythress, & Slobogin, 2007) and relatively similar proportions (taking population statistics into account) being ordered in countries with adversarial systems of justice, such as Canada, the United Kingdom, and Australia (Reed & Zapf, in press). Other types of forensic evaluations include assessments of a defendant’s mental state at the time of the offense, assessments of risk for general violence or specific forms of violence (e.g., intimate partner violence, sexual violence), and assessments of other issues relevant to either criminal or civil law (e.g., evaluation of competence to waive Miranda rights, evaluation of parenting capacity, evaluation of testamentary capacity, evaluation of capacity to provide informed consent to treatment). The common components of all forensic evaluations include the collection of data relevant to the issue at hand—typically including official (police) reports and third-party or collateral information sources relevant to the legal issue being evaluated, interview data provided by the individual being evaluated, and the results of psychological testing, if relevant to the legal inquiry—and the consideration and weighting of these various pieces of data, according to relevance and information source, to arrive at an opinion/conclusion regarding the legal issue being evaluated (see Neal & Grisso, 2014a for a survey of the types of tools and data used in various types of forensic evaluations). Forensic evaluation is distinct from clinical evaluation, which relies on limited self-report Understanding and Mitigating Bias in Forensic Evaluation 8 data from the individual being evaluated. Forensic evaluation places great importance on collecting and considering third party and collateral information in conjunction with an evaluee’s self-report data, and forensic evaluators are expected to consider the impact and relevance of the various pieces of data on their overall conclusions (see Melton et al., 2007). In addition, forensic evaluators are expected to strive to be as impartial, objective and unbiased as possible in arriving at their conclusions and opinions about the legal issue at hand. Indeed, forensic evaluators are trained to write reports that delineate the data that were considered, the inferences made on the basis of those data, and the opinions or conclusions arrived at on the basis of the data and the inferences made from those data (see, for example, Grisso, 2010; Otto, DeMier, & Boccaccini, 2014). Hence, it can be argued that forensic evaluations should aspire to be more similar to scientific investigations—where the emphasis is placed on using observations and data to test alternate hypotheses—than to unstructured clinical assessments, which accept an evaluee’s self-report at face value without attempts to corroborate or confirm the details of the evaluee’s account and with less emphasis on alternate hypothesis testing (additional details on the differences between forensic evaluation and clinical evaluation can be found in Goldstein (2003) and in Melton, Petrila, Poythress, & Slobogin (2007)). If we accept the premise that forensic evaluations should be more akin to scientific investigations than clinical evaluations, then forensic evaluators should conduct their work more like scientists than clinicians, using scientific methods to inform their conceptualization of the case and opinions regarding the legal issue at hand. Maximizing the accuracy and objectivity of observations and inferences is crucial to arriving at minimally biased conclusions regarding the legal issue being evaluated. We take the Understanding and Mitigating Bias in Forensic Evaluation 9 lessons from forensic science and apply these to forensic evaluation with the aim of making forensic evaluation as objective and scientific as possible within the confines and limitations of attempting to apply group-to-individual inferences (see Faigman, Monahan, & Slobogin, 2014 for a detailed discussion of group-to-individual inference in scientific expert testimony). We do so by developing the framework of a seven-level taxonomy delineating the various influences that might interfere with objective observations and inferences, potentially resulting in biased conclusions in forensic evaluation. The taxonomy starts at the bottom with innate sources that have to do with being human. As we ascend the taxonomy, we discuss sources related to nurture—such as experience, training, and ideology—that can cause bias and, as we near the top of the taxonomy, the sources related to the specific case at hand. So, the order of the taxonomy is from general, basic innate sources derived from human nature, to sources that derive from nurture, and then to those sources that derive from the specifics of the case at hand (see Figure 1). Influences that Might Interfere with Objective Observations and Inferences Factors that might interfere with objective observations and inferences, and thus lead to potentially biased decision making, exist at a number of levels and for various reasons. Figure 1 delineates the seven-level taxonomy that summarizes the various influences that might interfere with objective decision-making in forensic evaluation. We conceptualize this taxonomy as spanning from general influences that result from our basic human nature (bottom of the pyramid) to case-specific influences (top of the pyramid), denoted by the bi-directional arrow at the left side of the Figure. We discuss Understanding and Mitigating Bias in Forensic Evaluation 10 each of the levels of the taxonomy within the context of forensic evaluation, drawing from forensic science domains. Case Evidence H N a um a tu n re En vi & r on Ex m e pe nt rie , C nc ul e tu re , C S p as e ec ific Reference Materials Irrelevant Case Information Base Rate Expectations Organizational Factors Training and Motivation Cognitive Architecture & The Brain Figure 1: Influences that might interfere with accurate observations and objective inferences in forensic conclusions Cognitive Architecture and the Brain At the very base of the taxonomy are potentially biasing influences that result from our basic human nature and the cognitive architecture of the brain. These influences are akin to what Bacon considered idola tribus, idols of the tribe; that is, obstacles that derive from the very fact of our being members of the human species. These obstacles or influences result from the way in which our brains are built; “they result from our Understanding and Mitigating Bias in Forensic Evaluation 11 inability to look at the world from ‘outside’ of ourselves, to see things as they actually are, instead of seeing them as distorted by our own mental processes” (Dror, 2009, p. 95). The human brain has a limited capacity to represent and process all of the information presented to it and so it relies upon techniques such as chunking information (binding individual pieces of information into a meaningful whole), selective attention (attending to specific pieces of information while ignoring other information), and topdown processing (conceptually-driven processing that uses context to make sense of information) to efficiently process information (Lindsay & Norman, 1977). It is naïve to believe that we perceive the world objectively or that we use bottom-up processing (encoding and interpreting the nature of a stimulus based solely upon the properties of the object) to understand information. Indeed, many studies have demonstrated the power and pervasiveness of top-down, conceptually driven processing whereby we unconsciously weave our knowledge of the world into our processing of it (Nickerson, 1998; Nisbett & Wilson, 1977; Kassin, Dror & Kukucka, 2013). We actively process information by selectively attending to that which we assume to be relevant and interpret this information in light of that which we already know. This top-down processing allows us to efficiently process loads of information and occurs in an automatic manner that is largely outside of our conscious awareness. Ironically, this automaticity and efficiency—which serves as the bedrock for expertise—also serves as the source of much bias (Nickerson, 1998; Dror, 2011). That is, the more we develop our expertise in a particular area, the more efficient we become at processing information in that area, but this enhanced performance results in cognitive tradeoffs that result in a lack of flexibility and error. “Thus, superior abilities introduced Understanding and Mitigating Bias in Forensic Evaluation 12 through expertise—paradoxically—also introduce degradation in performance. These assets and these vulnerabilities are inherent to the cognitive architecture and constitute two inevitable sides of the same coin” (Dror, 2009, p. 96; see also Dror, 2011). The cognitive system is such that it processes information in an efficient and effective way; however, the shortcut mechanisms used to do so result in vulnerability to bias and error. Ample research evidence demonstrates the presence of various types of biases that result from our cognitive architecture (Festinger, 1957; Simon, 1956; Tversky & Kahneman, 1973, 1974). For example, information that we encounter first is more influential than information we encounter later. This anchoring bias (see Tversky & Kahneman, 1974) can result in a forensic evaluator being overly influenced by or giving greater weight to information that is initially presented or reviewed. Thus, initial information communicated to the forensic evaluator by the referring party will likely be more influential, and serve as an anchor for, subsequent information reviewed by the evaluator. We also have a tendency to overestimate the probability of an event or an occurrence when other instances of that event or occurrence are easily recalled. This availability bias (see Tversky & Kahneman, 1973) can result in a forensic evaluator overestimating the likelihood of a particular outcome on the basis of being able to readily recall similar instances of that same outcome. Confirmation bias (see Nickerson, 1998) results from our natural inclination to rush to conclusions that confirm what we want, believe, or accept to be true. Indeed, research in the fingerprint analysis domain has demonstrated this confirmation bias, indicating that latent fingerprint examiners are influenced by contextually irrelevant information when making fingerprint comparisons Understanding and Mitigating Bias in Forensic Evaluation 13 (Dror & Rosenthal, 2008). In the forensic evaluation domain, Neal and Grisso (2014b) warn that the confirmation bias can exert its influence on evaluators who share a preliminary opinion before the evaluation is complete by committing the evaluator in a way that makes it difficult to resist or overcome this bias in the final interpretation of the data. These and other biases that result from our cognitive architecture and the way in which our brains are built are within all of us and are simply a part of our human nature. What is important is that we recognize our limits and cognitive imperfections so that we might try to address them by using countermeasures. Training and Motivation Moving up the taxonomy, the next three sources of influences that can affect our perception and decision-making result from our environment, culture, and experience. First among these are those influences that are brought about by our upbringing—our training and motivations (e.g., Kunda, 1990). Bacon referred to these obstacles as idola spectus, idols of the den or cave, and conceptualized these influences as a function of one’s experiences, education, training, and other personal traits or characteristics. As distinguished from the idols of the tribe or those that result from our human nature, the idols of the den or cave are a result of our nurture. “Different people, based on their specific upbringing, life experiences and professional affiliations, have developed personal allegiances to groups, ideologies, disciplines, theories, or methodologies” (Dror, 2009, p. 99). Our personal motivations and preferences, developed through our upbringing, affect our perception, reasoning, and decision-making (Balcetis & Dunning, 2006). Understanding and Mitigating Bias in Forensic Evaluation 14 Motivation comes in many forms and motivational bias has been well documented as a source of erroneous decision making within the context of criminal investigators’ judgments of witness reliability (Ask & Granhag, 2007) and need for cognitive closure (Ask & Granhag, 2005). Just as the motivations of criminal investigators can lead to distortions in the way that they perceive and interpret information, so too can the motivations of forensic evaluators distort or cause bias in their perception and interpretation of relevant (and irrelevant) case information. Closely related to an individual’s motivations are how one sees oneself and with whom that individual identifies. One particularly salient and concerning influence in this realm for forensic evaluators is that of adversarial allegiance; that is, the tendency to arrive at an opinion or conclusion that is consistent with the side that retained the evaluator. Murrie, Boccaccini and colleagues have demonstrated that forensic evaluators working for the prosecution assign higher psychopathy scores to the same individual as compared to forensic evaluators working for the defense (Murrie, Boccaccini, Johnson, & Janke, 2008). Similarly, forensic evaluators assign higher scores on actuarial risk assessment instruments—known to be less subjective than other types of risk assessment instruments—when retained by the prosecution and lower scores when retained by the defense (Murrie at al., 2009). This was the case in both field studies as well as in a wellcontrolled experimental manipulation (Murrie, Boccaccini, Guarnera, & Rufino, 2013). In addition, these researchers also found that adversarial allegiance appears to influence norm selection and reporting practices such that defense-retained evaluators were more likely to endorse reporting practices that conveyed the lowest possible level of risk Understanding and Mitigating Bias in Forensic Evaluation 15 whereas prosecution-retained evaluators were more likely to endorse practices suggesting the highest possible level of risk (Chevalier, Boccaccini, Murrie, & Varela, 2015). In addition to the pull to affiliate with the side that retained the forensic evaluator is the issue of pre-existing attitudes that forensic evaluators hold and how these might impact the forensic evaluation process. Neal (2016) surveyed forensic evaluators to determine the impact of attitudes towards capital punishment on willingness to accept referrals from various adversarial parties. She found that evaluators who opposed the death penalty were more likely to decline referrals from any party or accept referrals from the defense only whereas those who supported the death penalty were more likely to accept referrals from any party. This raises the issue of a self-selection bias, as a result of preexisting attitudes, which might contribute to partisan participation in capital evaluations. These issues of pre-existing attitudes, affiliations, and the potential biases they create also relate to the next set of idols described by Bacon, which consider the impact of the communications and interconnections between people. Organizational Factors The third set of idols described by Bacon—idola fori, idols of the market—are a result of our social interactions, communications, and connections with others. These include organizational and cultural factors that reflect the ways in which we understand and communicate with each other, both within and outside of the organization. The primary focus of Bacon’s idols of the market is on the language we use to convey knowledge. Language has a profound effect on how we perceive and think about information. The words we use to convey knowledge—terminology, vocabulary, and even jargon—can cause errors in how we understand and interpret information when we Understanding and Mitigating Bias in Forensic Evaluation 16 use them without attention and proper focus on the true meaning, or without definition, measurable criteria, and quantification. It is important to consider the meaning and interpretation of the words we use and how these might differ by organization, discipline, or culture. It is easy to assume that we know what someone means when they tell us something—whether it be an evaluee, a retaining party, or a collateral informant—but we must be cautious about both interpreting the language of others and using language to convey what we mean. Language and Bacon’s idols of the market also relate to how systems communicate as well as to organizational structures and the flow of information within and between different systems, organizations, or entities. Different methods or procedures can result in different rates of accuracy. For example, differences in the eyewitness lineup identification process have resulted in different rates of accuracy in the conclusions reached (Wells, Steblay, & Dysart, 2015). In the forensic assessment domain, different methods of conducting risk assessments (using dynamic risk assessment measures versus static risk assessment measures) have been demonstrated to affect the predictive accuracy of the conclusions reached by evaluators (Chu, Thomas, Ogloff, & Daffern, 2011). Indeed, several meta-analyses comparing the predictive accuracy of clinical versus actuarial methods for risk assessment have shown that actuarial methods—that is, highly structured methods with explicit decision rules and little room for discretion—outperform unstructured clinical methods and show higher rates of reliability and less bias in the predicted outcomes (e.g., Aegisdottir et al., 2006; Dawes, Faust, & Meehl, 1989; Hanson & Morton-Bourgon, 2009). Structured methods improve reliability and reduce bias by limiting discretion on the part of evaluators; thus, methods Understanding and Mitigating Bias in Forensic Evaluation 17 allowing or requiring more clinical discretion are more susceptible to evaluator bias. Murrie and colleagues’ (2013) research demonstrated this by finding that an actuarial tool allowing for little discretion (STATIC-99R) demonstrated less bias than a structured tool that preserved some discretion on the part of evaluators/raters (PCL-R). Combating and safeguarding against these idols of the market should take place at an organizational level, and some institutional structures are better for scientific inquiry than others (see Koppl, 2005). For example, the NRC (2009) report recommended moving forensic laboratories out of police departments. Within existing organizational structures, using language with specific definition and meaning that serves to increase error detection and prevention is important for creating a more scientific discipline. Hawaii’s three-panel system for evaluating competency and insanity of criminal defendants, while not perfect, is one example of an organizational structure that serves to increase error detection and prevention. As Acklin, Fuger, and Gowensmith (2015) concluded after evaluating examiner agreement and judicial consensus within Hawaii’s three-panel system, procedural standardization, the use of forensic assessment instruments, and the application of structured professional methods can serve to reduce bias in assessments and improve the quality of forensic mental health opinions. Base Rate Expectations The final set of idols described by Bacon—idola theatri, idols of the theatre—are not inherent to our human nature but, rather, result from how we produce knowledge or draw inferences and conclusions. Bacon divided these idols into three types: “sophistical, based on just a few anecdotal observations (or even no experimental evidence); empirical, based on narrow research; and superstitious, based on unsupported or blind Understanding and Mitigating Bias in Forensic Evaluation 18 belief” (Dror, 2009, p. 106). In considering forensic evaluation in light of these idols of the theatre, we must ask what forensic evaluation is based on. That is, to what extent is forensic evaluation based on broad, systematic scientific research and to what extent is it based on anecdotal or narrow, in-house research and observations? To what extent do forensic evaluators accept the science without proper and sufficient questioning? How do they distinguish between what they actually know and what they merely believe? How much of what they do is a matter of belief and how much is a matter of scientific knowledge? These idols are perhaps the most difficult to confront, and defensiveness or a lack of openness to discussing and examining the bases of our knowledge can be serious obstacles that prevent forensic evaluators from broadening and strengthening the scientific basis for their discipline. The ways in which forensic evaluators produce knowledge within their discipline can serve as an impediment to accurate observations and objective inferences. Anecdotal observations or information based on unsupported or blind beliefs can serve to create expectations about conclusions or outcomes before an evaluation is even conducted. Similarly, using methods or procedures that have not been adequately validated or that have been based on narrow, in-house research for which generalizability is unknown can result in inaccurate conclusions. Drawing inferences on the basis of untested assumptions or base rate expectations can lead to erroneous outcomes. Early research on the risk and treatability of psychopaths that was narrowly focused on interventions such as milieu therapy with groups of serious, chronic offenders led some to conclude that psychopaths were untreatable (see Douglas, Nikolova, Kelley, & Edens, 2015). As research in this domain expanded to include degrees of psychopathy Understanding and Mitigating Bias in Forensic Evaluation 19 and various types of interventions the conclusion that psychopaths are untreatable is no longer widely accepted (see, for example, Salekin 2002). Consideration of that early body of narrowly-focused research sets an expectation for evaluators that individuals displaying psychopathic traits are a high risk to re-offend and have low treatment potential, a different expectation regarding risk and treatment potential than that set by the broader, more diverse body of research that has since accumulated. In either case, creating expectations that might result in potentially biased assessment outcomes for a particular individual. As we move up the pyramid we move from environmental, cultural, and experiential influences to those influences that are specific to the particular case at hand. That is, the top three levels of the pyramid describe influences that might interfere with accurate observations and objective inferences within a specific forensic evaluation. Irrelevant Case information Whereas the previous level deals with influences that come about as a result of the way in which we produce knowledge or draw inferences and conclusions from information that does not pertain specifically to the case at hand, this level deals with influences that result from information that is obtained or reviewed for a specific case but that is irrelevant to the referral question. Perhaps one of the most potentially biasing considerations at this level involves the inferences made by others. In any forensic evaluation, the evaluator is tasked with collecting, reviewing, and considering various sources of information and data, much of which is relevant to the referral question but a great deal of which is irrelevant. Detailed information about an evaluee’s criminal history (offenses committed prior to the index offense), in most instances, is irrelevant to the Understanding and Mitigating Bias in Forensic Evaluation 20 issue of his or her criminal responsibility, which is an inquiry that focuses on the mental state of the individual at the time of the index offense. This irrelevant information, however, can become biasing for an evaluator. Even more potentially biasing can be the inferences and conclusions that others make about an evaluee—including collateral informants as well as retaining and opposing parties—since evaluators typically do not have access to the data or the logic used by others in arriving at these inferences and conclusions. It is naïve to think that a forensic evaluator can only collect and consider relevant information, especially since many times it is not clear what is relevant and what is irrelevant until all collected materials have been reviewed; however, disregarding irrelevant information is nearly impossible. Just as a jury cannot un-hear evidence that has been stricken from the record (see, for example, Kassin & Sukel, 1997), it is difficult, if not impossible, to neglect potentially biasing irrelevant information obtained or reviewed as a component of a forensic evaluation. “Just as forensic examiners are well aware and take great steps to minimize physical contamination of the evidence, they also must be aware and take steps to minimize cognitive contamination. By focusing on the scientific data they need, and isolating themselves as much as possible from everything else, they minimize cognitive contamination and help achieve their role as scientists” (Dror, 2015, p. 4, emphases in original). As Neal and Saks (2016) highlighted, the decision task across different forensic evaluation types can be quite different; thus, the kind of information that is irrelevant, contextually biasing, etc. will be different for each forensic evaluation referral question and therefore the strategies used in an attempt to mitigate bias will likely vary by evaluation type or referral question. Understanding and Mitigating Bias in Forensic Evaluation 21 Attempting to limit, as much as possible, the irrelevant information that is reviewed or considered as part of a forensic evaluation is one means of mitigating bias. Having a third-party take an initial pass through documents and records provided for an evaluation to compile relevant information for the evaluator’s consideration is one way of potentially mitigating against biasing irrelevant information (Neal & Saks, 2016). Another potentially mitigating strategy might be to engage in a systematic process of review where clear and specific documentation of what was reviewed, when it was reviewed, in the order in which it was reviewed, and with the evaluator detailing his or her thoughts, formulations, and inferences after each round of review, beginning with the most explicitly relevant case information (e.g., the police report for the index offense in a criminal responsibility evaluation) and moving towards the least explicitly relevant case information (e.g., elementary school records in a criminal responsibility evaluation). This iterative process of systematic documentation of data, inferences, logic, and conclusions would require evaluators to consider how their inferences and conclusions evolve as a result of various data sources considered and how these relate to the various hypotheses that are being tested. Wills (2008) described a related method of opinion formulation; however, as per Neal and Grisso (2014b), we caution that evaluators should be careful to engage in a systematic review that includes rival hypothesis testing and documenting data and inferences that would both support and reject the various hypotheses being tested so as not to commit to an ultimate opinion prematurely. Hypothesis testing is an important component of the scientific method and the principle of falsifiability (see Popper, 1963) and several commentators have highlighted its utility and importance for forensic evaluation—confirming or disconfirming possible explanations for relevant behaviors Understanding and Mitigating Bias in Forensic Evaluation 22 and capacities (e.g., Grisso, 2003; Heilbrun, 2001; Melton et al., 2007; Milchman, 2015). Engaging in a process of systematic collection and consideration of relevant data to test alternate hypotheses, with careful documentation of what is being considered and how the various pieces of data either strengthen or weaken a particular hypothesis, can potentially mitigate the impact of bias in forensic evaluation. Reference Materials Just as irrelevant case material can be biasing, so too can contextual information included in the reference materials for a forensic evaluation. Whereas in the forensic identification sciences reference materials pertain to the target or to a known suspect (e.g., the tire print from a specific make and model of car; the fingerprint of a known suspect), in forensic evaluation reference materials might include the legal test or standard for the referral question (including jurisdictional statutes and relevant case law), the abilities required of a defendant to meet a specific legal test, or any other metric that an evaluator is to use in considering whether or to what extent the evaluee has certain characteristics or abilities, or meets the requirements for a legal test or issue in question. That is, reference materials would include whatever it is that the evaluator is supposed to be evaluating the evidence against and, of course, can include potentially biasing contextual information. Admittedly, the decision tasks faced by forensic scientists differ greatly from those faced by forensic evaluators. Whereas forensic scientists are faced with epistemic information that conveys a ‘right’ answer (such as a known suspect) and their decision task includes an ‘identification’ component, in forensic evaluation the decision task includes evaluating the extent to which a known individual’s characteristics, behaviors, Understanding and Mitigating Bias in Forensic Evaluation 23 or abilities meet a particular legal test, standard, or categorization. Just as fingerprint experts can “(re-)analyze the crime scene finger mark after analyzing the reference print, with the risk that the expert starts to ‘see’ features that he [or she] would not have seen otherwise, and that might not be there” (Stoel et al., 2015, p. 76) so too can forensic evaluators become influenced by the reference materials that they consult for a forensic evaluation. Precedent, case law, and the interpretations of legal statutes and decisions can shape or impact our interpretation of the evidence in a forensic evaluation (see Philipsborn, 2004 for a discussion of how interpretations can be shaped by case law). Reference materials provide the basis for the legal issue being evaluated but can also influence interpretations or inferences made about the relevant data. Biasing contextual information within the reference materials may direct the interpretation of the actual data being evaluated. Such might be the case when the defense attorney provides information to the competency evaluator regarding the abilities required for the defendant to proceed to trial. Even though the information provided by a defense attorney regarding the abilities required for his or her client to proceed to trial is a necessary and relevant component of the reference materials for a competency evaluation, this information can influence an evaluator’s observations and inferences. The reference materials also underpin the well-documented phenomenon of ‘rater drift,’ wherein one’s ratings shift over time or drift from standard levels or anchors by unintentionally redefining criteria (see Kaplan & Saccuzzo, 2013). This means that evaluators should be careful to consult the relevant legal tests, statutes, or standards for each evaluation conducted and not assume that memory for or conceptualization of the standard or reference material is accurate. As one of the reviewers of this manuscript Understanding and Mitigating Bias in Forensic Evaluation 24 noted, this point about rater drift appears to work well within the context of forensic mental health evaluation in light of Murrie and colleagues’ (2013) experimental findings about adversarial allegiance. Just as irrelevant or contextual information can influence our perceptions and interpretations, so too can our experiences and expectations influence our memories and conceptualizations, which might result in biased or inaccurate opinions or conclusions. Case Evidence In addition to irrelevant case information and contextual information included as part of the reference materials for a case, the actual case evidence itself might also include some irrelevant, contextual, or biasing information. Here we conceptualize case evidence as information germane to the focus of the inquiry that must be considered by any forensic evaluator in arriving at an opinion about the particular legal issue. For example, police reports and discovery materials for the index offense in a criminal responsibility evaluation, or information regarding the allegations involved for a defendant’s criminal charges in a competency to stand trial evaluation. In essence, case evidence would include any information that must be considered in evaluating the particular legal issue at hand (e.g., to proceed without the information would be considered inappropriate or sub-par practice by a majority of forensic evaluators). Whereas various background, educational, or treatment records may or may not be collected and considered in a criminal responsibility evaluation, to proceed without considering documented information regarding the index offense would be particularly problematic and generally viewed as sub-par or inappropriate practice for this type of evaluation. Influences at the case evidence level include biasing contextual information Understanding and Mitigating Bias in Forensic Evaluation 25 from the actual police reports or other data that must be considered for the referral question. Thus, contextual information that is inherent to the case evidence and that cannot be easily separated from it can influence and bias an evaluator’s inferences about the data. Handwritten notes in the margin of an official police report or witness statement, the questions being asked of a suspect in a videotaped interview of a confession or other interrogation, or an officer’s description of a witness’ report can all include irrelevant or contextual information that can influence the perceptions and inferences of a forensic evaluator. Information that might be relevant to other experts in the case at hand or relevant to the trier of fact but that should be considered as impertinent information for the referral question being addressed by the forensic evaluator can be potentially biasing. Data, inferences, and conclusions about an evaluee provided by the retaining party, information provided by defense counsel in competency evaluations, victim impact statements in risk assessments, and official reports are all sources of relevant case information that can nonetheless be potentially biasing. Irrelevant or contextual information can influence the way in which evaluators perceive and interpret data at any of these seven levels—ranging from the most basic aspects of human nature and the cognitive architecture of the brain, through one’s environment, culture, and experiences, and including specific aspects of the case at hand—but it is important to note that biased perceptions or inferences at any of these levels do not necessarily mean that the outcome, conclusion, or opinion will be biased. As West and Kenny (2011) make clear, bias direction and strength are independent. As explicated by Dror (2009), “If the bias is in the direction of the correct decision, then the Understanding and Mitigating Bias in Forensic Evaluation 26 evidentiary data may bring the decision considerations past the decision threshold, and the bias will only push the accumulated considerations further in the same direction” (p. 98). Even if the bias is in a contradictory direction from the correct decision, the evidentiary data might affect the considerations of the evaluator to some extent but not enough to impact the actual outcome of the evaluation or ultimate opinion of the evaluator. What appears important to the outcome is the degree to which the data are ambiguous; the more ambiguous the data, the more likely it will be that a bias will affect the actual decision or outcome (see, for example, Dror & Rosenthal, 2008). Ambiguous circumstances, therefore, are the most susceptible to biasing influences. The Path Forward: Using Science to Improve Forensic Evaluation In this paper we presented an overview of the various levels at which different factors can interfere with objective observations and inferences in forensic evaluation, potentially resulting in biased conclusions. Increased knowledge and understanding of the ways in which cognitive factors can impact a forensic evaluator’s perceptions and inferences is an important first step in attempting to mitigate the effect of bias in forensic evaluation; however, knowledge is not enough. The presence of a bias blind spot—the tendency of individuals to perceive greater cognitive and motivational bias in others than in themselves—has been well documented (see Pronin, Lin, & Ross, 2002). Neal and Brodsky (2014) demonstrated that forensic psychologists are occupationally socialized to believe that they can and do practice objectively (recall the discussion of training and motivational influences); however, emerging research on bias in forensic evaluation has demonstrated that this belief may not be accurate (see, for example, the discussion of Understanding and Mitigating Bias in Forensic Evaluation 27 adversarial allegiance). In addition, it appears that many forensic evaluators report using de-biasing strategies, such as introspection, which have been proven ineffective and some even deny the presence of any bias at all (Neal & Brodsky, 2016). What is clear, then, is that there is room for improvement in understanding and mitigating the effects of bias in forensic evaluation. As Dror (2009) noted, “the need for improvement by itself is by no means a sign of a lack of scientific rigor; on the contrary, a sign of any good science is its constant reflection, criticism, and examination” (p. 94). For forensic evaluation to advance and improve, we must behave as scientists. “Scientists continually observe, test, and modify the body of knowledge. Rather than claiming absolute truth, science approaches truth either through breakthrough discoveries or incrementally, by testing theories repeatedly” (NRC, 2009, p. 112). Identifying weaknesses and taking actions and countermeasures to avoid or limit those weaknesses must underpin any science (Dror, 2009). Countermeasures such as Linear Sequential Unmasking (LSU; Dror et al., 2015)—where irrelevant information is masked, and relevant information is sequentially revealed when needed —can be adapted for use in forensic evaluation. Approaching forensic evaluations like scientific inquiries and using rival hypothesis testing might place the necessary structure on the evaluation process to determine the differential impact of the various data considered (see Kassin, Dror & Kukucka, 2013 for further suggestions in the area of forensic science, and Dror, Kassin, & Kukucka, 2013 for debating the issues and objections). In addition to using principles of the scientific method to improve the way forensic evaluations are conducted, these same principles can be applied to test Understanding and Mitigating Bias in Forensic Evaluation 28 assumptions and hypotheses about forensic evaluation. Identifying weaknesses in forensic evaluation and conducting research and hypothesis testing on proposed counter measures to reduce the impact of bias will serve to improve the methods and procedures in this area. Being scientific about forensic evaluation and using scientific principles to understand and improve it appears to be a reasonable path forward for reducing and mitigating bias. Three considerations will be briefly highlighted. Evaluator characteristics. Recent research has demonstrated wide variability in the rates of evaluation outcomes as a function of evaluator, with some evaluators opining incompetence or insanity in a majority of evaluations whereas other evaluators almost never opine incompetence or insanity (see Murrie, Boccaccini, Zapf, Warren & Henderson, 2008; Murrie & Warren, 2005). The need for reliability among evaluators (as well as by the same evaluator at different times–inter and intra evaluator consistency) is a cornerstone for establishing forensic evaluation as a science (Dror, 2016). By understanding the characteristics of evaluators—including training, culture, and experience—that contribute to their opinions we can begin to propose and study different ways of limiting the impact of these characteristics on objective observation and inferences in forensic evaluation. Understanding and characterizing the abilities required for the specialized task of forensic evaluation will ultimately allow further development of profiles for selecting evaluators or for developing necessary evaluation skills. Evaluation methods. Research has demonstrated that reliability improves when standardized inquiries are used for competence evaluation (see Grisso, 2003; Zapf & Roesch 2009). Preliminary research has indicated that the use of clinical interview data in conjunction with file review data in risk assessments leads to different rating outcomes Understanding and Mitigating Bias in Forensic Evaluation 29 regarding relevant risk factors than when only file review data is considered (see Kropp & Hart, 2000). Conducting systematic research on the methods and procedures used in forensic evaluation and the impact of these on evaluation outcomes and bias will ultimately allow for development of the most effective strategies for forensic evaluation. In addition, examining the interrelations between evaluator characteristics and evaluation methods will further increase our understanding of the interplay between personal and situational variables in forensic evaluation. Training. Implementing professional training programs that address cognitive factors and bias in forensic evaluation and conducting systematic research on the impact of various training techniques for increasing understanding of these issues will likely improve the methods that forensic evaluators currently use to mitigate the impact of bias in their work (see Neal & Brodsky, 2016, for a discussion of strategies currently being used by evaluators). Indeed, Wettstein (2005) demonstrated that certification programs for forensic evaluators improve the reliability and quality of forensic evaluation reports and Gowensmith and colleagues (2015a) highlighted the importance of both initial as well as ongoing training. Although a trend toward state certification programs for forensic evaluators has been recently demonstrated, “most states still do not have a formal process for selecting or certifying” their forensic evaluators (Gowensmith, Pinals, & Karas, 2015b). In addition, a recent survey of forensic evaluators indicated that most evaluators are not provided any formal, direct training on the ways in which bias can impact their work (Zapf, Kukucka, Kassin, & Dror, 2017). Understanding the most effective ways of training evaluators to perform forensic evaluations in a consistent and reliable way while limiting the impact of bias will allow for the implementation of best Understanding and Mitigating Bias in Forensic Evaluation 30 practices, both with respect to the evaluations themselves as well as with respect to training procedures and outcomes. Applying scientific principles to test and compare various training strategies and outcomes will serve to improve knowledge and the practice of forensic evaluation. Summary & Conclusions Consideration of the various influences that might bias an evaluator’s ability to objectively evaluate and interpret data is an important component of forensic evaluation. We have delineated seven levels at which various influences might interfere with objective observations and inferences in forensic evaluation. Evaluators are encouraged to be mindful of these various influences and to consider the use of various bias mitigation strategies, as noted throughout. Knowledge about the ways in which bias can impact forensic evaluation is an important first step; however, the path forward also includes the use of scientific principles to test alternative hypotheses, methods, and strategies for minimizing the impact of bias in forensic evaluation. Using scientific principles to continue to improve forensic evaluation will bring us closer to the aspirational goal of objective, impartial, and unbiased evaluations. Understanding and Mitigating Bias in Forensic Evaluation 31 References Acklin, M.W., Fuger, K., & Gowensmith, W. (2015). Examiner agreement and judicial consensus in forensic mental health evaluations. Journal of Forensic Psychology Practice, 15, 318-343. Ægisdóttir, S., White, M. J., Spengler, P. M., Maugherman, A. S., Anderson, L. A., Cook, R. S., et al. (2006). The meta-analysis of clinical judgment process: Fifty-six years of accumulated research on clinical versus statistical prediction. The Counseling Psychologist, 34, 341-382. Ask, K., & Granhag, P. A. (2005). Motivational sources of confirmation bias in criminal investigations: The need for cognitive closure. Journal of Investigative Psychology and Offender Profiling, 2, 43-63. Ask, K., & Granhag, P. A. (2007). Motivational bias in criminal investigators’ judgments of witness reliability. Journal of Applied Social Psychology, 37, 561-591. Bacon, F. (1939). Novum organum. In J. Devey (Ed.), Novum organum. New York: P. F. Collier, 1902. Balcetis, E., & Dunning, D. (2006). See what you want to see: Motivational influences on visual perception. Journal of Personality and Social Psychology, 91, 612-625. Chevalier, C. S., Boccaccini, M. T., Murrie, D. C., & Varela, J. G. (2015). STATIC-99 reporting practices in sexually violent predator cases: Does norm selection reflect adversarial allegiance? Law and Human Behavior, 39, 209-218. Chu, C. M., Thomas, S. D. M., Ogloff, J. R. P., & Daffern, M. (2011). The short- to medium- term predictive accuracy of static and dynamic risk assessment measures in a secure forensic hospital. Assessment, 20, 230-241. Dawes, R. M., Faust, D., & Meehl, P. E. (1989). Clinical versus actuarial judgment. Aegisdottir Aegisdottir Aegisdottir 1668-1674. Douglas, K. S., Nikolova, N. L., Kelley, S. E., & Edens, J. F. (2015). Psychopathy. In B. L. Cutler & P. A. Zapf (Eds.), APA handbook of forensic psychology (Vol. 1, pp. 257-323). Washington, DC: American Psychological Association. Dror, I. E. (2009). How can Francis Bacon help forensic science? The four idols of human biases. Jurimetrics, 50, 93-110. Dror, I. E. (2011). The paradox of human expertise: why experts can get it wrong. In N. Kapur, A. Pascual-Leone, & V. S. Ramachandran (Eds.), The paradoxical brain (pp. 177-188). Cambridge, UK: Cambridge University Press. Understanding and Mitigating Bias in Forensic Evaluation 32 Dror, I. E. (2015). Cognitive neuroscience in forensic science: understanding and utilizing the human element. Philosophical Transactions of the Royal Society B, 370: 20140255. http://dx.doi.org/10.1098/rstb.2014.0255 Dror, I. E. (2016). A Hierarchy of Expert Performance. Journal of Applied Research in Memory and Cognition, 5 (2), 121-127. http://dx/doi.org/10.1016/j.jarmac.2016.03.001 Dror, I.E., Morgan, R., Rando, C. & Nakhaeizadeh, S. (2017). The bias snowball and the bias cascade effects: Two distinct biases that may impact forensic decision making. Journal of Forensic Sciences, 62 (3). Dror, I. E. & Hampikian, G. (2011). Subjectivity and bias in forensic DNA mixture interpretation. Science & Justice, 51, 204-208. Dror, I. E., Kassin, S. M., & Kukucka, J. (2013). New application of psychology to law: Improving forensic evidence and expert witness contributions. Journal of Applied Research in Memory and Cognition, 2, 78-81. Dror, I. E., & Rosenthal, R. (2008). Meta-analytically quantifying the reliability and bias ability of forensic experts. Journal of Forensic Science, 53, 900-903. Dror, I. E., Thompson, W. C., Meissner, C. A, Kornfield, I., Krane, D., Saks, M., & Risinger, M. (2015). Context management toolbox: A Linear Sequential Unmasking (LSU) approach for minimizing cognitive bias in forensic decision making. Journal of Forensic Sciences, 60, 1111-1112. Faigman, D. L., Monahan, J. & Slobogin, C. (2014). Group to individual (G2i) inference in scientific expert testimony. The University of Chicago Law Review, 81, 417480. Festinger, L. (1957). A theory of cognitive dissonance. Stanford, CA: Stanford University Press. Found, B., and Ganas, S. (2013). The management of domain irrelevant context information in forensic handwriting examination casework. Science and Justice, 53, 154-158. Goldstein, A. M. (2003). Overview of forensic psychology. In A. M. Goldstein (Ed.), Handbook of psychology, Volume 11: Forensic psychology (pp. 3-20). Hoboken, NJ: Wiley. Gowensmith, W. N., Pinals, D. A., & Karas, A. C. (2015). States’ standards for training and certifying evaluators of competency to stand trial. Journal of Forensic Psychology Practice, 15, 295-317. doi: 10.1080/15228932.2015.1046798 Gowensmith, W., Sledd, M., & Sessarego, S. (2015, August). Impact of stringent certification standards on forensic evaluator reliability. Paper presented at the Understanding and Mitigating Bias in Forensic Evaluation 33 American Psychology-Law Society annual conference, San Diego, CA. Grisso, T. (2003). Evaluating competencies: Forensic assessments and instruments (2nd ed.). New York, NY: Kluwer/Plenum. Grisso, T. (2010). Guidance for improving forensic reports: A review of common errors. Open Access Journal of Forensic Psychology, 2, 102-115. Hanson, R. K., & Morton-Bourgon, K. E. (2009). The accuracy of recidivism risk assessments for sexual offenders: A meta-analysis of 188 prediction studies. Psychological Assessment, 21, 1-21. Heilbrun, K. (2001). Principles of forensic mental health assessment. New York: Kluwer/Plenum. Kaplan, R. M., & Saccuzzo, D. P. (2013). Psychological testing: Principles, applications, and issues (8th ed.). Belmont, CA: Wadsworth. Kassin, S. M., Dror, I. E., & Kukucka, J. (2013). The forensic confirmation bias: Problems, perspectives, and proposed solutions. Journal of Applied Research in Memory and Cognition, 2, 42-52. Kassin, S. M., & Sukel, H. (1997). Coerced confessions and the jury: An experimental test of the ‘harmless error’ rule. Law and Human Behavior, 21, 27-46. Koppl, R. (2005). How to improve forensic science. European Journal of Law and Economics, 20, 255-286. Krane, D., Ford, S., Gilder, J., Inman, K., Jamieson, A., Koppl, R., et al. (2008). Sequential unmasking: a means of minimizing observer effects in forensic DNA interpretation. Journal of Forensic Science, 53, 1006–1007. Kropp, P. R., & Hart, S. D. (2000). The Spousal Assault Risk Assessment Guide: Reliability and validity in adult male offenders. Law and Human Behavior, 24, 101-118. Kunda, Z. (1990). The case for motivated reasoning. Psychological Bulletin, 108, 480– 498. Lindsay, P. H., & Norman, D. A. (1977). Human information processing: An introduction to psychology (2nd ed.). New York: Academic Press. Mattijssen, E., Kerkhoff, W., Berger, C., Dror, I., & Stoel, R. (2015). Implementing context management in forensic casework: Minimizing contextual bias in firearms examination. Science and Justice, 56, 113-122. Melton, G. B., Petrila, J., Poythress, N. G., & Slobogin, C. (2007). Psychological evaluations for the courts: A handbook got mental health professionals and Understanding and Mitigating Bias in Forensic Evaluation 34 lawyers (3rd ed.). New York, NY: Guilford. Milchman, M. S. (2015). The complementary roles of scientific and clinical thinking in child custody evaluations. Journal of Child Custody, 12, 97-128. doi: 10.1080/15379418.2015.1062453 Murrie, D. C., Boccaccini, M. T., Guarnera, L. A., & Rufino, K. A. (2013). Are forensic experts biased by the side that retained them? Psychological Science, 24, 18891897. Murrie, D. C., Boccaccini, M. T., Johnson, J. T., & Janke, C. (2008). Does interrater (dis)agreement on Psychopathy Checklist scores in sexually violent predator trials suggest partisan allegiance in forensic evaluations? Law and Human Behavior, 32, 352–362. doi:10.1007/s10979-007-9097-5 Murrie, D. C., Boccaccini, M. T., Turner, D., Meeks, M., Woods, C., & Tussey, C. (2009). Rater (dis)agreement on risk assessment measures in sexually violent predator proceedings: Evidence of adversarial allegiance in forensic evaluation? Psychology, Public Policy, and Law, 15, 19–53. doi:10.1037/ a0014897 Murrie, D. C., Boccaccini, M .T., Zapf, P. A., Warren, J. I., & Henderson, C. E. (2008). Clinician variation in findings of competence to stand trial. Psychology, Public Policy, and Law, 14, 177-193. Murrie, D. C., & Warren, J. I. (2005). Clinician variation in rates of legal sanity opinions: Implications for self-monitoring. Professional Psychology: Research and Practice 36, 519-524. National Research Council, Committee on Identifying the Needs of the Forensic Science Community. (2009). Strengthening forensic science in the United States: A path forward. Washington, DC: National Academies Press. Neal, T. M. S. (2016). Are forensic experts already biased before adversarial legal parties hire them? PLoS ONE, 11(4): e0154434. doi:10.1371/ journal.pone.0154434 Neal, T. M., & Brodsky, S. L. (2014). Occupational socialization’s role in forensic psychologists’ objectivity. Journal of Forensic Psychology Practice, 14, 24-44. http://dx.doi.org/10.1080/15228932.2013.863054 Neal, T. M. S., & Brodsky, S. L. (2016). Forensic psychologists’ perceptions of bias and potential correction strategies in forensic mental health evaluations. Psychology, Public Policy, and Law, 22, 58-76. http://dx.doi.org/10.1037/law0000077 Neal, T. M. S., & Grisso, T. (2014a). Assessment practices and expert judgment methods in forensic psychology and psychiatry: An international snapshot. Criminal Justice & Behavior, 41, 1406-1421. doi: 10.1177/0093854814548449 Neal, T. M. S., & Grisso, T. (2014b). The cognitive underpinnings of bias in forensic Understanding and Mitigating Bias in Forensic Evaluation 35 mental health evaluations. Psychology, Public Policy, and Law, 20, 200-211. doi:10.1037/a0035824 Neal, T. M. S., & Saks, M. J. (2016). The science of science offers solutions to cognitive bias in forensic psychology. Paper presented at the annual meeting of the American Psychology-Law Society, Atlanta, GA. Nickerson, R. S. (1998). Confirmation bias: A ubiquitous phenomenon in many guises. Review of General Psychology, 2, 175–220. Nisbett, R. E., & Wilson, T. D. (1977). Telling more than we can know: Verbal reports on mental processes. Psychological Review, 84, 231–259. Otto, R. K., DeMier, R. L., & Boccaccini, M. T. (2014). Forensic reports and testimony: A guide to effective communication for psychologists and psychiatrists. Hoboken, NJ: Wiley. Philipsborn, J. T. (2004). Searching for uniformity in adjudications of the accused’s competence to assist and consult in capital cases. Psychology, Public Policy, and Law, 10, 417-442. Popper, K. (1963). Conjectures and refutations. London: Routledge. Pronin, E., Lin, D. Y., & Ross, L. (2002). The bias blind spot: Perceptions of bias in self versus others. Personality and Social Psychology Bulletin, 28, 368-381. Reed, A., & Zapf, P. A. (in press). Criminal forensic assessment. In R. Roesch & A. N. Cook (Eds.), Handbook of forensic mental health services. New York: Routledge. Salekin, R. T. (2002). Psychopathy and therapeutic pessimism: Clinical lore or clinical reality? Clinical Psychology Review, 22, 79–112. doi:10.1016/ S02727358(01)00083-6 Simon, H. A. (1956). Rational choice and the structure of environments. Psychological Review, 63, 129-138. Stoel, R. D., Berger, C. E., Kerkhoff, W., Mattijssen, E., & Dror, I. (2015). Minimizing contextual bias in forensic casework. In K. Strom & M. J. Hickman (Eds.), Forensic science and the administration of justice (pp. 67-86). New York, NY: Sage. Tversky, A., & Kahneman, D. (1973). Availability: A heuristic for judging frequency and probability. Cognitive Psychology, 5, 207–232. doi: 10.1016/00100285(73)90033-9 Tversky, A., & Kahneman, D. (1974). Judgment under uncertainty: Heuristics and biases. Science, 185, 1124–1131. doi:10.1126/science.185 .4157.1124 Understanding and Mitigating Bias in Forensic Evaluation 36 Wells, G. L., Steblay, N. K., & Dysart, J. E. (2015). Double-blind photo lineups using actual eyewitnesses: An experimental test of a sequential versus simultaneous lineup procedure. Law and Human Behavior, 39, 1-14. West, T. V., & Kenny, D. A. (2011). The truth and bias model of judgment. Psychological Review, 118, 357-378. doi: 10.1037/a0022936 Wettstein, R. (2005). Quality and quality improvement in forensic mental health evaluations. Journal of the American Academy of Psychiatry and the Law, 33, 158–175. Wills, C. D. (2008). The CHESS method of forensic opinion formation: Striving to checkmate bias. Journal of the American Academy of Psychiatry and Law, 36, 535-540. Zapf, P. A., Kukucka, J., Kassin, S., & Dror, I. (2017). Cognitive bias in forensic mental health evaluation: A survey of evaluators. Paper presented at the Annual Meeting of the American Psychology-Law Society, Seattle, WA. Zapf, P. A., & Roesch, R. (2009). Evaluation of Competence to Stand Trial. New York, NY: Oxford.
© Copyright 2026 Paperzz