REGULATING INSCRUTABLE SYSTEMS Andrew D. Selbst* Solon Barocas† From scholars seeking to “unlock[] the black box” to regulations requiring “meaningful information about the logic” of automated decisions, recent discussions of machine learning—and algorithms more generally—have turned toward a call for explanation. Champions of explanation charge that algorithms must reveal their basis for decision-making and account for their determinations. But by focusing on explanation as an end in itself, rather than a means to a particular end, critics risk demanding the wrong thing. What one wants to understand when dealing with algorithms—and why one would want to understand it—can vary widely. Often, the goals that motivate calls for explanation could be better served by other means. Worse, ensuring that algorithms can be explained in ways that are understandable to humans may come at the cost of another value that scholars, regulators, and critics hold dear: accuracy. Reducing the complexity of a model, for example, may render it more readily intelligible, but also hamper its performance. For explanation to serve its intended purpose and to find its appropriate place among a number of competing values, its champions need to consider what they hope it to achieve and what explanations actually offer Existing law is replete with demands for explanation in service of other normative goals. The requirement that those affected by decisions be able to make sense of them is a central tenet of due process and administrative law. It is also embedded in specific laws like the Equal Credit Opportunity Act (ECOA), which requires that lenders give “statement[s] of reasons” for adverse determinations, and the General Data Protection Regulation (GDPR)—the recent data protection legislation in the European Union—that grants data subjects the right to seek “meaningful information about the logic” of automated decisions. Importantly, the type of explanation required differs between ECOA (specific reasons for a decision) and GDPR (the decision-making process itself), and that reflects the different purposes underlying each law. Similarly, those calling for explanations seek different results, whether a justification for those affected by a decision, means to assess the validity of decision-making procedures, ways to evaluate the fairness of processes, or even just mechanisms to anticipate an algorithm’s behavior and prepare accordingly. Those concerned with machine learning have many of these same concerns in mind, but new calls for explanation take issue not only with secrecy, opacity, or validity, but with inscrutability as well. Increasingly, demands for explanation are also demands for interpretability, reflecting a growing realization that certain algorithms may resist * † [insert author’s note] [insert author’s note] 2 Regulating Inscrutable Systems – DRAFT 3/20/2017 meaningful interrogation even when they are accessible, transparent, and perform well. Indeed, machine learning may produce models that differ so radically from the way humans make decisions that they resist sense-making. Older demands for explanation presupposed the possibility of a human who could provide an account of her decisions, even where those explanations amount to legal fictions. In this case, even when machines are able to explain themselves, they may not do so in ways that satisfy humans’ curiosity or concerns. This Article takes seriously the calls for regulation via explanation to investigate how existing laws implementing such calls fare, and whether interpretability research can fix the flaws. Ultimately, it argues that while machine interpretability may make compliance with existing legal regimes easier, or possible in the first instance, a focus on explanation alone fails to fulfill the overarching normative purpose of the law, even when compliance can be achieved. The paper concludes with a call to consider where such goals would be better served by other means, including mechanisms to directly assess whether models are fair and just. Introduction ................................................................................................... 2 I. What-How-Why and The Problem of Inscrutability ................................. 6 A. The What-How-Why Model ......................................................................... 7 B. The Effects of Inscrutability ......................................................................... 12 II. Existing Regulation of Inscrutable Systems ............................................ 16 A. Credit Scoring Regulations ......................................................................... 17 1. Introduction to FCRA, ECOA, and Regulation B ............................. 17 2. The Use of Explanation in Credit Scoring Regulation ....................... 21 B. General Data Protection Regulation ........................................................... 27 III. Interpretability in Machine Learning Systems ...................................... 32 A. Favoring Interpretable Methods .................................................................. 33 B. Global Explanations..................................................................................... 34 C. Specific Decisions ........................................................................................ 36 D. Task-Specific Techniques............................................................................ 38 IV. How Interpretability Can Assist Regulation ......................................... 38 A. Meeting Legal Requirements ...................................................................... 38 B. Vindicating Normative Principles ................................................................ 38 Conclusion ................................................................................................... 39 INTRODUCTION In Dan Solove’s 2001 Article, Privacy and Power, he argued that Franz DRAFT – Please ask permission before citing or quoting 3 Kafka’s The Trial serves as a better metaphor for society’s anxieties about today’s information culture than does George Orwell’s 1984.1 Though surveillance was and remains important, he wrote, Kafka offers “a more thoughtless process of bureaucratic indifference, arbitrary errors, and dehumanization, a world where people feel powerless and vulnerable, without any meaningful form of participation in the collection and use of their information.”2 As much prescient prediction as analysis, Solove’s replacement metaphor perfectly describes the state of anxiety over artificial intelligence and machine learning today. Complex, inscrutable algorithms regularly make consequential decisions about all our lives, with only minimal input from the people they affect and little to no explanation as to how they work. This worries people, and rightly so. The results of these algorithms could be unnerving,3 unfair,4 unsafe,5 unpredictable,6 and unaccountable.7 How can inscrutable algorithms be held accountable for bad results? As one commentator has observed, “we cannot effectively regulate what we do not understand.”8 This notion has led a number of scholars and policymakers to ask for Daniel J. Solove, Privacy and Power: Computer Databases and Metaphors for Information Privacy, 53 STAN. L. REV. 1393 (2001). 2 Id. at 1398. 3 See, e.g., Sara M. Watson, Data Doppelgängers and the Uncanny Valley of Personalization, THE ATLANTIC (June 16, 2014), https://www.theatlantic.com/technology/archive/2014/06/data-doppelgangers-and-theuncanny-valley-of-personalization/372780/; Omer Tene & Jules Polonetsky, A Theory of Creepy: Technology: Privacy and Shifting Social Norms 16 YALE J.L. & TECH. 59 (2013). 4 See, e.g., Solon Barocas & Andrew D. Selbst, Big Data’s Disparate Impact, 104 CALIF L. REV. 671, 677-692 (2016); Pauline T. Kim, Data-Driven Discrimination at Work, __ WM. & MARY L. REV. __ , (forthcoming 2017), available at http://papers.ssrn.com/sol3/papers.cfm?abstract_id=2801251; Andrew D. Selbst, Disparate Impact in Big Data Policing __ (forthcoming 2017), available at https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2819182 5 See, e.g., David Lazer, et al., The Parable of Google Flu: Traps in Big Data Analysis, 343 SCI. 1203 (2014). 6 See, e.g., Jamie Condliffe, Algorithms Probably Caused a Flash Crash of the British Pound, MIT TECHNOLOGY REV. (Oct. 7, 2016), https://www.technologyreview.com/s/602586/algorithms-probably-caused-a-flash-crashof-the-british-pound/; Curtis E.A. Karnow, The Application of Traditional Tort Theory to Embodied Machine Intelligence, in ROBOT LAW 51, 57-58 (Ryan Calo, A. Michael Froomkin & Ian Kerr, eds. 2016) (discussing unpredictability in robotics). 7 Joshua A. Kroll, et al., Accountable Algorithms, 165 U. PA. L. REV. __ (forthcoming 2017), available at https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2765268; Danielle Keats Citron & Frank Pasquale, The Scored Society: Due Process for Automated Predictions, 89 WASH. L. REV. 1, 18-27 (2014). 8 Kiel Brennan-Marquez, Plausible Cause 70 VAND. L. REV __, AT *8 (forthcoming 2017), available at https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2827733 1 4 Regulating Inscrutable Systems – DRAFT 3/20/2017 inscrutable algorithms to explain themselves.9 In a way, this makes sense. Explanation is at the core of law. Judges write opinions to explain themselves. Litigators explain their theories of the case. Administrative agencies write long explanations to accompany new rules and regulations. Explanations are the difference between authority vested in a decision and authority vested in a person or system.10 But the actual role of explanation in law is complicated, and without more specificity, the call can do more harm than good. Machines and humans have different strengths. People have “common sense” and evolutionary heuristics while machines lack human biases and can handle much more complexity.11 It is precisely the improvement in accuracy, often derived from complexity, that leads decision-makers to use machine learning systems in the first place.12 But accuracy often comes at a price. There is a commonly accepted tradeoff between the degree of interpretability and accuracy in machine learning systems.13 As a result, there will be cases in which asking for an explanation inherently makes predictions worse. And given what we understand about the flawed nature of human reasoning,14 we should be pretty clear on the normative rationales for any requirement that we reduce machine accuracy. At the same time, there is a vast and growing area of research on interpretability in machine learning, and the tradeoff is not necessarily a hard and fast rule. Therefore machine explanation might be useful to regulation. To be useful, machine explanation must correspond to the types of 9 General Data Protection Regulation (GDPR) (Regulation (EU) 2016/679) art 13(f)(2), 14(g)(2), 15 (1)(h) (requiring access to “meaningful information about the logic” of automated decisions); Brennan-Marquez, supra note 8, at 18-24 (drawing contrast between explanation and prediction); Kim, supra note 4, at (arguing for an employment discrimination standard of whether an adverse action was “because of” protected class membership, at 48-49 (arguing that employers should bear the burden of proving nondiscrimination based on the “meaning” of their models); 10 Frederick Schauer, Giving Reasons, 47 STAN. L. REV. 633, 636-37 (1995). 11 Andrew D. Selbst, A Mild Defense of Our New Machine Overlords, 70 VAND. L. REV. EN BANC __, at *14 (forthcoming 2017) (responding to Brennan-Marquez, supra note 8). 12 [cite] 13 Leo Breiman, Statistical Modeling: The Two Cultures (with comments and a rejoinder by the author) 16 STAT. SCI. 16, 199, 206-13 (2001). 14 DANIEL KAHNEMAN, THINKING FAST AND SLOW 199-200 (2011) (discussing the “narrative fallacy”), 224-25 (discussing the inconsistency of human judgment); Jerry Kang & Kristin Lane, Seeing Through Colorblindness: Implicit Bias and the Law, 58 UCLA L. REV. 465, 473 (2010) (discussing implicit bias). DRAFT – Please ask permission before citing or quoting 5 explanation that are useful for regulation. There can be infinitely many layered explanations for the same event. For example: a) The glass shattered because it hit the ground; b) It hit the ground because it was dropped; c) It was dropped because the holder was startled; d) It hit the ground because of gravity; e) It hit the ground because the ground is solid; f) It also shattered because glass is brittle; g) Glass is brittle because of its chemical composition, and so on. If someone asks for an “explanation” of why a glass shattered, it is necessary to infer the context of the question (e.g., someone is upset about their broken barware versus being in chemistry class) in order to give the correct answer. For the same reason, asking for machine explanation is incomplete without considering the specific role of explanation in regulation. If we understand that role and the emerging technology, then we might be able to design regulation to hold inscrutable systems accountable. This Article brings together the legal calls for explanation with the research on interpretability to examine how they align. The Article proceeds in four parts. Part I proposes a new model to understand the role of explanation in regulating decision-making systems. Called the What-How-Why model, it categorizes potential explanations for a decision into three levels: what happened, how it happened, and why the decision was made the way it was. In order to evaluate a decision-making system under a particular legal regime, it is necessary to access reasons from all three layers, but only those parts of the layers relevant to the particular normative concern. After explaining the What-How-Why model, the Part explains the role of inscrutability as breaking the model, and the role of regulation-by-explanation as seeking to repair it. Part II examines two existing legal frameworks of regulation by explanation. The first is credit scoring. Machine learning systems are not the first inscrutable systems. Statistical credit-scoring is similarly difficult to understand, and was regulated in the 1970s to require adverse action notices with “statement[s] of reasons” for the adverse action. The second is the portions of the General Data Protection Regulation (GDPR)—the recent data protection legislation in the European Union—that address automated decisionmaking. Unlike credit scoring law, the GDPR seeks “meaningful information about the logic” of automated decisions rather than what happened in an individual case. The Part argues that depending on interpretation, “meaningful information” could correspond perfectly to the failure of the What-How-Why model caused by inscrutability. Part III canvasses the computer science literature on interpretability. There are essentially three types of approaches: (1) purposefully selecting a sufficiently small number of variables and machine learning method to ensure that the resulting model is simple enough for a human to interpret; (2) adopting special methods that manage the trade-off between model 6 Regulating Inscrutable Systems – DRAFT 3/20/2017 complexity and performance or attempt reproduce complex models in simpler terms, and (3) forgoing any attempt a constraining the complexity of the model, but producing post hoc explanations for specific decisions by examining how changes to the input affect the model’s output. Part IV brings it all together. Individual technical approaches discussed in Part III and combinations of them may aid in compliance with the legal regimes of explanation. But there will be some types of algorithms that are unavoidably complex. If we require explanation, those algorithms will be entirely off limits. The Article concludes with a call for exploring types of regulation that do not require explanation in parallel to those that do. Without such regulation, we risk losing the societal benefits of some of the most advanced and valuable technology research to date. We should explore whether it is possible to avoid such a fate. I. WHAT-HOW-WHY AND THE PROBLEM OF INSCRUTABILITY Scholarly and policy debates over how to regulate a world controlled by algorithms have been mired in difficult questions about how to observe, audit, access, or understand those algorithms. This is at least partially because writing and speaking about this subject is tricky. Various terms have been used in the midst of describing the difficulties faced in regulating algorithms specifically: “black box,”15 “transparent/transparency,”16 “opaque/opacity,”17 “secret/secrecy,”18 “cause/causality/causation,”19 “foreseeability,”20 “explanation,”21 “intelligible/intelligibility,”22 See. e.g., FRANK PASQUALE, The BLACK BOX SOCIETY (2015). See, e.g., Citron & Pasquale, supra note 7, passim; Tal Z. Zarsky, Transparent Predictions, 2013 U. ILL. L. REV. 1503 (2013). 17 Jenna Burrell, How the Machine “Thinks”: Understanding Opacity in Machine Learning Algorithms, Big Data & Soc. 1, Jan.-Jun. 2016; Roger Allan Ford & W. Nicholson Price II, Privacy and Accountability in Black-Box Medicine, 23 MICH. TELECOMM. & TECH. L. REV. 1, 1112 (2016); Zarsky, supra note 16, passim. 18 See, e.g., Brenda Reddix-Smalls, Credit Scoring and Trade Secrecy: An Algorithmic Quagmire or How the Lack of Transparency in Complex Financial Models Scuttled the Finance Market, 12 U.C. DAVIS BUS. L. J. 87 (2011); Frank Pasquale, Restoring Transparency to Automated Authority, 9 J. TELECOMM. & HIGH TECH. L. 235, 237 (2011). 19 See, e.g., Kim, supra note 4, at *18. 20 See, e.g., Karnow, supra note 6, passim. 21 See, e.g., Bryce Goodman & Seth Flaxman, European Union Regulations On Algorithmic Decision-Making And A “Right To Explanation” (2016), https://arxiv.org/abs/1606.08813. 22 See, e.g., Brennan-Marquez, supra note 8, at *5. 15 16 DRAFT – Please ask permission before citing or quoting 7 “information/being informed,”23 “knowledge/knowing,”24 and “prediction/predictive” (used in opposition to explanation)25 among others. These terms all have overlapping meanings, whether described as synonyms or antonyms. For example, opacity is often seen as a synonym for secrecy,26 antonym for transparency,27 or antonym for understanding.28 The overlap in all these terms obscures the differences between the concepts, which need disentangling before the topic of regulation can even be addressed.29 In this Part, we attempt to avoid the confusion by introducing a new term30 and new abstraction to the debate. The new term is “inscrutability” and the new abstraction we call the What-How-Why model of understanding systems. The What-How-Why model is a way of thinking about the different layers of understanding necessary to regulate a decision system: what is does in the individual case, how it works generally, and why it works that way. Within that framework, we are able explain why the true difficulty in regulating machine learning algorithms comes from their inscrutability, and how that interferes with our normal regulatory processes. A. The What-How-Why Model Explanation is at the heart of the legal system. Judges write opinions. Litigators explain the facts. Administrative agencies give reasons for their rules. The concept of due process, complicated at it is, is essentially all about 23 See, e.g., Sandra Wachter, et al., Why a Right to Explanation of Automated Decision-Making Does Not Exist in the General Data Protection Regulation, at 26, https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2903469. 24 Mike Ananny & Kate Crawford, Seeing Without Knowing: Limitations of the Transparency Ideal and Its Application to Algorithmic Accountability, NEW MEDIA & SOCIETY 1 (2016); Mireille Hildebrandt, Profiling: From Data to Knowledge, 30 DATENSCHUTZ UND DATENSICHERHEIT 548 (2006). 25 Brennan-Marquez, supra note 8, at *18. 26 Burrell, supra note 17, at 3-4. 27 Zarsky, supra note 16; Ford & Price, supra note 17. 28 Burrell, supra note 17, at 4-5. 29 Moreover, most of these terms may be permanently confusing because they are tethered to the general meanings of the words. Some of this may be the ubiquity of sight metaphors in our vocabulary for knowledge and understanding. Even aside from “transparent” and “opaque,” some examples are “double blind study,” “in the dark,” “do you see what I mean?,” and even the cartoon image of a light bulb over someone’s head. The list goes on – virtually every way we describe knowledge or understanding turns out to be a sight metaphor. 30 While if the problem is that no one can agree on the right terminology, it is not clear that the solution is to provide a whole new term, cf. Randall Munroe, Standards, XKCD, https://xkcd.com/927/, we think it is important anyway. 8 Regulating Inscrutable Systems – DRAFT 3/20/2017 explanations, from predictability and vagueness doctrine,31 to reviewability of agency orders.32 Almost universally in regulating decisionmaking, and especially in liability regimes, the core questions turn on “why” an outcome occurred. In typical, human-driven decisions, a person makes a decision, then describes the one or more motives behind that decision, at which point another person evaluates the legality of that decision by testing it against the rationales permitted by statute.33 That chain of events depends on the second person understanding the first. If the first refuses to or cannot explain, then the ability of the second person to test for normative concerns is diminished. But asking for explanation alone is underspecified. An explanation is an infinitely varied concept, and there can be any number of valid explanations for a given phenomenon. In the famous story, a priest asked the notorious bank robber Willie Sutton: “Why do you rob banks?”, to which he replied “Because that’s where the money is.”34 The example is an illustration of the context-specific nature of explanation.35 Sutton’s explanation is in fact a valid explanation, but only to a question of the form “Why do you rob banks, not laundromats?” Surely the priest was asking something more like “Why do you rob banks, rather than obey the law?” This example demonstrates that explanations are tied to specific contexts. Law seeks not just an explanation, but the best explanation for its purposes.36 When the legal system seeks explanation of facts, those explanations are “inherently contrastive.”37 “Why X, rather than Y?” In casual conversation, this is true as well, but people often leave Y unspecified because it can be inferred from social context, or corrected after a misunderstanding. The specific explanations the law seeks are tied to the normative [cite] [cite] 33 See Andrew Verstein, The Jurisprudence of Mixed Motives, 126 YALE L.J. __, at App’x B (forthcoming 2017) (listing the vast breadth of mixed motive cases). 34 Philip Kitcher, Explanatory Unification and the Causal Structure of the World, in SCIENTIFIC EXPLANATION 410, 414 (Philip Kitcher, Wesley C. Salmon, eds. 1989) 35 W. Bradley Wendel, Explanation in Legal Scholarship: The Inferential Structure of Doctrinal Legal Analysis, 96 CORNELL L. REV. 1035, 1059 (2011). 36 Ronald J. Allen & Michael Pardo, Juridical Proof and the Best Explanation, 27 L. & PHIL. 223, 229-33 (2008); John Josephson, On the Proof Dynamics of Inference to the Best Explanation, 22 CARDOZO L. REV.. 1621, 1622-24 (2001). 37 Wendel, supra note 35, at 1049. 31 32 DRAFT – Please ask permission before citing or quoting 9 goals of the law. If a person is on trial for murder, the relevant question might be “Why did you, the defendant, kill the victim?,” by which the court means “…instead of not killing someone.” There, affirmative defenses can all be used as explanations that will satisfy the law. But if the defendant answers “because his brother was not available,” and that he has a vendetta against the family, it would be a valid explanation, but not a valid legal explanation. The availability or unavailability of legal explanation then, is tied directly to the normative goals enshrined in law. Now consider an arbitrary system that makes consequential decisions about people. The system could involve humans or machines. It could make decisions about credit scoring, lending, hiring, renting, policing, or any other item. At this high level, the important question for regulating it is whether it can offer an explanation of what it is doing and why that matches up with the normative concerns that the law in question seeks to vindicate. If the system were a black box, what would the legal system have to answer in order to regulate it? We argue that the questions one might ask of a black box system can be separated into three groups that we call the What-How-Why model of explanation: 1) What happened in an individual case? 2) How do the decisions get made generally? 3) Why are the decisions made that way? To answer the What question, the system would have to provide reasons for a specific outcome. For a given input, what led to the output? The kinds of explanations given in response to this question will vary. The simplest is to explain what features were examined. This helps a person verify that the input data is correct at least. For example, a person could verify her credit history after being denied credit. Another possibility is a list of factors that factored into the decision, or the most important factors by some measure. Such an explanation could be further elaborated by explaining a minimum standard or how the requirements were not met, if applicable. To continue with a credit example, possible explanations for one of the factors could be “length of residence” or “not in residence long enough” or “not in residence one year.” If no factor was clearly more important than the others, a possible explanation is that the applicant “just missed,” and an improvement of any of the variables taken into account would help. Overall, the important thing to consider about the What is that it is an accounting of a single result. The How level sits above the What. It is the set of rules that govern the decision process. Like at the What level, explanations at the How level are descriptive and take several forms. They could be descriptions of the logic of the decisionmaking process: filters, decision trees, regressions, or other decision logic. Another could be an explanation of how all or some the inputs map to their respective outputs, or how changes in inputs change 10 Regulating Inscrutable Systems – DRAFT 3/20/2017 the outputs; such as can be achieved through input/output testing. How level descriptions need not be full descriptions or mappings; partial ones can be useful as well. What is important about the How level is that it is functional information about more than the single instance. The What and How levels are related. Specifically, if someone knows the complete How level description, he will be able to determine all the individual outcomes on the What level. And the number of instances of individual explanations required to get from the What level to the How level is not fixed; the important point is whether there are enough instances to find a pattern or say something useful about the operation of the system. The top-most level of the abstraction is the Why. Explanations here may take the form of the ultimate goals of the decisionmaking process, decisions about what technology to implement, if any, intentional or unintentional assumptions made by the modelers, how input data to the decision are collected or chosen, and whether the designers of the black box considered certain normative goals laid out in law or policy, and how they addressed them.38 The Why level comprises everything outside the black box, about the black box. The Why level is often normative, but need not be so. For a machine vision system involved in facial recognition, the programmers will want to know a great deal of Why level information— goals and assumptions for example—just to test out and improve different detection methods. But the Why level also functions at the level of policy. The explanations provided here can test for society’s various normative requirements of decision systems: fairness, privacy, predictability, safety, and reviewability, among other concerns. As described above, though, to access the Why level, there much be a question in mind. Only once the normative concern is identified, then, can explanations be useful. Importantly, because the Why level is entirely outside the box looking in, nothing can be explained by reference to things going on inside the box. In a sense, that is the division implied by the very concept of the black box.39 There is a set of descriptive explanations internal to the 38 See, e.g., Solon Barocas & Andrew D. Selbst, Big Data’s Disparate Impact, 104 CALIF L. REV. 671, 677-692 (2016); 39 There is an irony to this use of the term “black box.” Engineering disciplines generally use the concept of a black box for almost exactly the opposite meaning. Whereas when referring to an algorithm as a black box implies mystery about its operation, see, e.g. PASQUALE, supra note 15, at 4, the other meaning is an abstraction of a system we so thoroughly understand that we do not care what is happening inside. As an example, consider the function “square(x).” In English, it means for every input x, the output is x2. In code, we do not know how square(x) is defined. It could be “square(x) DRAFT – Please ask permission before citing or quoting 11 operation of the box itself, and a set of (mostly) normative explanations about the design of the box, that by necessity are external. To illustrate with a simple example, imagine a black box, inside of which is an algorithm that filters résumés by whether or not the applicant went to Harvard Law School,40 or messes up her application. If the black box was only open at the What level, it means that a person who applies for a job would be able to tell that she was rejected because either “she went to Columbia” or “did not go to Harvard” Alternatively, she might be told she had an incomplete application or was rejected over a typo in the cover letter. All these reasons, singularly or in combination, give the applicant reasons for not getting the job, but they tell her very little of import about the system. This is true even in the case where she is told both “not Harvard” and “typo”, because she does not have reason to know that those are the entire basis of the decision. The How level explanations look different. In this case, the options are offering one or both of the minimum rules: Harvard only, and no sloppiness on the application. Someone learning one of those rules will both know and understand how the system works, at least partially, and someone knowing both knows and understands everything about the internal workings of the system. Why level explanations cannot be provided by the black box, but = x * x” or “square(x) = exp(double(log(x)))” When computer scientists say that “square(x)” is a black box, they mean that they are disciplining themselves not to ask what is inside the function, but instead just to take the function as it is. See HAROLD ABELSON & GERALD SUSSMAN WITH JULIE SUSSMAN, STRUCTURE AND INTERPRETATION OF COMPUTER PROGRAMS. Another example is treating a light bulb as a resistor inside a circuit. As long as the light bulb has a resistance R when running current through it, it will obey Ohm’s Law: V=IR, where the voltage V is the input, and the current I is the output, and the constant R is the function of interest. This is a black box, because it enables an engineer to entirely ignore the complicated physics and chemistry of a light bulb, as the function R describes everything we need to know. ANANT AGARWAL & JEFFREY H. LANG, FOUNDATIONS OF ANALOG AND DIGITAL ELECTRONIC CIRCUITS 6-7 (2005). In this meaning, we only open the black box if what is coming out is different than our supposedly complete understanding, and we need to troubleshoot. The Science, Technology, and Society literature generally understands “black box” this way too. See, e.g., WIEBE E. BIJKER, OF BICYCLES, BAKELITES, AND BULBS: TOWARD A THEORY OF SOCIOTECHNICAL CHANGE 77 (1997). But the meaning of “black box” in law, policy, and popular consciousness means a decisionmaking engine that is entirely mysterious. The differences between the two concepts are the same secrecy and inscrutability that are driving the scholarly and policy discussion about algorithms. See Part I.B, infra. 40 This is the “algorithm” deployed by the fictional law firm Pearson Specter Litt (f/k/a Pearson Hardman) in the TV show Suits. Suits: Pilot, (USA television broadcast June 23, 2011) (Harvey Spector to Mike Ross: “Unfortunately, we only hire from Harvard.”) References to this filter are made through the show’s run. 12 Regulating Inscrutable Systems – DRAFT 3/20/2017 they are about the black box. As discussed above, though “filter for Harvard and careful application” is the complete description of the system, knowing that fact offers no insight about why the system was designed that way or whether that is an acceptable rule. For that, it might need to know “why Harvard” or “why a single school” or “why put so much weight on typos to the exclusion of merits?” The Why layer connects the description of the black box from the How layer to the outside world. Suppose one wanted to test for discrimination, and had the full description. “Harvard only” will likely have a disparate impact, so an important explanation to get from the Why layer is whether the filter was chosen for a reason consistently with a business necessity defense.41 Or alternatively, to test for intentional discrimination, one would need to know whether someone chose the Harvard filter as a proxy for a discriminatory purpose. Suppose that from the How layer, one does not have the rule, but only has access to the inputs and outputs. Well, because the example is a simple one, it is possible to detect the rule. But even without that intermediate step, given enough input-output pairs, there would be enough information to discover whether there is prima facie disparate impact, and then from there, the problem moves on to business necessity as above. The Why reasons allow regulators to evaluate the legality and propriety of the box, and the How reasons give the tools to measure. One important thing to note is that there are an infinite number of possible explanations in this, and all, layers. The only way to determine the appropriate explanation is to start with the particular question that requires an answer and work from there. The relative simplicity of this example likely obscures the value of the What-How-Why model. We intuitively know how to ask all these questions and perform evaluations about a Harvard-only filter. But the purpose of the abstraction is to allow us to think specifically about the contributions of each layer of explanation, so when they break down, we can better understand why. It might not be surprising, then, that machine learning algorithms break this model. B. The Effects of Inscrutability Machine learning is a process of learning by example. Rather than programming computers by hand with explicit rules, describing, machine learning relies on pattern recognition algorithms and a large set of examples 41 See 42 U.S.C. 2000e-2(k). DRAFT – Please ask permission before citing or quoting 13 to independently uncover relationships. For example, when learning how to distinguish between spam and legitimate email, machine learning relies on pattern recognition algorithms and a large set of examples to independently uncover relationships between properties of email and whether these emails are spam. Machine learning is crucial to spam detection because the full set of rules that would be necessary to effectively instruct a computer to recognize spam would be impossible for a human to enumerate manually. And spam detection is among the simpler common tasks to rely on machine learning. Machine learning can learn many more—and far more complex—rules than humans would be able to work on our their own. The power of machine learning lies not only in its ability to relieve programmers of the difficult task of producing explicit instructions for computers, but it its capacity to learn subtle relationships in data that humans might overlook or cannot recognize. The property of machine learning algorithms that makes them difficult to regulate is that they are inscrutable. An inscrutable system is one in which a person could look at the code, or the entire corpus of the decision rules, and still not understand what the system is doing. What actually is inscrutable will be different for different people because of differences in knowledge specialization,42 but for the purposes of this discussion, we can focus on that which is inscrutable to the technically competent reasonable person. An inscrutable system defies human capabilities of reasoning and pattern matching. This inscrutability is a natural fact of machine learning systems derived from their inherent complexity.43 As Jenna Burrell has noted: Though a machine learning algorithm can be implemented simply in such a way that its logic is almost fully comprehensible, in practice, such an instance is unlikely to be particularly useful. Machine learning models that prove useful (specifically, in terms of the ‘accuracy’ of classification) possess a degree of Burrell, supra note 17, at 4 (discussing “opacity as technical illiteracy”). We mostly agree with Burrell’s way of thinking through the problem, but of her three definitions of opacity—secrecy, technical illiteracy, and some inherent property of algorithms that relates to complexity—we believe that secrecy is different in kind. What we call “inscrutability” maps on to the latter two. 43 Id. at 5. (“[T]here are certain challenges of scale and complexity that are distinctive to machine learning algorithms. These challenges relate not simply to total number of lines or pages of code, the number of team members on the engineering team, and the multitude of interlinkages between modules or subroutines. These are challenges not just of reading and comprehending code, but being able to understand the algorithm in action, operating on data.”) 42 14 Regulating Inscrutable Systems – DRAFT 3/20/2017 unavoidable complexity.44 Inscrutable systems are difficult to regulate because they break the What-How-Why model. In order for the model to work, the How layer must provide some functional way of understanding the system, either wholly or partially. This can be enough of a data mapping for humans to fashion a rough rule, or it could be a complete description of the system that is simple enough that humans can understand. If it is a partial mapping or description, it is only useful for regulation if it happens to answer a question that corresponds with what one seeks from the Why layer, but in those cases, it will work. By definition, though, an inscrutable system confounds description. While input-output testing is possible, and a complete list of the rules could even be available, humans cannot process them meaningfully, and thus cannot connect them to the normative concerns that motivate regulation. Compare the Harvard Law hypothetical to an inscrutable system. In the Harvard Law hypothetical, the How layer provided either a functional description of the logic or enough data for an observer to get the gist of what was happening. As a result, it was trivial to know what questions to ask regarding discrimination or other values. But if instead the results seemed incomprehensible, almost random, then it would be impossible to figure out which of the Why level explanations to even seek. How would one determine the propriety of the specific assumptions that went into the model, if one cannot understand the effects of those assumptions? Note, though, that the issue is not that there is no information available from an inscrutable system. For example, based on testing the data alone, it is possible to assess prima facie disparate impact based on the fourfifths rule.45 But that is not enough to find liability, because we always to want to know why the decisionmaker chose a specific rule. When faced with an inscrutable system, one might think that the burden is on the decisionmaker to usefully explain why, but that is often not the case.46 Thus, the typical operation of the What-How-Why model is disrupted by inscrutability. The How and the Why can no longer connect, and the Id. The “four-fifths rule” is, in most contexts, the showing required to demonstrate prime facie disparate impact: “A selection rate for any race, sex, or ethnic group which is less than four-fifths . . . of the rate for the group with the highest rate will generally be regarded . . . as evidence of adverse impact.” Uniform Guidelines on Employment Selection Procedures, 29 C.F.R. § 1607.4(D) (2015). 46 See Barocas & Selbst, supra note 38, at 701-12. 44 45 DRAFT – Please ask permission before citing or quoting 15 system cannot be tested against the normative concerns of the law. In the past, scholars and policymakers have called for algorithmic transparency to remedy the difficulties with regulating black-box algorithms, arguing we need to see the algorithm so that we can judge its fairness or safety.47 The What-How-Why model makes clear why transparency into systems of decisionmaking is important, but incomplete. Transparency requires information to be provided at the What level and How level. At the What level, transparency might require access to the inputs to an algorithm that bear on a specific decision. At the How level, transparency might make require access to the algorithm for input-output testing48 or even access to the code itself.49 But requirement of transparency assumes that information will lead to accountability. As Mike Ananny and Kate Crawford observe: “Transparency concerns are commonly driven by a certain chain of logic: observation produces insights which create the knowledge required to govern and hold systems accountable.”50 When the rules are inscrutable, the insights are missing, and therefore so is accountability. The problems with calling for algorithms to explain themselves is similarly demonstrated by the model. When the system is simple, and the explanation is therefore available but hidden, a call for machines to explain themselves is little more than a call for transparency at the How level. But with an inscrutable system instead, the call is either impossible by definition, if a total explanation is sought, or is underspecified. At what level should the system explain itself? At some level, asking the machine to “explain itself” is impossible, because it cannot answer questions about its own design, and without knowledge from the Why level, normative regulations cannot be met. But which of the various explanations is relevant at the Why level, and whether that can be satisfied by the available partial explanations of the How is determined by the particulars of the normative concern. Explanations in the abstract do no good. There are now two realities to reconcile. One is the explanatory What-How-Why model of regulating systems, and the other is inscrutability. The next Part addresses two attempts within existing law to reconcile them. See generally PASQUALE, supra note 15; see also Mikella Hurley & Julius Adebayo, Credit Scoring in the Age of Big Data, 18 YALE J.L. & TECH. 148, 196-98 (2016); Reddix-Smalls, supra note 18. 48 Citron & Pasquale, supra note 7, at 24. 49 Zarsky, supra note 16, at 1524, 1539. 50 Ananny & Crawford, supra note 24, at 2. 47 16 Regulating Inscrutable Systems – DRAFT 3/20/2017 II. EXISTING REGULATION OF INSCRUTABLE SYSTEMS The difficulties that arise in regulating automated decisionmaking comes from the systems’ complexity and resulting inscrutability, the fact of automation. And though much of the current concern over inscrutable systems stems from the growing importance of machine learning, inscrutable systems predate the technique. As a result, legislation already exists that seeks to regulate inscrutable systems by having the systems explain themselves. Credit scoring is an example of an inscrutable system that predates machine learning. Statistical credit scoring systems take information about consumers as inputs, give the inputs certain point values, to obtain a total score, and then make decisions based on that score. Credit scores and decisions made based on credit scores are regulated primarily by two statutes: the Fair Credit Reporting Act (FCRA)51 and the Equal Credit Opportunity Act (ECOA).52 While they have historically had different approaches, both statutes currently require adverse action notices that spell out the factors that lead to a credit denial or other adverse action. An adverse action notice prompts a consumer to check her credit report for errors,53 and a “[s]tatement of specific reasons”54 aims to prevent intentional discrimination and serves consumer education about how the processes work so they can better plan future behavior. Regarding automated decisions, the European Union tried a different approach. The General Data Protection Regulation (GDPR) requires data subjects to have access to “meaningful information about the logic involved” in any automated decisionmaking that significantly affects them.55 Yet, as the law has not yet taken effect, the import and proper interpretation of the requirement remain unclear. The Part argues that if interpreted correctly, the GDPR could mostly get it right. It goes farther than FCRA and ECOA do as far as using explanation itself as a tool for accountability, but the focus on “meaningful information” could place the 15 U.S.C. § 1681, et seq. Id. § 1691, et seq. 53 [cite] 54 12 C.F.R. § 202.9 (b)(2). 55 General Data Protection Regulation (GDPR) (Regulation (EU) 2016/679) art. 15 (1)(h) (GDPR); see also id. art. 13(e)(2), 14(f)(2). The Data Protection Directive, passed in 1995 also requires access to information about automated decisions, but the word “meaningful” was added to the GDPR. See [Data Protection Directive art. 12]. 51 52 DRAFT – Please ask permission before citing or quoting 17 onus on the exact kind of functional description missing from the How layer of inscrutable systems. But the GDPR also shares common features with the prior approaches that ay limit its effectiveness—specifically access only to the How layer, and not necessarily the Why. A. Credit Scoring Regulations 1. Introduction to FCRA, ECOA, and Regulation B Before the 1950s, credit was not a large part of daily life in the United States.56 During the 1950s and 1960s, as credit became more popular, small credit bureaus had begun to arise, amassing large quantities of information about prospective credit applicants.57 These bureaus would both “track peoples’ names, addresses, and loan information” and “scour newspapers for notices of arrests, promotions, and marriages.”58 Once credit became more common, credit decisions were guided largely by the “three C’s of credit”: capacity, character, and collateral.59 By the late 1960s, the modern credit industry had begun to emerge, relying on amassed information and statistical models to predict creditworthiness.60 While credit scoring was seen as a fairer, more objective way to make credit determinations, consumers were nonetheless worried. Credit reports often contained incorrect or outdated information that credit reporting agencies (CRAs; e.g., Experian, Transunion, and Equifax) had no incentive to correct.61 The industry was “secretive and enigmatic,”62 and consumers had no idea who had access to their information or to what uses it was being put.63 Thus, in 1970, Congress passed FCRA64 to begin to rein in the unregulated credit industry. The official purpose of FCRA is “to require [cite] ROBINSON + YU, KNOWING THE SCORE: NEW DATA, UNDERWRITING, AND MARKETING IN THE CONSUMER CREDIT MARKETPLACE 26 (2014), https://www.teamupturn.com/static/files/Knowing_the_Score_Oct_2014_v1_1.pdf. 58 Id. 59 Winnie F. Taylor, Meeting the Equal Credit Opportunity Act’s Specificity Requirement: Judgmental and Statistical Scoring Systems, 29 BUFF. L. REV. 73, 74 (1980). 60 ROBINSON + YU, supra note 57, at 26. 61 NATIONAL CONSUMER LAW CENTER, FAIR CREDIT REPORTING § 1.4.3 (8th ed. 2013) [hereinafter NCLC, FAIR CREDIT REPORTING]. 62 Lea Shepard, Toward A Stronger Financial History Antidiscrimination Norm, 53 B.C. L. REV. 1695, 1745 (2012). 63 NCLC, FAIR CREDIT REPORTING, supra note 61, §§ 1.4.2-1.4.4. 64 Pub. L. No. 91-508, tit. VI, 84 Stat. 1127 (1970) (codified as amended at 15 U.S.C. §§ 1681-1681x). 56 57 18 Regulating Inscrutable Systems – DRAFT 3/20/2017 that consumer reporting agencies adopt reasonable procedures for meeting the needs of commerce for consumer credit, personnel, insurance, and other information in a manner which is fair and equitable to the consumer, with regard to the confidentiality, accuracy, relevancy, and proper utilization of such information.”65 To accomplish this, FCRA requires that CRAs enlist “reasonable procedures to assure maximum possible accuracy” of the information it holds,66 allow consumers to access to their credit reports67 and investigate and resolve disputes with either a deletion, correction or annotation.68 FCRA also limits to whom and for what credit reports can be disclosed.69 Finally, FCRA requires that any person who takes an adverse action relying on information obtained from a CRA must issue the consumer an adverse action notice listing information about, among other things, the decision itself and her rights under the chapter.70 By 1974, credit was becoming more systematized and less characterdriven, but women were still being denied credit on the basis of prejudices related to a woman’s perceived role in the home and society. For example, creditors would often deny credit to a married woman in her own name.71 Thus, pushed by women’s groups, Congress passed ECOA in 1974.72 Coverage was expanded soon after, with the Equal Credit Opportunity Act Amendments of 1976,73 when it became immediately apparent that groups other than women also faced credit discrimination. ECOA now prohibits discrimination in credit decisions on the basis of race, color, religion, national origin, sex, marital status, age (for adults), receipt of public assistance income, or exercise in good faith of the rights guaranteed under the Consumer Credit Protection Act.74 As both a consumer protection and anti-discrimination statue, ECOA has two purposes.75 It was designed primarily to root out 15 U.S.C. § 1681(b). Id. § 1681e(b)(2). 67 Id. § 1681g. 68 Id. § 1681i. 69 Id. § 1681b. 70 Id. § 1681m. 71 Id. at 74 n.7 (citing NATIONAL COMM’N ON CONSUMER FINANCE, CONSUMER CREDIT IN THE UNITED STATES 152-53 (1972)). 72 Pub. L. No. 93-495, §§ 501-503 88 Stat. 1521 (1974) (codified at 15 U.S.C. § 1691(e)) 73 15 U.S.C. §§ 1691-1691f 74 15 U.S.C. § 1691. 75 John H. Matheson, The Equal Credit Opportunity Act: A Functional Failure, 21 HARV. J. LEG. 371, 372 (1984); Taylor, supra note 59, at 83. 65 66 DRAFT – Please ask permission before citing or quoting 19 discriminatory credit decisions,76 which should only be based on predictions of an applicant’s ability and willingness to repay a loan.77 Thus, ECOA is more like other anti-discrimination legislation than other credit laws than FCRA or the Truth-in-Lending Act.78 The second purpose was to provide a “valuable educational benefit.”79 “For the first time, federal legislation afforded rejected credit applicants an automatic right to discover why adverse action was taken.”80 That is, for the first time, rejected credit applicants were given a right to explanation.81 ECOA’s specific innovation was the requirement that a creditor that takes an adverse action against an applicant give a statement of “specific reasons” for the adverse action.82 As stated in the Senate Report accompanying the 1976 amendments, [t]he requirement that creditors give reasons for adverse action is . . . a strong and necessary adjunct to the antidiscrimination purpose of the legislation, for only if creditors know they must explain their decisions will they effectively be discouraged from discriminatory practices.83 The Senate Report further explains: Yet this requirement fulfills a broader need: rejected credit applicants will now be able to learn where and how their credit status is deficient and this information should have a pervasive and valuable educational benefit. Instead of being told only that they do not meet a particular creditor’s standards, consumers particularly should benefit from knowing, for example, that the reason for the denial is their short residence in the area, or their recent change of employment, or their already over-extended financial situation.84 Taylor, supra note 59, at 83. Id. at 76. 78 Susan Smith Blakely, Credit Opportunity for Women: The Ecoa and Its Effects, 1981 WIS. L. REV. 655, 662 (1981). 79 Id. 80 Taylor, supra note 59, at 82. 81 FCRA did not include anything like a statement of reasons until 2010, when the Dodd-Frank Act amended it to include up to four key factors that make up the credit score with serves as the basis for an adverse action. See Pub. Law 111-203, 124 Stat. 1376 (2010) (Sec. 1100F “Use of Consumer Reports”)(adding a requirement to disclose “the information set forth in subparagraphs (B) through (E) of section 609(f)(1)”); 15 U.S.C. § 1681g(f)(1)(C). (including among the disclosures “all of the key factors that adversely affected the credit score of the consumer in the model used, the total number of which shall not exceed 4”). 82 15 U.S.C. § 1691(d)(3). 83 S. REP. 94-589, 4, 1976 U.S.C.C.A.N. 403, 406. 84 Id. 76 77 20 Regulating Inscrutable Systems – DRAFT 3/20/2017 Thus, the statement of reasons links the explanation to a consumer’s future action, so that she will know how to improve her credit. This legislative requirement was subsequently implemented by the Federal Reserve Board in “Regulation B,”85 which states that the “statement of reasons . . . must be specific and indicate the principal reason(s) for the adverse action.” The regulation also notes that “[s]tatements that the adverse action was based on the creditor’s internal standards or policies or that the applicant . . . failed to achieve a qualifying score on the creditors credit scoring system are insufficient.”86 The Federal Reserve Board offered further guidance in an official interpretive document that describes how creditors should aim to comply.87 This document hints at some of the difficulties with requiring explanation, and is discussed further below. Thus, after ECOA, when a consumer is denied credit, she gets some form of “explanation” that is designed both to prove the reasons were not discriminatory and to educate her as to the reasons themselves.88 Following the ECOA model, in 2010, the Dodd-Frank Act updated FCRA to require adverse action notices to include “all of the key factors that adversely affected the credit score of the consumer in the model used,” for an adverse determination involving a credit score.89 An appendix to Regulation B offers a sample notification forms designed to satisfy both the rule’s and FCRA’s notification requirements. The statement of reasons section of Sample From 1 offers twenty-four different options.90 Though it is not necessary to use the form, most 12 C.F.R. § 1002.1, et seq. (“Regulation B”). Id. § 1002.9 (b)(2). 87 Id. Pt. 1002, Supp. I Para. 9(b)(2) (“Official Interpretations”) 88 The Senate Report gives one further reason, though it treats this third reason as an extension of the second: “In those cases where the creditor may have acted on misinformation or inadequate information, the statement of reasons gives the applicant a chance to rectify the mistake.” S. REP. 94-589, 4, 1976 U.S.C.C.A.N. 403, 406. This ability to correct the input data is a central motivation behind FCRA, passed several years earlier. See 15 U.S.C. § 1681. A statement of reasons is not actually required to accomplish this goal. FCRA, unlike ECOA, does not require any statement of reasons for adverse actions. It only requires notice that the decision was based in part on the credit report, id. § 1681m, and assorted other information designed to help the consumer correct errors according to procedures in another part of the statute. See id. § 1681i. FCRA is different from ECOA in that it never requires credit reporting agencies to reveal how they use the information they have. Citron & Pasquale, supra note 7, at 17. 89 15 U.S.C. § 1681m(a)(2)(B) (requiring disclosure “of the information set forth in subparagraphs (B) through (E) of section 1681g(f)(1) of this title); id. § 1681g(f)(1)(C). 90 12 C.F.R. Pt. 1002, App. C (“Sample Form”). The options are: __Credit application incomplete 85 86 DRAFT – Please ask permission before citing or quoting 21 creditors tend to report reasons contained on that form believing it to be a safe harbor.91 2. The Use of Explanation in Credit Scoring Regulation The original version of FCRA did not seek explanations for inscrutable credit scores. This is likely because FCRA was not concerned with how decisions were being made, but rather with the relatively new phenomenon of amassing large quantities of information in the first place. FCRA was, after all, “the first information privacy legislation in the United States.”92 Moreover, inscrutable credit scoring systems were not as dominant in 1970, with the majority of credit decisions still being made by human credit managers.93 Thus, the original goals of FCRA were more modest: ensuring that the information collected would only be used for permitted purposes, and that the information was accurate and up to date.94 __Insufficient number of credit references provided __Unacceptable type of credit references provided __Unable to verify credit references __Temporary or irregular employment __Unable to verify employment __Length of employment __Income insufficient for amount of credit requested __Excessive obligations in relation to income __Unable to verify income __Length of residence __Temporary residence __Unable to verify residence __No credit file __Limited credit experience __Poor credit performance with us __Delinquent past or present credit obligations with others __Collection action or judgment __Garnishment or attachment __Foreclosure or repossession __Bankruptcy __Number of recent inquiries on credit bureau report __Value or type of collateral not sufficient __Other, specify: ___ 91 [From conversations with CFPB; need source]. This is particularly ironic because it can sometimes lead to violations. For example, a creditor might use education information as a tool to infer income. The creditor will then report income, rather than education, because it is listed. 92 PRISCILLA M. REGAN, LEGISLATING PRIVACY: TECHNOLOGY, SOCIAL VALUES, AND PUBLIC POLICY 101 (1995). 93 [cite] 94 NCLC, FAIR CREDIT REPORTING, supra note 61, § 1.4.3. 22 Regulating Inscrutable Systems – DRAFT 3/20/2017 But four years later, ECOA started to aim at the question of how credit decisions were actually made. By 1974, forcing hidden intentions into the open was a common technique for addressing discrimination. Just one year before ECOA’s passage, McDonnell Douglas Corp. v. Green laid out the canonical Title VII burden-shifting framework, which requires a defendant to rebut a prima facie case of employment discrimination with a nondiscriminatory reason, and allows a plaintiff a chance to prove that the proffered reason is pretextual.95 Just two years before that, the Court in Griggs v. Duke Power Co.96 invented disparate impact doctrine, the purpose of which was arguably also to smoke out intentional discrimination where intent was hidden.97 Thus, ECOA sought the same goal—to force decisionmaking into the open in order to prevent discrimination. This approach can work for the most flagrant forms of intentional discrimination but less well in other cases. Forcing stated reasons into the open captures the most egregious form of intentional discrimination. Thus, no longer can creditors intentionally discriminate based on protected class directly, but they can use known proxies to get at the same result, if desired.98 And although Regulation B bars collection of protected class information,99 for credit systems that rely on face-to-face contact, that information is readily available.100 Even without the face-to-face contact, it can often be inferred from other data, such as a person’s name.101 If a creditor wants to discriminate, stating a made up reason is not that difficult for “judgmental” systems that rely on human credit managers rather than scoring.102 But that also makes judgmental credit systems easily subject to 95 411 U.S. 792, 805 (1973). The Supreme Court later found that a jury may presume that if all the employer had was pretext, that itself is evidence of discrimination. St. Mary’s Honor Ctr. v. Hicks, 509 U.S. 502, 511 (1993) (“The factfinder’s disbelief of the reasons put forward by the defendant (particularly if disbelief is accompanied by a suspicion of mendacity) may, together with the elements of the prima facie case, suffice to show intentional discrimination.”). 96 Griggs v. Duke Power Co., 401 U.S. 424, 431 (1971). 97 Richard A. Primus, Equal Protection and Disparate Impact: Round Three, 117 HARV. L. REV. 494, 520-21 (2003) (discussing the “evidentiary dragnet” theory of disparate impact). 98 Matheson, supra note 75, at 388. 99 12 C.F.R. § 1002.5 100 Matheson, supra note 75, at 387. 101 See Marianne Bertrand & Sendhil Mullainathan, Are Emily and Greg More Employable than Lakisha and Jamal? A Field Experiment on Labor Market Discrimination, 94 AMER. ECON. REV. 991 (2004). 102 “Judgmental” credit systems are older, human-driven, and subjective. Regulation B defines them as anything that is not a “statistical” system. 12 C.F.R. § 1002.2. DRAFT – Please ask permission before citing or quoting 23 lawsuits. Over the years, scoring systems have become more of the norm, partially because they are seen as more objective.103 Creditors using quantitative methods that do not expressly consider protected class membership are likely not engaged in intentional discrimination,104 thus satisfying the most basic purpose of ECOA. But the scoring systems might very well evince a disparate impact. Disparate impact doctrine, as developed under Title VII, states that if a facially neutral decision has a disproportionate adverse effect on a protected class, liability can be found for that reason, unless the decisionmaker can provide a legitimate business reason for the scoring system and no equally effective, but less discriminatory alternative exists.105 While ECOA does not expressly provide for a disparate impact theory of discrimination, it is very likely available.106 What is interesting, then, about ECOA’s affirmative statement of reasons, is that while for intentional discrimination, a consumer only needs to know that the decision was not made for an improper reason, knowing specifically which reasons it was made for becomes important for a disparate impact case. Thus, in a world of credit scoring, the demand for explanation can serve three different purposes. 1) preventing intentional discrimination, 2) permitting testing for disparate impact and subsequent correction, and 3) enabling consumers to determine how to improve their credit.107 So how does the statement of reasons work in practice? The advent of credit scoring systems introduced an element of inscrutability that was not present in the judgmental systems before it. Credit scoring can almost be thought of as proto-machine learning,108 in that the systems exhibit the See Taylor, supra note 59, at 119. Though they could indeed still use proxies. 105 [cite] [[This ignores the word “refuse,” but is probably the more common reading.] 106 The Supreme Court has not ruled that it is available, but most circuit courts that have considered it have permitted it. Hurley & Adebayo, supra note 47, at 193 (citing Golden v. City of Columbus, 404 F.3d 950, 963 (6th Cir. 2005)). In addition, the Supreme Court ruled in 2015 that disparate impact theory was cognizable in the Fair Housing Act, which also does not expressly provide for it. Texas Dep’t of Housing & Cmty. Affairs v. Inclusive Communities Project, Inc., 135 S. Ct. 2507, 2518 (2015). 107 Error-checking could be considered a fourth purpose. If the consumers knew which factors mattered, they could check that the records were correct. But because a credit history has only a relatively small number of variables, a consumer only needs notice that the credit history was accessed for error-checking purposes, not notice about which factors mattered, because she can check the whole report. See supra note 88. In the machine learning context, error-checking/validation becomes another reason for explanation. See Part III, infra. 108 This is just as inversion of the argument that data mining “scores” everything. See generally Citron & Pasquale, supra note 7. 103 104 24 Regulating Inscrutable Systems – DRAFT 3/20/2017 same reliance on statistics and sufficient complexity that they are hard to explain usefully. If an explanation given by the statement of reasons is too complex, the average credit applicant will not understand it, and the explanation’s purpose is defeated. The Federal Reserve Board recognized this problem, observing in its official interpretation that although all the principal reasons must be disclosed, “disclosure of more than four reasons is not likely to be helpful to the applicant.”109 The difficulty is that there will be situations where the complexity cannot be avoided in a faithful representation of the scoring system.110 Winnie Taylor described the problem well.111 She imagined a hypothetical credit scoring system with eight different variables, each assigned different point values, including whether an applicant owns or rents, whether he has a home phone, and what type of occupation he has, among other things. In a system like that, someone who comes up one point short will find himself with every factor listed as a “principal reason” for the denial. In one sense this has to be right, because a positive change in any factor at all would change the outcome. But in another sense, “can it be said that credit is denied because the applicant is a 48 year old farm worker who rents? Probably not,” even though the scoring system arrives at that output.112 Complexity comes from more than just the number of variables, however. Currently, creditors “may say ‘length of residence’ rather than “too short a period of residence.”113 That does not do much for consumer education. But the disclosure has to be structured this way because creditors cannot easily disclose a minimum standard. Some variables are nonmonotonic or otherwise fail to conform to clear patterns. In a real system Taylor described, applicants who have lived at their present address for less than six months are awarded 39 points, a level which they could not reach again until they had maintained the same residence for seven and one-half years. Furthermore, 109 12 C.F.R. Pt. 1002, Supp. I Para. 9(b)(2) (“Official Interpretations”). FCRA codified the same limitation. 15 U.S.C. 1681g(f)(1)(C). 110 The document also states that the “specific reasons . . . must relate to and accurately describe the factors actually considered or scored by a creditor,” “[a] creditor need not describe how or why a factor adversely affected an applicant, and “[i]f a creditor bases the . . . adverse action on a credit scoring system, the reasons disclosed must relate only to those factors actually scored in the system.” Id. 111 Taylor, supra note 59, at 105-107. 112 Id. at 107 (emphasis added). 113 12 C.F.R. Pt. 1002, Supp. I Para. 9(b)(2). DRAFT – Please ask permission before citing or quoting 25 applicants who have been residents for between six months and 1 year 5 months (30 points) are considered more creditworthy than those who have been residents for between 1 and 1/2 years and 3 years 5 months (27 points).114 If the creditor tried to explain simply, it would leave information out, but if the creditor were to explain in complete detail how the system worked, it would likely overwhelm a credit applicant. This is an equivalent problem to just providing transparency into a machine learning algorithm; transparency is not understanding.115 Thus, there are times where it is impossible to truly comply using the type of checklist that Regulation B models, based on complexity alone. The checklist aims to disclose the reason for a specific decision without explaining the system functionality, but the two are not always easily separable, and a reason given for the former will often be unsatisfying without the context provided by the latter. Suppose a credit scoring system exists that weighs home ownership incredibly highly. Suppose also that an applicant named Alice is denied credit, when she has above average point values for every attribute, but does not own a home. If Alice is given a list of reasons per ECOA, two possibilities exist. She could either be told that home ownership was the principal reason, or be told that attributes that are not home ownership are the principal reason, such as not being at her job long enough, or needing to pay down some debt. In some sense, the latter approach is just as correct. Tweaking all these levers could result in credit approval, and they are probably easier to fix than buying a home, so they serve the consumer education purpose. Both lists of reasons should therefore satisfy ECOA. But the latter missing the point of the question. Here, the primary factor was that the system was designed in such a way that home ownership was incredibly heavily weighted. But weighting of variables and other concerns of overall system functionality are not the kinds of explanation that ECOA seeks, so the real explanation will not fit the question asked.116 This is not to say that an explanation is always useless. As Taylor points out, some “reasons, such as ‘no credit file,’ ‘unable to verify income,’ and ‘we do not grant credit to any applicant on the terms and conditions you request’ are self explanatory, and appear to simultaneously provide the threshold educational benefit Congress intended for consumers and the Taylor, supra note 59, at 123. See Ananny & Crawford, supra note 24, at 7 (“Transparency can intentionally occlude.”) 116 [cite] 114 115 26 Regulating Inscrutable Systems – DRAFT 3/20/2017 brevity Congress intended to ease a creditor’s burdens.”117 There may be ways to design scoring systems or explanations so that at least in certain cases, the explanation will make sense to the applicant. Analogues to these kinds of constrained systems and case-specific explanations are the subject of Part III. But while some fixes are possible, it is doubtful that explanation will ever work to solve the entire problem. The discussion so far has addressed only descriptive concerns— What or How level. But ECOA sets out some specific normative concerns as well: anti-discrimination and predictability. ECOA envisions enforcement of these normative concerns through litigation.118 With a good enough understanding of the system, a person (in this case judge or jury) can test whether it violates the normative concerns. As discussed in Part I, relying on people to connect the normative concerns is the only way to scrutinize a system normatively. No decision system, whether machine- or human-driven, can self-examine; it cannot know anything about the choices that went into its own design. It cannot examine why certain features were chosen or emphasized, how the scoring was determined, or if the results are truly bizarre. But reasons provided under ECOA and FCRA are at the What level. The statement of reasons alone cannot ensure the normative outcomes the statute seeks to vindicate. But the complexity of this credit scoring system, even though it is low when compared to machine learning, makes the How layer impenetrable. Where the descriptive explanation is simple, it is easier for an external observer to relate it to the normative concern. For example, if a single variable is dominant in the outcome, rather than many interacting with each other, it is easier to check whether that variable is also a proxy for a protected class. Or if a variable is monotonic, rather than seemingly arbitrary or random, it is easier to explain to an applicant how to improve her credit score based on that variable in the future. But neither is true here, and the increased complexity therefore creates a gap in understanding that frustrates the ultimate purposes of the statute. Returning to Alice’s credit application, imagine that the reason home ownership is weighted so heavily was because in the 1970s, fewer women owned homes, and a designer of the system wanted to deny credit to women, who he thought should be getting their husbands’ approval for loans. This gets to the core of the intentional discrimination ECOA sought to curb. In this case, it would be correct to say that the reason for the 117 118 Taylor, supra note 59, at 97. [cite] DRAFT – Please ask permission before citing or quoting 27 ultimate result was discrimination, because it was the reason for the weighting that ultimately cost Alice the loan. ECOA, by itself, can never get this type of explanation from the system, because it only asks internal questions. But the complexity further compounds the problem. If ECOA’s statement of reasons does not even mention home ownership—recall that both the version with and without probably satisfy ECOA—then there would be no way for an outside observer to investigate whether home ownership was included due to prejudice. The statement of reasons only addresses one version of how the machinery got to its conclusion, not how the machinery works generally, or why it works that way. Thus, the statement of reasons could fail even for an intentionally discriminatory design.119 B. General Data Protection Regulation In 2016, the European Union passed the General Data Protection Regulation (GDPR),120 which goes into effect in May 2018. The GDPR is an EU-wide regulation that replaces the federalist system of data protection governed by the 1995 Data Protection Directive (DPD).121 Both laws regulate automated decisionmaking,122 but in the 22 years of the Directive’s existence, little jurisprudence has developed around that particular aspect of the law.123 The GDPR’s discussion of automated decisions is contained in Article 22, Article 13(2)(f), Article 14(2)(g), and Article 15(1)(h). Article 22 is the primary piece and states, in relevant part, as follows: Article 22. Automated individual decision making, including profiling 1. The data subject shall have the right not to be subject to a decision based solely on automated processing, including profiling, which produces legal effects concerning him or her or similarly significantly affects him or her. Note that ECOA does have a substantive nondiscrimination provision, separate from the statement of reasons, which may allow it to get to these types of questions in litigation, given proof. Thus, the law overall can capture intentional discrimination, but the law’s approach of using explanation alone cannot. 120 Regulation (EU) 2016/679 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing Directive 95/46/EC (General Data Protection Regulation) [2016] OJ L119/1. 121 1995 Directive (95/46/EC) (“Data Protection Directive”) 122 GDPR art. 22(1) (“The data subject shall have the right not to be subject to a decision based solely on automated processing, including profiling, which produces legal effects concerning him or her or similarly significantly affects him or her.”; Data Protection Directive art. 123 Wachter, et al., supra note 23, at 19. 119 28 Regulating Inscrutable Systems – DRAFT 3/20/2017 2. Paragraph 1 shall not apply if [exceptions (a)-(c)]. 3. In the cases referred to in points (a) and (c) of paragraph 2, the data controller shall implement suitable measures to safeguard the data subjects rights and freedoms and legitimate interests, at least the right to obtain human intervention on the part of the controller, to express his or her point of view and to contest the decision. 4. [omitted] Articles 13-15 spell out a data subject’s right to be informed about the data that data controllers have about them.124 Articles 13 and 14 describe the obligations of data processors to affirmatively notify data subjects about the uses of their information,125 and Article 15 delineates the affirmative access rights that data subjects have to information about their own data is used.126 Under Articles 13 and 14, the data controller shall provide, and under Article 15 the data subject has the right to access, “the existence of automated decision-making, including profiling, referred to in Article 22(1) and (4) and, at least in those cases, meaningful information about the logic involved, as well as the significance and the envisaged consequences of such processing for the data subject.”127 After the passage of the GDPR, scholars have begun to debate whether these requirements amount to a “right to explanation.”128 Bryce Goodman and Seth Flaxman have argued that by combining Article 22(3)’s requirement that any automated decisionmaking include “safeguards for the rights and freedoms of the data subject” with the Article 13-15 requirement that data subjects have “meaningful information about the logic involved” in automated decisions, one can infer a right to explanation.129 Sandra Wachter, Brent Mittelstadt, and Luciano Floridi disagree. They note that non-binding Recital 71, which explains the purpose of Article 22,130 repeats the safeguard language of Article 22, but includes a “right . . . to obtain an explanation of the decision reached,” which is missing in the operative text. Wachter, et al., supra note 23, at 14. See GPDR art.13-14. 126 See GDPR art. 15. 127 GPDR art.13(2)(f), art. 14(2)(g), art. 15(1)(h). 128 See Goodman & Flaxman, supra note 21. But see, Wachter, et al., supra note 123. 129 Goodman & Flaxman, supra note 21, at 6. 130 Recitals in EU law are explanations of the purpose of operative provisions and are nonbinding. See Tadas Klimas & Jurate Vaiciukaite, The Law of Recitals in European Community Legislation, 15 ILSA J. INT’L & COMP. L. 1, 32 (“Recitals have no positive operation of their own.”) 124 125 DRAFT – Please ask permission before citing or quoting 29 The authors then look at prior drafts and surrounding commentary of the GPDR to argue convincingly that the right to explanation was intentionally removed from Article 22.131 Then they somewhat less convincingly argue that the same right to explanation cannot be derived from Articles 13-15 or anywhere else.132 If the right to explanation was removed from Article 22, it could have been because it was already captured by “meaningful information about the logic,” in Articles 13-15, for example. Whether one calls it a right to explanation or not, requiring that data subjects have meaningful information about the logic has to mean something related to explanation. The phrase “meaningful information” is the major difference between the DPD access right’s treatment of automated decisions and the parallel language in the GDPR. Article 12 of the DPD granted data subjects a right of access to “knowledge of the logic involved” in automated decisions,133 whereas Article 15 of the GDPR grants access to “meaningful information about the logic involved.”134 The word “meaningful” is then central to this new right. If one considers the possibility of information that is not meaningful—that is, literally contains no meaning to the data subject—then the addition of the word “meaningful” implies that data processors must go beyond mere provision of the information to include something else to provide meaning. That something else could be explanatory, as explanation is a method of infusing otherwise unintelligible information with meaning. But perhaps there was a different intent. Another possibility could be simplification of the automated decision-making process itself such that the available information would naturally be meaningful. But Wachter et al.’s demonstration that Article 22 does not include the right among its “safeguards” makes this interpretation unlikely. Article 22 is written as a prohibition on automated decisions, except in certain circumstances.135 If the drafters intended “meaningful information” to imply a limit on the complexity of automated processing itself, why would it be part of the Article describing a data subject’s right to access information rather than the Article delineating the limitations on permissible automated decisions? As Wachter, et al., supra note 23, at 9-11. See Andrew D. Selbst, Calling Foul on the “Right to Explanation” Debate (draft on file with authors). Wachter et al. argue that the GDPR creates a “right to be informed,” which they put in opposition to a “right to explanation,” all without ever engaging with the word “meaningful.” Wachter, et al., supra note 23, at 26-27 (misreading Article 15 to state that it “requires only information about the ‘existence of’ automated decision-making.”). 133 DPD art. 12(a). 134 GPDR art. 15(1)(h). 135 It could also be a right of data subjects to object, see Wachter, et al., supra note 23, at 35-37, but this does not change the analysis. 131 132 30 Regulating Inscrutable Systems – DRAFT 3/20/2017 Wachter et al. point out, the drafters expressly removed such a right from the safeguards in Article 22. Thinking about “meaningful information” as a safeguard on the automated decision itself therefore does not comport with the structure of the regulation. Articles 13-15 instead take the processing as a given, and allow data subjects certain rights to be informed about it. In this case that has to be some additional requirement to make the provided information “meaningful.” So whatever the right is called, what does it mean in practice? According to the regulation, the “meaningful information” must be about the “logic” of the decisions.136 The logic of the decisions must mean something different than the outcomes of the decisions. In Wachter et al.’s terminology, the difference is an explanation of the results in a specific decision versus the overall system functionality, or what we have called the What and the How. Whereas ECOA and FCRA seek the factors that lead to a certain result, the “logic involved” in the “decision-making” seems to refer directly to the How. This is, in principle, a broader right than the one provided by ECOA and FCRA, and likely more useful to data subjects. To demonstrate this, return to Taylor’s hypothetical system where a person missed on her credit application by one point, after the creditor totaled the point values from eight factors. It might be impossible to point to four factors or fewer that were “principal reasons,” but the explanation of what the eight factors were, that they were all assigned point values, and that the hypothetical applicant just missed by a point would be much more useful to the rejected applicant. Or consider Alice in the heavily home-weighted credit system. Telling her that home ownership was weighted heavily in this particular system would be meaningful, even if other factors contributed more directly to her rejection, in the sense that she could more easily improve those. The logic of the system would allow her to do what is necessary to improve her credit, but also, to potentially expose the home-ownership centered system as discriminatory in litigation. Within the What-How-Why framework, “meaningful information about the logic” appears to precisely target inscrutability. As discussed earlier, inscrutability destroys the ability of the How layer to offer a functional description or pattern from which a rule can be inferred. Read in what appears to be the most natural way, “meaningful information” seems to require that for automated decisions, data processors need to find a way to bring back the capability of the How layer to serve its function. 136 GPDR art.13(2)(f), art. 14(2)(g), art. 15(1)(h). DRAFT – Please ask permission before citing or quoting 31 “Meaningful information” is a flexible standard, but that is because the kinds of functional information required will be different for every problem. This flexibility fits within the overall purposes of the GDPR, which are much broader than FCRA and ECOA. The EU treats data protection as a fundamental right,137 and Article 5 lists the following as principles the GDPR seeks to vindicate with respect to personal data: lawfulness, fairness and transparency, purpose limitation, data minimization, accuracy, storage limitation, integrity and confidentiality, and accountability. Several of these principles are a restatement of the familiar Fair Information Practice Principles (FIPPs) that have shaped privacy policy for decades.138 But considered as a whole, including “lawfulness” and “fairness,” they begin to sound like the general purpose of due process in all its expansiveness. To achieve such a result, the standard needs to target exactly the part of the reasoning process that inscrutability interferes with. And it does just that.139 While requiring an explanation at the system functionality level will be more useful, it still does not require that the data processors explain the choices that led to the decision model. ECOA and FCRA ask for the What, the GDPR asks for the How and neither one asks Why. Like ECOA and FCRA, then, the GDPR relies on private rights of action140 and other provisions141 to test the propriety of the algorithms from a normative perspective. But if “meaningful information” does the work to repair the What-How-Why model, then perhaps algorithms can be regulated. The GDPR’s demand for meaningful information requires either that systems be designed so that the algorithm is simple enough to understand, or can provide enough functional information about the logic of the system that it can be tested. Part III canvasses the computer science GDPR art. 1. Kate Crawford & Jason Schultz, Big Data and Due Process: Toward A Framework to Redress Predictive Privacy Harms, 55 B.C. L. REV. 93, 106-7 (2014). While different lists of FIPPs conflict, one prominent example is the OECD’s list of eight: Collection Limitation Principle, Data Quality Principle, Purpose Specification Principle, Use Limitation Principle, Security Safeguards Principle, Openness Principle, Individual Participation Principle, and Accountability Principle. OECD, THE OECD PRIVACY FRAMEWORK 14-15 (2013), http://www.oecd.org/sti/ieconomy/oecd_privacy_framework.pdf 139 Here, we have only discussed the provisions in Articles 13-15 that require meaningful information. But the GDPR has an enormous loophole that limits its applicability: Article 22, and by extension Articles 13(f)(2), 14(g)(2), and 15(h)(1) only apply to decisions “solely based on automated processing,” GPDR art. 22(1). The ultimate meaning of the clause is unclear, as humans are always involved to a degree, but it could turn out to eviscerate the right. See Wachter, et al., supra note 23, at 30-33. Nonetheless, that limitation does not affect the substance of Articles 13-15, when they do apply. 140 See GDPR art. 79. 141 See GDPR art. 77, 83(5) (authorizing potentially enormous administrative fines). 137 138 32 Regulating Inscrutable Systems – DRAFT 3/20/2017 research on interpretability, and Part IV considers whether it might prove helpful to regulation. III. INTERPRETABILITY IN MACHINE LEARNING SYSTEMS The demand for explanation in machine learning is not new. Early research recognized and tried to grapple with the challenge of explaining the decisions of machine learning models such that people using these systems would feel comfortable acting upon them.142 Interpretability has continued to receive considerable attention in the field, generating a surprisingly vast body of scholarship, despite a lack of consensus regarding the precise technical meaning of interpretability or the exact problem to which improved interpretability actually responds.143 Much of this work has taken place in the context of an ongoing debate about the value of interpretability, given the apparent trade-off between performance and interpretability in many practical applications of machine learning. Notably, Leo Breiman, a significant figure in the field, argued in the early 2000s that there are entire classes of problems where the accuracy of predictions is paramount and where the additional insights that practitioners might draw from a more interpretable model will not compensate for the loss of accuracy.144 The success of neural networks, for example, seems to depend on a willingness to forgo interpretability, or an acceptance that the improved accuracy of the resulting predictions provides more information than might be gleaned by inspecting a simpler model. As a result, much of the research on interpretability in machine learning has focused on managing the trade-off between performance and interpretability, with the goal of increasing interpretability while sacrificing the minimum amount of performance.145 At the same time, however, the experience of applying machine 142 Bruce G. Buchanan & Edward H. Shortliffe. Rule-Based Expert System, in THE MYCIN EXPERIMENT OF THE STANFORD HEURISTIC PROGRAMMING PROJECT __ (1984). 143 Zachary C Lipton, The Mythos of Model Interpretability, arXiv preprint arXiv:1606.03490 (2016); Doshi-Velez, Finale, and Been Kim. "A Roadmap for a Rigorous Science of Interpretability." arXiv preprint arXiv:1702.08608 (2017); David Bamman (2016), "Interpretability in Human-Centered Data Science," CSCW Workshop on HumanCentered Data Science 144 Breiman, supra note 13, at __. 145 Michael Gleicher, A Framework For Considering Comprehensibility In Modeling. 4 BIG DATA 4, 75 (2016). DRAFT – Please ask permission before citing or quoting 33 learning to real-world problems has lead to widely held beliefs among practitioners about the relative interpretability of the kind of models that result from different learning methods. Henrik Brink and Joshua Bloom capture much of this informal knowledge in a chart that plots learning methods along two axes: interpretability and accuracy.146 In the top left corner (highly interpretable; relatively low performance) are methods like linear regression; in the bottom right corner (difficult to interpret; relatively high performance) are methods like deep learning. Though this is a common perception of the trade-offs involve with different methods, researchers, including Brink and Bloom themselves, point out that such contrasts do not rest on a rigorous definition of interpretability.147 And yet such beliefs routinely guide practitioners’ decisions when applying machine learning to different kinds of problems.148 Over the past decade, however, other methods have emerged that attempt to side-step these difficult choices altogether. Rather than aiming to build interpretable models, recent methods provide mechanisms to generate explanations for specific decisions; they forgo any attempt at providing a global description of the model in favor of furnishing explanations on a caseby-case basis. The demand for explanations in machine learning can thus be met with at least three different responses: (1) purposefully choosing a feature set, learning method, or parameters that are likely to lead to an intelligible model; (2) adopting specialized methods that attempt to minimize any loss in performance when building interpretable models; and (3) applying post hoc methods to account for individual decisions. A. Favoring Interpretable Methods Where complexity might cause a model to become uninterpretable, practitioners have a number of different levers at their disposal to ensure greater simplicity. First, they may choose to consider only a limited set of all possible features. By limiting the analysis to a smaller set of variables, the total number of relationships uncovered in the learning process might be sufficiently limited to be intelligible to a human. A model with five features, for example, is far more likely to be interpretable than a model with five hundred. Such choices are sometimes forced on practitioners to avoid the Henrik Brink & Joshua Bloom, Overcoming the Barriers to Production-Ready MachineLearning Workflows, STRATA (2014) 147 Alex A. Freitas, Comprehensible Classification Models: A Position Paper. 15 ACM SIGKDD EXPLORATIONS NEWSLETTER 1 (2014). 148 [Cite] 146 34 Regulating Inscrutable Systems – DRAFT 3/20/2017 so-called “curse of dimensionality,” whereby exploring the total number of potential relationships between variables is too large to be computationally tractable.149 These natural limits on the complexity of the desired computation may result in a less than ideal selection of variables and discoveries with less predictive accuracy, but they may also improve the interpretability of these discoveries.150 Second, practitioners might elect to use a learning method that outputs a model that can be more easily parsed than the output of other learning methods. For example, decision tree algorithms learn nested rules that can be represented visually as a tree with subdividing branches. To understand how the model would process any particular case, practitioners need only walk through the relevant branches of the tree; to understand the model overall, practitioners can explore all the branches to develop a sense of how it would apply to all possible cases. Finally, practitioners might set the parameters of the learning process to ensure that the resulting model is not so complex that it defies human comprehension. Even decision trees will become unwieldy for humans at some point if they involve an exceedingly large number of branches and terminal leaves. Practitioners routinely set an upper bound on the number of terminal leaves to constrain the potential complexity of the model. B. Global Explanations A number of more specialized methods exist that allow practitioners to achieve greater model interpretability while maintaining a level of performance that would be impossible with less sophisticated learning methods. These methods generally fall into two camps: regularization and rule extraction. Much like setting an upper limit on the number of branches in a decision tree, regularization methods allow model complexity to be taken into account during the learning process by assigning a cost to excess complexity. In other words, regularization allows practitioners to make model simplicity an additional objective alongside model performance—and the learning process can be set in such a way to find an optimal trade-off between these sometimes competing objectives.151 It can help to automate difficult choices about feature selection, reducing the overall complexity of Breiman, supra note 13, at 208. David Martens & Bart Baesens, Building Acceptable Classification Models, in 8 DATA MINING AND INFORMATION SYSTEMS: QUO VADIS? 53, (Robert Stahlbock, Sven F Crone, & Stefan Lessmann, eds. 2009). 151 Gleicher, supra note 145. 149 150 DRAFT – Please ask permission before citing or quoting 35 the model while still trying to maximize performance. One such method is Lasso.152 Lasso was originally designed to increase accuracy by avoiding overfitting, which occurs when a model learns patterns that are peculiar to the training data and not representative of the patterns the model will encounter when deployed in the real world.153 Lasso increases accuracy by forcing the learning process to ignore relationships that are relatively weak, and therefore more likely to be artifacts of the specific examples that happened to be in the training data, rather than real relationships. Because regularization works by strategically removing unnecessary features, in many real-world applications, the technique can simultaneously improve interpretability (by reducing complexity) and increase performance (by helping to avoid overfitting). Where it applies, this demonstrates that improved interpretability might not always come at the cost of performance. Where potential overfitting is not a danger, however, regularization methods will indeed result in degradations in performance, and these methods will attempt to keep those losses to a minimum. In contrast, rule extraction is a common technique that works by allowing practitioners to learn a model using any method they like and then replicating the discovered rules using simpler methods. Simpler models cannot always reliably discover as many relationships. Take a decision tree, for example. Once the computer makes a decision to take a particular branch, the method does not permit walking back up the branch. Therefore, if there is a relationship between items on two different branches, it will not be discovered.154 More complex algorithms lack the same limitation, but they do not result in models as interpretable as decision trees. Rule extraction takes advantage of the fact that rules that can’t be learned with simpler methods can still be represented effectively by simpler models. Simpler models are often expressive enough to represent—with high fidelity—the complex relationships captured by more sophisticated learning methods like support vector machines or neural networks.155 The Robert Tibshirani. Regression Shrinkage and Selection Via the Lasso, J. ROYAL STAT. SOC’Y 267. Series B (Methodological) (1996). 153 Machine learning is only effective in practice when it successfully identifies robust patterns in the training data while also ignoring patterns that are just artifacts of the particular sample of cases assembled in the training data. 154 A decision tree is an example of a “greedy” algorithm. At each stage, it makes what it perceives to be an optimal choice, but it will often not be a global optimum. STUART RUSSELL & PETER NORVIG, ARTIFICIAL INTELLIGENCE: A MODERN APPROACH 92 (3d. ed. 2014). 155 Bart Baesens, Rudy Setiono, Christophe Mues, & Jan Vanthienen. Using Neural Network Rule Extraction and Decision Tables for Credit-Risk Evaluation, 49 MANAGEMENT 152 36 Regulating Inscrutable Systems – DRAFT 3/20/2017 problem with simpler learning methods is that they might miss certain relationships. With the discoveries of complex learning methods behind them, rule extraction methods allow simple models to “cheat” because the answer is known ahead of time. Where rule extraction works well, the resulting simpler model can be as accurate or nearly as accurate as the more complex model it aims to replicate. Because the goal of rule extraction is to generate models that are interpretable to humans, they must sometimes sacrifice part of the complexity that allowed other learning methods to achieve better performance. But even when rule extraction involves a notable loss in performance, the resulting models frequently perform far better than what would have been learned with simple methods alone. This helps to explain why such methods are widely used in practice.156 C. Specific Decisions A new class of tools have emerged over the past decade that attack the problem of interpretability from a different direction. Rather than attempting to ensure that machine learning generates an intelligible model, these new tools furnish more limited explanations that only account for the relative importance of different features in making particular determinations. At a high level, most of these methods adopt a similar approach: they attempt to establish the importance of any feature to a particular decision by iteratively varying the value of that feature while holding the value of other features constant.157 In so doing, these methods SCIENCE 3 (2003); David Martens, et al., Comprehensible Credit Scoring Models Using Rule Extraction From Support Vector Machines, 183 EUROPEAN J. OPERATIONAL RESEARCH , 1466 (2007); David Martens, et al., Performance of Classification Models from a User Perspective. 51 DECISION SUPPORT SYSTEMS 782 (2011). 156 Johan Huysmans, Bart Baesens, & Jan Vanthienen. Using Rule Extraction To Improve The Comprehensibility Of Predictive Models. (2006). 157 David Baehrens, et al., How to Explain Individual Classification Decisions.11 J. MACHINE LEARNING RESEARCH 1803 (2010); Andreas Henelius, et al.. A Peek Into the Black Box: Exploring Classifiers by Randomization, 28 DATA MINING AND KNOWLEDGE DISCOVERY 1503 (2014). Philip Adler, et al., Auditing Black-box Models for Indirect Influence, ___ ; Marco Tulio Ribeiro, Sameer Singh, & Carlos Guestrin. "Why Should I Trust You?: Explaining the Predictions of Any Classifier." In PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 1135. ACM, 2016; Anupam Datta, Shayak Sen, & Yair Zick, Algorithmic Transparency Via Quantitative Input Influence: Theory And Experiments With Learning Systems, in SECURITY AND PRIVACY (SP), 2016 IEEE SYMPOSIUM on, pp. 598-617. IEEE, 2016. DRAFT – Please ask permission before citing or quoting 37 can uncover the relative contribution of different features to specific determinations, but also the features whose values would have to change the least or the most to change the determination.158 Such methods are particularly popular when trying to explain the decisions of a deep learning model developed for object detection because they can help to pinpoint what part of an image the model ascribes greatest significance when making its determination (e.g., what in this image explains why the model has identified a cat). Note that such explanations do not provide a global description of the model, which, in the case of deep learning, would almost certainly defy human interpretation; they instead describe what gives the model greatest confidence about its classification or what would have to be different in a specific image to trigger a different classification, which can be easily rendered visually for a human observer. Although these methods are especially well suited to models that take raw signal as input (e.g., the pixel values in an image), they work equally well when applied to models with a much more limited number of semantically meaningful features. A credit scoring model that considers a handful of discrete features can be subject to the same test,159 establishing which features most influence a specific outcome. If practitioners know the distribution of feature values that the model is likely to encounter in the real world, they can also determine which features are most important on average for a specific outcome, which are most important to different outcomes, and which account for a disparity between groups. Unfortunately, these methods will not work well in cases where, to stick with the example, credit scoring takes a much larger set of features into account. Should many features each contribute a small amount to a particular determination, listing each of them in an explanation for a particular decision is not likely to be terribly helpful. Indeed, the number of features identified as influential might be sufficiently large that the explanation would simply reproduce the problem of inscrutability that it aims to address. This might come as a surprise, given how well this approach works when applied to deep learning models, but recall that explanations in the case of object recognition took the form of visualizations. Unfortunately, post hoc explanations for credit and other similarly important decisions are likely to be most attractive precisely when they do These methods are generally sensitive to interactions among variables and are able to measure indirect as well as direct influence. See, e.g., Adler, et al., supra note 157; Datta, et al., supra note 157; JULIUS ADEBAYO, FAIRML: AUDITING BLACK-BOX PREDICTIVE MODELS (2017), http://blog.fastforwardlabs.com/2017/03/09/fairml-auditing-black-boxpredictive-models.html. 159 Anupam Datta, Shayak Sen, & Yair Zick, supra note 158. 158 38 Regulating Inscrutable Systems – DRAFT 3/20/2017 not seem to work well—that is, when the only way to achieve a certain level of performance is to vastly expand the range of features under consideration. D. Task-Specific Techniques [still to come] IV. HOW INTERPRETABILITY CAN ASSIST REGULATION [What follows is an outline of how the research on interpretability can assist with regulatory compliance, but why that might not go far enough.] A. Meeting Legal Requirements • • • Global descriptions can fulfill GDPR Specific decisions can fulfill ECOA o Existing practice of using rule extraction for ECOAcompliance o But these will not work where we want them most (e.g., deep learning) Combining global descriptions with specific decisions might be a way forward o Simplify the model enough to be able to give a limited number of reasons B. Vindicating Normative Principles • • But let’s imagine we can fulfill both ECOA and GDPR; would this get us what we want? The writing on this presupposes that having the explanation will automatically get us to the goal of accountability. o The what is not enough to answer the how and the how is not enough to answer the why o The concerns we have in mind when we ask for the why are rarely generic questions about why; they are questions to make sure the basis for decision is not something we would find objectionable (e.g., faulty reasoning; weak proxies; discriminatory intent; non-volitional factors, etc…) DRAFT – Please ask permission before citing or quoting • • • • 39 o Explanation mismatch: is the explanation going to be the one we need? Interpretability is useful because we don’t know in advance what bases for decisionmaking we’ll find objectionable in general, or in the future. o Interpretability gives us something to work with and makes us ask the question. o But still a weaker solution to things we do know we care about than targeted ones (e.g. discrimination-aware data mining). The real problem is just category error; we need to think about why we ask “why”? o The honest answer to why would be “because these are the patterns in the data” § On what basis should we assess whether these patterns are a legitimate basis for decisions making? o Perhaps we should look for regulation that does not ask “why” § Environmental law, nuisance… Machine explanation cannot address disagreement in the why layer itself o All the work is being done at the What and How o If there’s disagreement about the normative goal (e.g., in discrimination, what counts?), then fixing the What and How layers to self-explain doesn’t solve the problem. Other styles of explanation-regulation include asking at the Why layer. o Describing what has been set as the target variable and why § Credit: creditworthiness or maximum profit? § Employment: all the things in BDDI o Can we get specific about asking for aspects of modeling as justification? CONCLUSION [still to come]
© Copyright 2025 Paperzz