Safety and Decision-making Niklas Möller STOCKHOLM 2006 This licentiate thesis consists of the following introduction and: Möller N., Hansson S. O., Peterson M. (2005), “Safety is More Than the Antonym of Risk”, forthcoming in Journal of Applied Philosophy. Möller N. (2005), “Should We Follow the Experts’ Advice? On Epistemic Uncertainty and Asymmetries of Safety”, submitted manuscript. Möller N. and Hansson S. O. (2005), ”Interpreting Safety Practices: Risk versus Uncertainty”, submitted manuscript. Niklas Möller, Division of Philosophy, Department of Philosophy and the History of Technology, Royal Institute of Technology (KTH), SE-100 44 Stockholm, Sweden. ii Abstract Möller, N., 2006. Safety and Decision-making. Theses in Philosophy from the Royal Institute of Technology n 12. 92 + viii pp. Stockholm ISBN 91-7178-272-9. Safety is an important topic for a wide range of disciplines, such as engineering, economics, sociology, psychology, political science and philosophy, and plays a central role in risk analysis and risk management. The aim of this thesis is to develop a concept of safety that is relevant for decision-making, and to elucidate its consequences for risk and safety research and practices. Essay I provides a conceptual analysis of safety in the context of societal decision-making, focusing on some fundamental distinctions and aspects, and argues for a more complex notion than what is commonly given. This concept of safety explicitly includes epistemic uncertainty, the degree to which we are uncertain of our knowledge of the situation at hand. It is discussed the extent to which such a concept may be considered an objective concept, and concluded that it is better seen as an intersubjective concept. Some formal versions of a comparative safety concept are also proposed. Essay II explores some consequences of epistemic uncertainty. It is commonly claimed that the public is irrational in its acceptance of risks. An underlying presumption in such a claim is that the public should follow the experts’ advice in recommending an activity whenever the experts have better knowledge of the risk involved. This position is criticised based on considerations from epistemic uncertainty and the goal of safety. Furthermore, it is shown that the scope of the objection covers the entire field of risk research, risk assessment as well as risk management. Essay III analyses the role of epistemic uncertainty for principles of achieving safety in an engineering context. The aim is to show that to account for common engineering principles we need the understanding of safety that has been argued for in Essays I-II. Several important principles in engineering safety are analysed, and it is argued that we cannot fully account for them on a narrow interpretation of safety as the reduction of risk (understanding risk as the combination of probability and severity of harm). An adequate concept of safety must include not only the reduction of risk but also the reduction of uncertainty. Keywords: conceptual analysis, safety, risk, epistemic uncertainty, epistemic values, values in risk assessment, risk analysis, risk management, safety engineering © Niklas Möller 2006 ISSN 1650-8831 ISBN 91-7178-272-9 iii iv Acknowledgements I am very grateful to my supervisors Sven Ove Hansson and Martin Peterson for their valuable comments, suggestions and support without which this licentiate thesis would not have come about. Thanks also to my colleagues at the Division of Philosophy for valuable suggestions and comments on earlier version of the papers. A special thanks goes to Lars Lindblom, Kalle Grill and Per Wikman-Svahn, for their friendship, inspiration, careful reading and lengthy discussions. And Eva, for her love and our never-ending deliberation. This work has been financially supported by Krisberedskapsmyndigheten, The Swedish Emergency Management Agency. The support is gratefully acknowledged. Stockholm Niklas Möller January 2006 v vi Contents Abstract Acknowledgements Introduction Aim and scope of the thesis Decision theoretical background Preview of Essays I-III References Essay I “Safety is More Than the Antonym of Risk” Forthcoming in Journal of Applied Philosophy. Essay II “Should We Follow the Experts’ Advice? On Epistemic Uncertainty and Asymmetries of Safety” Submitted manuscript Essay III ”Interpreting Safety Practices: Risk versus Uncertainty” Submitted manuscript. vii viii Introduction Safety considerations are important in a wide range of disciplines. Risk assessment, risk management, risk perception and risk communication directly revolve around questions of safety. There is also a wide range of other disciplines where safety concerns are important: engineering, economics, sociology, psychology, political science and philosophy. It should therefore not come as a surprise that there are many different conceptions of safety. This is as it should be, since different aspects are more or less important depending on what our goals are. Depending on whether we are interested in how non-experts think about (“perceive”) risk and safety, or what role such concepts have in forming governmental policies, or how we should construct a bridge, we might expect quite different – yet hopefully related – conceptualisations of safety. That does not mean that all conceptualisations are equally good. Quite the opposite, since an understanding of a concept in one field often has a bearing on another, related field. This is especially true in such an interdisciplinary field as risk and safety research. Here, influences from many different disciplines come together, and it is of the outmost importance that the concepts we use are clear and adequate. When we set safety as a goal for our systems and processes, or claim that one type of technology is safer than another, we must ensure that the concepts that we use are the ones we should use. Otherwise we are not addressing the right question, and then the answer does not have the right significance. This is not only a logical possibility. In this thesis I claim that common conceptions of safety used in risk research and risk practice are deficient, and I develop a more adequate concept. Also, I point out some of the consequences of a better understanding of safety for risk and safety research and practices. 1 Aim and scope of the thesis The aim of this thesis is to develop a concept of safety that is relevant for decision-making, and to elucidate its consequences for risk and safety research and practices. Arguably, the concept of safety is important in decision-making, since concern about safety is a significant factor in many decision processes. The concerns on which I focus are primarily normative. In developing and defending a concept of safety, I argue for a certain understanding of how we should look at safety as opposed to “simply” how we do look at it. The conceptual analysis of safety is closely related to the use of this concept in decision-making. There are two basic approaches to this relation. According to one of them, safety is a factual concept much like length or mass. Whether something is more or less safe is analogous to whether one person is taller than the other: with careful measurement we can find the answer that is there. On such an understanding of safety, safety concerns can be treated as a factual input in the overall decision process. Risk and safety assessment can then be successfully isolated from the normative dimension of decision-making.1 According to the other approach to the relation between safety and decision-making, safety cannot be separated from the normative aspects of decision-making, since there are essential normative aspects already in the safety assessment. I argue for this second way of looking at the relation between safety and decision-making.2 Decision theoretical background Several philosophical subdisciplines are relevant for questions regarding safety and decision-making. Since many questions about safety concern technological systems, hazardous substances etc, they involve scientific 1 2 Cf. e.g. Ruckelshaus (1983) and the treatment in Mayo (1991), 252. Mainly in Essay II. 2 knowledge and are therefore of interest for philosophy of science.3 Moral philosophy deals with the “narrow” question of how to act, viz. how to act morally right. The growing sub-field of practical reasoning deals with how we should reason about what to do in more general terms (not exclusively dealing with the moral question).4 However, decision-making has most systematically been studied in decision theory. Several decision theoretical themes and concepts are highly relevant for the essays in the thesis, notably the conceptualisation of probability, epistemic uncertainty, and utility, as well as decision rules such as Maximising Expected Utility. In this section I will give a short introduction of these topics. Decision theory is an interdisciplinary field of inquiry of interest for mathematics, statistics, economics, philosophy and psychology. It covers a broad spectrum of questions involving decision-making, from experimental studies and theories of how we do in fact make decisions (descriptive decision theory) to theories about how we should make them (normative decision theory).5 Even though the foundations were laid in the seventeenth and eighteenth centuries by authors such as Pascal, Bernoulli and Condorcet, modern decision theory dates to the twentieth century and gained in influence in the 1950s and onwards.6 A fundamental concern in decision theory is information on which decisions are made. Classical decision theory divides decision problem into different categories. One category is called decision under certainty. This is when we know what the outcome will be when we choose an alternative. Many decisions that we make are – or can at least be approximated as – decisions under certainty. If I have the desire for a mango and contemplate whether 3 Even though most philosophers of science have been mainly interested in basic sciences such as physics – e.g. Kuhn ([1962] 1970), Lakatos and Musgrave (1970) – in the last decades there has been an interest in more applied areas as well. See for example Mayo and Hollander (1991). 4 For a recent anthology, c.f. Millgram (2001). 5 Cf. Resnik (1987) for an introduction to the field, and Gärdenfors and Sahlin (1988) for a comprehensive anthology focusing on Bayesian decision theory. 6 Pascal (1670); Bernoulli (1738); Condorcet ([1793] 1847). 3 to go to my local supermarket to buy one, the decision may not be characterised as a decision under certainty, since the store is sometimes out of mangos. If I, on the other hand, call my friend that works there and get a positive answer that the store is full of them, I know the outcomes of my alternatives: a walk to the store, less money and a mango if I decide to go, and no walk, more money but no mango if I do not. There are many ways of categorising situations with less than certainty about the outcome. The modern point of departure here is the seminal work from 1921 by the economist Frank Knight.7 He made a distinction between on the one hand “measurable uncertainty” and on the other “something distinctly not of this character”.8 For the first kind he reserved the term “risk”. This kind, he claims, “is so far different from an unmeasurable one that it is not in effect an uncertainty at all”.9 For the other, “unmeasurable” kind, he reserved the term “uncertainty”. The entities referred to as measurable or unmeasurable are the probabilities of the different outcomes.10 In effect, this categorisation remains the basic distinction still used. In their classical textbook, Duncan Luce and Howard Raiffa defines decisionmaking under risk as when “each action leads to one set of possible outcomes, each outcome occurring with a known probability” and decisionmaking under uncertainty “if either action or both has as its consequence a set of possible specific outcomes, but where the probabilities of these outcomes are completely unknown or not even meaningful”.11 Decision under certainty, risk and uncertainty are the three basic categories in classical decision theory. As they have been described by Luce and Raiffa above, however, they are not exclusive. Our knowledge of the probability can be partially known. We may, for example, know that the probability that a chess game between Gary Kasparov and Viswanathan Anand will end by a Knight ([1921] 1957). Ibid, 20. 9 Ibid. 10 Ibid, ch. VII. 11 Luce and Raiffa (1957). 7 8 4 tie or a win for Kasparov when he is holding the white pieces is, say, 60-70 percent, without pretending that we know what they are exactly.12 Then we are not at all completely ignorant of the probabilities, but they are not known with total precision, either. Some textbooks, like David Resnik’s Choices, reserve the third category for when the probability of the outcomes is unknown or only partly known.13 I will use the concept decision under uncertainty in this latter construal, including partial as well as no knowledge of the probabilities. This distinction between decision under risk and decision under uncertainty is fundamental in classical decision theory, where the probability referred to is thought to be an objective concept, a property of the world itself.14 An alternative is to construe probability as a subjective concept. In Bayesian decision theory, probability is conceived of as a measure of the degree of belief that an agent has in a proposition or a state of affairs (such as, say, “it will rain tomorrow”). This is combined with a notion of utility into a sophisticated decision system. Frank Ramsey was the first to show that it is possible to represent the beliefs of an agent by a unique probability measure based on some rationality assumptions and assumptions on ordering of utilities.15 Authors such as de Finetti, von Neumann and Morgenstern, and Leonard Savage have suggested alternative axiomatisations and developments of Ramsey’s work.16 On a Bayesian construal, all (rational) decisions are decisions under risk (known probabilities), since the rational decision-maker always, at least How to understand these types of statements is highly controversial in decision theory. Part IV of Gärdenfors and Sahlin (1988) deals with some suggestions as to conceptualise “unreliable probabilities”. 13 Resnik (1987), 13-14. However, he somewhat confusingly calls this category for “decision under ignorance” and expression that is more commonly used to mark out the case when there is not even partial knowledge of the probability. I will stick to the more common term “uncertainty”. 14 Relative frequencies or logical (”a priori”) probability. E.g. Knight ([1921] 1957), 214216, 223-224. 15 Ramsey (1931). 16 de Finetti (1937), von Neumann and Morgenstern (1944), and Savage ([1954] 1972). 12 5 implicitly, assigns a probability value to an outcome. Faced with new information, the agent may change her probability assessment (in accordance with Bayes’ theorem), but she always assigns determinable probabilities to all states of affairs. Critics, internal as well as external to the Baysian framework, have challenged the plausibility of this view. Daniel Ellsberg, Henry Kyburg, Isaac Levi, Gärdenfors and Sahlin and others have pointed out that there seems to be a significant difference between some decision situations that should be given the same probability distribution according to classical Bayesianism.17 The amount of knowledge may vary with the situation at hand, and this seems to be relevant for decisionmaking. Judging situations such as coin-tossing and the likelihood of picking a red ball from an urn with a known configuration seems very different from judging whether a bridge will hold or what the weather will be like in Rome a month from now. Thus, there seems to be a difference in how certain we can be of the likelihood of an outcome depending on the information and situation at hand, i.e. there is an epistemic uncertainty that it may not be reasonable to reduce to a unique probability value. With the inclusion of epistemic uncertainty into the Bayesian framework the gap between classical decision theory and its Bayesian counterpart is considerably narrowed. One common problem is how to specify this epistemic uncertainty, the entity Knight described as “unmeasurable”. There has been a significant amount of work on trying to specify the notion in the literature, and in Section 4 of Essay I we provide a short overview. A central tenet in Bayesian decision theory, dominating also in classical decision theory, is the notion of Maximising Expected Utility. The common idea is that each outcome can be assigned a numerical value signifying the goodness of the outcome (the “utility”) as well as a probability value, and that the decision-maker should pick the alternative that has the largest sum Ellsberg (1961), Kyburg (1968), Levi (1974). Gärdenfors and Sahlin (1982). These articles and others dealing with ”unreliable probabilities” are collected in Gärdenfors and Sahlin (1988). 17 6 of the products of the utilities and the probabilities of outcomes. With the inclusion of epistemic uncertainty into the framework, this decision criterion has been questioned and others have been proposed.18 There is however a fundamental problem with utility ascriptions, namely the strong assumptions that must be in place in order for any criterion even remotely similar to the principle of maximising utility to be meaningful. It is quite easily shown that the utility numbers that are assigned to an outcome must conform to a cardinal scale. This means that we should not only be able to rank an outcome as better than another, but also tell how much better, viz. the magnitude of distances between them. This is an assumption that has received a great deal of criticism.19 Most problems where safety is an issue should be considered as decision under various degrees of uncertainty. The possibility of a harmful event taking place is at the heart of the safety consideration. Therefore, it should not come as a surprise that probability as well as the comparison of utility (severity of harm) is important in risk and safety analysis. Preview of Essays I-III Essay I. The first essay of this thesis is written in collaboration with Sven Ove Hansson and Martin Peterson. It provides a conceptual analysis of safety in the context of societal decision-making, focusing on some fundamental distinctions and aspects, and argues for a more complex notion than what is commonly given. Although safety is a fundamental concept in societal decision-making, it is heavily under-theorised, seldom given any explicit definition or clear characterisation in the literature. When it is defined, however, it is often as the antonym of risk, an analysis we deem unsatisfactory. We propose an understanding of safety that explicitly Cf. Gärdenfors and Sahlin (1982) for an example as well as comparisons with earlier attempts. 19 Apart from criticism regarding the possibility for one agent of comparing outcomes, there is the question of how to compare utilities among different agents. This is known as the problem of interpersonal comparisons. Cf. Harsanyi (1976), Sen (1970) and Weirich (1984) for some influential views on the subject. 18 7 includes epistemic uncertainty, the degree to which we are uncertain of our knowledge of the situation at hand. We also discuss the extent to which such a concept may be considered an objective concept, and conclude that it is better seen as an intersubjective concept. We end by proposing some formal versions of a comparative safety concept. By discussing the distinctions between absolute and relative safety and that between objective and subjective safety, an initial clarification of the concept is reached. All combinations of these have their usage and our conclusion is rather that it is more important to keep them apart than to use only one of them. In most societal applications, however, reference seems to be made to the objective safety concept (often in its relative form, allowing safety to be a matter of degree). We then analyse safety in terms of another important concept, risk. Safety is often characterised as the antonym of risk, such that if the risk is low then the safety is high and vice versa, where risk is construed as the combination of probability of a harmful event and the severity of its consequences.20 We criticise this understanding of safety by pointing to the importance of epistemic uncertainty for safety – which is construed broadly, as an uncertainty of the probabilities as well as the uncertainty of the severity of the harmful events – concluding that a concept of safety relevant for decision-making should acknowledge this aspect and thus include it in the characterisation of safety. We also give an overview of the discussion of uncertainty about probability ascription. Our characterisation of safety thus includes three aspects: harm, probability and epistemic uncertainty. Another potential component of the safety concept is then considered, viz. control. There is a connection between the degree to which the agent is able to control an outcome and her perceived safety. However, emphasis on the objective safety concept for societal decision-making reveals that there is no general connection between 20 Normally as the product of the probability and the severity of the harmful event. 8 control and safety in the relevant interpretations of the concept: sometimes less control enhances safety, sometimes the opposite is true. Control is thus only an apparent aspect of safety. A strictly subjective notion of safety cannot be very useful in societal decision-making, but on the other hand an objective safety concept is not easy to construe. Particularly regarding the aspect of harm, with an essential value-ladenness, but also regarding probability and epistemic uncertainty, it is hard if not impossible to reach an objective standard. What we can argue for is instead an intersubjective concept, with an independence from pure agent-subjectivity, relying on intersubjective values (in the case of harm) and the best available expert judgements (in the case of probability and uncertainty). We end the essay with some formal definitions showing a way of including an interpretation of uncertainty as degrees of confidence. Essay II. The second essay of the thesis explores consequences of epistemic uncertainty for the areas of risk and safety. It is a common opinion in risk research that the public is irrational in its acceptance of risks. An underlying claim, implicit or explicit, is that the public should follow the experts’ advice in recommending an activity whenever the experts have better knowledge of the risk involved. (I call this the Expert Argument.) I criticise this claim based on considerations from epistemic uncertainty and the goal of safety. I also argue that many of the attitudes of the public are reasonable for the same reasons that make the argument invalid. Furthermore, I show that the scope of the objection covers the entire field of risk research, risk assessment as well as risk management. After introducing the Expert Argument, I review some of the empirical results from studies of risk perception, sorting out the type of beliefs and dispositions in public perception that signal a discrepancy from what is know about risks. For several findings I agree that they signal a discrepancy. However, from these cases it is premature to conclude that the dispositions 9 of the public in matters of risk in general are unreliable. I claim in particular the reasonableness of two alleged discrepancies, Consequence Dominance (the perceived severity of the harmful consequence is the predominant factor that determines whether an activity is accepted or rejected) and Value Asymmetry of Safety (avoiding bad outcomes is much more important than receiving good ones). To evaluate the claim of the Expert Argument, I use the common understanding of risk as the statistical expectation value of the severity of a harmful outcome. Such an interpretation depends on very strong premises. Here I can grant them, however, since I show that even with such strong premises, the Expert Argument is invalid. I argue that decisions regarding risks in the relevant context are decisions regarding safety, and as such considerations about epistemic uncertainty are vital (a main conclusion of Essay I). Even if the risk is judged to be small, it is not safe unless the epistemic uncertainty is sufficiently small as well. The vital concern is not whether the expert knowledge of the risk is the best one available, but whether that knowledge is good enough. I then show that the invalidity of the expert argument is more than a logical possibility by defending the Consequence Dominance. I argue for the plausibility of the Consequence Dominance trough connecting it to (and arguing for) what I call the Knowledge Asymmetry of Safety: In general, the lower the probability of a harmful event is estimated to be, the higher is the epistemic uncertainty about the severity of harm as well as its probability. However, the uncertainty regarding probability increases more than the corresponding uncertainty about the severity of harm. Finally, I show that the scope of the Expert Argument does not limit itself to risk management only but is evident also in risk assessment. The natural objection against this conclusion is that the Expert Argument explicitly mentions the recommendation of the experts and the purpose of risk 10 assessment is not to make any recommendations but merely to state the scientific facts of the matter. However, epistemic uncertainty is an important decision factor in the scientific process of risk assessment and this has a different relevance in risk assessment than for science in general since these two activities have different goals. The main goal of science is to gain knowledge of the world, and it may be argued that there are acknowledged methods of handling different grades of epistemic uncertainty, since there are certain epistemic values integral to the entire process of assessment in science. For risk assessment, however, the goal is safety. There is reason to consider a potential harm relevant to risk assessment, even if the “purely” scientific case cannot (as of yet) be made. Therefore, since there is no possibility of “insulating” risk assessment from epistemic uncertainty and thus the kind of normative aspects we recognised in risk management, the objections to the Expert Argument based on considerations of epistemic uncertainty are relevant also for risk assessment. Essay III. The final essay, written in collaboration with Sven Ove Hansson, analyses the role of epistemic uncertainty for principles and techniques of achieving safety in an engineering context. The aim of this esssay is to show that to account for common engineering principles we need the understanding of safety that has been argued for in the previous essays. On a narrow risk reduction interpretation of safety (understanding risk as the combination of probability and severity of harm) we cannot fully account for these principles. This is not due to deficiencies of those principles, however, but due to a shortcoming in the capability of the theoretical framework to capture the concept of safety. An adequate concept of safety must include not only the reduction of risk but also the reduction of uncertainty. After giving an initial theoretical background, we analyse the principles and methods put forward in the engineering literature (giving a list of 24 11 principles of various levels of abstraction in Appendix 1). These principles are divided into four categories: (1) Inherently safe design. Minimizing inherent dangers in the process as far as possible. This means that potential hazards are excluded rather than just enclosed or otherwise coped with. Hence, dangerous substances or reactions are replaced by less dangerous ones, and this is preferred to using the dangerous substances in an encapsulated process. (2) Safety reserves. Constructions should be strong enough to resist loads and disturbances exceeding those that are intended. A common way to obtain such safety reserves is to employ explicitly chosen, numerical safety factors. (3) Safe fail. The principle of safe fail means that the system should fail “safely”; either the internal components may fail without the system as a whole failing, or the system fails without causing harm. For example, failsilence (also called “negative feedback”) mechanisms are introduced to achieve self-shutdown in case of device failure or when the operator loses control. (4) Procedural safeguards. Procedures and control mechanisms for enhancing safety, ranging from general safety standards and quality assurance to training and behaviour control of the staff. One example of such procedural safeguards is regulation for vehicle operators to have ample time between actual driving in order to prevent fatigue. Frequent training and checkups of staff is another. Procedural safeguards are important as a ‘soft’ supplement to ‘hard’ engineering methods. Focusing mainly on the first three (more technical) covering principles, it is concluded that the principles and methods of engineering safety are best understood as methods that reduce uncertainty as well as risk. This has important implications for the role of probabilistic risk analysis in 12 engineering contexts: it may be an important tool for safety, but it is not the final arbitrator since it does not deal adequately with issues of uncertainty.21 I would like to thank Sven Ove Hansson, Martin Peterson, Rikard Levin, Lars Lindblom, Kalle Grill and Per Wikman-Svahn for their helpful comments on drafts of this introduction. 21 13 References Bernoulli, D., "Exposition of a New Theory on the Measurement of Risk", Comentarii Academiae Scientiarum Imperialis Petropolitanae, as translated and reprinted, Econometrica 22 ([1738] 1954), 23-36. Condorcet, M., “Plan de Constitution, presenté a la convention nationale les 15 et 16 février 1793”, Oeuvres 12 ([1793] 1847), 333-415. de Finetti, B. ”La prevision: see lois logiques, ses sources subjectives”, Annales de l’Institut Henri Poincaré 7 (1937). Gärdenfors, P. and Sahlin, N.-E., Decision, Probability and Utility, Cambridge: Cambridge University Press (1988). Gärdenfors and Sahlin, ”Unreliable Probabilities, Risk Taking, and Decision Making”, Synthese 53 (1982), 361-386. Harsanyi, J., Essays on Ethics, Social Behavior, and Scientific Explanation, Dordrecht: Reidel (1976) Knight, F., Risk, Uncertainty and Profit, New York: Houghton Mifflin ([1921] 1957). Kuhn, Thomas, S., The Structure of Scientific Revolutions, Chicago: The University of Chicago Press, ([1962] 1970). Lakatos, I. and, Musgrave, A. (eds), Criticism and the Growth of Knowledge, Cambridge: Cambridge University Press (1970). Luce, R. D. and Raiffa, H., Games and Decision, New York: Wiley (1957). Mayo, D. “Sociological Versus Metascientific Views of Risk Assessment”, in Mayo and Hollander (1991). Mayo, D. G. and Hollander, R. D. (eds.), Acceptable Evidence, Science and Values in Risk Management. Oxford: Oxford University Press (1991). Millgram, E. (ed), Varieties of Practical Reasoning, Cambridge MA: MIT Press (2001). Pascal, B., Pensées, Paris: Garnier ([(1670] 1961). Ramsey, ”Truth and Probability”, in R. B. Braithwaite (ed), The Foundations of Mathematics, London: Routledge and Kegan Paul (1931), 156-198. 14 Resnik, M. D., Choices: An Introduction to Decision Theory, Minneapolis: University of Minnesota Press (1987). Ruckelshaus, W. D., “Science, Risk and Public Policy”, Science 221 (1983). Savage, L. J., The Foundation of Statistics, New York: Dover ([1954] 1972). Sen. A., Collective Choices and Social Welfare, San Francisco: Holden-Day (1970). von Neumann, J and Morgenstern, O., Theory of Games and Economic Behavior, Princeton: Princeton Univerity Press (1944). Weirich, P, “Interpersonal Utility in Principles of Social Choice”, Erkenntnis 21 (1984), 295-317. 15 Forthcoming in Journal of Applied Philosophy SAFETY IS MORE THAN THE ANTONYM OF RISK Niklas Möller, Sven Ove Hansson and Martin Peterson Division of Philosophy, Royal Institute of Technology ABSTRACT. Even though much research has been devoted to studies of safety, the concept of safety is in itself under-theorised, especially concerning its relation to epistemic uncertainty. In this paper we propose a conceptual analysis of safety. The paper explores the distinction between absolute and relative safety, as well as that between objective and subjective safety. Four potential dimensions of safety are discussed, viz. harm, probability, epistemic uncertainty, and control. The first three of these are used in the proposed definition of safety, whereas it is argued that control should not be included in a reasonable definition of safety. It is shown that strictly speaking, an objective safety concept is not attainable. Instead, an intersubjective concept is proposed that brings us as close as possible to an objective concept. Keywords: conceptual intersubjectivity. analysis, safety, risk, uncertainty, control, 1. Introduction Even though much research has been devoted to studies of safety, the concept of safety is in itself under-theorised.1 In most research on safety the meaning of the term is taken for granted. A closer scrutiny will show that its meaning is often far from well defined. The Oxford English Dictionary divides its definition of safety into eleven denotations with several branches.2 In technical contexts, safety is frequently defined as the inverse of risk: the lower the risk, the higher the safety.3 In this article we are going to show that The authors would like to thank Nils-Eric Sahlin, the members of the Risk Seminar at the Department of Philosophy and the History of Technology at the Royal Institute of Technology, and an anonymous referee for their helpful criticism. 2 Oxford English Dictionary (1989, 2nd edition). 3 In Oxford English Dictionary, the corresponding definition would be 5a, “The quality of being unlikely to cause or occasion hurt or injury”. 1 17 such a definition of safety is insufficient, since it leaves out the crucial aspect of deficiencies in knowledge. Safety-as-the-antonym-of-risk certainly captures important dimensions of safety, but it does not give an exhaustive understanding of the concept. The aim of this article is to provide a conceptual analysis of safety. We purport to offer an analysis that captures what experts in risk and safety research as well as ordinary laypersons would be liable to include in the concept. Without an in-depth understanding of the concept of safety, the subject matter of risk and safety research remains fuzzy and it is not clear what the objectives of reducing risk and achieving safety really means. This article contributes to such an understanding by analysing several aspects of safety – mainly focusing on the neglected aspect of epistemic uncertainty – as well as by drawing central distinctions between different types of safety. In Section 2 we introduce distinctions between absolute and relative safety and between objective and subjective safety. In Section 3, we investigate the relation between safety and risk, in Section 4 its relation to uncertainty, and in Section 5 its relation to control. In Section 6 we put into question that an objective safety concept is attainable. In Section 7, the results from the previous sections are summarised in the form of a proposed definition of safety, and in the concluding section, Section 8, some more general conclusions are offered. 2. Two basic distinctions It is fruitful to distinguish between an absolute and a relative concept of safety. According to the absolute concept, safety against a certain harm implies that the risk of that harm has been eliminated. According to a relative concept of safety, safety means that the risk has been reduced or controlled to a certain (tolerable) level. Some writers take an absolute concept of safety for granted. For example, in the context of aviation safety it has been claimed that “[s]afety 18 means no harm”4 and that “[s]afety is by definition the absence of accidents”.5 However, the absolute concept of safety is problematic in this and many other contexts, since it represents an ideal that can never be attained.6 Therefore, operationalisations of the safety concept that have been developed for specific technological purposes typically allow for risk levels above zero. Such a relative concept of safety is also expressed in the US Supreme Court’s statement that “safe is not the equivalent of ‘risk free’”.7 According to a relative safety concept, a statement such as “this building is fire-safe” can be interpreted as a short form of the more precise statement “the safety of this building with regard to fire is as high as can be expected in terms of reasonable costs and preventive actions, and the risk of a fire spreading in the building is very low”. Another example, taken from the American Department of Defense, states that safety is “the conservation of human life and its effectiveness, and the prevention of damage to items, consistent with mission requirements.”8 For our present purposes there is no need to denounce either the absolute or the relative safety concept. They can both be retained, but must be carefully kept apart. Another, equally important distinction is that between objective and subjective concepts of safety. According to the subjective concept of safety, “X is safe” means that “S believes that X is safe”, where S is a subject from whose viewpoint the safety of X is assessed.9 According to the objective concept of safety, the truth of the claim “X is safe” depends on S’s beliefs only insofar as these beliefs have an influence, through S’s action, as to whether or not any harm will occur. For example, if you are about to drink a Miller, C.O. (1988) ‘System Safety’, in Wiener, E.L. & Nagel, D.C. (eds.), Human factors in aviation (San Diego, Academic Press), 53-80. 5 Tench, W. (1985) Safety is no accident (London, Collins). 6 On the use of “utopian” goals that cannot be attained, see Edvardsson, K. and Hansson S. O. (2005) ‘When is a goal rational?’, Social Choice and Welfare, 24, 343-361. 7 Miller, ‘System Safety’, 54. 8 Ibid. 9 We will here leave aside the problem of selecting an appropriate subject S for a particular safety issue. 4 19 large glass of something you believe to be gin, but which is in fact a lethal drug, then this is a situation of subjective rather than objective safety. As the example shows, when we talk about safety, the objective safety concept is indispensable. If we only use the subjective safety concept we will not have a language accustomed to dealing with the dangers of the real world. On the other hand, we also need to be able to talk about (different persons’) subjective concepts of safety. We propose to deal with this terminological issue by reserving “safe” for the objective concept and using appropriate belief-ascribing phrases to denote subjective safety. The objective safety concept constitutes a terminological ideal that may be difficult to realise. If our knowledge about every determinant of safety cannot be considered objective, it may be impossible to construct a fully objective concept of safety. We will return to this issue in Section 6 after having introduced the dimensions of the safety concept. 3. Safety as the antonym of risk In most technical contexts, safety is defined as the antonym of risk. We may call this the standard theory of safety. Here, safety is conceived of as a state of low risk: the lower the risk, the higher the safety.10 This definition is, however, complicated by the fact that ‘risk’ is in itself not a very clear concept. This term holds numerous well-established meanings; below are five examples of how risk can be defined: (1) risk = an unwanted event which may or may not occur. (2) risk = the cause of an unwanted event which may or may not occur. (3) risk = the probability of an unwanted event which may or may not occur. (4) risk = the statistical expectation value of unwanted events which may or may not occur. 10 Note that the standard theory of safety presupposes a relative conception of safety. 20 (5) risk = the fact that a decision is made under conditions of known probabilities (“decision under risk”).11 The last definition, (5), is the interpretation of risk used in decision theory and is of little interest in the present context. (4) may be considered as a compound of (1) and (3), where the unwanted event has been assigned a value (disutility) and a probability. However, there are aspects of (1) and (3) that are not very easily included in the representation of (4). Notably, (4) rests on the presupposition that an unwanted event can be given a precise value, which is a strong assumption not needed in (1) and (3). For the purpose of this article, we shall assume that probability and harm are the major components of risk, but we do not need to assume that they can be combined into a one-dimensional measure of risk as in definition (4).12 It should be obvious that safety increases as the probability of harm, or its severity, decreases. One of the definitions of safety in the Oxford English Dictionary is fully in line with the interpretation of risk in terms of probability and harm: “The quality of being unlikely to cause or occasion hurt or injury”.13 A typical example taken from a safety application states: “[R]isks are defined as the combination of the probability of occurrence of hazardous event and the severity of the consequence. Safety is achieved by reducing a risk to a tolerable level”.14 If an unwanted event (harm) associated with a given risk is small, it may be incorrect to talk about safety. For example, drawing a blank ticket in a lottery would be an unwanted event, but the avoidance of this would not be described as a matter of safety (unless, of course, the lottery was about something severe, such as when a person participates in a game of Russian roulette). Thus, the nature of the unwanted event is relevant here: if its Hansson, S. O. (2004) ‘Philosophical Perspectives on Risk’, Techne, 8,1. A notion in line with this is put forward by, for example, Slovic, his definition of risk being “a blend of the probability and the severity of the consequences”. See Slovic, P. (2000) ‘Do Adolescent Smokers Know the Risks?’, in P. Slovic, The Perception of Risk (London, Earthscan), 365. 13 Oxford English Dictionary, Safety-definition 5 a. 11 12 21 severity is below a certain level it does not count as a safety issue.15 We will use the term “harm” that (contrary to “unwanted effect”) implies a nontrivial level of damage. There are many types of harm that may be relevant to safety. We will assume that all statements about safety refer to an either explicitly or implicitly delimited class of harms (”safety against accidents”, “safety against device failure” etc.). 4. Safety and epistemic uncertainty The probabilities referred to in a risk or safety analysis are in most cases not known with certainty, and are therefore subject to epistemic uncertainty. This aspect is paramount for the notion of safety, but is often neglected in the safety discourse. The relevance of uncertainty for decision-making has been shown in empirical studies, e.g. by Daniel Ellsberg.16 Ellsberg showed that people have a strong tendency in certain situations to prefer an option with low uncertainty to one with high uncertainty, even if the expectation value is somewhat lower in the former case. Below we show a similar point using a different example.17 Suppose that you are walking in the jungle, and are about to cross an old wooden bridge. The bridge looks unsafe, but the innkeeper in a nearby village has told you that the probability of a breakdown of this type of bridge is less one in ten thousand. Contrast this example with a case in which you are accompanied by a team of scientists at the bridge, who have just examined this particular bridge and discovered that the actual probability that it will break is one in five thousand. Even though the probability is now Misumi, Y. and Sato, Y. (1999) ‘Estimation of average hazardous-event-frequency for allocation of safety-integrity levels’, Reliability Engineering & System Safety, 66, 2, 135-144. 15 The term unwanted event also refers to the subject’s desires in an unfortunate way. If you, for some reason, had a desire to hurt yourself, and therefore engaged in a stunt act in which the probability of a severe accident is very high, we might say that you chose a safe way to kill yourself, meaning certain, but we would never say that you were safe. 16 Ellsberg, D. (1961) ‘Risk, Ambiguity and the Savage axioms’, Quarterly Journal of Economics, 75, 643-69. 17 The example is inspired by a similar one in Gärdenfors, P. and Sahlin, N.-E. (1988 [1982]) ‘Unreliable probabilities, risk taking, and decision making’, in P. Gärdenfors and N.-E Sahlin (eds.), Decision, Probability, and Utility (Cambridge, Cambridge University Press), 313-334. 14 22 judged as higher, the epistemic uncertainty is far smaller and it is not unreasonable to regard this situation as preferable to the first one in terms of safety. This kind of example indicates that safety should be a function that decreases with the probability of harm, with the severity of harm, and with uncertainty. In a discussion of safety against one particular type of harm, the severity-of-harm factor is constant, and we can schematically summarize the effects of probability and uncertainty as in figure 1. A shift from a to b illustrates how we may have the same level of safety for different estimates of probability, if there is a corresponding difference in uncertainty. Likewise, the same probability estimate gives rise to different levels of safety when there is a difference in uncertainty, which is illustrated by a shift from b to c. Uncertainty Safety c a z b y x Probability Figure 1. Safety as a function of probability and uncertainty. x, y and z are levels of safety such that x>y>z. The relevance of uncertainty for safety is also evident from the common engineering practice of adding an “extra” safety barrier even if the probability that this barrier will be needed is estimated to be extremely low, e.g. in the context of nuclear waste facilities. Such extra barriers that make 23 the construction fail-safe are best argued for in terms of the possibility that the probability estimate may be incorrect.18 It should thus be clear that epistemic uncertainty is a necessary aspect of the concept of safety. A fundamental question is how epistemic uncertainty should be characterised more in detail. This is a controversial area in which no consensus has been reached. The most extensive discussions have been in decision theory regarding how to express the uncertainty of probability assessments. In order to give an indication of possible directions to develop the concept we will conclude this section with an overview of this discussion. Two major types of measures of incompletely known probabilities have been proposed. Let us call them binary and multi-valued measures. A binary measure divides the probability values into two groups, possible and impossible values. In typical cases, the set of possible probability values will form an interval, such as: ”The probability of a major earthquake in this area within the next 20 years is between 5 and 20 per cent.” Binary measures have been used by Ellsberg, who refers to a set of ”reasonable” probability judgments.19 Similarly, Levi refers to a ”permissible” set of probability judgments.20 Kaplan has summarised the intuitive appeal of this approach as follows: As I see it, giving evidence its due requires that you rule out as too high, or too low, only those values of con [degree of confidence] which the evidence gives you reason to consider too high or too low. As for the values of con not thus ruled out, you should remain undecided as to which to assign.21 Contrast this with another type of safety barrier, serving a somewhat different purpose: river dikes in the Netherlands, a case where the frequency data for water levels in different seasons are well known, but different layers of dikes are built to contain different water levels (summer dike, winter dike and sleeper dike). 19 Ellsberg, ‘Risk, Ambiguity and the Savage axioms’. 20 Levi, I. (1986) Hard Choices: Decision Making under Unresolved Conflict (Cambridge, Cambridge University Press). 21 Kaplan, M. (1983) ‘Decision theory as philosophy’, Philosophy of Science, 50, 570. 18 24 Multivalued measures generally take the form of a function that assigns a numerical value to each probability value between 0 and 1. This value represents the degree of reliability or plausibility of each particular probability value. Several interpretations of the measure have been used in the literature, of which we will mention (1) second-order probability, (2) fuzzy set membership, and (3) epistemic reliability: 1. Second-order probability. The reliability measure may be seen as a measure of the probability that the (true) probability has a certain value. We may think of this as the subjective probability that the objective probability has a certain value. Alternatively, we may think of it as the subjective probability, given our present state of knowledge, that our subjective probability would have had a certain value if we had ”access to a certain body of information”.22 As was noted by Brian Skyrms, it is ”hardly in dispute that people have beliefs about their beliefs. Thus, if we distinguish degrees of belief, we should not shrink from saying that people have degrees of belief about their degrees of belief. It would then be entirely natural for a degree-of-belief theory of probability to treat probabilities of probabilities.”23 In spite of this, the attitude of philosophers and statisticians towards second-order probabilities has been mostly negative, due to fears of an infinite regress of higher-and-higher orders of probability. David Hume expressed strong misgivings against second-order probabilities.24 Similar doubts are expressed in a modern formulation: ”merely an addition of second-order probabilities to the model is no real solution, for how certain are we about these probabilities?”25 Baron, J. (1987) ‘Second-order probabilities and belief functions’, Theory and Decision, 23, 27. 23 Skyrms, B. (1980) ‘Higher order degrees of belief’, in DH Mellor (ed.), Prospects for Pragmatism (Cambridge, Cambridge University Press), 109. 24 Hume, D (1888 [1739]) A Treatise of Human Nature, ed. by LA Selby-Bigge (Oxford, Oxford University Press), 182-183. 25 Hansson, B. (1975) ‘The appropriateness of the expected utility model’, Erkenntnis, 9, 189. 22 25 This is not the place for a discussion of the rather intricate regress arguments against second-order probabilities.26 It should be noted, however, that similar arguments can also be devised against the other types of measures of incomplete probability information. The basic problem is that a precise formalization is sought for the lack of precision in a probability estimate. 2. Fuzzy set membership. In fuzzy set theory, uncertainty is represented by degrees of membership in a set. In common set theory, an object is either a member or not a member of a given set. A set can be represented by an indicator function (membership function, element function) µ. Let µY be the indicator function for a set Y. Then for all x, µY(x) is either 0 or 1. If it is 1, then x is an element of Y. If it is 0, then x is not an element of Y. In fuzzy set theory, by contrast, the indicator function can take any value between 0 and 1. If µY(x) = 0.5, then x is ”half member” of Y. In this way, fuzzy sets provide us with representations of vague notions. Vagueness is different from randomness. In fuzzy decision theory, uncertainty about probability is taken to be a form of (fuzzy) vagueness rather than a form of probability. Consider an event about which the subject has partial probability information (such as the event that it will rain in Oslo tomorrow). Then to each probability value between 0 and 1 is assigned a degree of membership in a fuzzy set A. For each such probability value x, the value µA(x) of the membership function represents the degree to which the proposition ”it is possible that x is the probability that the event occurs” is true. In other words, µA(x) is the possibility of the proposition that x is the probability that a certain event will happen.27 The difference between fuzzy membership and second-order Skyrms ‘Higher order degrees of belief’. Cf. Sahlin, N.-E. (1983) ‘On second order probability and the notion of epistemic risk’, in B.P. Stigum and F. Wenztop (eds.), Foundations of Utility Theory with Applications (Dordrecht, Reidel), 95-104. 27 On fuzzy representations of uncertainty, see Unwin, S. (1986) ’A Fuzzy Set Theoretic Foundation for Vagueness in Uncertainty Analysis’, Risk Analysis, 6, 27-34, and Dubois, D. and Prade, H. (1988) ‘Decision evaluation methods under uncertainty and imprecision’, in J. Kacprzyk and M. Fedrizzi (eds.), Combining Fuzzy Impression with Probabilistic Uncertainty in Decision Making (Berlin, Springer Verlag), 48-65. 26 26 probabilities is not only of a technical or terminological nature. Fuzziness is a non-statistical concept, and the mathematical laws of fuzzy membership are not the same as the laws of probability. 3. Epistemic reliability. Gärdenfors and Sahlin take a different approach than the traditional Bayesian approach. They use a set of possible probability distributions and assign to each probability representation a real-valued measure ρ that represents the ”epistemic reliability” of the probability representation in question.28 The specific mathematical properties of ρ are kept open. As should be obvious, a binary measure can readily be derived from a multivalued measure.29 The latter carries more information, but this is an advantage only to the extent that such additional information is meaningful. Another difference between the two approaches is that binary measures are in an important sense more operative. In most cases it is a much simpler task to express one's uncertain probability estimate as an interval than as a real-valued function over probability values. For our present purposes, we do not have to assume that epistemic uncertainty is expressible in a particular format, but in Section 7 we will use a multivalued measure to illustrate the relation between probability and uncertainty. 5. Control and safety Extensive psychological research indicates that people’s attitudes towards risk depend on controllability. Everything else being equal, more controllability implies more perceived safety, and vice versa. For example, air travel is controllable (by the passenger) only to a low degree and therefore perceived of as unsafe, whereas driving one’s own car is controllable to a much higher degree, and therefore perceived as safer.30 Gärdenfors and Sahlin, ‘Unreliable probabilities’. Let M1 be the multivalued measure. Then a binary measure M2 can be defined as follows, for some real number r: M2(p) = 1 if and only if M1(p) ≥ r, otherwise M2(p) = 0. Such a reduction to a binary measure is employed by Gärdenfors and Sahlin, ibid. 30 See e.g. P. Slovic (2000), The Perception of Risk (London, Earthscan). 28 29 27 To say that a certain risk is more controllable for an agent than another risk means, in the present context, that there is a more reliable causal relationship between the acts of the agent and the probability and/or the severity of the harm.31 Controllability can be conceived either as an objective or a subjective concept. Objective controllability is hard if not impossible to measure, but subjective controllability can be measured (on an ordinal scale) by asking respondents to indicate to what degree they can control various processes, for instance by ticking the appropriate box on a Lickert-scale. At first sight one might believe that, everything else being equal, more controllability implies more safety. However, this does not seem to hold in general. Consider two nuclear power plants, one of which runs almost automatically without the intervention of humans, whereas the other is dependent on frequent decisions taken by humans. If the staff responsible for the non-automatic plant is poorly trained, it seems reasonable to maintain that the automatic plant is safer than the non-automatic one because the degree of controllability is lower in the automatic plant. Arguably, even if they are excellently trained a high degree of control may lower the safety. Consider a third nuclear plant that is much less automated than the ones we have today. Due to cognitive and mechanical shortcomings, a human being could never make the fast adjustments that an automatic system does, but this plant has excellent staff that perform as well as any human being can be expected to do under the circumstances. In spite of this increased control this too would be an unsafe plant. The reason is that the probability of accidents due to human mistakes would be much greater than in a properly designed, more automatic system. The effects of control on safety in our nuclear plant example can be accounted for in terms of the effects of control on probability. If increased human control decreases the probability of an accident, then it leads to higher safety. If, on the other hand, it increases the probability of accidents, then it decreases safety. Hence, this example does not give us reason to add This is an extension of the psychometrical concept, which only regard the subjective dimension. 31 28 control to the three dimensions of safety that we have already listed; it acts here via one of the dimensions we already have. However, there are other cases in which such a reduction is not equally easy to perform. For example, consider a person who is sitting in a car next to the driver, driving around in heavy traffic. Even if she judges the driver to be just as skilled as herself, she may feel less safe than if she were herself the driver, i.e. in control. She may agree that the probability of an accident is the same in the two cases, but nevertheless feel much safer when able to control the situation. Another example would be a person who is afraid of being robbed, and therefore carries a gun. Even if she were convinced by statistical evidence showing that robbery victims who carry arms run a greater risk of injury or death, she would feel safer when carrying a gun. For the last two examples, the distinction between subjective and objective safety is essential for the analysis. Being the driver or carrying a gun makes the person feel safer. However, in both cases it would be sensible for someone else to say: “You only feel safer. In fact, you are exposed to a greater risk and therefore you have lost rather than gained in safety.” Given our choice in Section 2 of an objective safety concept (or rather, a safety concept that is as objective as possible) these and similar cases can be dealt with by showing that the control factor is relevant only for a subjective concept of safety and should therefore be excluded from the analysis. In some safety issues, the effects of control can be reduced to uncertainty rather than probability. A preference for driving rather than being a passenger may be a case of this. I may judge the other to be just as good a driver as I am, but I am more certain about my own skills than about those of the other. Hence, I am safer (from a rational subjective point of view) if I am driving, even if I judge our skills to be equal. Here, the control dimension is reduced to the uncertainty dimension. However, for an external observer, the degree of epistemic certainty about their skills might be equal; hence from that viewpoint the level of safety is equal in the two cases.32 The intersubjectively best estimation possible, however, would require us to do extensive testing of the two drivers to obtain a better knowledge of their skills. 32 29 6. The limits of objective safety With the aim of elucidating important aspects of the safety concept for technical and scientific use, we have identified three dimensions: severity of harm, probability and uncertainty. As noted in Section 2 our aim is a safety concept that is as objective as possible, and we have eliminated subjective elements wherever possible. It is now time to evaluate how far we have been successful in eliminating subjective elements from each of these three dimensions. The severity of harm is, of course, essentially value-laden. Even if we are able to compare some harms, such as a broken finger vs. a broken finger and a broken leg, this comparison is in general a subjective evaluation. For example, in a comparison between two accidents, one of which caused a few serious injuries and the other a larger number of less severe injuries, it might be far from clear which is the most severe accident. To some extent we may use intersubjective methods for judging degrees of harm. Attempts such as QALY (quality adjusted life years) in medical ethics have been made to solve such issues of comparison.33 In some variants of risk-benefit analysis, harms such as deaths and injuries are assigned monetary values. However, whatever currency the severity of harm is measured in – be that euros or healthy years – rational persons can disagree about the comparative severity of different harms and have no access to objective means, no independent standard, by which their disagreement can be resolved. It is hard to see how the subjective aspect of the value-ladenness of severity of harm can be eliminated. What we can often achieve, however, is an intersubjective assessment that is based on evaluative judgments that the vast majority of humans would agree on. As can be seen from studies of the QALY concept, there is indeed agreement on a wide range of such judgments. Nord, E. (1999) Cost-Value Analysis in Health Care: Making Sense out of QALYs (Cambridge, Cambridge University Press). See also the review by Hansson, S. O. (2001) of Erik Nord, ‘Cost-Value Analysis in Health Care: Making Sense out of QALYs’ in Philosophical Quarterly, 51, 132-133. 33 30 There is a well-established distinction in probability theory between subjective and objective interpretations of probability.34 According to the objective interpretation, probability is a property of the external world, e.g. the propensity of a coin to land heads up. According to the subjective interpretation, to say that the probability of a certain event is high means that the speaker’s degree of belief that the event in question will occur is strong. When we are dealing with the repetition of technological procedures with historically known failure frequencies, it may be possible to determine (approximate) probabilities that can be called objective. However, in most cases when a safety analysis is called for, such frequency data are not available, unless perhaps for certain parts of the system under investigation. Therefore, (objective) frequency data will have to be supplemented or perhaps even replaced by expert judgment. Expert judgments of this nature are not, and should not be confused with, objective fact. Neither are they subjective probabilities in the classical sense, since by this is meant a measure of a person’s degree of belief that satisfies the probability axioms but does not have to correlate with objective frequencies or propensities. They are better described as subjective estimates of objective probabilities. However, what we aim at in a safety analysis is not a purely personal judgment but the best possible judgments that can be obtained from the community of experts. Therefore, this is also essentially an intersubjective judgment. Procedures for expressing and reporting uncertainties are much less developed than the corresponding procedures for probabilities.35 However, the aim should be analogous to that of probability estimates, namely to obtain the best possible judgment that the community of experts can make on the extent and nature of the uncertainties involved. Objective knowledge The subjective approach to probability theory was introduced by Ramsey 1926 in his paper ‘Truth and probability’ which can be found in P. Gärdenfors and N.-E Sahlin (eds.) (1988), Decision, Probability, and Utility (Cambridge, Cambridge University Press), 19-47. See also Savage, L. (1972, [1954]) The foundations of statistics, 2nd ed. (New York, Dover). 35 Levin, R., Hansson, S. O., and Rudén, C. (in press) ‘Indicators of Uncertainty in Chemical Risk Assessments’, Regulatory Toxicology and Pharmacology. 34 31 about uncertainties is at least as difficult to obtain as objective knowledge about probabilities. In summary, then, an objective safety concept is not attainable, but on the other hand we do not have to resort to a subjective safety concept that is different for different persons. The closest we can get to objectivity is a safety concept that is intersubjective in two important respects: (1) it is based on the comparative judgments of severity of harm that the majority of humans would agree on, and (2) it makes use of the best available expert judgments on the probabilities and uncertainties involved. This intersubjective concept of safety should be our main focus in technical and scientific applications of safety.36 7. The definition of safety Many of the concepts that we need to define come in clusters of closely related concepts, and ”safety” belongs to one such cluster. Any serious definition work should start with a careful investigation of the relevant cluster, in order to determine if and how the concepts in the cluster can be defined in terms of each other, and on the basis of that, which concept should be chosen as the primary definiendum. Here, the essential choice is between using a monadic concept such as “safe” and a comparative concepts such as “at least as safe as” as the primary definiendum. In order to define a comparative concept such as “at least as safe as” we only need to determine the quality. Therefore, it is expedient to begin with the comparative concept, and define it as precisely as we can before we proceed to deal with the monadic concept.37 As we have seen there are three factors that need to be included in a comprehensive definition of “at least as safe as”, namely the severity of harm, the probability that harm will occur, and the uncertainty of our Even those who, in analogy with the case of probability, deny the existence of anything like objective safety could accept an intersubjective usage of the concept. 37 In this we follow the convention of preference logic, in which “at least as good as” is taken as primitive rather than “better”, for analogous reasons. 36 32 knowledge about the harm. For expository reasons, we will first propose a definition that only takes severity and probability into account, and then add uncertainty. One way to deal with the combination of severity and probability is to consider only safety against classes of events that are so narrowly defined that all elements in each class have the same degree of severity. This would mean for instance that we would not talk about “safety against industrial accidents” but about “safety against an industrial accident killing exactly one person”, etc. For each of these categories, a state is then at least as safe as another, with respect to that category, if and only if the probability of an event in that category is at most as large. However, this way of speaking does not correspond to how we in practice talk about safety. For our definition to be at all compatible with common usage, it must be applicable to categories of harm that contain events of different severity, such as “safety against industrial accidents”. In order to avoid unnecessary complications we will assume that the intersubjective relation “at least as severe as”, as introduced in Section 5, is complete.38 The two-dimensional comparative safety concept can then be defined as follows: (1) A state A is at least as safe as a state B with respect to a class of unwanted events E if: For each e ∈ E, the probability that either e or some event that is at least as severe as e will take place is at most as high in state A as in state B. There are at least two ways to deal with cases when it is not. One is to replace “at least as severe as” by “not more severe than” in the definitions that follows. This relation is complete. Another is to consider all reasonable relations of severity, and replace “at least as severe as” by the more demanding “at least as severe as according to all reasonable relations of severity”. In the latter case, incompleteness in severity may (but need not, depending on the probability distribution) give rise to incompleteness in the comparative “at least as safe as”. 38 33 Probability Probability B A Harm Harm Figure 2. Probability vs. severity of harm in a two-dimensional definition of safety. The horizontal axis represents severity of harm, and the vertical axis probability. The probability assigned by a curve to a certain level of harm represents the probability that a harm with at least that level of severity will occur. In the left diagram the situation represented by curve A is at least as safe as that represented by curve B. In the right diagram, there is a stalemate according to definition (1). Definition (1) resolves those conflicts between considerations of probability and severity of harm that have an obvious resolution. However, examples can be constructed in which (1) gives rise to a stalemate, as figure 2 illustrates. For a practical example, let X represent the use of a chemical substance that gives rise to a 1% risk of contact allergy and no risk of cancer, and Y the use of a chemical with no risk of allergy and 0.01% risk of cancer. Then Y is associated with a higher probability than X for a high level of harm, whereas the relation is reversed with respect to a lower level of harm. Such stalemates can of course be “solved” by the adoption of a stronger rule, such as expected utility maximization, that implies but is not implied by (1).39 We will refrain from the introduction of such stronger rules, since it is far from clear that they can in general be motivated, as indicated by our discussion above. The introduction of uncertainty into this definition is not trivial. Although we presented uncertainty and probability as orthogonal in figure 1, they are in actual fact not independent of each other, since the uncertainty that we are concerned with is largely uncertainty about probability. Uncertainty in relation to safety comes in many shapes. We can have uncertainty about probabilities, about our values, about what adverse events Note that whereas (1) only requires a relation “at least as severe as”, the expected utility rule requires a cardinal measure of severity. 39 34 should be taken into account and how they should be grouped, etc.40 Arguably the most important of these is probability. We will focus on uncertainty about probabilities. The assumption that we need is that there are at least two well-defined “degrees of confidence” with which a probability can be stated. Hence, instead of dealing with statements of the form The probability of (adverse event) Z is p. we will operate with statements of the form There is confidence-to-degree-c that the probability of (adverse event) Z is at most p. The degrees of confidence may be thought of as confidence intervals of a second-order plausibility (or probability) measure. Alternatively they may be thought of as a small set of discrete levels of confidence (“fairly certain”, etc) that need not have any exact mathematical correlate. We can introduce degrees of confidence into (1) as follows: (2) A state A is at least as safe as a state B with respect to a class of unwanted events E if it holds for each e ∈ E and each degree c of confidence that: if there is confidence-to-degree-c that the probability that either e or some event that is at least as severe as e will take place is at most p in state B, then there is also confidence-to-degree-c that the probability that either e or some event that is at least as severe as e will take place is at most p in state A. Hansson, S. O. (1996) ‘Decision-Making Under Great Uncertainty’, Philosophy of the Social Sciences, 26, 369-386. 40 35 It should be noted that definition (2) does not require more than an ordering of severities, and does not require even comparability of degrees of confidence.41 Clearly, (2) introduces incomparable situations, namely when A is safer than B with respect to one level of confidence whereas B is safer than A according to another level of confidence. These levels of confidence might for example represent the shift from a 50% confidence interval to a 1% interval that includes uncertain threats that concern mainly B. See figure 3. A B Probability Probability Harm Harm Figure 3. A case in which A is at least as safe as B holds with respect to (1), but not (2). In the two graphs to the left the probabilities as a function of different severities of harm are shown for the two states A (thin lines) and B (thick lines). The full lines represent the most likely probability estimate, whereas the two dotted lines in each graph represent the upper probability limit for different degrees of confidence. The dotted line with the highest probabilities (square dots) represent the “worst case” probability assessment in regard to the knowledge at hand. The graphs in the rightmost part show comparisons of each of the three degrees of confidence. The reason that the relation A is as least as safe as B holds with respect to (1), but not in respect to (2) is that whereas (1) holds both for the It is in fact also compatible with replacement of numerical probabilities by an ordering, but that will not be pursued here. 41 36 most likely probability assessment (full line; lowest graph) and for the next level of confidence (middle graph), it does not hold for the lowest level of confidence, the “worst case scenarios”. Given a relation “at least as safe as”, a monadic predicate representing the adjective “safe” can be constructed at different levels of stringency.42 The logical technique for this is the same as that developed for other pairs of dyadic and monadic predicates, such as “at least as long as” vs. “long”, etc.43 The following basic criteria should be satisfied by any monadic concept “safe” that conforms with a given relation “at least as safe as”. Positivity: If X is safe* and Y is at least as safe as X, then Y is safe*.44 Clearly, there may be many predicates safe* that satisfy this criterion. Different such predicates may differ in terms of stringency, defined as follows: The predicate safe** is at least as stringent as the predicate safe* if and only if it holds for all X that if X is safe*, then X is safe**. Furthermore, safe** is more stringent than safe* if and only if: safe** is at least as stringent as safe* but safe* is not at least as stringent as safe**. The term absolutely safe can be constructed as the most stringent safety predicate. Other terms such as reasonably safe and relatively safe can be constructed as safety concepts at lower levels of stringency. However, not all predicates that satisfy the positivity property correspond to the term “safe” in natural language. Predicates with a low degree of stringency (such as “safer than the least safe alternative”, which can be substituted for safe* in It should be noted that the relative ”at least as safe as” is not defined in terms of a monadic predicate “safe” (such a definition would be impossible) and therefore a definition of “safe” in terms of that relation is not circular. 43 Hansson, S. O. (2001) The Structure of Values and Norms (Cambridge, Cambridge University Press), 114-126 and 161-164. 44 For the concept of positivity, see Hansson, S. O. (1990) ‘Defining “good” and “bad” in terms of “better”’, Notre Dame Journal of Formal Logic, 31, 136-149. 42 37 the definition, and satisfies positivity) would not be so described. There seems to be a socially constructed level, perhaps best specified as “acceptable level of safety”, below which it is misleading to use the term “safe”. This leads us to a second criterion for a concept of safety: Minimal stringency: safe* is at least as stringent as the predicate that represents the lowest level of safety that is socially acceptable. We propose to call a predicate a safety predicate if it satisfies positivity and minimal stringency with respect to a relation “at least as safe as” that satisfies our three-dimensional definition (2). 8. Conclusion When analysing the concept of safety, it is paramount to go beyond the simple view of safety as the antonym of risk, even if risk is understood in the two-dimensional way adopted above. A full understanding of epistemological uncertainty is of great importance when discussing safety and safety matters. It is also important to distinguish between safety concepts with respect to how demanding they are. We did this first in Section 2 with the division between absolute and relative concepts of safety, and then more elaborately in Section 7 with the distinction between predicates of safety that use the same underlying criterion of safety but draw the limit between safety and unsafety at different levels of safety. Furthermore, a distinction should be drawn between on one hand a concept of perceived or (individually) subjective safety and on the other hand the concept of intersubjective safety as delineated in Section 6. Closer attention to these distinctions will make clearer what we mean when we say that a technology or a technological practice is “safe”. Communication between different social and technological fields that use divergent safety concepts should be facilitated if the differences are stated as 38 precisely as possible in terms of the defining constituents of the respective safety concepts. 39 Submitted manuscript SHOULD WE FOLLOW THE EXPERTS’ ADVICE? On Epistemic Uncertainty and Asymmetries of Safety. Niklas Möller Division of Philosophy, Royal Institute of Technology ABSTRACT. It is a common opinion in risk research that the public is irrational in its acceptance of risks. Many activities that are claimed by experts to be safe are not deemed to be safe by the public, and vice versa. The aim of this article is to put forward a normative critique against a common argument, viz. the claim that the public should follow the experts’ advice in recommending an activity whenever the experts have better knowledge of the risk involved. Even after making plausible limitations, the claim remains invalid. The importance of safety in risk acceptance together with the phenomenon of epistemic uncertainty highlights the vital concern: not whether the expert knowledge of the risk is the best one available, but whether that knowledge is good enough. The scope of the objection from epistemic uncertainty covers not only risk management but also risk assessment. Keywords: safety, risk perception, epistemic uncertainty, epistemic values, values in risk assessment 1. Introduction It is a common opinion in risk research that the public is irrational in its perception and acceptance of risks.1 Many activities that are considered safe by experts, such as storage of nuclear waste, are considered to be unsafe by the public. Other activities that are known by all to be hazardous – such as driving a car without wearing a safety belt – are avoided by the public only I would like to thank Sven Ove Hansson, Martin Peterson, Nils-Eric Sahlin, Birgitte Wandall, Lars Lindblom and Kalle Grill for their detailed and constructive criticism, as well as all members of the risk seminar whose help has guided me along the way. 1 41 when forced by regulation. A paradigmatic example of the discrepancy between the views of experts and the public is nuclear power production: a large number of people regard nuclear power as dangerous even though most experts regard it as being a safe method of energy production. A more recent example is GMO, genetically modified organisms, which is considered safe to a much higher degree by experts than by laymen. The influence of laypeople if often characterised as harmful,2 and the analysis of the experts characterised as complete in itself, leaving no more than the implementation of their recommendation to the decision makers.3 Reasoning in risk communication frequently rests on the premise that people are too ignorant of the real risk and tend to make uninformed judgments concerning it.4 Therefore, it is asserted, a major objective is to adjust the perceptions of laypersons in order to narrow this gap.5 In this article, normative objections are put forward against the argument that the public should follow the experts’ advice because the experts have better knowledge of the risk. Let us refer to this argument as the Expert Argument: (1) Experts recommend X.6 (2) Experts have better knowledge than the public of the risk involved in X. --(3) Therefore: the public should accept X. The aim of this article is threefold. First and foremost, I will argue that even reasonably restricted, the Expert Argument is not valid. It is not valid because there is always epistemic uncertainty involved apart from estimations of the risk. This highlights a vital concern: the question is not whether the Durodié (2003a, 2003b). Ackerman and Heinzerling (2002). These examples are found in Hansson (in press). 4 Leiss (2004), for example, claims that the essential problem in risk management is the public distrust of the risk assessments and point to information made accessible to the public as the solution. 5 This aim is expressed in Kraus, Malmfors, & Slovic (1992/2000), 312. 6 ‘X’ may be an activity or technology like nuclear energy production of energy, or a substance like a medical drug. 2 3 42 expert knowledge of the risk is the best available, but whether it is good enough. Second, I will show that the objection from epistemic uncertainty is more than a logical possibility. Central tendencies from the empirical study of risk perception, though criticised by experts, may be considered reasonable in light of considerations from epistemic uncertainty. Third, I will show that the scope of the objection from epistemic uncertainty covers the entire field of risk research, risk assessment as well as risk management. In Section 2, some results from empirical studies of risk perception are presented in order to distinguish between discrepancies not defended and “discrepancies” defended for the same reasons that render the Expert Argument invalid. Section 3 consists of some specifications and limitations of the Expert Argument. In Section 4, the basic objection to the Expert Argument is presented, and in Section 5 I argue that the invalidity of the Expert Argument is more than a mere theoretical possibility. In Section 6, the scope of the objections is explored and in the final section, Section 7, it is concluded that focusing on ways of describing and communicating epistemic uncertainty is necessary to make the best possible decisions regarding risks. 2. The perception of risk The claim that the risk perception of laypersons is not well calibrated with the real risks has been supported by results from empirical studies of risk behaviour. I will not dispute that the risk perception of the public contains several idiosyncrasies, and certainly not argue against giving the public more information about all areas of risk related activities. What is argued, however, is that some important general tendencies of the risk perception of the public, often criticised as “discrepancies”, may be considered reasonable for the same reasons that renders the Expert Argument invalid. Thus, taking these “discrepancies” seriously will point to the reasons for the invalidity of the Expert Argument as well as show the practical relevance of this invalidity. 43 Let us start discussing some claims regarding discrepancies that will not be disputed in this paper. Many of the findings on risk perception show tendencies that might be regarded as irrational. One type of discrepancy may be labelled value incoherence. For example, when asked about acceptable levels of risk for an activity, different framings of the question – whether stated in lives saved or lives lost – render different answers from the respondents.7 Likewise, ordinary laypersons may assess a program that saves, for example, 4500 out of 11000 lives as more valuable than one of equivalent cost that saves 10000 out of 100000, even though the total number of lives saved is much higher in the second case.8 This is also mirrored in the fact that actual spending per lives saved varies with several magnitudes; Slovic (1997) shows a variation from US$ 500 up to 10 millions for life saved by various interventions, and Ramsberg & Sjöberg (1996) show a similar variation.9 Another type of discrepancy is epistemic, caused by insufficient knowledge of risks. One such example is the perception of the relation between dose/exposure (of radiation and chemicals) and risk. Kraus et al (1992) show that the public tend to be much less sensitive to dose and exposure considerations than are experts.10 Chemicals, for example, tend to be perceived as either safe or unsafe, regardless of quantities such as dose or exposure.11 In many areas where the connection between exposure and harmful effect is well documented, this perception may be regarded as a genuine lack of knowledge on part of the public. From cases of discrepancies like these it may be tempting to conclude that the dispositions of the public in general in matters of risk are unreliable. This conclusion is however premature. One important tendency of the public that is tempting but – as will become clear – wrong to regard merely as a discrepancy is the tendency to base acceptance or rejection of risk mainly on the possible harmful consequences of the risk. Several studies have Slovic (1997/2000), 394. Fetherstonhaugh, Slovic, Johnson, & Friedrich (1997/2000). 9 Slovic (1997/2000) and Ramsberg & Sjöberg (1996). 10 Kraus, Malmfors, & Slovic (1992/2000). 11 Ibid, 309. 7 8 44 shown that the severity of consequences is by far the strongest parameter for explaining acceptance of an activity that involves risk. Starr (1969) showed that the accepted level of risk was inversely related to the number of people exposed.12 In the seminal psychometric study by Fischhoff et al (1978), seven different influencing aspects where examined, and severity of consequences was by far the most influential factor regarding whether the risk was considered acceptable or not.13 More recently, Sjöberg (1999, 2003, 2004) has demonstrated this tendency in various contexts.14. Let us call this the Consequence Dominance: (Consequence Dominance) The perceived severity of the harmful consequence is the predominant factor that determines whether an activity is accepted or rejected. For the expert eye, this tendency is deplorable, since an improbable outcome such as an airplane crash may be used as an argument for preferring going by car even if the expected outcome of the latter action is much worse, or for preferring old means of energy production instead of nuclear power in light of the catastrophic potential of a meltdown regardless of how unlikely this outcome may be. Therefore, this looks like yet another misconception that should be corrected, another instance of the application of the Expert Argument. In Section 5, however, this tendency will stand out as reasonable once the problem with the Expert Argument is revealed. There is another empirical result that provides a link for understanding why the Consequence Dominance is telling against the Expert Argument. Several studies by Sjöberg et al (1993, 1999, 2001) have shown an asymmetry between safety and (positive) utility, which we will call the Value Asymmetry of Safety.15 Starr (1969). Fischhoff, Slovic, Lichtenstein, Read & Combs (1978/2000). The seven aspects were Voluntariness of risk, Immediacy of effect, Knowledge about risk, Control over risk, Newness, Chroniccatastrophic, Common-dread and Severity of consequences. (p. 86). 14 Cf. Sjöberg, (1999), Sjöberg (2003) and Sjöberg (2004). 15 Sjöberg & Drottz-Sjöberg (1993), Sjöberg (1999), Sjöberg & Drottz-Sjöberg (2001). 12 13 45 (Value Asymmetry of Safety) Avoiding bad outcomes is much more important than receiving good ones. This value asymmetry states that in choosing between an activity that provides a utility for sure and another that most likely provides a higher utility but also has a probability for a bad outcome, there is a tendency to prefer the former activity. For example, if choosing between having your own old car and an alternative with a high chance of getting a new expensive car but also a small risk of losing the car you have, the tendency is to go for the “safety first”-alternative and keep your car.16 If the Value Asymmetry of Safety is used as a decision principle it may be criticised for being too conservative or even inconsistent, since there may always be a small probability for a bad outcome whatever we do.17 But even if strong versions of this principle – not accepting any risk of a bad outcome – may be unacceptable (or even inconsistent), there is a strong case for some version of such a principle from both legislative praxis and common sense moral arguments.18 More importantly, the principle is actually an implicit premise in the Expert Argument. The Expert Argument mentions only the risks involved – the question of benefits does not enter into the equation (other than in the implicit way that a certain risky activity would not even be considered if it did not bring expected benefits). Thus, the knowledge and evaluation of benefits are not deemed relevant for the argument. The The seminal paper Ellsberg (1961) shows a similar tendency. This tendency implies, of course, not that there is no breaking point: for example if your own car is breaking down in any minute anyway, if the value of the new car is very high for you or if the risk of losing your own car is very small. 17 In its strongest interpretation it is an instance of the maximin rule in decision theory, a rule saying that one should choose the alternative that maximises the worst possible outcome. 18 In present risk management, the critique of pure risk-benefit analysis is a sign of the general acceptance of the principle of the Value Asymmetry of Safety: it is not acceptable for an activity to be too risky, even if the gains are significant (c.f. Hansson (1993)). Of course, we may accept an activity that includes risks even though we could avoid it. Automobile driving is a classic example of an activity we seem to accept even though it is to blame for a significant percentage of the annual unnatural deaths. Other voluntary activities like rock climbing and parachuting are examples of situations where we are prepared to increase risks in exchange for something we desire. Therefore, we have an asymmetry here rather than “safety at all costs”. 16 46 question of safety is therefore the primary concern implicit in the explicit mentioning only of the risk. The Value Asymmetry of Safety is thus not only an empirical fact of public perception, but also (albeit implicitly) accepted by proponents of the Expert Argument. This principle points to a shift from considerations of risk to considerations of safety, which in Section 4 will be shown to have important consequences for our evaluation of the Consequence Dominance and the Expert Argument. 3. Specifications of the Expert Argument The Expert Argument starts with a descriptive premise and ends with a normative conclusion. As was famously pointed out by David Hume, such argument is dependent on a bridge premise that links descriptive facts to normative evaluations.19 The relevant bridge premise can be formulated as follows: (φ) If A has better knowledge of the risk involved in X than B, and there is no agent with better knowledge of the risk than A, then B should follow the recommendation of A.20 The main critique of the Expert Argument proposed here focuses on this bridge premise. Even if experts recommend X, and experts have better knowledge of risk than the public, it is not at all certain that (φ) is valid. Therefore, what has been characterised as a disagreement between experts and the public is not sufficient grounds for the normative conclusion that the public should accept X. This is the thesis called Hume’s law, stating that a moral conclusion cannot be validly inferred from non-moral premises. In this case, (1) and (2) are descriptive, non-normative (and thus non-moral) claims, whereas (3) is normative. Some interpretations of the thesis are controversial for moral philosophers, but the weak logical interpretation I make use of here is not much disputed. Cf. Hume (1967 [1739-40]), 469-470, for the original formulation. 20 We need the subordinate clause, since in the general case, there could be a third agent having even better information and both agents should follow the recommendation of this third agent rather than anything else’s. Thus, we interpret (1) as entailing that there is no third agent with better knowledge than the experts in question. 19 47 For the Expert Argument to offer any initial plausibility, we have to state certain limitations. Firstly, it will be assumed that what is broadly captured by the term “experts” are the relevant technical and scientific experts specialised in X as well as the non-technical experts on managing risks, i.e. professionals preparing and making risk decisions.21 Secondly, we assume that there are no specific morally relevant considerations involved, such as considerations of justice.22 This is indeed idealised, but may be granted since the aim of this paper is to show that even under these circumstances the Expert Argument is invalid. 4. The objection from epistemic uncertainty Risk is the central concept in the Expert Argument. In different contexts, ‘risk’ is used to refer to quite different things. It may mean the probability of a negative outcome, or a negative outcome itself, or the cause of a negative outcome.23 All relevant conceptions of risk refer in some way to harmful events and/or some sort of estimate of their likelihood. The standard technical approach to quantifying risks is to consider the risk to be the statistical expectation value of the severity of the harmful event.24 Thus, to get a measure of the risk of a certain activity we should use the best estimation possible of the probability for a harmful event multiplied with its severity (e.g. the expected number of casualties).25 This technique needs very strong premises, since it assumes not only that it is meaningful to ascribe values to harmful events and compare them, but that we should take the If we only include the scientific experts it may be argued that the Expert argument doesn’t even get off the ground, since the prevailing opinion is that scientific experts should produce scientific facts, not decide courses of action. This argument will be met in Section 6, where the question of science and values will be touched upon. 22 What is morally relevant is of course a matter of debate, since anything affecting the well being of people is a moral matter according to utilitarians. “Relevant according to common sense morality” may be a somewhat more precise statement. 23 In Möller et al (2005), we mention five different interpretations of risk. 24 For example, the International Organisation for Standardization (2002) defines risk as “the combination of the probability of an event and its consequences” which may reasonably be interpreted as referring to the expectation value. See also Cohen (2003) for a recent example. 25 More correctly, we should sum up the values for all different harmful outcomes, since there is often more than one harmful event to take into consideration. 21 48 product of these values as the measure of the risk.26 However, for our purposes we may concur that the risk associated with X is the statistical expectation value of the severity of the harmful event. We may here accept this strong conception of risk, even if weaker conceptions may be considered more reasonable, since this gives a maximum of force to the Expert Argument. Weakening the assumption only gives us further reasons against it.27 In decision theory, decisions under known probabilities are called “decisions under risk”.28 Known probabilities are always an idealization, of course, since even a perfectly normal coin-tossing situation may be biased in some subtle way, not to mention any normal complex decision situation. In reality, there is always an epistemological primacy to our ascriptions of probabilities and outcome evaluations: the probability and degree of severity that we assign to an outcome are always based on the estimations we have access to. The objective probabilities – if such an entity exists beyond our epistemic limitations – are in general different from our estimations. This almost trivial point is always a matter of concern, but in general decision-making, it may sometimes be considered a matter of gaining on the swings what you lose on the roundabouts: the estimates may be just as often too high as too low. This symmetric relation between over- and underestimating the probabilities involved may be argued to hold for most outcomes: using the best estimations we have may be the most rational way to go when making decisions. However, in the cases we are considering here, we have no such symmetry: instead the background assumption of the For the statistical expectation value to be meaningful we must be able to compare the utilities of outcomes not only on a ordinal scale, like grading them from best to worst, but on a cardinal (interval) scale, like the measurement of temperature in Celsius or Fahrenheit. (Cf. e.g. Resnik (1987), especially 81-85, for an introduction of utility scales). When the only harmful effect we consider is number of casualties this assumption may be reasonable, but how much worse is e.g the loss of a leg compared to a whiplash injury? That these effects may be measured on a cardinal scale is very much a matter of controversy. 27 For example, if there is no justified way of scientifically comparing the value of the harmful effects, and thus no way of using an expectation value, referring to expert knowledge of the risk may be considered incorrect. 26 49 entire Expert Argument is what we above have called the Value Asymmetry of Safety. In the context in which we are interested, decisions involving risk are decisions regarding safety, and then swing-and-roundabout arguments are not valid. The reasons for this are many. Firstly, from the Value Asymmetry of Safety follows that we cannot “compensate” a higher risk with a higher utility if safety really is our primary concern. Secondly, we are not helped by arguments saying that the risk of other activities are probably overestimated even if activity X turns out to be underestimated, so that for the society at large it “evens out”. For what is at stake in the Expert Argument is explicitly the safety of activity X. Finally, and most importantly, when dealing with safety more than the best estimations of risk are involved, which the following paragraphs show. Often, safety is used as the antonym of risk: the less risky an activity is, the safer it is, and vice versa. The Value Asymmetry of Safety, however, suggests that something more must be taken into consideration than the best available knowledge of the risk.29 This is so because in the context of risky activities, most of our decisions are made with different levels of uncertainty about the probability and severity of harm. In those cases, shifting focus from risk to safety includes an additional epistemic aspect, as the following example illustrates: Suppose you are on the way to make a parachute jump in a poor country far away from home. You have never been to this beautiful place before and do not have much knowledge of the exact equipment you are about to use, but have previously read from a reliable source that the probability of equipment malfunctioning in the air is low, less than one in a hundred thousand on a normal jump. The equipment does look somewhat old and worn, however, and you cannot help feeling worried. Contrast this with a case in which a team of experts flown in from home tells you that this exact equipment (in light of its long usage) actually has a somewhat higher probability of malfunctioning, and that a better probability estimate is one in “Risk” here thus refers to the type of information at hand (known probabilities) and not whether the possible outcome is in any way harmful. 28 50 thirty thousand per jump. Even though the probability is now judged as higher, the uncertainty whether the information is trustworthy is significantly lower than in the previous case and it is not unreasonable to regard this situation as preferable to the first one in terms of safety. Whereas in the first case the knowledge at hand was such that it allowed for a rather large range of actual probabilities – it could be much less safe than your best estimate tells you – the estimation of the second case is much more certain. Different levels of epistemic justification are thus significant for decisions regarding safety. What is referred to in this example is the aspect of epistemic uncertainty. The safety of an activity in the context of decision-making is not only a function of risk in terms of the estimated probability and severity of harm, but also of the epistemic uncertainty of these entities. Now we may see that the vital premise in the Expert Argument that one should follow the recommendation of the agent with the best knowledge of the risk involved is far from evident. For us to say that the safety is high requires not only that the expected value of the harm is small, but also that the epistemic uncertainties involved are small enough.30 The premises in the Expert Argument are insufficient for the conclusion it wants to make; hence, the argument is invalid. 5. The objection from the Knowledge Asymmetry of Safety The invalidity of the Expert Argument is more than a mere logical possibility. The strong correlation between consequence and risk acceptance that we called Consequence Dominance (Section 2) may be interpreted as reasonable due to epistemic uncertainty. This goes contrary to expert opinion and gives us a counter-example to the Expert Argument. For a more elaborated treatment of the difference of the concepts of risk and safety than what follows here, see Möller et al (2005). 30 Note here that we have accepted, for the sake of argument, the expectation value conception of risk. However, the argument in this section is not dependent on this interpretation but is valid as long as risk is understood as some combination of the estimations of the severity of harm and its probability. 29 51 Consequence Dominance states that the predominant factor for accepting/rejecting a risky activity is the severity of the consequence of this activity. If, for example, activities A and B have the same probability for a lethal accident, but for activity A the number of casualties in an accident are higher than for activity B, then A is normally considered less safe than B and thus less likely to be accepted.31 This is consistent with the expectation value approach to risk. However, the tendency of the public is to prefer B to A as soon as the possible harm of A is considered significantly higher, even if the statistical expectation value of harm for A is significantly less than that of B. Focusing only on the outcome thus goes against general expert opinion and is judged as irrational. The modern classic is the nuclear power debate. Compared to the harmful effects of several competing methods of energy production (e.g. using coal, oil etc) most experts consider nuclear energy to be much safer. Still many people are reluctant to accept nuclear energy in view of the catastrophic potential of a meltdown, even though the estimated probability for such an event is deemed as very low. There are many important psychological factors such as control, affect, dread etc put forward in the literature to account for Consequence Dominance. Many of these have a restricted normative force in terms of accepting or rejecting an activity. In the case of dread as well as affect and control a relevant question is whether these attitudes are reasonable in light of the evidence. They seem to be descriptive claims only and we still need an analysis of why they carry any normative force. If such an explanation cannot be given, one might argue that it is the attitudes rather than the activities that are in need of changing. After a person has claimed that she feels more in control when driving a car than going by airplane, and thus safer, we may still ask what relevance this has in light of the amount of statistical data telling another (general) story. The epistemic uncertainty aspect of safety, however, provides a case for the normative force of the Consequence Dominance that the mere Let us say, for the sake of argument, that the severity of an event is measured in number of casualties. 31 52 descriptive factors fail to do. This case is built on a very plausible assumption concerning epistemic uncertainty that we may call the Knowledge Asymmetry of Safety. (Knowledge Asymmetry of Safety) In general, the lower the probability of a harmful event is estimated to be, the higher is the epistemic uncertainty about the severity of harm as well as its probability. However, the uncertainty regarding probability increases more than the corresponding uncertainty about the severity of harm.32 The plausibility of this comes from the nature of estimations of probability and harm. Starting with probability, we may say that the lower probability we ascribe to an event the more unusual the event should be, and therefore the less frequency data there normally is to justify the estimate. Thus, the lower our probability estimate, the less justification – hence, the greater the epistemic uncertainty. True, there are several other methods for justification of a probability ascription than relative frequency data, such as various forms of theoretical reasoning.33 For example, even if the probability for a hundred consecutive “tails” from coin tossing involves just as little epistemic uncertainty as a single one, this is due to the “ideal” character of the example. Some large systems may have parts where the case for theoretical models of probability estimations may be good,34 but societal activities involving human action and decision-making are just too complex and uncertain, and we have to rely also on frequency data. Hence, in general it is reasonable to expect the epistemic uncertainty of probability to increase when the probability estimate decreases. There is an implicit assumption here, namely that that all probability estimates are comparably low - as is the case with relevant societal activities and substances. For high probability events such a knowledge asymmetry is less likely: if the probability for an event is 40% and 50% respectively, there is probably no evident difference in uncertainty, ceteris paribus. 33 I.e. the classical interpretation (Laplacean view) of probability. Cf. Resnik (1987), ch. 3 (especially 3-3) for an introduction to various views on probability. 34 Various propensity interpretations of probability use the relevant theoretical laws as source of justification in addition to actual frequencies. 32 53 For harm, there is an open question regarding individuation. On the one hand, a harmful event may be an end-state outcome such as number of casualties.35 Then there is no uncertainty about the severity of harm and the first part of the claim of the Knowledge Asymmetry is invalid. However, the truth of the important second statement – that the increase in uncertainty about probability is higher than the corresponding increase in uncertainty about harm – is trivially true. Hence, the conclusion of the argument is still valid. On the other hand, a harmful event may be a non end-state harm. In these more interesting cases – such as “a plane crash” or a “a nuclear meltdown” – the uncertainty about the effect in human lives is evident. Analogous to the case of probability estimates, the lower the reasonable estimate, the less knowledge do we have of the harmful outcome of the event. In the beginning of nuclear energy production, for example, there was comparably little knowledge of the human effect of radiation. Thus, the estimations of the severity of harm had higher uncertainty than current ones. However, even though there is a correlation between probability estimates and the uncertainties about the severity of harmful events, this correlation is much weaker than the essential one mentioned above between probability estimates and the uncertainty of that estimate. This is because the severity of a harmful event is less dependent on the probability estimate for a certain activity: what is of importance if the total knowledge of that type of harm. Thus, if the harmful event is a certain amount of radioactivity, the cause of this harmful event is not important for the estimation severity of the harm. The less the knowledge of the type of event depends on the type of activity in question, the less correlation is there between the probability estimate and the uncertainty of the severity of an event. This means that the increase in uncertainty for lower probability estimates is, ceteris paribus, higher for probability than for severity of harm. In terms of our parachute illustration: even if the estimated probability for the malfunctioning of the 35 Assuming, of course, that our measure of harm is number of casualties. Otherwise, he uncertainty of how to value a certain outcome may be difficult, raising questions we 54 equipment decreases, the uncertainty about the severity of the harmful event of malfunctioning during a jump stays the same because we have good enough knowledge of what happens when falling to the ground from high altitudes. From the considerations of the Knowledge Asymmetry of Safety, the initial discussion of the section is presumably seen in a different light. When the possible harm of A was considered significantly higher than that of B but with lower expectation value of harm, at first glance it seems irrational to prefer B to A. However, when the epistemic uncertainty is included this is not at all certain, since the difference in epistemic uncertainty about the probabilities may be many times greater than the differences in the statistical expectation value of harm. The Knowledge Asymmetry of Safety is of course not a complete justification for a strict use of consequence reasoning alone when it comes to risky activities. Indeed it cannot be, since it describes a tendency on theoretical grounds for what on each occasion (to a significant degree) is an empirical matter. What the above considerations do, however, is to show the reasonability in putting a strong emphasis on the possible harmful outcome even if the probability is low, which is the criticised tendency of the Consequence Dominance. Expert criticism of such a tendency thus emerges as an instance of the invalidity of the Knowledge argument. Based on this knowledge asymmetry, the factors of control, affect and dread, if considered at least loosely connected to consequences, may be considered as reasonable dispositions to have since they make us avoid situations with low but uncertain (estimated) probabilities of great harm. The same applies to low acceptance of new and ‘unnatural’ activities.36 Thus, epistemic uncertainty provides a normatively relevant explanation of many common psychological tendencies surrounding risk perception. mentioned earlier about the possibility of comparing different harmful effects. However, this is not the outcome-uncertainty of which we are referring to here. 36 Fischhoff, Slovic, Lichtenstein, Read & Combs (1978/2000). 55 In summary, the Knowledge Asymmetry of Safety thus shows that rejection of the best estimation of the risk involved need not be a sign of irrationality but may in many cases be reasonable in matters of safety. Even if there is no party with better knowledge of the risk of a certain activity, that fact is not sufficient for accepting it. Depending on the epistemic uncertainty, the activity may be much more risky than the best estimations state, and that possibility may not be a price we are willing to pay.37 6. The scope of the objection from epistemic uncertainty Traditionally a distinction is made between risk assessment and risk management.38 Risk assessment is considered as the scientific stage where the estimations of the risks at hand are produced. Risk management uses the output of the risk assessment as input for making a decision about the risk, ultimately whether to accept or reject it. It may thus seem as if the Expert Argument does not affect risk assessment, since it explicitly mentions the recommendation of the experts and the purpose of risk assessment is not to make any recommendations but merely to state the scientific facts of the matter. However, the possibility of upholding this distinction between risk management and risk assessment has been severely questioned.39 At least when it comes to complex activities on a societal level, the borderline between scientific expert and decision maker is considerably blurred. The toxicologist analysing the structure and receptor In light of the reasoning in this section, the public rejection of what experts recommend need not be considered a sign of distrust of the experts. However, that is the way it is often interpreted especially in the risk communication field, with ”public relation” consequences such as focusing on how to receive better trust from the public in a way that does not have anything remotely to do with science and justification. In many cases, distrust may perhaps be reasonable, since the acclaimed experts may be stakeholders in the very same activity they are recommending. (In resource demanding areas, like parts of the chemical industry, the tests as well as the main part of the analysis is made by the industry itself. In such cases, the public scepticism may be especially reasonable. See Hansson (2004), 357-58). In general, however, the public rejection of expert recommendations need not and should not be seen as distrust for experts (Cf. Drottz-Sjöberg (1996), Sjöberg & Drottz-Sjöberg (2001) and Sjöberg (2001)). It may fruitfully be interpreted as distrust of the level of scientific knowledge. 38 Cf. National Research Council (1983) and European Commission (2003) for standard classifications. 39 Cf. Mayo (1991). 37 56 binding capacity of a molecule in a cell-culture may not be considered a decision maker in the relevant sense any more than the mathematician working with the basic theoretical models underlying such an analysis. But what about the toxicological experts responsible for the complex evaluation giving the outcome of the risk assessment of this molecule? In this section, it is argued that epistemic uncertainty is an important decision factor in the scientific process of risk assessment, having consequences for the scope of the objections to the Expert Argument. Until now we have been focusing mainly on the impact of the best estimations of risk on decisions involving safety, i.e. on what is traditionally called risk management. The presence of epistemic uncertainty questions the claim that we should follow those with the best knowledge of the risk. It may be perfectly reasonable to accept the estimations of the risks involved and still, due to epistemic uncertainty, reject the recommendation of the experts.40 The experts and the public are then using different methods for handling this uncertainty, and only referring to the best available estimation of the risk is not sufficient for settling the issue. However, epistemic uncertainty is not a product of risk management but of risk assessment. Among the data relevant for risk assessment there are likely to be parts that are based on a core of scientific theories and observational data with relatively high evidential justification.41 But the more models and approximations of limited justification and application are included in the final inferences, the more these evaluations – and thus the output of the process – contain epistemic uncertainty. This may be illustrated in figure 1. The layers in the triangle to the left illustrate different Note that ’experts’ here is understood in a broad sense (as stated in Section 2), including not only scientific experts but also ‘risk management experts’, i.e. experts with understanding of the relevant risk assessment data that are responsible for making recommendations for decisions. 41 The following description of scientific knowledge as a core of deeply justified knowledge and a less deeply justified periphery should be seen as in line with Willard van Orman Quine’s holistic “web of knowledge” approach. Cf. Quine (1951). The necessity of theory in making observations (a reason for including theoretical statements as well as observational statements in the “core”) is commonly labelled theory-laden observation; 40 57 degrees of justification for the different scientific data and relations. The bottom layer of the figure consists of the core data of direct observations and theoretical knowledge. The higher levels illustrate scientific hypotheses of less direct justification.42 An example is carcinogenic effects of chemical substances. Our primary interest is the effects on humans, but for obvious reasons we mainly perform animal testing and extrapolate the results. The direct observations of animal effects may here represent lower levels in the figure, whereas the extrapolations to humans have a higher degree of epistemic uncertainty and are represented in the upper levels of the figure. The data serving as input to the risk assessment consists of complex inference chains with different levels of justification and, hence, different degrees of epistemic uncertainty. High uncertainty * * * * * * * * Low uncertainty Figure 1. An illustration of levels of scientific knowledge and their corresponding uncertainties. The arrows in the left triangle represent inferences from hypotheses at the lower levels to the upper levels. Estimates of risks are often derivates on the upper levels and thus more uncertain. This is illustrated by the width of the grey triangle to the right: epistemic uncertainty is small in the lower levels and higher in the upper ones. Hanson (1958) coined the phrase. Cf. Churchland (1979), Scheffler (1982) and Hesse (1974) for important contributions in support of this claim. 42 Even the “core statements” are, in the holistic view, susceptible to change in view of the “peripheral statements”. Arguably, due to the theory-ladenness of observation mentioned above, observations come with epistemic uncertainties partly from the uncertainties of the higher-level statements. The uncertainty dimension in the figure is thus rather simplified to stress the point made here. The ”tip” of the right triangle being pointed, e.g., should not be taken to mean that there are data with no uncertainty whatsoever. 58 In science in general, it may be argued that there are acknowledged methods of handling these different grades of epistemic uncertainty, since there are certain epistemic values “integral to the entire process of assessment in science”.43 Epistemic values are normally taken to be internal to science, i.e. they are exactly the kind of values that the scientific experts are competent to apply in their scientific enterprise.44 Thus, in science there exist standards for when to acknowledge a hypothesis and regard it as scientific knowledge as well as when to discard it, i.e. ways to handle epistemic uncertainty. However, risk assessment does not have the same goal as science. The main goal of science is to gain knowledge of the world, and the function of epistemic values is to accept as scientific knowledge only what has been sufficiently “confirmed”.45 For risk assessment, however, as has been argued throughout this article, the goal is safety. Everything there is reason to consider a potential harm is relevant to risk assessment, even if the “purely” scientific case cannot (as of yet) be made.46 Therefore, epistemic values made for the binary goal of inclusion/exclusion do not provide the full story for risk assessment. The aggregation of the data to assessment output is thus made with reference – implicit or explicit – to different methods for aggregation of uncertainties; thus, with reference to other values than those internal to science.47 An illustrating example by Nils-Eric Sahlin and Johannes Persson shows how different estimations of the epistemic uncertainty, not different scientific data, results in different safety levels regarding dioxin in fish for McMullin (1982), 6. Cf. ibid for a standard characterisation of epistemic values as opposed to other types of values. In the last decades of philosophy of science, there has been a lively discussion as to the extent of values in science, especially after Thomas Kuhn’s seminal essay (1962) and Lakatos & Musgrave (1970). However, the minimal view is to acknowledge what is called epistemic values (sometimes called cognitive values) in science as opposed to ethical and other values. 44 This view is criticised by many (Cf. Mayo (1991), 254-255). Brian Wynne, e.g., claims that scientific rationality is “intrinsically conditioned by social commitments” (Wynne (1982), 139). For our purposes, however, we need not argue against this claim and may allow the view that epistemic values are intra-scientific. 45 Sometimes, it is expressed that some of the epistemic values are the goals of science, e.g. empirical adequacy and explanatory power (Cf. McMullin (1993)). 46 For example, a substance may have been proven harmful to rats but due to restrictions on human testing it is uncertain whether it is potentially harmful to us. 43 59 the United States and the Nordic countries.48 Different estimations on how the sensitivity of humans compare to the test animals and different safety factors resulted in a difference of almost 1000 in magnitude for the safety levels of dioxin intake (5 pg/kg body weight and day in the Nordic countries compared to 0.0064 pg/kg for the EPA, American Environmental Protection Agency).49 The conclusion to be drawn from the above considerations is that even if risk assessment does not conclude in an explicit recommendation whether or not to accept a certain activity or substance, it is susceptible the same type of objections due to epistemic uncertainties. Just as was the case when we focused on the risk management process, which weight we should put on different cases of uncertain knowledge in the risk assessment stage reflects our values in regard to uncertain knowledge and is thus an open question not to be settled only by those with the best knowledge of the risk. Therefore, depending on where we draw the line for the scientific process we get different “Expert Arguments”. If we say that the scientific process actually does conclude in statements of safety, such that “activity X is safer than Y”, we may in practice consider this very close to a recommendation indeed, and the step to our Expert Argument is short. If, on the other hand, the outcome of the scientific process is only the most basic scientific data, the core level in figure 1, we get considerably further away from the Expert Argument, since the gap from data presented at its fullest to recommendation will be quite large. But then the question emerges how this data itself may be of help to risk management, since it will probably be in a form that is in great need of interpretation and merging to be possible to manage – i.e., of “risk assessment”. It seems reasonable to conclude that when the outcome of the risk assessment is such that it is possible to use as basis for a decision regarding a complex societal activity, it is on such a high level that it includes, explicitly A similar point is made in Wandall (2004). Sahlin & Persson (1994), 38 ff. For a more general analysis of different health standards and the limits of science, cf. Hansson (1998), especially chapters 1 and 4. 47 48 60 or implicitly, recommendations or proto-recommendations. Therefore, since there is no possibility of “insulating” risk assessment from epistemic uncertainty and thus the kind of normative aspects we recognised in risk management, the objections of the Expert Argument based on considerations of epistemic uncertainty are relevant also for risk assessment. 7. Summary and conclusion The vital premise (φ) of the Expert Argument stated that one should follow the recommendation of the agent with the best knowledge of the risk involved. Our discussion has given us reasons for not accepting the validity of (φ): reasons from safety considerations and epistemic uncertainty. Even where there is expert agreement about a risk estimate, the presence of epistemic uncertainty makes it an open question whether the risk at hand should be accepted. The Expert Argument is thus invalid. This invalidity is not only a logical possibility, but is evident in the debate over empirical results. One such result is the Consequence Dominance, the tendency to perceive risks as unacceptable mainly based on the potential harmful outcome. Such a tendency is considered irrational by experts, but in view of the presence of epistemic uncertainty and the specific relation of the Knowledge Asymmetry of Safety, it was interpreted as a reasonable disposition to have and thus as an instance of the very invalidity of the Expert Argument. The last section considered the scope of the objection to the Expert Argument and involved the very basis for risk estimation: science itself. Every complex risk estimation has inferential elements involving aggregations as well as extrapolations with less than “certain” scientific status, and the different goals of science in general and risk research in particular makes it evident that other than epistemic values are involved in handling the uncertainties involved. Thus, the objections to the Expert argument are relevant for risk assessment as well as risk management. 49 Sahlin & Persson (1994), 40-41. 61 Trying to improve the scientific basis and reaching better estimations of the risk involved in hazardous activities, technologies and/or substances is certainly a major aim for risk assessment and risk management. However, epistemic uncertainty will always be present where risk is involved and the focus is on safety. The quality of knowledge differs for different activities, technologies and substances, and it is far from evident that what on the surface looks like the same degree of risk should be handled the same way. Therefore, we must actively focus on ways of measuring and communicating epistemic uncertainty, giving societal decision-making the best foundation possible. There are other important aspects outside the scope of the argument from epistemic uncertainty and a focus on safety, such as moral aspects of risk taking and questions of whether there is a just distribution of risks and benefits among the population.50 These aspects should be considered in addition to the epistemic uncertainty factor. Even without them, what has here been called Expert Argument is heavily questioned. Having the best knowledge of the risk is not enough. Epistemic uncertainty is a basic fact relevant to risk and safety decisions and this should be reflected in the entire field of risk research. 50 Mentioned e.g. in Slovic (1997/2000), 392. 62 References Ackerman, F. and Heinzerling, L. (2002) Pricing the priceless: cost-benefit analysis of environmental protection, University of Pennsylvania Law Review 150:1553-1584. Möller N., Hansson S. O., Peterson M. (2005), Safety is more than the antonym of risk, forthcoming in Journal of Applied Philosophy. Cohen, B. (2003) Probabilistic Risk Analysis for a High-Level Radioactive Waste Repository, Risk Analysis 23:909-915. Churchland, P. (1979) Scientific Realism and the Plasticity of Mind. Cambridge: Cambridge University Press. Drottz-Sjöberg, B.-M. (1996) Stämningar i Storuman efter Folkomröstningen om ett Djupförvar (Projekt Rapport No. PR D-96-004). Stockholm: SKB. Durodié, B. (2003a) The True Cost of Precautionary Chemicals Regulation, Risk Analysis 23:389-398. Durodié, B. (2003b) Letter to the Editor Regarding Chemical White Paper Special Issue, Risk Analysis 23:427-428. Ellsberg, D. (1961) Risk, Ambiguity and the Savage axioms, Quarterly Journal of Economics 75:643-669. European Commission (2003) Technical Guidance Document in support of Commission Directive 93/67/EEC on Risk Assessment for new notified substances, Commission Regulation (EC) No 1488/94 on Risk Assessment for existing substances and Directive 98/8/EC of the European Parliament and of the Council concerning the placing of biocidal products on the market. Luxembourg: Joint Research Centre, EUR 20418 EN, Office for Official Publications of the EC. Fetherstonhaugh, D., Slovic, P., Johnson, S. M., & Friedrich, J. (1997/2000) Insensitivity to the Value of Human Life: A Study of psychophysical Numbing, in Slovic (2000), 372-389. Fischhoff, B., Slovic, P., Lichtenstein, S., Read, S., & Combs, B. (1978/2000) How Safe Is Safe Enough? A Psychometric Study of 63 Attitudes Toward Technological Risk and Benefits, in Slovic (2000), 80103. Hanson, N. (1958) Patterns of Discovery. Cambridge: Cambridge University Press. Hansson, S.O. (1993) The false promises of risk analysis. Ratio 6:16-26. Hansson, S.O. (1998) Setting the limit: Occupational Health Standards and the Limits of Science. Oxford: Oxford University Press. Hansson, S.O. (2004) Fallacies of risk, Journal of Risk Research 7:353-360. Hansson, S.O. (in press), Seven Myths of Risk, Risk Management. Hesse, M. (1974) The structure of Scientific Inference. Berkeley: University of California Press. Hume, D. (1967 [1739-40]) A Treatise of Human Nature. Ed. Selby-Bigge, L. A. Oxford: Clareton Press. International Organization for Standardization (2002) Risk Management – Vocabulary – Guidelines for use in standards, ISO/IEC Guide 73:2002. Kuhn, T. (1962) The Structure of Scientific Revolutions. Chicago: University of Chicago Press. Kraus, N., Malmfors, T. & Slovic, P. (1992/2000) Intuitive Toxicology: Experts and Lay Judgements of Chemical Risks, in Slovic (2000), 285315. Lakatos, I. and Musgrave, A., eds. (1970) Criticism and the Growth of Knowledge, London: Cambridge University Press. Leiss, W. (2004) Effective risk communication practice, Toxicology Letters 149:399-404. Mayo, D. (1991) Sociological Versus Metascientific Views of Risk Assessment, in Acceptable Evidence, Science and Values in Risk Management, 249-280. Eds. Mayo, D. G., and Hollander, R. D. Oxford: Oxford University Press. McMullin, E. (1982), Values in Science, PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association 2:3-28. 64 McMullin, E. (1993). Rationality and paradigm change in science. In World Changes: Thomas Kuhn and the Nature of Science, ed. P. Horwich, 55-78. Cambridge: The MIT Press. National Research Council (1983) Risk assessment in the federal government – managing the process. Washington: National Academy Press. Quine, W. (1951) Two dogmas of empiricism, Philosophical Review 60: 20-43. Ramsberg, J. & Sjöberg, L. (1996) The cost-effectiveness of lifesaving interventions in Sweden, Rhizikon: Risk Resesarch Report No. 24, 271-290. Stockholm: Center for Risk Research. Resnik, M. (1987) Choices: An introduction to decision theory. Minneapolis: University of Minnesota Press. Sahlin, N.-E. & Persson, J. (1994) Epistemic Risk: The Significance of Knowing What One Does Not Know, in Future Risks and Risk Management, 37-62. Eds. Brehmer, B. & Sahlin, N.-E. Berlin: Springer. Scheffler, I. (1982, 2nd ed.) Science and Subjectivity. Indianapolis: Hackett. Slovic, P. (1997/2000) Trust, Emotion, Sex, Politics and Science: Surveying the Risk-assessment Battlefield, in Slovic (2000), 390-412. Slovic, P. (2000) The Perception of Risk. London: Earthscan. Sjöberg, L. (1999) Risk Perception in Western Europe, Ambio 28 (6):543549. Sjöberg, L. (2001) Limits of Knowledge and the Limited Importance of Trust, Risk Analysis 21:189-198. Sjöberg, L. (2003) Risk Perception, Emotion, and Policy: The case of Nuclear Technology”, European Review 11:109-128. Sjöberg, L. (2004) Explaining Individual Risk Perception: The Case of Nuclear Waste, Risk Management: An International Journal 6 (1):51-64. Sjöberg, L. & Drottz-Sjöberg, B.-M. (1993) Attitudes to Nuclear Waste, Rhizikon: Risk Research Report No. 125. Stockholm: Center for Risk Research. Sjöberg, L. & Drottz-Sjöberg, B.-M. (2001) Fairness, risk and risk tolerance in the siting of a nuclear waste repository, Journal of Risk Research 4:75101. 65 Starr, C. (1969) Social benefit versus technological risk, Science, 165:12321238. Wandall, B. (2004) Values in science and risk assessment, Toxicology Letters, 152:265–272. Wynne, B. (1982) Institutional mythologies and dual societies in the management of risk, in The risk analysis controversy, 127-143. Eds. Kunreuther, H. & Ley, E. Berlin Springer-Verlag. 66 Submitted manuscript INTERPRETING SAFETY PRACTICES: Risk versus Uncertainty Niklas Möller and Sven Ove Hansson Division of Philosophy, Royal Institute of Technology ABSTRACT. There are many principles and methods recommended for the engineer as means to ensure safety, as well as methods for assessing the safety of a system. In Probabilistic risk analysis (PRA) and Probabilistic safety analysis (PSA), risk is interpreted as a combination of the probability of an adverse consequence and the severity of the consequence, and safety is seen as the antonym of risk. This account is insufficient on theoretical grounds. Furthermore, common practices in safety engineering supply evidence of its insufficiency. Putting forward a number of principles referred to in the literature, and focusing on three important covering principles of safety engineering, Inherently safe design, Safety reserves and Safe fail, we show that engineering principles for achieving safety cannot be fully accounted for on a probabilistic risk reduction interpretation. General theoretical reasons and safety practices alike show that an adequate concept of safety must include not only the reduction of risk but also the reduction of uncertainty. Keywords: safety, risk, uncertainty, probabilistic risk analysis, safety engineering, probabilistic safety analysis, inherently safe design, safety factors, safe fail 1. Introduction Safety is a concern in virtually all engineering processes and systems.1 There are many principles and methods recommended for the engineer as means to ensure safety. There are also methods for assessing the safety of a system. A dominating quantitative method is probabilistic risk and safety analysis. Safety is here conceived as the antonym of risk, and risk is interpreted as a combination of the probability of an adverse consequence and the severity 67 of the consequence. On such an understanding, achieving safety is conceived as risk reduction. Risk reduction is indeed an important aspect of gaining safety. We will argue, however, that it is insufficient as a total understanding of safety: ensuring safety is not only a matter of reducing the risk. The purpose of this paper is to show that common engineering principles, if taken seriously, cannot be fully accounted for on a narrow risk reduction interpretation. This is not due to deficiencies of those practices, however, but to a shortcoming in the capability of the theoretical framework to capture the concept of safety. We will propose that an adequate concept of safety must include not only the reduction of risk but also the reduction of uncertainty. (Uncertainty differs from risk in the absence of well-determined probabilities.) Only with such a broadened concept of safety can we adequately account for the success of important engineering principles in achieving safety. 2. Safety as risk reduction Contemporary risk and safety assessment is dominated by quantitative, probabilistic methodology. Methods such as Probabilistic risk analysis (PRA) and Probabilistic safety analysis (PSA) are frequently employed. The standard interpretation of risk is here in terms of the possible adverse consequences and their probabilities: the larger the probability of a hazardous event and the more severe this event is, the larger the risk at hand.2 The expected value, the product of the probability and a measure of the severity of consequences is used as a measure of the risk. Quantitative risk and safety analysis aims at measuring the risk of a system – in this sense of risk – and then strives to minimize it; thus reaching the best possible safety. A characteristic example is the following statement of the aims of nuclear safety engineering: “Nuclear power plants shall be designed, constructed and operated to an appropriate quality to minimize the likelihood of failure which could lead to the escape of significant quantities of radioactive material.”3 Likewise, addressing the issue of acceptance criteria, the classical probabilistic perspective is put forward: “Radiological acceptance criteria for the safety of 68 a nuclear power plant follow the principle that plant conditions associated with high radiological doses or releases shall be of low likelihood of occurrence, and conditions with relatively high likelihood shall have only small radiological consequences.”4 Stated in such a vague manner, the identification of safety with risk reduction appears uncontroversial. Clearly, if the probability or the severity of an accident has been reduced, then this is an advantage in terms of safety. However, it does not follow that a complete characterisation of safety can be obtained with reference only to probability and severity. In particular, an account restricted to these variables assumes that they are both known (or knowable). Hence, there is an underlying premise of knowability regarding the probabilities and consequences involved. In decision theory, such a situation is called decision under risk.5 Events like coin-flipping or roulette spinning are paradigmatic examples of such decision situations: if the coin or roulette is correctly made, the probabilities as well as the possible outcomes are known. Decisions in most complex areas of life are not based on probabilities that are known with certainty. Strictly speaking, the only clear-cut cases of decision under risk (known probabilities) seem to be idealized textbook cases such as the abovementioned coin-tossing. Even statistical data based on an abundance of experience, such as the probability of rainfall in June in a given European city or the accident frequency of a Boeing 747, are not fully known in this idealized way. They provide important information to be used in decisions, but it is perfectly reasonable to doubt whether the values apply to the situations at hand; they are characterized by epistemic uncertainty and should not be treated as probabilities known with full precision. Hence, almost all decisions are decisions under uncertainty. To the extent that we make decisions under risk, this does not mean that these decisions are made under conditions of completely known probabilities. Rather, it means that we have chosen to simplify our description of these decision problems by treating them as cases of known probabilities. 69 The presence of epistemic uncertainty specifically applies to engineering contexts. An engineer performing a complex design task has to take into account a large number of hazards and eventualities. Some of these eventualities can be treated in terms of probabilities; the failure rates of some components may for instance be reasonably well-known from previous experiences. However, even when we have a good experiencebased estimate of a failure rate, some uncertainty remains about the correctness of this estimate and in particular about its applicability in the context to which it is applied. In addition, in every system there are uncertainties for which we do not have good or even meaningful probability estimates. This may include the ways in which humans will interact with as yet untested devices. It may also include unexpected failures in new materials and constructions or complex new software, and many other types of more or less well-defined hazards. It always includes the eventuality of new types of failures that we have not been able to foresee. In the presence of epistemic uncertainty the identification of safety with risk reduction is problematic. The best estimate may state that the probability that a particular patient survives with a left ventricular support device by, say, 15 years is 50 percent. Yet, we have good reasons to be uncertain of this estimate in a way we do not have to be in the case of coin tossing. We have far more reason to believe in the probability estimates of some failures than others, and as soon as there is uncertainty it is an open question whether the probability of a harmful consequence is in fact higher than what our best estimates say. If we can choose between two components such that the best estimate of the probability of malfunction is the same in both cases but the uncertainty about the first is greater than about the second, then we have good reasons to use the second. In terms of safety, ceteris paribus, an old computer system that has been running in the environment for years may be preferable to a new one, even if the best estimate of the probability of malfunction for the latter is the same (or even somewhat lower). 70 In summary, engineering safety always has to take into account uncertainties that can be meaningfully expressed in probabilistic terms as well as eventualities for which this is not possible. In the following sections we will argue that a reasonable interpretation of several common engineering practices is to view them not only as methods for reducing the risk but also as methods for reducing the uncertainty. 3. Principles and methods for engineering safety Although there are many specific treatments of safety considerations in the field of safety engineering, we are not aware of any fully general account. The most common approach in the literature is to list a number of different principles and practices on different levels without the pretence of giving a general account.6 In the table in Appendix 1 we have identified a large number of such principles and methods recommended in the engineering literate (broadly construed). As can easily be seen, the principles are on different levels of abstraction and many are very closely related. Attempts at generalised accounts of safety measures are naturally abstract in their taxonomy. An example is Koivistio (1996), who divides safety considerations into three different types: (1) Adherence to good practice, (2) Safety analysis and (3) Inherently safe design.7 Bahr (1997) – based on NASA (1993) – gives a more substantial taxonomy and puts forward a stronger claim, categorising hazard reduction into a lexicographic ordering of importance:8 1) “Designing out” the hazard 2) Safety devices 3) Warning devices 4) Special procedures and training Firstly, Bahr writes, we should “design out” the hazard from the system.9 If that is not possible, we should control the hazard using various fail-safe devices, e.g. pressure valves relieving the system of dangerous pressure 71 build-up.10 When designing out or controlling is not an option, warning devices (e.g. smoke alarm) and procedures (e.g. emergency shutdown) and training should be used.11 We have divided the principles listed in Appendix 1 into four categories or covering principles:12 (1) Inherently safe design. A recommended first step in safety engineering is to minimize the inherent dangers in the process as far as possible. This means that potential hazards are excluded rather than just enclosed or otherwise coped with. Hence, dangerous substances or reactions are replaced by less dangerous ones, and this is preferred to using the dangerous substances in an encapsulated process. Fireproof materials are used instead of inflammable ones, and this is considered superior to using inflammable materials but keeping temperatures low. For similar reasons, performing a reaction at low temperature and pressure is considered superior to performing it at high temperature and pressure in a vessel constructed for these conditions. (2) Safety reserves. Constructions should be strong enough to resist loads and disturbances exceeding those that are intended. A common way to obtain such safety reserves is to employ explicitly chosen, numerical safety factors. Hence, if a safety factor of 2 is employed when building a bridge, then the bridge is calculated to resist twice the maximal load for which it is intended. (3) Safe fail. There are many ways a complex system may fail. The principle of safe fail means that the system should fail “safely”; either the internal components may fail without the system as a whole failing, or the system fails without causing harm.. One common example is fail-silence mechanisms: fail-silence (also called “negative feedback”) mechanisms are introduced to achieve self-shutdown in case of device failure or when the operator loses control. A classical example is the dead man’s handle that stops the train when the driver falls asleep. One of the most important safety measures in the nuclear industry is to ensure that reactors close down automatically in critical situations. 72 (4) Procedural safeguards. There are several procedures and control mechanisms for enhancing safety, ranging from general safety standards and quality assurance to training and behaviour control of the staff. Procedural safeguards are especially important in identifying new potential harms (audits, job studies) and controlling employee behaviour that cannot be “designed out” from the process (warnings, training). One example of such procedural safeguards is regulation for vehicle operators to have ample time between actual driving in order to prevent fatigue. Frequent training and checkups of staff is another. Procedural safeguards are important as a ‘soft’ supplement to ‘hard’ engineering methods. In light of the general considerations in the last section, we propose that each of the principles listed in the Appendix is best understood as a method that reduces the uncertainty as well as the risk. In what follows we will focus on covering principles (1) to (3) above. Two of these three principles, inherently safe design and safe fail, are covering principles into which many other methods and principles on lower abstraction levels can be categorised. This is especially so in the case of safe fail: more than a third of the listed principles can be categorised as applications of this principle. In fact, most of the principles in the list are included in one of these covering principles. We have two main reasons for focusing on these three covering principles. First, they represent three different perspectives on safe design: “removal” of the hazards to prevent failure, overdimensioning of the structure to prevent failure, or designing for failure not to be total failure. Secondly, they are (if not theoretically then at least in practice) ‘hard’ principles, focusing on technical solutions to safety problems. The methods of category (4), procedural safeguards, focus primarily on human action such as safeguards, training of staff, audits, behaviour control, etc. If, as we will show, the case for exclusively probabilistic analysis in technological cases looks bleak, there is not much hope to succeed when dealing with human behaviour and 73 interaction. Probabilistic treatment of human agent-hood is even more problematic than the cases we are explicitly dealing with in what follows. 4. Inherently safe design The most basic way of protecting a system is to remove the source of the hazard in question. If flammable substances are used in the system process, taking away such substances also takes away the potential for disaster. Inherently safe design thus refers to the practice of excluding the potential hazard entirely instead of controlling it. In practice, there are many forces working against the concept of inherently safe design. For reasons of efficiency and cost effectiveness, production plants, transport systems and other human designs have strived to control hazards rather than excluding them altogether. Furthermore, the dangerous element is often a fundamental part of the system, either as the very substance motivating the system (e.g. nuclear power plants), or as in other ways necessary for the system (e.g. fuel in various transport systems). Focusing on the principle of inherently safe design, however, has been a way of questioning the ‘necessity’ of many of these arrangements. For example, even if gaseous nitrogen is needed for a process step, the nitrogen bottles do not have to be located in the operators’ work area but can be moved outside, thus reducing the risk of asphyxiation.13 Often, it is possible to replace a hazardous substance with another, less hazardous one that can do the job. Well-known examples include using helium instead of hydrogen in balloons and reducing pesticide use by changing agricultural technologies. In the context of processing plants, Kletz (1991) has suggested five main principles – and several additional ones – for inherently safe(er) design: Intensification, Substitution, Attenuation, Limitation of Effects and Simplification.14 Above we have given examples of intensification (availability of nitrogen only in the necessary process step, not for storing in operator proximity) as well as substitution (hydrogen replaced by helium) and limitation of effects (pesticide). Attenuation refers to the method of attenuating substances to make them less flammable or toxic. A more 74 general principle includes avoidance of all types of hazardous “concentrations” such as accumulation of energy and storage of large quantities of hazardous substances in one place.15 An example of simplification is the method of making incorrect assembly of a system part impossible, for example by making asymmetric parts.16 Instead of training the staff to avoid incorrect assembly, failure is simply made practically impossible.17 The principle of inherently safe design aims at eliminating the sources of harm. The natural interpretation of such a method is that we ensure that the harmful event will not take place: if we remove the flammable substance, fire will not occur. Naturally, we may say that the probability of a harmful consequence is reduced. However, in most cases of inherently safe design we are dealing with issues that are hard to give probabilistic treatment. The principle of inherently safe design is best viewed as a method for protection against the unforeseen: it is not mainly the worker at the assembly line putting together the same parts a hundred times a day that is in need of asymmetric parts, but the passenger in a burning lower deck compartment trying to assemble the fire extinguisher.18 The principle of inherently safe design is a way to decrease the uncertainty about whether harmful events will take place. The soundness of the principle even in absence of any (meaningful) probability estimates is an indication that something is lacking from the traditional conception of safety as risk reduction. 5. Safety reserves Humans have presumably made use of safety reserves since the origin of our species. They have added extra strength to their houses, tools, and other constructions in order to be on the safe side. However, the use of numerical factors for dimensioning safety reserves seems to be of relatively recent origin, probably the latter half of the 19th century. The earliest usage of the term recorded in the Oxford English Dictionary is from WJM Rankine’s book A manual of applied mechanics from 1858. In the 1860s, the German railroad engineer A. Wohler recommended a factor of 2 for tension.19 The 75 use of safety factors has been well established for a long time in structural mechanics and its many applications in different engineering disciplines. Elaborate systems of safety factors have been developed, and specified in norms and standards. A safety factor is typically specified to protect against a particular integrity-threatening mechanism, and different safety factors can be used against different such mechanisms. Hence one safety factor may be required for resistance to plastic deformation and another for fatigue resistance. As already indicated, a safety factor is most commonly expressed as the ratio between a measure of the maximal load not leading to the specified type of failure and a corresponding measure of the applied load. In some cases it may instead be expressed as the ratio between the estimated design life and the actual service life. In some applications safety margins are used instead of safety factors. Although closely linked concepts, a safety margin differs from a safety factor in being additive rather then multiplicative. In order to keep airplanes at a sufficiently long distance from one another a safety margin in the form of a minimal distance is used. Safety margins are also used in structural engineering, for instance in geotechnical calculations of embankment reliability.20 According to standard accounts in structural mechanics, safety factors are intended to compensate for five major categories of sources of failure: 1) higher loads than those foreseen, 2) worse properties of the material than foreseen, 3) imperfect theory of the failure mechanisms, 4) possibly unknown failure mechanisms, and 5) human error (e.g. in design).21 The first two of these refer to the variability of loads and material properties. Such variabilities can often be expressed in terms of probability distributions. However, when it comes to the extreme ends of the 76 distributions, lack of statistical information can make precise probabilistic analysis impossible. Let us consider the variability of the properties of materials. Experimental data on material properties are often insufficient for making a distinction between e.g. gamma and lognormal distributions, a problem called distribution arbitrariness.22 This has little effect on the central part of these distributions, but in the distribution tails the differences can become very large. This is a major reason why safety factors are often used for design guidance instead of probabilities, although the purpose is to protect against failure types that one would, theoretically, prefer to analyze in probabilistic terms. As Zhu (1993) puts it: Theoretically, design by using structural system reliability is much more reasonable than that based on the safety factor. However, because of the lack of statistical data from the strength of materials used and the applied loads, design concepts based on the safety factor will still dominate for a period.23 The last three of the five items in the list of what safety factors should protect against all refer essentially to errors in our theory and in our application of it. They therefore are clear examples of uncertainties that are not easily amenable to probabilistic treatment. The eventuality of errors in our calculations or their underpinnings is an important reason for applying safety factors. This uncertainty is not reducible to probabilities that we can determine and introduce into our calculations. (It is difficult to see how a probability estimate could be accurately adjusted to compensate selfreferentially for the possibility that it may itself be wrong.) It follows from this that safety factors cannot be accounted for exclusively in probabilistic terms. 6. Safe fail The concept of safe fail is that of safety even if parts of the system or even the whole system fails. There are many concepts in engineering safety for 77 which the principle may be applicable; several of the principles in the Appendix may indeed be seen as safe fail applications. Sometimes in the literature safe fail is put in contrast to fail-safe: a safe fail system, then, is a system designed to safely fail whereas a fail-safe system is one designed not to fail. (Put differently: it is safe from failing rather than safe when failing.) The point is somewhat polemical but rightly used an instructive one, and it carries an important lesson to which we will return later (the Titanic lesson). That said, however, it should be noted that this polemic distinction between safe fail and fail-safe does not work on most explications of ‘fail-safe’ in the literature, since those explications of ‘fail-safe’ would rightly be interpreted as referring to safe fail rather than fail-safe. Thus, both terms would refer to the same thing. Hammer (1980), for example, says: “Fail-safe design tries to ensure that a failure will leave the product unaffected or will convert it to a state in which no injury or damage will occur.”24 Similarly, the IAEA (1986) states that “the principle of ‘fail-safe’ should be incorporated […] i.e. if a system or component should fail the plant should pass into a safe state”.25 Rather, the distinction seems to be one of perspective: for any level of a system/component, we may ask whether it will fail as well as whether its failure will result in danger beyond the system/component (e.g. harm to humans). Using the concept of safe fail means ultimately focusing on the latter question. We will therefore use this term to cover the entire spectrum including this latter concern of being safe when the system fails. Fail-safe. The concept of fail-safe is mainly used for specific methods and principles for keeping the system safe in case of failure, such as shutting down the components or the entire system. Basically there are two modes of fail-safe (in this narrower construal), fail-silence and fail-operational. Fail-silence means that the system is stopped when a critical failure is detected, prohibiting any harmful event from occurring (the expressions “negative feedback” and “fail-passive”26 are also used).27 An electrical fuse is a paradigmatic example of a fail-silence application, as is the dead man’s handle that stops the train when the driver falls asleep. Fail-operational means that the system will continue to work despite the fault.28 In aviation, 78 fail-operational systems are paramount: airborne failures may lead to partial operational restrictions, but system shutdown is normally not a particularly safe option. A safety-valve is another paradigmatic fail-operational device: if the pressure becomes too high in a steam-boiler, the safety-valve lets out steam from the boiler (without shutting down the system). Safety barriers. Another application of the safe fail principle is the usage of several safety barriers. Some of the best examples of the use of multiple safety barriers can be found in nuclear waste management. The proposed subterranean nuclear waste repositories all contain multiple barriers. We can take the current Swedish nuclear waste project as an example: The waste will be put in a copper canister that is constructed to resist the foreseeable challenges. The canister is surrounded by a layer of bentonite clay that protects the canister against small movements in the rock and “acts as a filter in the unlikely event that any radionuclides should escape from a canister”.29 This whole construction is placed in deep rock, in a geological formation that has been selected to minimize transportation to the surface of any possible leakage of radionuclides. The whole system of barriers is constructed to have a high degree of redundancy, so that if one of the barriers fails the remaining ones will suffice. With usual PRA standards, the whole series of barriers would not be necessary. Nevertheless, sensible reasons can be given for this approach, namely reasons that refer to uncertainty. Perhaps the copper canister will fail for some unknown reason not included in the calculations. Then, hopefully, the radionuclides will stay in the bentonite, etc. In nuclear power plant safety the principle of safety barriers has been generalised to the concept of Defence in depth.30 Even if physical layers of safety barriers is the foundation of the concept, Defence in depth includes a large superstructure of procedural and technical safeguards for events including and going beyond emergency preparedness in a worst-case scenario. Reliability, Redundancy, Segregation, Diversity. Several of the principles listed in the appendix can be seen as applications of the safe fail principle. Four 79 closely related principles are reliability, redundancy, segregation and diversity. Reliability is here the key concept which the latter three may be seen as means of achieving.31 Reliability and safety are related concepts but important to keep apart. A system can be very unreliable and yet perfectly safe, as long as the failures in question are minor. A regular office PC is an example of a (relatively) unreliable but safe system: even if you have to boot the system now and then, the (physical) safety consequences are not severe.32 In the context of engineering safety, reliability issues mainly concern system functions important for safety. Redundancy is an important means of achieving reliability: if one component fails, there are alternative ways to achieve the function in question and the system performance is unaltered.33 Having more engines than needed for flight is an example of system redundancy. Sending a piece of information two independent ways through a system is another. Segregation is a related concept. Two parts of a system may both fail if they are physically (or temporally) too close; there may for example be a common cause of their failure due to their proximity, or the failure of one part may cause the failure of the other, for instance if overload of an engine causes a fire that makes another engine fail. Yet another related concept is diversity: redundant parts should have different realisations of the same function in order to avoid common cause failures, i.e. failures resulting from one common cause. Redundant software systems, for example, should not be based on the same algorithms.34 The principle of safe fail in all its different applications (far from all of which have been mentioned here) is yet another example of how the uncertainty aspect is inherent in the concept of safety. Meaningful probability estimates may perhaps, in some context, be possible for most of the sub-principles here mentioned. Yet, as should now be clear, even so, the purpose of these methods is to prevent the unforeseeable. And here the tools of probabilistic risk and safety assessment are insufficient. We may call this the Titanic lesson. We now know that the Titanic was far from unsinkable. But let us consider a hypothetical scenario. Suppose that tomorrow a ship80 builder comes up with a convincing plan for an unsinkable boat. A probabilistic risk analysis has been performed, showing that the probability of the ship sinking is incredibly low. Based on the PRA, a risk-benefit analysis has been performed. It shows that the cost of life-boats would be economically indefensible. Because of the low probability of an accident, the expected cost per life saved by the life-boats is above 1000 million dollars, way above what society at large is prepared to pay for a life saved in areas such as that of road traffic. The risk-benefit analysis therefore clearly shows us that the ship should not have any lifeboats. How should the naval engineer respond to this proposal? Should she accept the verdict of the economic analysis and exclude lifeboats from the design? Our proposal is that a good engineer should not act on the risk-benefit analyst’s advice in a case like this. The reason should now be obvious: it is possible that the calculations may be wrong, and if they are, then the outcome may be disastrous. Therefore, the additional safety barrier in the form of lifeboats (and evacuation routines and all the rest) should not be excluded, in spite of the probability estimates showing them to be uncalled for. 7. Conclusions We have seen that major strategies in safety engineering are used to deal not only with risk – in the standard, probabilistic sense of the term – but also with uncertainty (in a way that is not reducible to risk). From this either of two conclusions can be drawn: either these principles are inadequate to deal with safety; or the concept of safety as the antonym of risk is insufficient. We propose that the latter conclusion is the more plausible one. If so, then this has important implications for the role of probabilistic risk analysis in engineering contexts. PRA is an important tool for safety, but it is not the final arbitrator since it does not deal adequately with issues of uncertainty. Although engineers calculate more than members of most other professions, the purpose of these calculations is to support, not to supplant, the engineer’s judgment. Safety is a more complex matter than what can be 81 captured in probabilistic terms, and our understanding of the concept must mirror this fact. 82 Appendix 1 Principles of engineering safety. In the third column, they are divided into different categories: (1) inherently safe design, (2) safety reserves, (3) safe fail and (4) procedural safeguards. Principle, reference Brief description Cat Inherently safe Potential hazards are avoided rather than 1 35 design controlled. Safety factor36 The system is constructed to resist loads 2 and stresses exceeding what is necessary for the intended usage by multiplying the intended load by a factor (>1). Margin of safety37 An (additative) margin is used for 2 acceptable system performance as a precautionary measure. Stress margins38 The system is designed so that statistical 2 variations in stresses do not lead to failure.39 Screening40 Control measure to eliminate 3, components that may pass operating tests 4 for specific parameters but show signs of possible future failure (or reduced sustainability).41 Safety barriers42 Physical barriers providing multiple layers 3 of protection; if one layer fails, the next will protect from system failure. Reliability43 A measure of system failure rate. High 3 reliability against certain types of failures is necessary for system safety. Redundancy44 Method of achieving reliability for 3 important system functions. Redundant parts protect the system in case of failure of one part. Diversity45 Redundant 83 system parts are given 3 different design characteristics to avoid failures from a common cause to cause failure in all redundant parts. Segregation Redundant (Independence, dependent on each other. Malfunction in 46 Isolation) parts should not be 3 part should not have any consequences for a redundant part. One way to avoid this is to keep the parts physically apart. Fail-safe design47 Method to ensure that even if a failure of 3 one part occurs the system remains safe, often by system shut down or by entering a “safe mode” where several events are not permitted. Proven design48 Relying on design that has been proven 3 by the ”test of time”, i.e. using solutions or materials that have been used on many occasions and over time without failure. Single failure criterion Design criteria stating that a failure of a (Independent single system part should not lead to malfunction)49 system failure. System failure should only 3 be possible in case of independent malfunction. Pilotability (safe The system operator should have access information load)50 to the control means necessary to prevent 3 failure, and the work should not be too difficult to perform.51 Quality52 Reliance on materials, constructions etc 3,4 of proven quality for system design. Operational interface Focusing on controlling the interface control53 between humans and (the rest of) the 3 system and equipment. For example, using interlocks to prevent human action to have harmful consequences. Environmental The environment should be controlled so 54 84 4 control54 that it cannot cause failures. Especially, neither extremes of normal environmental fluctuations nor energetic events such as fire should be able to cause failures. Operating and Automatic as well as manual procedures maintenance are used as a defence against failures. procedures55 Training in order to follow procedures is 4 a part of such safety procedures. Job study Identifying potential causes through observations56 collecting data from observations and 4 audits, e.g. interviewing staff about potential or existent hazardous practices. Controlling Controlling certain types of behaviour 57 (e.g. alcohol and drug abuse, lack of behaviour 4 sleep), e.g. by tests and audits. Standards58 Standardised solutions of system design, 1- material usage, maintenance procedures 4 etc. Standards may be applied to all areas of safety engineering. Timed replacement59 Replacing components performance has before decreased their as 4 a precautionary procedure. This can be done regularly without any signs of decreased performance, or by using indicators of potential failure such as component degradation or drift. Procedural 60 safeguards Procedures such as instructions to 4 operators to take or avoid specific actions in general or in special circumstances. Warnings61 Warning devices and information are provided when control measures are insufficient (or in addition to them). 85 4 Notes The authors would like to thank Martin Peterson, Kalle Grill and the members of the Risk Seminar at the Department of Philosophy and the History of Technology at the Royal Institute of Technology for their helpful criticism. 2 For example, the International Organisation for Standardization (2002) defines risk as “the combination of the probability of an event and its consequences”. Green (1982), 3, and Cohen (2003) are two examples of this usage. Conceptualising safety as the inverse of the risk is frequently shown in the literature, e.g. Harms-Ringdahl (1987-1993), 3. Misumi and Sato (1999), 135-144, writes “[R]isks are defined as the combination of the probability of occurrence of hazardous event and the severity of the consequence. Safety is achieved by reducing a risk to a tolerable level”. Koivisto (1996), I/5, defines safety as a function of probability and consequence, as do Roland & Moriarty (1983), 8-9. 3 International Atomic Energy Agency (1986), 2 (my emphasis). 4 International Atomic Energy Agency (1986), 3. 5 Decisions under perfect deterministic knowledge is called “decision under certainty”. 6 E.g. International Atomic Energy Agency (1986), Hammer, W. (1980); Nolan (1996). 7 Koivisto (1996), 18. 8 N. J. Bahr (1997), 14-17. NASA (1993), 1-3. 9 Bahr (1997), 14. 10 Bahr (1997), 15. 11 Bahr (1997), 16-17. 12 The categorization is tentative; as should be clear from the above discussion, there are several options for which categories to choose as well for which category the principles fit (since they are not exclusive). 13 Example from Bahr (1997), 14-15. 14 Kletz (1991). 15 Gloss & Wardle (1984), 174. 16 Hammer (1980), 108-109. 17 C.f. D. Gloss & Wardle (1984), 171-172, Hammer (1980), 108-109. 18 The worker in the assembly line also needs help not to make mistakes, of course; primarily in the cases where she is tired, ill or absent-minded. I.e. the cases hardest to treat probabilistically. 19 Randall (1976). 20 Duncan (2000). 21 Knoll (1976). Moses (1997). 22 Ditlevsen (1994). 23 Zhu, T.L. (1993). 24 Hammer (1980), 115. 25 International Atomic Energy Agency (1986), 9. 26 E.g. Hammer (1980), 115. 27 Jacobsson, Johansson, Lundin (1996), 7. 28 Jacobsson, Johansson, Lundin (1996), 7. Some authors distinguishes between partial operational (“fail-active”) and fully operational; c.f. Hammer, 115. 29 http://www.skb.se/templates/SKBPage____8762.aspx. 30 C.f. International Atomic Energy Agency (1986). 31 C.f International Atomic Energy Agency (1986). 32 In an information safety context, however, the system is most unsafe. 33 C.f. Hammer (1980), 71-75, for different types of redundancy: decision redundancy, standby systems, serial redundancy etc. 34 A parallel procedural safeguard is here the classical rule for aviation pilots not to eat at the same restaurant before flight in order to avoid food poison. 1 86 Nolan (1996), 22; Gloss & Wardle (1984), 171-174, gives, several principles in this category; Koivisto (1996), 25-28; Bahr (1997), 14. 36 Hammer (1980), 66, 71 (derating). 37 Hammer (1980), 67. 38 International Atomic Energy Agency (1990), 34. 39 International Atomic Energy Agency (1990), 34 and Hammer (1980), 67. 40 Hammer (1980), 76. 41 Hammer (1980), 76. 42 International Atomic Energy Agency (2000); International Atomic Energy Agency (1986), 4. 43 International Atomic Energy Agency (1986), 7; Nolan (1996), 22. 44 International Atomic Energy Agency (1986), 7; Nolan (1996), 23; Hammer (1980), 71, 74-76. 45 International Atomic Energy Agency (1986), 8. 46 International Atomic Energy Agency (1986), 8. In International Atomic Energy Agency (1990), 34, the term “segregation” is used (a better term perhaps if including also temporal “isolation”). Gloss & Wardle (1984), 174, use the term “isolation” here. C.f. also Nolan (1996), 23, 117. 47 Jakobson & Johansson (1996), 7-8; IAEA (1986), 9; Nolan (1996), 22, 119; Bahr (1997). 48 Koivisto (1996), 19-24; International Atomic Energy Agency (1986), 10; International Atomic Energy Agency (1990), 32. 49 Jakobson & Johansson (1996), 10; International Atomic Energy Agency (1990); Hammer, 97-99; Nolan (1996), 23. 50 Carnino & Nicolet (1990), 24-28. 51 Carnino & Nicolet (1990), 24 ff. 52 International Atomic Energy Agency (1990), 32; Nolan (1996), 23. 53 International Atomic Energy Agency (1990), 33. 54 International Atomic Energy Agency (1990), 32. 55 International Atomic Energy Agency (1990), 33. 56 Gloss & Wardle (1984), 167. 57 Gloss & Wardle (1984), 175-176. 58 Bahr (1997), 16. 59 Hammer (1980), 77. 60 Hammer (1980), 97-99; Nolan (1996), 23. 61 Bahr (1997), 16. 35 87 References Bahr, N. J. (1997), System Safety Engineering and Risk Assessment: A Practical Approach, Washington, DC: Taylor & Francis. Carnino, A. Nicolet, J.-L. , Wanner, J.-C. (1990) Man and risks: technological and human risk prevention, New York: M. Dekker, Inc. Cohen, B. (2003) Probabilistic Risk Analysis for a High-Level Radioactive Waste Repository, Risk Analysis 23:909-915. Ditlevsen, O. (1994) Distribution arbitrariness in structural reliability, in Schuëller, G. Shinozuka, M. and Yao, J. (1994) Proc. of ICOSSAR'93: Structural Safety & Reliability, 1241-1247. Duncan, J.M. (2000) Factors of safety and reliability in geotechnical engineering, Journal of Geotechnical and Geoenvironmental Engineering 126:307–316. Gloss, M. Gayle Wardle (1984), Introduction to safety engineering, New York: Wiley. Green, A.E. (1982), High risk safety technology, Chichester: Wiley. Hammer, W. (1980) Product Safety Management And Engineering, Englewood Cliffs, New Jersey: Prentice-Hall. Harms-Ringdahl, Lars (1993), Safety analysis: principles and practice in occupational safety, London: Taylor & Francis. International Atomic Energy Agency (1986), General design safety principles for nuclear power plants: a safety guide, Vienna: International Atomic Energy Agency. International Atomic Energy Agency (1990), Application of the single failure criterion, Vienna: International Atomic Energy Agency. International Atomic Energy Agency (2000), Safety of Nuclerar Power Plants: Design, Vienna: International Atomic Energy Agency. 88 International Organization for Standardization (2002) Risk Management – Vocabulary – Guidelines for use in standards, ISO/IEC Guide 73:2002. Jacobsson, J., Johansson, L.-Å., Lundin, M. (1996), Safety of Distributed Machine Control Systems, Swedish, National Testing and Research Institute: SP Report 1996:23. Kletz, T.A. (1991), Plant Design for Safety, a user-friendly approach, New York: Hemisphere Pub. Corp. Knoll, F. (1976) Commentary on the basic philosophy and recent development of safety margins, Canadian Journal of Civil Engineering 3:409-416. Koivisto (1996), Raija, Safety-conscious process design, Espooo: VTT Offsetpaino. Misumi, Y., and Sato, Y. (1999) Estimation of average hazardous-event-frequency for allocation of safety-integrity levels, Reliability Engineering & System Safety, 66(2): 135-144. Moses, F. (1997) Problems and prospects of reliability-based optimisation, Engineering Structures 19:293-301. NASA (1993), Safety Policy and Requirements Document. NHB 1700.1 (V1-B). Washington DC: NASA, 1-3. Nolan, D. (1996), Handbook of fire and explosion protection engineering principles for oil, gas, chemical, and related facilities, New Jersey: Noyes Publications. Randall, F.A. (1976) The safety factor of structures in history, Professional Safety January:1228. Roland H., Moriarty B. (1983), System safety engineering and management, New York: John Wiley & Sons. Zhu, T.L. (1993) A reliability-based safety factor for aircraft composite structures, Computers & Structures 48: 745-748. 89 Theses in Philosophy from the Royal Institute of Technology 1. Martin Peterson, Transformative Decision Rules and Axiomatic Arguments for the Principle of Maximizing Expected Utility, Licentiate thesis, 2001. 2. Per Sandin, The Precautionary Principle: From Theory to Practice, Licentiate thesis, 2002. 3. Martin Peterson, Transformative Decision Rules. Foundations and Applications, Doctoral thesis, 2003. 4. Anders J. Persson, Ethical Problems in Work and Working Environment Contexts, Licentiate thesis, 2004. 5. Per Sandin, Better Safe than Sorry: Applying Philosophical Methods to the Debate on Risk and the Precautionary Principle, Doctoral thesis, 2004. 6. Barbro Björkman, Ethical Aspects of Owning Human Biological Material, Licentiate thesis, 2005. 7. Eva Hedfors, The Reading of Ludwig Fleck. Sources and Context, Licentiate thesis, 2005. 8. Rikard Levin, Uncertainty in Risk Assessment – Contents and Modes of Communication, Licentiate thesis, 2005. 9. Elin Palm, Ethical Aspects of Workplace Surveillance, Licentiate thesis, 2005. 10. Jessica Nihlén Fahlquist, Moral Resonsibility in Traffic Safety and Public Health, Licentiate thesis, 2005. 11. Karin Edvardsson, How To Set Rational Environmental Goals: Theory and Applications, Licentiate thesis, 2006. 12. Niklas Möller, Safety and Decision-making, Licentiate thesis, 2006. 91
© Copyright 2026 Paperzz