הטיות בהערכת סיכונים Bias in Risk Assessment יניב מרדכי אוניברסיטת תל-אביב הפקולטה להנדסה רקע • ניהול סיכונים בפרויקטים מצריך הערכת סיכון איכותית ,לצורך קבלת החלטות מבוססות נתוני סיכון. • הערכת הסיכון היא בד"כ סובייקטיבית ,כלומר מבוססת על הערכותיהם של אנשי מקצוע ע"ס ניסיון ואינטואיציה ,ולא אובייקטיבית או סטטיסטית. • ישנם גורמים שונים המביאים להטיות בהערכת סיכון ע"י אנשי מקצוע. 2 סיכון = התשובה לשלוש שאלות )(ועוד אחת • What can go wrong? • What is the likelihood? • What are the consequences? • Kaplan and Garrick (1981) • What is the time domain? • Haimes (1998) 3 הודאות-רצוי של אי-סיכון = החלק הלא Probability Density Function h(t) Opportunity T ~ h(t) 4 Risk is about this part of uncertainty T c Target Value t :הערכת סיכון סובייקטיבית/אובייקטיבית “(Objective) Probability does not exist!”, De-Finetti (1937) Subjective Objective Art Science Intuition Evidence Relative Subjectivists 5 Vs. Absolute Frequentists סיכון רב-מימדי עם חלופות להמרת סיכון בין יעדי הפרויקט 6 Bias in Risk Assessment 7 Sources of Bias • • • • 8 Information and knowledge gaps Modeling Bias Cognitive Bias Strategic Bias Information and Knowledge gaps • One assessor knows more than the other about a certain topic. • Information gaps are themselves risks, and information acquisition is a means to reduce risk • Information does not necessarily reduce uncertainty, but does reduce Uncertainty About Uncertainty 9 Methodological Bias • The Assessment Method leads to assessment bias. • Method components: – Set space partitioning. – Resolution. – Diversity (over results, risks, objectives) – Predetermined aspects. – Focus. – Utility/Likelihood mixup 10 Risk Modeling • Move from target domain to risk domain. • Goal: Minimize overall Project Risk, by selecting Risk handling strategies. • Measures of Risk: – Estimated Risk (NxN) – Expected Risk – Expected Disutility of Risk 11 Estimating Risk with 5x5 Risk Impact Grids (is wrong…) Likelihood 12 5 5 10 15 20 25 4 4 8 12 16 20 3 3 6 9 12 15 2 2 4 6 8 10 1 1 2 3 4 5 1 2 3 4 Severity 5 • The most popular risk assessment technique • Assess Likelihood Score. • Assess Severity Score. • Impact = Product of Scores • Color provides additional classification (low,medium,high,extreme) • Other variation: 3x3, 10x10 • Scores are sometimes replaced by real values Estimating Risk with 5x5 Risk Impact Grids (is wrong…) • Why this method is flawed: Likelihood 13 5 5 10 15 20 25 4 4 8 12 16 20 3 3 6 9 12 15 2 2 4 6 8 10 1 1 2 3 4 5 1 2 3 4 Severity 5 – “Bernoulli” Fallacy – How much is “1”,”2”,”3”,“4”,”5”? – What are we assessing (mean, mode, median)? – Score product is algebraically invalid. – Loss of information. • [Pennock & Haimes (2002), Williams (1996)] Partitioned Multi-Objective Risk Analysis 14 The Fallacy of Averages: What Would Life Be Like If… • Highways were constructed to accommodate the average traffic load of vehicles or average weight. • Communication networks were sufficient to handle only the average data transfer rate. • Airports were designed to handle the average airliner traffic. • Emergency services were staffed to attend to the daily average number of emergency calls. 15 Eliciting Subjective Probabilities: The Fractile Method - Raiffa (1968) Continue identifying the median for each half-range until Construct Identify x25 the Identify and assessment Identify x75Interpolate -xmedians and as amedian xof CDF, each not half-range as a PDF 50 x-0 the 100 sufficient precision is achieved 1.00 0.75 0.50 0.25 0.00 1 16 2 3 4 5 6 7 8 9 10 Risk Disutility • A Disutility Function quantifies the adverse effect of each possible deviation (from target value). • Disutility is similar to Utility – Complies with all Utility Theory axioms – Increases as deviation increases • Expected Disutility – The expectation of the disutilityfunction, the smaller the better. • The objective is to minimize Expected Disutility of Risk 17 5 4 3 2 1 100000 120000 140000 160000 180000 200000 Cognitive Bias • Stems from (mostly unaware) perceptional deviations, fallacies and deceptions. • Research led by Tversky & Kahneman (1970’s-80’s). • Some Cognitive Biases may be observed and fixed. 18 Cognitive Bias - examples • Availability – Human tendency to base assessment on vivid memory. – Memorable experiences and impressions lead to exaggerated assessments. – Examples: car accidents, airplane crashes. • Size – We have difficulty grasping absolutely or relatively large numbers, and think more clearly in smaller numbers. – We tend to ignore the effects of the Law of Large Numbers, and refer to estimates from smaller samples as equally representative as estimates from larger ones, or even as better ones 19 Cognitive Bias - examples • Anchoring – Taking an initial assessment based on historical information, memory or cues, and adjusting it according to circumstances of the new assessment – Frequently insufficient to neutralize the initial anchor • Representativeness – Person relies on one estimate to make an assessment of some other correlated estimate. – Bias due to the intuitive attempt to rely on the inverse conditional probability as an anchor, and ignore the absolute probabilities of the events. (Bayes' Theorem) 20 Strategic Behavior • Stems from personal cost/benefit considerations. • Appears in collaborative but non-cooperative settings. • Applications in Social Choice: Voting Scheme, Point Estimates and Single Peaked Preferences. • Assessment combination methods like Averaging and Bayesian Inference are vulnerable to Strategic Behavior. 21 Strategy-Proof Assessment • The Median of a set of scores is a strategy-proof assessment (Moulin 1980). • Game Theoretically, once the decision maker chooses the Median, no one gains benefit from being untruthful. • For a continuous/discrete distribution, the median assessment of the cumulative probability distribution per each point/quantile is strategy-proof • This is the MEDAS setting: Median Distribution Assessment Scheme. 22 Illustration: Multiple Assessments (CDF) group_id 265 type (All) function_id 264 1014 1764 2514 3264 F(X) 1.1 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 23 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5 x 0.55 0.6 0.65 0.7 0.75 0.8 0.85 0.9 0.95 1 Illustration: Multiple Assessments +MEDAS (CDF) group_id 265 type (All) function_id 0 264 1014 1764 2514 3264 F(X) 1.1 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 24 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5 x 0.55 0.6 0.65 0.7 0.75 0.8 0.85 0.9 0.95 1 Illustration: MEDAS group_id 265 type (All) function_id 0 F(X) 1.1 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 25 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5 x 0.55 0.6 0.65 0.7 0.75 0.8 0.85 0.9 0.95 1 Illustration: PDF of MEDAS function_id (All) group_id 265 Total Total 0.3 f(x)_ 0.25 0.2 0.15 0.1 0.05 0 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5 x 26 0.55 0.6 0.65 0.7 0.75 0.8 0.85 0.9 0.95 1 MEDAS - Illustration (1/3) G1(x) 0.9 0.8 F3(x) 0.7 F2(x) 0.6 F1(0.4) higher than strategy: F2(0.4) is Assessor best Assessor 1 is1’s confronted with the median, F2(0.4) On x=0.4, but>F1(0.4) causes G1(x)greater is > below G1(0.4) disutility the is following situation and has to provide F * (0.4) ≤ F (0.4)=M(0.4) 1 candidate 2 will1:be pointless. F2(0.4) to assessor median range median. F1(x) 2 Weakly2>[F2(0.4)-G1(0.4)] Sincere. [M(0.4)-G1(0.4)] 0.5 0.4 0.3 0.4 0.5 G1(x) F2(x) F3(x) 0.6 F1=G1(x) Median 27 MEDAS - Illustration (2/3) G1(x) 0.9 0.8 F3(x) 0.7 F2(x) 0.6 Assessor best strategy: Assessor F1(0.6)< 1F2(0.6) is1’s confronted is median, withbut the F3(0.6) On x=0.6, >F1(0.6) G1(x) is > above G1(0.6) the is following causes greater situation and has to assessor to provide 1: F1*(0.6) ≥disutility Fmedian 3(0.6)=M(0.6) pointless. candidate F3(0.6) will be range median. 2>[F2(0.6)-G1(0.6)]2 [M(0.6)-G1(0.6)] F1(x) Weakly Sincere. 0.5 0.4 0.3 0.4 0.5 G1(x) F2(x) F3(x) 0.6 F1=G1(x) Median 28 MEDAS - Illustration (3/3) Assessor best strategy: F1(0.5)≠G1(0.5) is still median, On x=0.5, 1’s G1(x) is within thebut causes greater=disutility to assessor 1: candidate range. F1*(0.5) Gmedian 1(0.5)=M(0.5) F1(0.5)=G1(0.5) is the 2median! [F1(0.5)-G1(0.5)] >0 Strongly Sincere. 0.9 0.8 G1(x) F3(x) 0.7 F2(x) 0.6 0.5 0.4 0.3 0.4 0.5 G1(x) F2(x) F3(x) 0.6 F1=G1(x) Median 29 MEDAS - characteristics • Works on CDF form (not PDF) • Relaxes “Single-Peakedness”. • Works for single point, quantile and continuous cases. • Always induces sincere behavior as dominant strategy • “Winning” (having one’s assessment elected to represent) is not the goal. • Does not require payoff considerations. • Does not repair cognitive bias. 30 Challenges • Definition – Sources, targets, objectives, utility functions. • Assessment – Let each assessor assess the effect of each source on each target under each strategy – Conduct a strategy-proof assessment process • Decision making – Evaluate trade-off according to predefined preference relations. – Select optimal risk handling strategies 31 • Implementation – Translate decisions to practical risk mitigation steps. – Monitor progress and effectiveness. • Iteration – Reassess and reevaluate – Repeat Risk Management Process LEAN Risk Management • The challenge: To found an effective, successful and easy risk management process. • The response: – Conduct Risk Management workshops, based on the LEAN philosophy and methodology. – Bring all stakeholders together, extract all knowledge and information about uncertainty. – Conduct methodological analysis and assessment of risks. – Produce and execute a risk handling plan, with immediate effects. 32 Summary • Risk Assessment - not what you thought! • High Fidelity Risk Assessment requires an initial effort, and continuous maintenance, but has clearly significant ROI. • Risk Assessment requires appropriate and valid methods and sometimes even professional risk analysis escort. • Risk management workshops are a constructive and effective way to produce quality risk assessments. 33 תודה! 34
© Copyright 2026 Paperzz