PROBABILITY OF PRECIPITATION Assessment and Enhancement of End-User Understanding by Susan Joslyn, Limor Nadav-Greenberg, and Rebecca M. Nichols Three psychological studies show that many people misunderstand traditional probability of precipitation forecasts and icons, although adding phrases specifying the chance of “no precipitation” reduces misunderstanding. P robability of precipitation (PoP) has appeared in public forecasts in the United States from the late 1960s (National Research Council 2006). Since then however, very few additional probabilistic forecasts have been made publicly available. This is in part due to lingering questions about how well the general public will understand them. Recent evidence suggests that many people do not understand PoP, after decades of exposure (Gigerenzer et al. 2005). The main source of misunderstanding appears to be the class of events to which the probability refers (Gigerenzer et al. 2005). Take, for example, a forecast for “80% chance of precipitation.” According to a AFFILIATIONS: Joslyn , Nadav-Greenberg , and N ichols —University of Washington, Seattle, Washington CORRESPONDING AUTHOR: Susan Joslyn, Department of Psychology, University of Washington, Box 351525, Seattle, WA 98195 E-mail: [email protected] The abstract for this article can be found in this issue, following the table of contents. DOI:10.1175/2008BAMS2509.1 In final form 17 June 2008 ©2009 American Meteorological Society AMERICAN METEOROLOGICAL SOCIETY recent survey, approximately 35% of New Yorkers and larger percentages of Europeans do not understand of what it is 80% (Gigerenzer et al. 2005). It is not surprising that this issue is difficult for the general public, given that it is debated even within the scientific community (DeElia and Laprise 2005). Some propose a “frequentist” interpretation: there will be at least a minimum amount of rain on 80% of days with weather conditions like they are today (DeElia and Laprise 2005; Gigerenzer et al. 2005). Although preferred by many scientists, this explanation may be particularly difficult for the general public to grasp because it requires regarding tomorrow as a class of events, a group of potential tomorrows. From the perspective of the forecast user, however, tomorrow will happen only once. A perhaps less abstract interpretation is that PoP reflects the degree of confidence that the forecaster has that it will rain (DeElia and Laprise 2005). In other words, an 80% chance of rain means that the forecaster strongly believes that there will be at least a minimum amount of rain tomorrow. The problem, from the perspective of the general public, is that when PoP is forecasted, none of these interpretations is specified. There are clearly some interpretations that are not correct. The percentage expressed in PoP neither february 2009 | 185 refers directly to the percent of area over which precipitation will fall nor does it refer directly to the percent of time precipitation will be observed on the forecast day. Although both interpretations are clearly wrong, there is evidence that the general public holds them to varying degrees (Gigerenzer et al. 2005). Such misunderstandings are critical because they may affect the decisions that people make. If people misinterpret the forecast as precent time or percent area, they may be more inclined to take precautionary action than are those who have the correct probabilistic interpretation, because they think that it will rain somewhere or some time tomorrow (Gigerenzer et al. 2005). The negative impact of such misunderstandings on decision making, both in terms of unnecessary precautions as well as erosion in user trust, could well eliminate any potential benefit of adding uncertainty information to the forecast. It is important to understand the psychological basis of these misunderstandings and how to overcome them because these issues may be crucial to successful communication of forecast uncertainty in general. That is the purpose of the experimental work reported here. The research reported here asks whether people are indeed missing the probabilistic aspect of the forecast. It also tests whether PoP forecasts can be improved to clarify the essential uncertainty involved and contradict the erroneous proportional interpretations. One potential enhancement that was tested is visual imagery. Increasingly, PoP in television and Web forecasts is accompanied by visual representations, icons with rain and sun imagery. However, as far as we know, peoples’ interpretation of such icons Fig. 1. Icons used in experiment 1. (from left to right) Question mark icon, pie icon, bar icon. 186 | february 2009 has never been tested systematically. In other words we do not know whether this imagery actually helps. The three experiments reported here were designed to determine whether enhancements including icons using visual imagery (see Fig. 1) and explicit verbal explanations promote a more accurate understanding of PoP forecasts. Each experiment approached the question in a slightly different way. Experiment 1 probed participants’ understanding of the reference class with multiple-choice questions. Experiment 2 investigated understanding of PoP by asking participants to provide an explanation in their own words. Experiment 3 compared the icon that led to the fewest errors in experiment 1 to two verbal descriptions for probability of precipitation. In addition, all three experiments also investigated the effect of the forecast on a binary decision. All three experiments were conducted in Seattle, Washington, among participants presumably very familiar with rain forecasts. The overarching question for these experiments was as follows: How can we express probability of precipitation to enhance understanding and to improve decision making? Experiment 1. Experiment 1 investigated three PoP icons, each showing 25% chance of precipitation and 75% chance of precipitation (see Fig. 1). Like most icons in use today, they included rain and/or cloud imagery. The icons were based on icons currently in use, but simplified and standardized for the purpose of the experiment, to neutralize extraneous factors (e.g., color, size, text). The icon in the leftmost column of Fig. 1 was based on a cartographic convention in which greater fuzziness or fogginess indicates more uncertainty (MacEachren 1992). This icon showed a cloud with raindrops and a question mark that was pale gray in the 25% condition when precipitation was unlikely and darker gray in the 75% condition in which precipitation was more likely. The question mark disappeared when precipitation was highly likely, a condition not tested here. This imagery may appeal to the “confidence that it will rain” interpretation described above because haziness and the question mark indicate lack of confidence (i.e., the rain pictured here is questionable). The other two icons represented probability as a proportion of a whole: the portion filled by rain imagery corresponded to the percent chance of rain. We tested both a bar icon and a pie icon (see similar imagery at www.weather.com). Notice that this imagery may suggest a frequentist interpretation, rain during a proportion of days with similar weather conditions. The proportion of the icon with no raindrops may indicate that a “no-rain” outcome is also a possibility. The question for this experiment was whether any of these icons, when viewed casually, as they would be when shown on a TV or Web site weather forecast, do a better job of clarifying PoP than the others? Method. One hundred and eighty-three psychology undergraduates participated in this study as part of a course requirement, filling out a one-page anonymous questionnaire during a mass-testing session. One of the three experimental icons was shown in the upper-left-hand corner of each questionnaire. The numerical chance of rain was written above the icon. The likely amount of rain, 0.0 in. in the 25% condition and 0.03 in. in the 75% condition, was written below it. The ensemble forecast from which the probability was derived also produced a likely amount that constituted an average over all possible outcomes for the probability. The instructions in the upper-right-hand corner read, “The picture to the left displays the rain forecast for the Seattle–Tacoma Airport for today. Please use it to answer the following questions.” Two of the questions were designed to uncover the reference class misunderstandings of PoP. One asked, “Over approximately what area of the Puget Sound region will it likely rain today?” The next question asked, “How much of the time will it likely rain today?” For both questions there were four check-box answer options, “None of the [area/time],” “Less than half of the [area/time],” “More than half of the [area/time],” and the correct answer, “Can’t tell from this forecast.” Another question, designed to determine the impact of the forecast on decision making, asked whether participants would take an umbrella or wear a hooded jacket. Each participant saw only one icon and forecast and answered this set of questions only once. Results. The correct answer to both of the reference class questions was “can’t tell from this forecast” because no information about percent of area or percent time was provided. Only 43% (i.e., 79/183) selected the correct response to both questions. Approximately 32% of participants instead selected one of the incorrect area-related answers.1 Although the pie icon led to the fewest errors, in a logistic regression analysis there were no statistically significant differences by icon type. Table 1 Table 1. Expt 1: Percent of reference class errors by icon type. Reference class errors Icon type % area errors % time errors Pie 19/63 (30%) 21/65 (32%) Bar 21/59 (36%) 24/59 (41%) Question mark 18/59 (31%) 25/59 (42%) Total 58/181 (32%) 70/183 (38%) lists the mean number of error responses in each icon type.2 We predicted that if participants thought the percentage referred directly to area they would select “less than half the area” in the 25% condition and “more than half the area” in the 75% condition. Indeed, the majority of wrong answers fell into these categories (i.e., 21/34, 62% and 14/24, 58%, respectively). Moreover, the pattern of error responses was significantly different in the 25% condition compared to the 75% condition [i.e., χ2 (2, N = 58) = 17.25, p < 0.001)]. Overall, these results suggest that quite a few participants believed that PoP had something to do with percentage area. A substantial proportion of participants (38%) selected incorrect responses in answer to the percent time question as well. A logistic regression analysis revealed no significant differences due to icon type. However, as Table 1 shows, the pie icon again yielded slightly fewer error responses overall. We expected those who held the “percent time” misconception to select “less than half the time” in the 25% condition and indeed, 56% of errors (22/39) fell into this category. In the 75% condition the percent time misconception would lead participants to select “more than half the time,” and 39% (12/31) of error responses fell into this category. Again, the pattern of error responses was significantly different at the two levels of uncertainty, χ2 (2, N = 70) = 19.13, p < 0.0001, suggesting that quite a few participants believed that PoP had something to do with percent of time. It is important to note that the group of participants who selected wrong answers to the time question in the predicted categories (i.e., “less than half” for 25% and “more than half” for 75%) was largely independent of the group who selected wrong answers in the predicted categories to the area question. In all, 59 people, 32% of the sample, selected an answer to Two participants failed to answer this question so the n for these analyses is 181. Because “exactly half” was not an option, we do not know whether some participants might have chosen it. Also, “none” is logically included in the category “less than half.” Some who thought it meant none might have selected that category instead of none. 1 2 AMERICAN METEOROLOGICAL SOCIETY february 2009 | 187 either the time or area questions that suggested that they thought the percent provided referred directly to either time or area. Examination of the umbrella–hooded jacket decision revealed that, as one might expect, a smaller proportion (i.e., 28%, 26/94) of participants chose to take an umbrella in the 25% chance condition than did those in the 75% chance condition (i.e., 72%, 65/88). Interestingly, however, participants’ decisions were related to their interpretation of PoP. Pairwise comparisons in a logistic regression analysis revealed that the odds of choosing to take an umbrella or wear a hooded jacket were significantly smaller among participants who chose the correct answer “can’t tell” for both the time and area question (β = 0.15, p = 0.001) as compared to those who chose “more than half of the time” or “more than half of the area.” The same was true of participants who chose “none” or “less than half” (i.e., β = 0.08, p < 0.001). This suggests that misinterpreting the forecast as more than half the area or time affects decision making by increasing the tendency to take precautionary action. Discussion of experiment 1. A surprising percentage of participants misinterpreted PoP despite the visualizations that accompanied the numeric probabilities. The pattern of responses suggested that a substantial proportion of participants thought PoP included information about percentage of area or percentage of time.3 Participants choose “less than half” more often in the 25% condition and “more than half” more often in the 75% condition for both the area and the time questions. Those who selected percent-aligned wrong answers, suggesting that they thought that the percentage referred directly to time or area rather than to the chance of precipitation, comprised a third of participants overall. Furthermore, misunderstandings affected decision making, increasing precautionary action among those with the “more than half” misinterpretation. This combined evidence suggests that at least some were indeed interpreting PoP as a deterministic forecast for rain. An alternative explanation for the substantial percentage of incorrect answers in experiment 1 concerns the multiple-choice procedure used to probe reference class understanding. The correct answer In fact, these data may underestimate the proportion of respondents with erroneous understanding because it is possible that some may have selected the correct answer to the area question simply because of the mismatch between the point forecast provided and the area in the question, not because they knew the correct reference class. 3 188 | february 2009 to both the area and time questions was “can’t tell from this forecast” because the information was not specified in the forecast. Some participants may have avoided this answer because they thought that experimenters would not construct questions for which this was correct answer. Thus, some responses may have been partly an artifact of the experimental procedure (Orne 1962). Indeed, participants would only resort to this strategy if they did not understand the forecast in the first place. Moreover, we would not have seen the systematic impact on the umbrella decision if all participants’ error responses had been the result of such an artifact. Nonetheless, it is important to rule out this explanation. One way to overcome this problem is to ask participants to explain PoP in their own words. Experiment 2. Experiment 2 probed participants understanding of PoP using open-ended questions. Due to the minimal differences between icon types observed in experiment 1, only the question mark icon and the pie icon were used in this study because the imagery in these icons may give rise to different interpretations of uncertainty and hence, might lead to different explanations of PoP. They were compared to a “no icon” condition. Methods. One hundred and sixty-nine psychology undergraduates participated in this study as part of a course requirement. Participants filled out a onepage anonymous questionnaire during a 30-min testing session. The instructions indicated that the questionnaire provided the “rain forecast for the Seattle–Tacoma Airport for tomorrow.” Two levels of uncertainty were expressed: 25% and 75%. In the icon conditions, the forecasts were identical to those used in experiments 1. In the no-icon condition only the probability of precipitation phrase and the likely amount were presented. Below the forecast were four open-ended questions. The first three questions were designed to prove participants’ understanding of the reference class. Pilot testing had revealed that participants were reluctant to specify a reference class, so three different questions were included in an attempt to encourage specificity. The first question asked participants to “Explain the meaning of the forecast.” The second question asked participants to “Explain the forecast to someone who did not understand the concept of probability of precipitation.” The third question asked, “Of what is it 25% (or 75%)?” A final question asked whether participants would take an umbrella or wear a hooded jacket. Each participant saw only one forecast and answered only one set of questions. Results. We sorted the 428 answers participants gave to the three open-ended questions into give general categories (see the appendix). Intercoder reliability using this system was 87%.4 The first two categories included answers that directly addressed the issue of interpretation. One included responses that specified a reference class (i.e., 6%, 23/428). There was also a small group who provided a “confidence in the forecast” explanation (i.e., 1%, 5/428). A larger category (i.e., 25%, 107/428) included answers that provided clarification to the numeric aspect of the probability phrase (e.g., 25% of 100%, 1 in 4). A fourth category included a small proportion (i.e., 3%, 13/428) of answers that added information about other parameters such as cloud cover or temperature. The fourth and largest category of responses (i.e., 65%, 279/428) included those that simply repeated the probability phrase presented or summarized it verbally, without adding in any information or clarification. Then, we summarized responses by participant, giving each individual credit for the most specific response given. If participants gave the reference class or confidence in the forecast explanations they were put into an “interpretation” category (i.e., 17%, 28/169). If participants did not mention a reference class or confidence explanation but gave a numeric clarification they were put into that category (i.e., 47%, 80/169). If they did not give any of the first three explanations but added information about other parameters they were put into that category (i.e., 2%, 3/169). If participants gave none of the other explanations but merely repeated or summarized the probability phase, they became part of that category (i.e., 34%, 58/169). Figure 2 shows these proportions. We will focus here on the group that provided interpretation explanations (i.e., reference class or confidence). The majority (i.e., 52%, 12/23) of participants who addressed the reference class issue provided an answer, such as “days with atmospheric conditions like today,” which was considered correct. If we included those who provided a “confidence in the forecast” explanation in the category of correct responses, then 61% (17/28) of those providing an explicit interpretation were correct. Of the 39% (11/28) error responses, the majority were answers explaining that the forecast indicated the percent of time that precipitation would fall (i.e., 91%, 10/11). Although the number of participants who gave interpretation explanations was small, the proportion of correct Intercoder reliability was calculated from the percentage of matching codes from a randomly selected sample of 10 questionnaires by 3 independent coders. 4 AMERICAN METEOROLOGICAL SOCIETY Fig. 2. Proportions of responses per response category in experiment 2. to error responses was similar to that observed in experiment 1. The decisions about whether to take an umbrella or wear a hooded jacket were examined to determine whether they were related to erroneous reference class interpretations. Indeed, a greater proportion of those in experiment 2 who held erroneous reference class beliefs decided to take an umbrella (i.e., 72%, 8/11) than did those who provided an appropriate interpretation (i.e., 65%, 11/17). However, due to the small number of participants who provided explicit interpretations, this difference did not reach statistical significance. There were no significant differences due to icon type (e.g., question mark icon, pie icon, or no icon) for the category of explanation, error rates, or decisions. Discussion of experiment 2. Remarkably, when asked to explain PoP in their own words, few participants provided information about the reference class, despite the pointed and probing questions that were asked. There are at least two possible explanations for this omission. Participants may not have believed this aspect required explanation. On the other hand they may not have known what the explanation was and did not want to expose this deficit in understanding. The important thing to note, however, is that among those who attempted to explain the reference class (perhaps those most confident in their understanding), the error proportion was similar to that observed in experiment 1. This suggests that the results of experiments 1 were not simply an artifact of how the question was asked, but were instead an accurate reflection of participants misunderstanding of the reference class issue. february 2009 | 189 As with experiment 1, the icon visualizations provided here did not reduce errors and did not influence participants’ explanation of the forecast. Perhaps some people’s interpretation of PoP is now well entrenched, developed, and maintained over many years of exposure. If their entrenched interpretation involves a misunderstanding, it may be particularly resistant to correction. These users may expect the proportional imagery in the icons to be compatible with their misunderstanding (i.e., as representing percent time or percent area). Experiment 3. The combined evidence from experiments 1 and 2 established that a proportion of college students mistake probability of precipitation for information about the percent of time or area. This is true regardless of whether they are asked in multiple-choice or open-ended format. Moreover, the association between incorrect choices and the tendency to take precautionary action suggests that at least some of them are converting the probabilistic forecast into a deterministic forecast for precipitation with additional information about percent time or area. Hence, the solution may be to contradict this notion explicitly in the forecast. Including a statement of the probability of no precipitation may make it clear that PoP is a forecast indicating that it might not rain at all. This solution was tested in experiment 3. Method. One hundred and two psychology undergraduates participated as part of a course requirement. The procedure, instructions, and questions were similar to that described in experiment 1. The forecast displayed in the upper-left-hand corner of the questionnaire included a phrase describing the chance of rain. Two conditions included only this verbal–numeric information. In one, the chance of no rain appeared beneath the chance of rain (e.g.; “Chance of rain: 25%”; “Chance of no Rain: 75%”); in the other it did not. In the third condition the pie icon appeared beneath the chance of rain statement. Results. We combined the percent area and percent time answers in experiment 3 to create a single reference class variable because this experiment included only about half the number of participants as in experiment 1. As with experiment 1, approximately half of participants made one or both reference class errors (i.e., 48%, 49/102). As Table 2 shows, a smaller proportion of participants answered these questions incorrectly when the chance of no rain was explicitly expressed in the forecast. Pairwise comparisons in a logistic regression analysis revealed that the odds 190 | february 2009 Table 2. Expt 3: Percentage of error responses to reference class questions by visualization type. Visualization type % reference class errors “Chance of rain” 21/33 (64%) “Chance of rain” and “Chance of no rain” 12/33 (36%) Pie icon and “Chance of rain” 16/36 (44%) Total 49/102 (48%) of making an error were significantly lower if the chance of no rain was included as compared to the condition in which only the chance of rain was provided (i.e., β = 0.33, p < 0.03). The odds of making an error in the pie condition were also lower than in the control condition and approached significance (i.e., β = 0.46, p = 0.11). No other significant differences were detected. Discussion of experiment 3. Again, a large proportion of participants indicated that they thought that the PoP forecast included information about percent time or percent area. In this study for the first time, however, we have evidence that this misconception can be significantly attenuated if the forecast makes explicit the fact that there is also a chance that no rain will be observed. Including the chance of no rain along with the chance of rain cuts the reference class error in half. It apparently clarifies for some that the percentage expressed in the forecast refers to the chance of precipitation rather than to the proportion of area or time. There is also evidence that the pie icon provided some improvement in understanding, although the difference observed here did not reach significance. General Conclusions. These three studies demonstrate that probability of precipitation, which has been included in public forecasts for decades, is still confusing to people. This is strongest evidence to date that misunderstandings arise from confusion about the reference class. Errors effected subsequent decisions and were reduced only when the chance of no rain was made explicit. Taken together these results suggest a tendency to construe PoP as a deterministic forecast that indicates something about proportion of time or area. If such deep-seated misunderstanding is evident among this college-educated sample in the rain-experienced Pacific Northwest, we can assume that it exists in similar or larger proportions among the general public. Importantly, this research also demonstrates that targeted enhancements can overcome this misunderstanding among many end users. A reduction in errors with the pie icon was demonstrated in two of the three studies reported here. Although none of these differences reached significance, this icon may have some effect because, in addition to a proportion that includes rain, it illustrates a proportion that includes no rain, clarifying that this too is a possible outcome. Given the widespread use of icons in the media, this is an issue that demands further research. Visualization such as this may be an important tool for successful communication of forecast uncertainty if psychological research is part of the design process. In the research reported here, the biggest reduction in errors was observed when the phrase describing the chance of no rain was provided. These combined results suggest that understanding of PoP forecasts can be improved with targeted enhancements: those that describe or illustrate both possible outcomes. These results have important implications for extending probabilistic forecasts to other weather parameters. Although information about forecast uncertainty could be useful to the general public for weather-related decision making, misinterpretations like those reported here could reduce or even eliminate the potential benefit. While the consequences of carrying around an unnecessary umbrella for a day may be small, for other weather-related decisions the cumulative costs of unnecessary school closures, road treatments, or crop-protection procedures could be substantial. In addition, there is a potential cost to user trust. If the user misinterprets the probabilistic forecast as deterministic and no precipitation is observed, it could be regarded as a false alarm, reducing trust in subsequent forecasts. Although the consequences of reduced user trust in everyday precipitation forecasts are not great, the same effect extended to severe weather warnings could be extremely costly leading to a reduction in compliance and a threat to safety. Much additional research is needed in this area. The results reported here identify some of key psychological issues to be explored. Why do people misinterpret PoP as a deterministic forecast? This may be part of a general human tendency to avoid the complications of incorporating uncertainty in the decision process by ignoring it or turning it into certainty (Slovic 2007). People have shown the same tendency when being told about the side effects of Prozac, thinking that side effects occur in a percentage of sexual encounters of every individual who takes it (Gigerenzer 2002). AMERICAN METEOROLOGICAL SOCIETY In addition, evidence suggests that those living in flood planes appeal to perceived cyclical patterns to predict the size of the next flood (Burton and Kates 1964) rather than regarding it as uncertain. Moreover, weather forecasters, finding that a numerical model prediction is not reliable, tend to make corrections and assume that they have eliminated the element of uncertainty (Joslyn and Jones 2008). Granted, each of these strategies is slightly different. Some of them may have deliberate components. For others the shift from regarding the situation as uncertain to certain may be unconscious. However, all of these strategies strive to reduce or eliminate uncertainty, perhaps because it is easier than incorporating it. Incorporating uncertainty directly into the decision process requires the reasoner to simultaneously consider several hypothetical outcomes, their corresponding levels of uncertainty, and their consequences. This is particularly difficult because many weather-related decisions are made quickly, relying solely on what psychologists refer to as “working memory,” roughly synonymous with consciousness (Baddeley 1987). A major undisputed conclusion of decades of psychological research is the severe limitation that human information processing has in this regard (Miller 1956). There are only so many things we can hold and manipulate in consciousness simultaneously. Because uncertainty constitutes an additional and complex piece of information it may simply be easier to reduce the cognitive load by committing oneself to a single outcome and proceeding as though the uncertainty has been resolved. In many cases, people may not be aware of making this simplification. As in the case with PoP, if there is an interpretation that does not involve uncertainty (e.g., percent refers to time or area) it may seem preferable to people, without their understanding why. As a result, it is crucial when providing the general public with uncertainty information to make the uncertain aspect of the forecast crystal clear. It is also important to provide a forecast in the simplest possible terms to avoid further increase in the cognitive load. In experiment 3 reported here, uncertainty was made explicit by providing both the chance of rain and the chance of no rain. The important thing to realize about this solution is that although “25% chance of rain” implies “75% chance of no rain” to those with the correct interpretation of this phrase, the same implication is NOT clear to those who are misinterpreting the phrase as percent area or time. In other words, some people are simply not interpreting it as an expression of uncertainty. february 2009 | 191 Given the potentially serious consequences, it is important for communicators of forecast information to understand the psychology of the end user, to consider not only what the forecast means, but also how the user is “hearing” it. We have demonstrated one kind of misunderstanding here. However, other expressions (e.g., frequency, predicative intervals, etc.) and visualizations of uncertainty may give rise to different misinterpretations and call for different kinds of clarification. Until the research is done that specifically targets these expressions, the details of their psychological impact will remain unknown. The bottom line is that people can understand and make good use of forecast uncertainty if it is communicated in a manner that takes into account the psychological characteristics of the human information processing system. Armed with this knowledge, forecast providers will be able to communicate uncertainty in a manner that will be understandable and beneficial to a wide range of end users. APPENDIX. RESPONSE CATEGORIES WITH EXAMPLE RESPONSES Category Example response Reference class Percent/proportion days like today “. . . based on the current conditions, 75 days out of 100 days with these conditions will see rain…” Percent/proportion agreement between forecasters “. . . it’s like 3 weathermen say it will rain while 1 say it won’t . . .” Percent/proportion time it will rain tomorrow “. . . 25% of the day (tomorrow) would rain . . .” Percent/proportion area over which it will rain “75% of the entire area mentioned will have rain . . .” Confidence in forecast Degree of confidence of forecaster “…that there will probably be some rain, but they don’t know for sure . . .” Numeric clarification Switching to frequency format “. . . 3 times in 4 it will rain . . .” Switching to ratio format “There is a 1/4 chance of precipitation . . .” Describing 25%/75% as a part of 100% “It is 25% out of 100% that it will rain tomorrow . . .” Confusing the amount of rain with the chance of rain “If it does precipitate there is a very low chance of rain ~0.0” Adding information Adding information “It will be cold tomorrow.” Repeat or summarize Repeating the original probability phrase “There is a 75% chance of rainfall and a predicted amount of 0.03 in. of rain.” Correct verbal summary of the information “There is a very good chance that it will rain . . .” Wrong verbal summary of the information In the 25% condition: “It is more likely to rain.” Vague or general explanation “Probability of precipitation is how likely it is that it is going to rain . . .” REFERENCES Baddeley, A. D., 1987: Working Memory. Clarendon Press, 289 pp. Burton, I., and R. W. Kates, 1964: The perception of natural hazards in resource management. Nat. Resour. J., 3, 412–414. DeElia, R., and R. Laprise, 2005: Diversity in interpretations of probability: Implications for weather forecasting. Mon. Wea. Rev., 133, 1129–1143. 192 | february 2009 Gigerenzer, G., 2002: Calculated Risks: How to Know When Numbers Deceive You. Simon & Schuster, 310 pp. —, R. Hertwig, E. van den Broek, F. Fasolo, and K. V. Katsikopoulos, 2005: “A 30% chance of rain tomorrow”: How does the public understand probabilistic weather forecasts? Risk Anal., 25, 623–630. Joslyn, S., and D. Jones, 2008: Strategies in naturalistic decision-making: A cognitive task analysis of naval weather forecasting. Naturalistic Decision Making and Macrocognition, J. M. Schraagen, Ed., Ashgate Publishing, in press. MacEachren, A. M., 1992: Visualizing uncertain information. Cartogr. Perspect., 13, 10–19. Miller, G. A., 1956: The magical number seven, plus or minus two: Some limits on our capacity for processing information. Psychol. Rev., 63, 81–97. National Research Council, 2006: Completing the Forecast: Characterizing and Communicating Uncertainty for Better Decisions Using Weather and Climate Forecasts. The National Academies Press, 191 pp. Orne, M. T., 1962: On the social psychology of the psychological experiment: With particular reference to demand characteristics and their implications. Amer. Psychol., 17, 776–783. Slovic, P., 2007: Perception of risk from asteroid impact. Comet/Asteroid Impacts and Human Society, P. Bobrowsky and H. Rickman, Eds., Springer–Verlag, 369–382. AMERICAN METEOROLOGICAL SOCIETY february 2009 | 193
© Copyright 2026 Paperzz