General enquiries on this form should be made to: Defra, Science Directorate, Management Support and Finance Team, Telephone No. 020 7238 1612 E-mail: [email protected] SID 5 Research Project Final Report Note In line with the Freedom of Information Act 2000, Defra aims to place the results of its completed research projects in the public domain wherever possible. The SID 5 (Research Project Final Report) is designed to capture the information on the results and outputs of Defra-funded research in a format that is easily publishable through the Defra website. A SID 5 must be completed for all projects. 1. Defra Project code 2. Project title This form is in Word format and the boxes may be expanded or reduced, as appropriate. 3. ACCESS TO INFORMATION The information collected on this form will be stored electronically and may be sent to any part of Defra, or to individual researchers or organisations outside Defra for the purposes of reviewing the project. Defra may also disclose the information to any outside organisation acting as an agent authorised by Defra to process final research reports on its behalf. Defra intends to publish this form on its website, unless there are strong reasons not to, which fully comply with exemptions under the Environmental Information Regulations or the Freedom of Information Act 2000. Defra may be required to release information, including personal data and commercial information, on request under the Environmental Information Regulations or the Freedom of Information Act 2000. However, Defra will not permit any unwarranted breach of confidentiality or act in contravention of its obligations under the Data Protection Act 1998. Defra or its appointed agents may use the name, address or other details on your form to contact you in connection with occasional customer research aimed at improving the processes through which Defra works with its contractors. SID 5 (Rev. 3/06) Project identification PS2301 Web-integrated framework for addressing uncertainty and variability in pesticide risk assessment (WEBFRAM1) Contractor organisation(s) Central Science Laboratory Sand Hutton York YO41 1LZ UK 54. Total Defra project costs (agreed fixed price) 5. Project: Page 1 of 21 £ 278,035 start date ................ 01 January 2003 end date ................. 31 December 2006 6. It is Defra’s intention to publish this form. Please confirm your agreement to do so. ................................................................................... YES NO (a) When preparing SID 5s contractors should bear in mind that Defra intends that they be made public. They should be written in a clear and concise manner and represent a full account of the research project which someone not closely associated with the project can follow. Defra recognises that in a small minority of cases there may be information, such as intellectual property or commercially confidential data, used in or generated by the research project, which should not be disclosed. In these cases, such information should be detailed in a separate annex (not to be published) so that the SID 5 can be placed in the public domain. Where it is impossible to complete the Final Report without including references to any sensitive or confidential data, the information should be included and section (b) completed. NB: only in exceptional circumstances will Defra expect contractors to give a "No" answer. In all cases, reasons for withholding information must be fully in line with exemptions under the Environmental Information Regulations or the Freedom of Information Act 2000. (b) If you have answered NO, please explain why the Final report should not be released into public domain Executive Summary 7. The executive summary must not exceed 2 sides in total of A4 and should be understandable to the intelligent non-scientist. It should cover the main objectives, methods and findings of the research, together with any other significant events and options for new work. This project created a set of computer models for assessing pesticide risks to birds, mammals and aquatic organisms, using improved approaches that account for variability and uncertainty in factors that influence risk. The models produced by the project are freely available for use on the internet at www.webfram.com. Before a pesticide is approved for use, the risks to non-target organisms are assessed. These risks depend on two main factors: the toxicity of the pesticide to the organisms, and the amounts of pesticide they are exposed to. Most current methods for assessing pesticide risks are deterministic – they use a single value for toxicity, a single value for exposure, and produce a single estimate of risk. In the real world, toxicity and exposure are not fixed; instead they vary both between species and between individuals. Furthermore, many aspects of risk assessment involve uncertainty – for example, when extrapolating toxicity from species tested in the laboratory to those exposed in the wild. Consequently, the effects of pesticides are both variable and uncertain. Deterministic methods cannot represent this variability and uncertainty. Instead, they provide a snapshot of one possible outcome. This may be sufficient for simple screening assessments, but for refined assessments a more complete picture of the possible outcomes and their uncertainty would be desirable. Probabilistic methods take account of variability and uncertainty using probability distributions. Probability distributions show the range and relative likelihood of possible values for each input and output, and thus provide a fuller picture of the possible outcomes. This is a major advantage compared to deterministic methods, which use only single values for inputs and outputs. However, uptake of probabilistic methods has been hindered by a number of factors including lack of reliable data for input distributions, lack of standardised methods for probabilistic calculations, the complexity of the methods, and concerns about the validity of assumptions. This project (Webfram1) is one of a set of inter-linked projects that sought to address these obstacles by developing a suite of refined risk assessment models for addressing uncertainty and variability in higher tier assessments and making them available to users via the internet. The specific models that were implemented in this project address risks to aquatic organisms (models from project PS2302, Webfram2), birds and mammals (models from project PS2303, Webfram3). The sister projects designed the models, collated, evaluated and processed the relevant data, evaluated the models by applying them to worked examples (case studies), and provided the algorithms and data for the finished models to this project for implementation on the internet. SID 5 (Rev. 3/06) Page 2 of 21 This project, Webfram1, had an integrating role. First, it reviewed the available approaches for addressing uncertainty and variability in risk assessment, and developed a general framework for applying these approaches across the Webfram suite of projects. This work was closely integrated with preceding and concurrent initiatives on probabilistic risk assessment in Europe and the USA, especially the EUFRAM project (www.eufram.com). Second, this project implemented the models produced by Webfram2 and Webfram3 as web-enabled software freely available for use on the internet at www.webfram.com. Third, this project produced additional, generic models for quantifying variability and uncertainty in toxicity, exposure and risk, to extend the potential applicability of the Webfram software to other taxonomic groups in addition to aquatic organisms, birds and mammals. Finally, this project provided overall coordination for the whole suite of projects, including the organisation of annual meetings with potential end-users and stakeholders. This report provides an overview of the results of this project, including a review of alternative methods for addressing uncertainty and variability, a general framework of principles for using them in risk assessment, and a summary of the models and how they are organised on the project website. The models together with extensive supporting information and user instructions are available at www.webfram.com. Project Report to Defra 8. As a guide this report should be no longer than 20 sides of A4. This report is to provide Defra with details of the outputs of the research project for internal purposes; to meet the terms of the contract; and to allow Defra to publish details of the outputs to meet Environmental Information Regulation or Freedom of Information obligations. This short report to Defra does not preclude contractors from also seeking to publish a full, formal scientific report/paper in an appropriate scientific or other journal/publication. Indeed, Defra actively encourages such publications as part of the contract terms. The report to Defra should include: the scientific objectives as set out in the contract; the extent to which the objectives set out in the contract have been met; details of methods used and the results obtained, including statistical analysis (if appropriate); a discussion of the results and their reliability; the main implications of the findings; possible future work; and any action resulting from the research (e.g. IP, Knowledge Transfer). SID 5 (Rev. 3/06) Page 3 of 21 WEB-INTEGRATED FRAMEWORK FOR ADDRESSING UNCERTAINTY & VARIABILITY IN PESTICIDE RISK ASSESSMENT (WEBFRAM) Andy Hart and Willem Roelofs Central Science Laboratory, York, YO41 1LZ, UK Introduction EU and UK regulations require that, before a pesticide is approved for use, the risk to non-target organisms including birds and mammals has to be assessed. EU Directive 91/414/EEC specifies criteria to be used in an initial “first-tier assessment”, and states that pesticides which fail to meet these criteria may not be authorised for use unless an “appropriate risk assessment” shows that it will cause no unacceptable impact. Various options for these refined, higher tier risk assessments are identified in existing EU Guidance Documents, including both deterministic and probabilistic approaches. Most current methods for assessing pesticide risks are deterministic – they use fixed values for toxicity and exposure, and produce a single measure of risk (e.g. a toxicity-exposure ratio). In the real world, toxicity and exposure are not fixed, but variable. Furthermore, many aspects of risk assessment involve uncertainty – for example, when extrapolating toxicity from test species to humans or wildlife. Consequently, the effects of pesticides are both variable and uncertain. Deterministic methods cannot incorporate variability and uncertainty directly. Instead, uncertain or variable factors are fixed to worst-case values, or dealt with subjectively using expert judgement, or simply ignored. Probabilistic methods take account of variability and uncertainty by using probability distributions to represent them in risk assessment. Although this is a major advantage, uptake of probabilistic methods has been hindered by a number of factors including lack of reliable data for input distributions, lack of standardised methods for probabilistic calculations, the complexity of the methods, and concerns about the validity of assumptions. This project (Webfram1) is one of a set of inter-linked projects that sought to address these obstacles by developing a suite of refined risk assessment models for addressing uncertainty and variability in higher tier assessments and making them available to users via the internet. The specific models that were implemented in this project address risks to aquatic organisms (models from project PS2302, Webfram2), birds and mammals (models from project PS2303, Webfram3). The sister projects designed the models, collated, evaluated and processed the relevant data, evaluated the models by applying them to worked examples (case studies), and provided the algorithms and data for the finished models to this project for implementation on the internet. This project, Webfram1, had an integrating role. First, it reviewed the available approaches for addressing uncertainty and variability in risk assessment, and developed a general framework for applying these approaches across the Webfram suite of projects. This work was closely integrated with preceding and concurrent initiatives on probabilistic risk assessment in Europe and the USA, especially the EUFRAM project (www.eufram.com). Second, this project implemented the models produced by Webfram2 and Webfram3 as web-enabled software freely available for use on the internet at www.webfram.com. Third, this project produced additional, generic models for quantifying variability and uncertainty in toxicity, exposure and risk, to extend the potential applicability of the Webfram software to other taxonomic groups in addition to aquatic organisms, birds and mammals. Finally, this project provided overall coordination for the whole suite of projects, including the organisation of annual meetings with potential end-users and stakeholders. This report provides an overview of the approaches and results of this project. The models together with extensive supporting information and user instructions are available at www.webfram.com. Scientific objectives The scientific objectives agreed for the project are listed below. All have been fully achieved. 1. Review, evaluate and recommend the most appropriate existing mathematical/statistical approaches to addressing uncertainty and variability. 2. Develop and evaluate a generic model framework that incorporates uncertainty and variability into the assessment of pesticide risks to nontarget species. 3. Develop fully web-integrated software, by web-enabling the generic framework and a suite of specific models (algorithms and data for the latter being provided by sister projects), freely accessible via a browser interface with options to provide varying levels of access to different stakeholders. SID 5 (Rev. 3/06) Page 4 of 21 4. Establish a Coordinating Committee and effective means of electronic communication to ensure efficient interaction and exchange of results between this project and the sister projects which will provide the specific models and data for web-enabling in objective 3. Objective 1. Review of approaches for addressing uncertainty and variability Variability is defined as real variation in factors that influence risk. For example, toxicity varies between species, and exposure varies in time and space. Variability matters because risk assessment usually needs to address a range of relevant species and exposures, not just one particular species and one exposure. Uncertainty is defined as limitations in knowledge about the factors that influence risk. For example, there is uncertainty when we extrapolate toxicity from a small number of tested species to other, untested species, and uncertainty when we extrapolate from mathematical models of exposure to the real world. Uncertainty matters because decision-makers and stakeholders need to know the range of possible outcomes and their relative likelihoods. An important practical difference between uncertainty and variability is that obtaining further data can often reduce uncertainty, whereas variability can be better quantified but not reduced by further data. Uncertainties can be classified in various ways (e.g. Morgan & Henrion 1990, Cullen & Frey 1999). Important types of uncertainty include: Uncertainty about distribution shape – often, several different distributions may be plausible for the same model input, and may show similar goodness of fit to sample data. Sampling uncertainty – when a sample is used to estimate distribution parameters or to derive an empirical distribution, there is uncertainty about their relationship to the true parameters or distribution for the larger population from which the sample was drawn. Measurement uncertainty – various factors may cause random errors or bias in measurements or experimental data. Extrapolation uncertainty – when it is necessary to extrapolate beyond the range of a dataset, or from one type of data to another (surrogacy), there is uncertainty about how closely the extrapolated values match the true values that are being estimated. Model uncertainty – there is often uncertainty about which of several alternative model structures best represents real mechanisms and processes. Uncertain dependencies – there may be uncertainty about the nature, direction and magnitude of dependencies between the model inputs. Ignorance – the risk may be influenced by factors of which we are unaware. Quantitative methods for dealing with variability and uncertainty can be divided into two primary categories: Deterministic methods use point estimates (i.e. single values) to represent factors influencing risk. Probabilistic methods use probability distributions to represent variability and/or uncertainty in factors influencing risk. It is important to note that individual assessments often use deterministic methods to represent some factors and probabilistic methods to represent others. There is thus a continuum between assessments that are entirely deterministic and those that are entirely probabilistic. Deterministic methods Deterministic methods use point estimates to represent one or more factors in a risk assessment and treat them as if they were fixed and precisely known. Point estimates represent a measured or estimated quantity by a single number (e.g. the minimum, mean or maximum value), rather than a distribution. When all of the inputs to a risk assessment are deterministic, the output will also be deterministic. The key advantage of deterministic methods is simplicity. Their key limitation is that point estimates do not reflect the variability and uncertainty that, in reality, affect almost all inputs to every assessment. Consequently, the output is one possible value for the risk or impact: it does not reflect the range of possible values, nor give any indication of their relative likelihood. This is potentially a very serious limitation, as it provides the decision-maker with no information on the relative likelihood of different outcomes. The common way of countering this limitation is to make the inputs conservative: i.e. to choose values that will tend to exaggerate the risk. Decision-makers can then act on the result and assume that the actual outcome is unlikely to be worse. Choice of input values for deterministic assessments Choosing suitably conservative input values to use for deterministic calculations is often difficult. SID 5 (Rev. 3/06) Page 5 of 21 Sometimes there are absolute lower and upper bounds for the possible values of an assessment input. However, these are often much wider than the range of values which is really plausible, leading to extremely unrealistic risk estimates. If measurements are available, it must be remembered that the minimum and maximum recorded values will generally underestimate the true range of possible values because the true minimum and maximum are very unlikely to be observed in a limited sample. Often, data will be lacking, limited, poor in quality, or only indirectly relevant (e.g. measured for a different population or in different conditions). In these cases it may still be possible to identify a plausible range of values, based on the available data, expert judgement and common sense. In all cases, the rationale and justification for the values chosen should be reported in sufficient detail so that others can review and evaluate the judgements that have been made. Combining multiple conservative assumptions can quickly lead to a scenario that is extremely conservative or even beyond the bounds of possibility. It is therefore important to consider which combinations are reasonably plausible. The most objective approach would be to use probabilistic methods (see later) to estimate the probabilities of different combinations. Otherwise, the plausibility of different combinations of input values can be assessed subjectively, taking account of any relevant data or evidence, expert judgement and common sense. The highly subjective nature of this assessment makes it essential to report the rationale and justification for the selection of combinations, so that others can review and evaluate the judgements that have been made. Choosing values for deterministic inputs, and deciding which combinations of values are appropriate to consider in the assessment, involves two distinct types of judgement. First, it involves evaluating how conservative (probable) the values are, which is a scientific judgement. Second, it involves deciding how conservative (precautionary) to be, which is a policy judgement. Therefore, some interaction between scientists and policymakers is needed to decide what values are appropriate to use in a deterministic assessment. In many areas of regulation this is achieved, in effect, by the establishment of regulations and/or guidance documents containing standard sets of assumptions for use in deterministic assessments (e.g. EC 2002a,b). Consultation between scientists and policy-makers is then only required for those cases that deviate from the standard procedure. Sensitivity analysis for deterministic assessments Given the difficulty of deciding on appropriate values for deterministic inputs it may be helpful to repeat the assessment for a number of different scenarios, ranging from more typical or probable values to more conservative or improbable ones. A set of alternative scenarios should provide the decision-maker with more insight into the effect of variability and uncertainty in the assessment, and into the range of possible outcomes, compared to what they would learn from a single result. This is a simple form of sensitivity analysis; a broad term used to describe the diverse range of methodologies available for analysing the sensitivity of calculation outputs to their inputs. It may be helpful to summarise the results of sensitivity analysis graphically, showing the relationship between the value of a particular input and the resulting assessment outcome. If ranges of values are being considered for several different assessment inputs, each can be varied in turn while keeping the others constant. More complex methods for sensitivity analysis provide various ways of systematically exploring multiple combinations of inputs. Detailed descriptions, examples and evaluations of these methods may be found in the literature (e.g. Saltelli et al. 2000, Mokhtari & Frey 2005). Another important benefit of sensitivity analysis is that it helps to identify which inputs contribute most to uncertainty and/or variability in the output. Those inputs that contribute most uncertainty may be good targets for research, if it is desired to refine the assessment, whereas those factors that contribute most to variability may be good targets for risk management, if it is desired to reduce the risk. Sensitivity analysis is also useful in conjunction with probabilistic methods (see later). Interval analysis Interval mathematics can be used in cases when it is possible to specify not just a selection of different possible values but the absolute minimum and maximum, forming for each input an interval representing the full range of possible values. Correspondingly, the model output is also an interval: the range of possible outcomes but not their probabilities. Special arithmetic procedures for calculation of functions of intervals used in this method are available in computational tools that are capable of performing interval arithmetic on functions of variables that are specified by intervals (e.g. Ferson 2002). The primary advantage of interval mathematics is that it can address problems for which the probability distributions of the inputs cannot be specified. In addition, interval analysis is capable of handling any kind of uncertainty, no matter what its nature or source. However, this advantage is counterbalanced by the fact that the method does not characterise the nature of the output interval, as both variability and uncertainty are forced into one interval, nor does it estimate the probabilities of different values within that interval. Furthermore, as SID 5 (Rev. 3/06) Page 6 of 21 discussed earlier in relation to deterministic methods, it is frequently difficult to identify truly absolute upper and lower bounds for assessment inputs. Probabilistic methods Probabilistic methods use distributions for one or more inputs to a risk assessment, to represent variability and/or uncertainty in those inputs. If at least one of the inputs to a risk assessment is probabilistic, the output will also be probabilistic. The key advantage of probabilistic methods is that they provide an objective way to account for variability and uncertainty in the factors influencing risk, using formal statistical methods. In principle this provides the means to characterise not only the range of possible outcomes or impacts, but also their relative frequency or likelihood. A variety of approaches exist for probabilistic analysis; those most commonly considered for risk assessment are outlined in the following sections. The approaches differ in their underlying theory, in the way distributions are specified, and in the way distributions are combined or propagated in the assessment, and in the type of output they produce. Some approaches (e.g. 2-dimensional or 2D Monte Carlo) account for variability and uncertainty separately, whereas others do not. Fuzzy arithmetic Fuzzy arithmetic is a generalization and refinement of interval analysis in which the bounds vary according to the level of confidence or belief that one has that the value is within an interval. These confidence bounds are described by so-called membership functions, which can be seen as stacked intervals, where the bounds become wider when moving down from possibility level 1 to 0. At the highest possibility level, everyone would agree that the values within that interval could contain the true, but unknown value. Close to the lowest possibility level, the interval is much wider, indicating that only a few people would expect that the true, unknown value could be as high or as low as the interval bounds. In other words, the possibility level at a certain value is an estimate of our degree of belief of whether the parameter could have that value. Just as with confidence intervals, arithmetic operations could be applied to fuzzy numbers. The output of fuzzy calculations, however, will be a distribution rather than a single range. Another advantage of fuzzy arithmetic is that it does not require any information about correlations between parameters. The main disadvantage of fuzzy arithmetic is that it still mixes variability and uncertainty. In addition, many assumptions are necessary to define the fuzzy numbers, particularly when information suggests that there are no outer bounds or when the information on possible bounds is scarce. Probability bounds Probability bounds is a method of combining uncertain distributions that does not require any assumptions about either distribution shape or dependencies among variables, although such information can be used to some extent if available. It first produces absolute (outer) bounds for the cumulative distribution of each input variable, and uses these to derive absolute (outer) bounds for the output distribution (illustrated in Figure 1). However, an important limitation is that this method does not provide any information on where the true distribution might lie between the bounds. The method can work with many types of information, and each piece of information can itself be uncertain. Probability bounds copes well with most types of uncertainty but has difficulties with sampling uncertainty, which is important for the small sample sizes that are common in risk assessment. Probability bounds may be very wide, but this is appropriate if there is genuinely little information about the inputs. Software for probability bounds is provided by Ferson (2002). Frequency of effect (how often) 1 a Uncertainty (how sure) 0 .5 0 -2 0 -1 0 0 10 20 Magnitude of effect (how bad) Figure 1. Example of output from probability bounds analysis. This method provides absolute bounds, within which the cumulative distribution of the assessment endpoint must lie. SID 5 (Rev. 3/06) Page 7 of 21 1D Monte Carlo Monte Carlo simulation combines distributions by taking large numbers of samples from each distribution at random. A good introduction is provided by Vose (2000). In 1D (one-dimensional or first order) Monte Carlo, all the input distributions are sampled together, and produce a single distribution for the output. The following points are rather complex, but important for the construction and interpretation of 1D Monte Carlo assessments. When all the input distributions exclude uncertainty and represent only variability, the output distribution represents only variability in the assessment endpoint. It can be used to estimate specific percentiles of the output distribution (e.g. the 5th percentile), but provides no confidence intervals and may give a false impression of certainty. When input distributions representing variability and uncertainty are combined by 1D Monte Carlo, the output distribution represents the combined effect of variability and uncertainty on the assessment endpoint. The output distribution can be interpreted as representing our uncertainty about the assessment endpoint for a single member of the statistical population or ensemble, selected at random. The difference between 1D Monte Carlo including and excluding uncertainty is illustrated in Figure 2. The distribution excluding uncertainty (labelled MLE in Figure 2) represents a central estimate of the CDF, whereas the distribution combining both variability and uncertainty (labelled 1D) spans a wider range of possible values for the assessment endpoint. Monte Carlo methods can take account of dependencies between input variables, if information is available to characterise them, but require strong assumptions about distribution shape. A number of software packages exist for 1D Monte Carlo and are fairly easy to use (e.g. @Risk© and CrystalBall©). 100% Cumulative frequency % of species 80% 60% 95% Conf. Interval Median EC50 2D 2D 40% 20% 0% 100 1D 1D MLE MLE 104 1000 105 Magnitude Figure 2. Example of the difference between the output of different forms of Monte Carlo simulation for the same input data. 2D Monte Carlo can produce a median estimate and confidence intervals for each percentile of the distribution. 1D Monte Carlo combining variability and uncertainty produces a single distribution representing uncertainty about a single member of the statistical population selected at random. Monte Carlo based on the maximum likelihood (MLE) estimates of the mean and variance excludes uncertainty and produces a distribution close, but not identical, to the median curve from 2D Monte Carlo. 2D Monte Carlo Two dimensional (2D) Monte Carlo is very similar to 1D Monte Carlo except that it separates uncertainty and variability. Again, a good introduction is provided by Vose (2000). In 2D Monte Carlo, distributions for uncertainty and variability are sampled separately, so that the combined effect of the uncertainties can be shown as a confidence interval around the output distribution (illustrated in Figure 2). The ability to show the effect of uncertainties separately as confidence intervals makes 2D Monte Carlo especially useful. Statistical estimates can be used for some types of uncertainty (e.g. measurement and sampling uncertainty, and extrapolations calibrated by regression analysis). Other types of uncertainty must be estimated subjectively, although this has been criticised as lacking mathematical rigor (Ferson, 2002). Facilities for 2D simulation are included in some Monte Carlo software (e.g. CrystalBall©), although the number of samples that can be drawn is often limited. Like 1D Monte Carlo, 2D methods can take account of known or assumed dependencies between input variables but require strong assumptions about distribution shape. SID 5 (Rev. 3/06) Page 8 of 21 Bayesian methods Bayesian methods use a subjective concept of probability that is designed to make use of subjective information (e.g. expert judgements, which are often used in risk assessment), as well as objective data. Bayes’ theorem is used to combine existing (“prior”) knowledge with new data, a process sometimes called “updating”. This produces “posterior” distributions for the assessment endpoint. For variables where no prior information is available, “uninformative priors” are used. Confidence intervals can be generated for different percentiles of the assessment endpoint, if required, so Bayesian methods can generate the same types of output as 1D and 2D Monte Carlo (Figure 2). The concept of updating fits neatly with the process of tiered risk assessment, and can also be used to combine different types of information (e.g., lab and field). Several different schools of Bayesian approaches exist, e.g. robust Bayes uses families of uninformative prior distributions to counter concerns that choosing between them may have undue influence on the results. Some of the theory and computations involved in Bayesian methods are complex and require specialist expertise (for a simple introduction see Vose 2000). However, this need not be an obstacle to their use if they can be provided in a pre-packaged form that is known to be appropriate to the problem in hand (e.g. customised software or lookup tables). Sensitivity analysis for probabilistic assessments Sensitivity analysis can be applied to probabilistic assessments, although the practicalities depend on the type of probabilistic method used. For 1D Monte Carlo, relationships between inputs and outputs can be examined graphically or by computing correlation coefficients, both of which are provided in the commonly-used Monte Carlo software packages @Risk© and CrystalBall©. The same can potentially be done for 2D Monte Carlo, with the addition that sources of variability and uncertainty can be analysed separately, although this is more cumbersome computationally and is not included in the 2D Monte Carlo function in CrystalBall©. The contributions of different sources of variability and uncertainty in probability bounds can be analysed by repeating the calculations and omitting each source in turn. Simple forms of sensitivity analysis can also be used to examine model assumptions, e.g. by conducting multiple analyses with different assumptions for distribution shapes and dependencies. This helps to increase the robustness of the assessment as a whole but quickly becomes cumbersome if there are many such assumptions to consider. More detailed accounts of sensitivity analysis are given by Saltelli et al. (2000) and Mokhtari & Frey (2005). Choosing between probabilistic methods Each probabilistic approach has strengths and weaknesses, and there is no general consensus on which approaches should be preferred. In a report on quantitative microbial risk assessment, the European Commission’s former Scientific Steering Committee (SSC) stressed the need for further evolution of the methodology of quantitative risk assessment, and of criteria for the appropriate use of different probabilistic approaches. The SSC concluded that “a quick harmonisation at the present state-of-the-art should be avoided” (European Commission 2003). The lack of a generally preferred approach implies that prospective users need to consider for each assessment which of the alternative approaches is appropriate. Issues that users might consider in choosing between alternative approaches include: Robustness. Users may reasonably prefer methods that are robust in the sense of not requiring questionable assumptions. This may lead them to favour interval analysis or probability bounds, since these do not require information about distribution shape and dependencies between variables, which is often limited. However, see next issue. Bounds versus probabilities. Probability bounds (and also interval analysis) provide upper and lower bounds on the outcome, but cannot estimate the probability of different outcomes. Bounds are useful if the nature of the risk is such that it is sufficient to know whether a particular outcome is inside or outside the bounds of possibility. This may apply to severe outcomes (e.g. extinction of a species) where decisionmakers wish to find management options that ensure a probability of (very close to) zero. But for most risk problems, the occurrence of adverse outcomes cannot be ruled out and decision-makers want to know their probabilities, which cannot be estimated from bounding methods. This need may outweigh the preceding issue of robustness: approximate probabilities may be more useful than no probabilities. Separation of uncertainty and variability is useful because it distinguishes the main sources of uncertainty (which may be good targets for research) from the main sources of variability (which are not reducible by research but may be options for risk management, if they are controllable). Furthermore, methods that separate uncertainty and variability are needed if decisions hinge on particular percentiles of a variable, e.g. the probability that an impact occurs in 1% of fields, rather than the probability that an impact occurs in a randomly-selected field. 2D Monte Carlo, Bayesian methods and probability bounds separate uncertainty and variability, whereas 1D Monte Carlo, interval analysis and fuzzy arithmetic do not. Use of subjective information. If objective data is lacking for key inputs, then it may be desirable to use subjective estimates (e.g. expert opinion). Bayesian approaches provide a formal theoretical framework for this, although subjective information can also be used less formally in 2D Monte Carlo and probability bounds. Ease of use is an important practical consideration. 1D Monte Carlo is relatively easy to use and available in many software packages; 2D Monte Carlo is of intermediate difficulty and available in fewer software packages. Probability bounds and Bayesian methods are (or are perceived as) more difficult to use, but may SID 5 (Rev. 3/06) Page 9 of 21 provide better solutions depending on the problem. Any of these methods can be made easy to use if readyto-use models are provided within user-friendly software, although the more complex methods will pose greater challenges to user understanding and acceptance. Choice of approaches for addressing uncertainty and variability in Webfram 2D Monte Carlo was selected as the primary approach for use in the Webfram framework. This choice was based primarily on two considerations: The desire to quantify both variability and uncertainty with distributions, so as to characterise not only the range of alternative outcomes but also their relative frequency (variability) and likelihood (uncertainty). This argues in favour of Monte Carlo or Bayesian approaches rather than deterministic approaches, interval analysis or probability bounds. The desire to separate variability and uncertainty, so that distributions (and hence percentiles) for variable outcomes can be provided together with confidence or probability intervals to represent their uncertainty. This was considered more relevant and interpretable for decision-making than a distribution for randomly-chosen outcomes (e.g. for a random field or random individual). This argues for a 2D rather than 1D Monte Carlo approach. It was decided to adopt a 2D Monte Carlo approach rather than fully Bayesian equivalent because the former was simpler to develop, simpler to implement as online software, and easier to communicate. However, this does not preclude the use of Bayesian methods to model input distributions for use in (propagation by) 2D Monte Carlo, where appropriate. For example, Bayesian methods were used for normally- and log-normally distributed parameters and for a linear regression relationship in the Webfram3 (bird and mammal) models. Monte Carlo approaches require more assumptions than probability bounds, especially concerning distribution shapes and dependencies. One possibility is to use Monte Carlo and probability bounds in parallel, and compare the results (e.g. Regan et al. 2002): the Monte Carlo results provide an estimate of the median distribution of outcomes and its probability density under specific assumptions regarding distribution shapes and dependencies, while the probability bounds (which should be wider) provide a measure of the additional uncertainty associated with those assumptions. That was not done in Webfram partly because the extra work would have reduced the number of models that could be developed, and partly because of difficulties encountered in implementing probability bounds (for representing sampling uncertainty, and for modelling variables that must sum to 1 such as dietary composition in the Webfram3 models). It was recognised that, whichever approach is adopted, assumptions will be required, only some of the relevant uncertainties will be quantified, and some parameters that influence risk may be omitted altogether (e.g. the Webfram3 models consider only exposure via the dietary route). Such limitations are unavoidable in risk assessment and apply equally (or more so) to deterministic methods. They result in additional uncertainty about the relation between the assessment outcome and the “true” risk. It is therefore essential to evaluate the potential impact of these additional uncertainties as far as is possible, so that they can be taken into account in decisionmaking. A qualitative approach for doing this was developed in Webfram3, as a way of responding to questions raised in the EUFRAM project about the reliability of model outputs. The approach involves systematically identifying unquantified uncertainties and tabulating them together with an evaluation of their impact on the assessment outcome. The approach is potentially applicable to all types of assessment, and is therefore included as a standard part of the Webfram framework (see below). It has also been adopted by the European Food Safety Authority for dealing with uncertainty in human dietary exposure assessments (EFSA 2006). Objective 2. General framework for incorporating uncertainty and variability in the assessment of pesticide risks to non-target organisms DEFRA’s specification for the Webfram projects stated that the selected mathematical/statistical approaches to addressing uncertainty and variability should be developed into a generic framework, and that modules for different groups (aquatic, terrestrial vertebrates, terrestrial invertebrates) should be made compatible with the framework. It was recognised that the specifics of the models (their structure as well as data) would need to differ between groups. The aim of the framework is to identify those things that can be done in a consistent way between groups. This is desirable to ensure best practice and avoid unnecessary divergence of approaches. It should also lead to savings in effort required to develop model software, as some components will be transferable between modules. As anticipated in the project proposal, development of the general framework within Webfram proceeded in close consultation with parallel work in the EUFRAM project (www.eufram.com). The resulting framework is presented in the form of a concise description of the main steps that should be taken in developing a risk assessment, focussing on where and how uncertainty and variability are taken into account, and identifying key issues that should be considered. In addition, a shorter checklist of the key points is provided, which is intended as an aide memoire for those developing and reporting risk assessments, and for those evaluating them. SID 5 (Rev. 3/06) Page 10 of 21 General framework for incorporating uncertainty and variability in risk assessment The following sections outline a general framework for conducting quantitative risk assessments. The main elements are the same for both deterministic and probabilistic approaches. Aspects that are specific to one approach or the other are identified in the text. The framework can be used when assessing the risk for a particular pesticide, but in Webfram they are applied to developing scenario-based models that can be applied to a number of pesticides with similar uses (crop, date of application). Problem definition Developing a risk assessment model, whether for a specific pesticide or for repeated use in a given scenario, should start with a process of problem definition (sometimes called problem formulation) to ensure the assessment will meet the regulatory need. Problem definition should first define the scope and objectives of the assessment and then specify what form of output will meet those objectives: these steps are discussed in the following two sections. Problem definition is a vital task that requires effective interaction between risk assessors and decision-makers (sometimes called risk managers) to ensure the risk assessment is fit for purpose, i.e. addresses the relevant question and communicates the results effectively. Effort will be saved if a similar assessment has been done previously, or if a standardised problem definition has been established, but a check should always be made to confirm that the problem definition is truly appropriate for the issue under consideration. Assessment scope and objectives The first step is to define the objectives and scope of the assessment, starting with the basic elements that should also be defined for qualitative assessments: Define the scenario to be considered (crop, time of year, non-target species, etc.) Define the types of effect to be considered (e.g. direct/indirect, mortality, reproduction, etc.). Define the timescale of the assessment (often implied by the type of effect, e.g. acute/chronic). Define the spatial boundaries of the assessment (e.g. regional, national, EU). These choices should be made to reflect as closely as possible the protection goal – what the decision-maker wants to protect, and what types of effect they wish to limit or prevent. If it is too difficult to assess the types of effect that ultimately concern the decision-maker, then the assessment can be directed at a simpler measure of effect (e.g. effects on a species rather than an ecosystem). However, if this is done then it will be necessary to consider the relation between these two measures of effect when interpreting the results. If the decision-maker is concerned about particular sources of variability and uncertainty in the assessment inputs, then it is helpful to identify these as part of the problem definition. However, this should not preclude additional sources being examined later, if it becomes apparent during the assessment that they are important. Finally, any limitations on the scope of the assessment should be specified, e.g. if the assessment is restricted to considering particular routes of exposure or types of impact. Assessment endpoint (measure of risk) The next step is to define a measure of risk that will address the assessment objectives and provide a suitable basis for decision-making: we refer to this key output as the assessment endpoint. Assessment endpoints should provide a meaningful and readily interpretable measure of impact on something the decision-maker wishes to protect or prevent, i.e. one of the types of effect listed in the preceding step. Each distinct type of effect will require a separate assessment endpoint. The assessment endpoint measure may be a point value, e.g. percentage mortality for individuals exposed to a pesticide. Here the measure is a single value (though it will be uncertain, see later). On the other hand, it may be important to the decision-maker to know how the measure of impact is distributed over one or more dimensions, e.g. how the level of mortality varies between species, or the proportion of species in which it exceeds some specified value. Defining the dimensions over which variability is to be assessed is very important in probabilistic assessments, for two reasons: to ensure that decision-maker and assessor are both clear about what output is to be produced; and because it affects the structure of the model and the types of input distributions that are needed. For example, if the assessment endpoint is the proportion of species that exceed some specified level of mortality, the assessment is likely to require input distributions for variation between species of any factors that influence mortality (toxicity, exposure, etc.). Two elements are needed to define each dimension over which the endpoint varies: the units of the distribution, and the population or ensemble of units to which the distribution refers. For example, the units are species for a distribution representing variation between species, and the population or ensemble might be all species found in the exposed habitats, or a subset of these (e.g. all fish species). SID 5 (Rev. 3/06) Page 11 of 21 Defining the dimensions over which the assessment endpoint varies is also important even if all parts of the assessment will be treated deterministically. This is because consideration needs to be given to how to choose appropriate point values to represent each input to the assessment. Otherwise, (a) it is difficult to judge where the assessment outcome lies in the true distribution (i.e. how conservative the assessment is), and (b) inconsistencies may arise in the construction of the model, e.g. if an average in space for one input is combined with an average over time for another. It is also important to discuss in advance with the decision-maker how uncertainty affecting the assessment is to be characterised. For example, if uncertainty is to be expressed as a distribution or confidence interval for the assessment endpoint, this would require a probabilistic method separating variability and uncertainty. If confidence intervals are required, it is helpful to agree in advance what level they should be (e.g., 90 or 95% confidence intervals). Finally, we emphasise that it is important to define the output first, and then use that to infer what the inputs should be. It is tempting to do the opposite – i.e. allow the available input datasets to determine the nature of the output. This can cause under- or over-estimation of variability or produce output distributions that have no realworld interpretation1. Developing and implementing the assessment Having defined the output that is required, the next step is to develop and implement the an appropriate assessment model to generate the desired output from the available data. Developing a model may require substantial effort if the assessment scenario is new. Effort will be saved if a similar assessment has been done previously, or if a standardised assessment has been established, but a check should always be made to confirm that the assessment is truly appropriate for the issue under consideration. The key steps in developing and implementing the assessment are outlined below. 1. Identify the factors and mechanisms that influence the assessment endpoint, and how they interact. It may help to draw a diagram representing the various factors and mechanisms and showing their interactions, e.g. as a flow chart (this is sometimes referred to as a conceptual model). An example from Webfram3 is shown in Figure 3. BW: body weight DEE: daily energy expenditure GE: gross energy content of food FIR: Food intake per day (wet wt) M: moisture content of food PT: Fraction of diet obtained in treated areas AE: assimilation efficiency AppRate: application rate RUD: residue per unit dose C: Initial concentration PD: Fraction of food type in diet Exposure (mg/kg bw/day) C: Concentration on food type LD50 (mg/kg bw) AV: Avoidance omitted Slope of dose- response %Mortality Figure 3. Example of a conceptual model diagram, for acute risks to birds (from project PS2303, Webfram3). 2. Consider each part of the model in turn, and decide which factors and mechanisms might contribute significantly to variability in the assessment endpoint. These will need to be represented by suitable distributions or point estimates in the model. 3. For each variable in the assessment, identify the appropriate population or ensemble. This depends on the ensemble of the assessment endpoint (see earlier) and the model structure. For example, if the assessment endpoint is effects in water bodies adjacent to treated fields, then water bodies adjacent to treated fields will be the relevant ensemble for spatial variation of exposure, and the set(s) of species found in such water bodies will be the relevant ensemble for between-species variation in toxicity. 1 For example, combining within-species variation for one parameter with between-species variation for another would create an output distribution of entities that are all one species with respect to the first parameter but represent multiple species with respect to the second parameter. SID 5 (Rev. 3/06) Page 12 of 21 4. Identify what data are available that can help in quantifying each factor and mechanism. 5. Decide whether any extrapolations or adjustments are needed to model the key factors and mechanisms from the available data. Extrapolation or adjustment may be required to deal with various issues including: Lab to field extrapolation. Non-random sampling. Surrogacy – where one type of data is used as a surrogate for another type. Incompatible data populations or ensembles, e.g. a model may require data on variation over time, but the only available data may be on variation between sites. Inappropriate averaging/incomplete information. Scientific publications frequently report just averages, or averages and standard deviations without the raw data or any information on distribution shape, which are necessary to properly estimate variation and uncertainty. If issues of this type do require the use of adjustment or extrapolation factors, then they become, in effect, part of the assessment model and should be identified as such in the model plan. Also, consideration should be given to whether these factors should be represented by point values or by distributions (if they are subject to substantial variability or uncertainty). 6. Consider each part of the model in turn to identify possible sources of uncertainty. Decide which uncertainties might contribute significantly to uncertainty in the assessment endpoint, and which of these will be quantified using distributions (if any). Decide what to do about important uncertainties that cannot or will not be quantified (e.g. use best estimates or conservative values). Different types of uncertainty, and methods for dealing with them, are discussed under objective 1 (above). 7. Consider carefully for each input distribution whether it contributes uncertainty or variability to the output. This again depends partly on the ensembles. For example, a species sensitivity distribution contributes variability if the output population comprises multiple species, but uncertainty if the output population comprises individuals of a single species. Distributions estimated from data generally contain both variability and measurement uncertainty, which can be separated if the distribution for measurement uncertainty can be specified (e.g. Zheng & Frey 2005). If a distribution includes both variability and uncertainty and they cannot be separated, it may be better to treat the distribution as uncertainty as this reflects the state of knowledge more accurately than if it were treated as variability. 8. Identify potential dependencies affecting the assessment. Dependencies occur where the value of one variable depends upon the value of another variable (e.g. food intake may be positively related to body weight). It is important to consider dependencies carefully, as they can have a major impact on the assessment outcome. 9. Decide on which type of deterministic or probabilistic method to use. The choice of method will depend on various factors including whether bounds or probabilities are required for the output, whether it is desired to separate variability and uncertainty, ease of use, and what expertise is available. It may also be influenced by the preceding steps e.g. the availability of data for estimating distributions and dependencies. For reasons discussed earlier in this report, 2D Monte Carlo was adopted as the primary approach for propagating variability and uncertainty in Webfram. 10. Define the distributions or point estimates to be used for variable and uncertain inputs and dependencies. This is a critical step, as inappropriate choices will give misleading or invalid results for both deterministic and probabilistic outputs. The methods used for representing distributions and dependencies depend on the type of probabilistic or deterministic approach that has been chosen. For example, probability bounds uses p-boxes to represent distributions, and assumes either independence or perfect correlation between each pair of variables (Ferson, 2002). There is a wide range of options for specifying distributions for Monte Carlo and Bayesian approaches, including parametric and non-parametric methods, and subjective approaches based on expert opinion. The most familiar form of dependency is linear correlation, but there are many others. Whatever methods are used, special attention should be paid to ensuring the tails of distributions are reasonable, as they are frequently critical to the outcome of risk assessment but poorly represented by data except in very large samples. An extensive literature exists on these topics, including substantial chapters in publications on probabilistic risk assessment (e.g. Vose 2000, Cullen & Frey 1999, US EPA 1997, 2001). Deterministic approaches use point estimates to represent variability, uncertainty and dependencies. Ideally, these should be chosen based on a careful statistical analysis similar to that required for a probabilistic treatment, so that the relation of the point estimate to the distribution is known. In practice, however, deterministic point estimates are often chosen in a rather arbitrary way – e.g. by using the average or maximum value from a small set of measurements, without any statistical analysis. The disadvantage of this is that it provides very poor control over the conservatism of the assessment. SID 5 (Rev. 3/06) Page 13 of 21 It should be emphasised that specifying appropriate distributions and dependencies requires a high level of statistical expertise, as well as a clear understanding of the assessment model, and expert knowledge of the parameters in question (e.g. ecology, ecotoxicology, environmental chemistry, etc.). 11. Express the entire assessment model as a set of mathematical equations, so that they can be used for calculations. 12. Consider conducting sensitivity analyses. These may help in various ways, e.g. exploring alternative assumptions or scenarios, and analysing the relative importance of different sources of variation and uncertainty. 13. Select appropriate software to carry out the computations and generate outputs. The selection of software will depend in part on the selected method for conducting the quantitative analysis, as some methods may only be possible with specific software. 14. Program and run the model including any sensitivity analysis, and generate the outputs. 15. Check the outputs of the assessment. Results should be evaluated carefully for any indication of model mis-specification or programming errors, or assumptions that have led to unrealistic results. 16. Qualitatively evaluate sources of variability, uncertainty and dependency that have not been quantified. Ultimately it is never possible to quantify all uncertainties affecting an assessment. It is therefore essential to state clearly in the assessment, (a) which uncertainties have been quantified, and (b) what uncertainties remain unquantified and their potential implications for the assessment. The following steps may assist in evaluating unquantified uncertainties: Systematically list or tabulate all unquantified sources of uncertainty. Qualitatively evaluate the direction and relative magnitude of each unquantified uncertainty in terms of its impact on the assessment outcome Consider re-running the assessment with different assumptions to assess the influence of potentially important uncertainties or dependencies (a form of sensitivity analysis) Consider using probability bounds to explore the impact of potentially important uncertainties about distribution shape and dependencies (in effect, putting conservative outer bounds around the assessment result, e.g. Regan et al. 2002) Qualitatively evaluate the possible combined effect of all the quantified and unquantified uncertainties and dependencies on the predicted outcome (how different could the “true” outcome be?). An example of qualitative assessment of uncertainties using a tabular format is included in the report on the sister project, Webfram3 (birds and mammals, project number PS2303). 17. Consider other lines of evidence. Assessors should seek to provide an overall characterisation of risk, combining the outcome of the quantitative assessment with any other relevant lines of evidence, e.g. direct measurement of the assessment endpoint in field studies, data from monitoring of commercial use, or evidence from different but related pesticides. However, it should be remembered that such types of information are also subject to uncertainty and bias. Therefore, if other lines of evidence are to influence the characterisation of risk, then they should be subject to the same standards of assessment as the quantitative assessment. Sometimes, different lines of evidence can be compared quantitatively, e.g. using Bayesian updating to combine model predictions with field measurements. No line of evidence should be discounted unless there is convincing evidence that it is invalid. 18. Consider the wider ecological consequences. Full characterisation of risk may require consideration of wider ecological consequences of effects indicated by the assessment. This is often necessary because the quantitative assessment is constrained (for reasons of practicality) to simple measures of risk that are only indirect or partial measures of the assessment objectives or protection goal. For example, quantitative estimates may relate to effects on a local population of one species, but decision-makers may be more interested in impacts at the level of the community or landscape. Any conclusions about such extrapolations should be subject to the same standards of evidence as the quantitative assessment itself. For example, arguments about the ability of the ecological systems to absorb or recover from impacts should be supported by data. 19. Communicate the assessment and results to decision-makers and others as appropriate. This will include presenting results for the assessment endpoints that were agreed at the outset, together with explanation and interpretation to assist understanding. In addition, it is essential to document the assessment in detail, including the justification for each choice made in planning the assessment model, and each component of the resulting equations. It may aid communication if the primary output of the assessment is a concise summary focussing on the key results, but fuller details should be accessible to support the credibility of the assessment and facilitate peer review. SID 5 (Rev. 3/06) Page 14 of 21 Summary checklist The following list is intended as a concise aide memoire when developing risk assessments, when drafting assessment reports, or when evaluating reports of assessments produced by others, to ensure that all the major issues are covered. Introduction. Is the purpose and context of the assessment clearly explained? Does the assessment as a whole follow an established approach accepted by the relevant authorities, or is it novel? Problem definition. Are the protection goals identified? Are the assessment scenario and endpoints appropriate and consistent with the protection goals? Assessment model. Is the assessment model appropriate? Does it include all relevant routes of exposure? Does it include all relevant factors and combine them in appropriate ways? Probabilistic methods and software. Are any probabilistic methods (e.g. 1D or 2D Monte Carlo, etc.) and software appropriate, and are they correctly used? Have they been accepted by relevant authorities or independent peer review? For computational methods, is it demonstrated that the number of iterations is sufficient to produce stable outputs? Input data. Consider every model input in turn. Are the data used for each input appropriate? Are they consistent with the statistical populations needed to generate the assessment endpoint? If extrapolation is required, has this been justified and are the associated uncertainties considered? Distributions. Are appropriate distributions used to represent variability and uncertainty? Are they compatible with the nature of the parameter concerned? Is goodness of fit evaluated graphically and statistically? Are the tails of the distribution appropriate? Is any truncation justified and correctly applied? Uncertainties. Are potential uncertainties of all types sufficiently considered? Are potentially significant uncertainties quantified, or addressed through conservative assumptions, or is their influence evaluated qualitatively? Deterministic methods. Where point estimates are used to represent variable or uncertain inputs, is the choice of values justified? Is their effect on the conservatism of the assessment output evaluated? Dependencies. Is the possibility of dependencies between variables been sufficiently considered? Are potentially significant dependencies quantified, or is their influence evaluated qualitatively? Risk characterisation. Are exposure and effects combined in an appropriate way? Does it quantify the frequency, magnitude and duration of effects? What is the probability of exceeding any relevant decisionmaking criteria? What types of organisms are at risk? Is sensitivity analysis used and, if so, which sources of variability and uncertainty are most influential? Are the strengths and weaknesses of the assessment, and their overall effect on the reliability of the assessment, evaluated? Is it useful to compare probabilistic and deterministic results? Are other relevant sources of evidence critically evaluated? Are the wider ecological implications considered? What is the applicability of the results to different scenarios, regions etc.? Conclusion. Is the overall conclusion fully supported by the assessment? Are options for further refinement identified, if relevant? Are sufficient details reported to repeat the assessment? Objective 3. Develop web-enabled software for a suite of models provided by the sister projects The specific models that were implemented in this project address risks to aquatic organisms (models from project PS2302, Webfram2), birds and mammals (models from project PS2303, Webfram3). The sister projects designed the models, collated, evaluated and processed the relevant data, evaluated the models by applying them to worked examples (case studies), and provided the algorithms and data for the finished models to this project for implementation on the internet. This project, Webfram1, implemented the models produced by Webfram2 and Webfram3 as web-enabled software freely available for use on the internet at www.webfram.com. In addition, this project produced a set of generic models for quantifying variability and uncertainty in toxicity, exposure and risk, to extend the potential applicability of the Webfram software to other taxonomic groups in addition to aquatic organisms, birds and mammals. The organisation of the models on the website is illustrated in Figure 4. SID 5 (Rev. 3/06) Page 15 of 21 Figure 4. Diagram of Webfram internet site including main models and user functions. Welcome & User login New model Bird & mammal Saved models Aquatic Instructions & info Generic Bird/mammal Aquatic Tox data entry Scenarios (choice of 5) Risk modelling (acute or chronic) Exposure Herbest model (recovery potential) LIfe history database Tox data entry & goodness of fit Exposure data entry Effects Output options Exposure model Default data User RUDs User DT50 Spray drift 7 scenarios: - 4 arable - 3 orchard SSD Drainflow 3 soils 3 climates 10 crops Single value (eg lowest or mesocosm) SSD Outputs SSD Dose response Exposure User data (eg monitoring) Overlay graph Food intake Exposure Overlay plot TERs % Mortality Risk graphs (JPC & FA) Single value (eg FOCUS) Details of the models for birds, mammals and aquatic organisms are described in the final reports of projects PS2302 and PS2303 (Webfram2 and Webfram3), including data, assumptions, modelling methodologies and outputs. Full details are provided on the website, via the “instructions and info” page and as help pages attached to specific steps within each model. The generic model allows the user to fit lognormal distributions to toxicity and exposure distributions and produce a number of outputs including species sensitivity distributions (SSDs), exposure distributions, overlay plots (SSD and exposure distribution plotted together) and 3 types of joint probability curves (JPCs). Examples of these outputs are illustrated in Figures 5 – 8 below. The use of SSDs, exposure distributions, overlay plots and joint probability curves is well established in ecotoxicology literature (e.g. Posthuma et al. 2002 and chapters therein), although only sparingly used in regulatory assessment in the past. New features introduced in the Webfram generic model are (a) estimation of confidence intervals for the whole of each distribution (rather than just selected percentiles, e.g. HC5), and (b) a new format for the JPC, plotted as an exceedance distribution to show the proportion of exposures that exceed any given effect level (Figure 8). SID 5 (Rev. 3/06) Page 16 of 21 100% 95% Conf. Interval Median EC50 90% 80% % of species 70% 60% 50% 40% 30% 20% 10% HC5 0% 0.1 1 10 100 EC50 [g / L] 1000 104 Figure 5. Example of using a species sensitivity distribution produced by the Webfram generic model, using hypothetical data. Points represent the EC50 for 9 tested species (species names can be shown as labels). The arrows show the HC5 (hazardous concentration for 5% of species), together with confidence intervals. In this hypothetical case, the HC5 is 3.9 g/L (95% confidence interval 0.1 – 18.0 g/L). 100% 90th %-ile 90% % of exposures 80% 70% 60% 95% Conf. Interval Median Data 50% 40% 30% 20% 10% 0% 0.1 1 10 100 Concentration [g / L] 1000 104 Figure 6. Example of a cumulative distribution for exposure produced by the Webfram generic model, using hypothetical data on concentrations in surface water. The arrows show the 90 th percentile concentration. In this hypothetical case, the 90th percentile concentration is 22 g/L (95% confidence interval 13 – 50 g/L). SID 5 (Rev. 3/06) Page 17 of 21 % of exposures 100% 80% 80% Median - Exposure 95% Conf. Interval Median - Toxicity 95% Conf. Interval Exposure Data EC50 60% 40% 60% 40% 20% 20% 0% -6 10 10-4 0.01 1 100 104 106 % of species exceeding toxicity endpoint 100% 0% Concentration [g / L] Figure 7. Example of risk characterisation using an overlay graph, produced by the Webfram generic model. The distribution on the left is the same data as in Figure 6, but plotted as a complementary cumulative distribution or exceedance function. The distribution on the right is the same SSD as in Figure 5. The degree of overlap between the two curves gives a visual impression of the level of risk. 100% 90% % of exposures 80% 70% 95% Conf. Interval Median 60% 50% 40% 30% 20% 10% 0% 0% 10% 20% 30% 40% 50% 60% 70% 80% 90%100% % of species exceeding toxicity endpoint Figure 8. Example of an exceedance risk distribution or exceedance profile plot. This is one of three types of joint probability curve (JPC) that can be produced with the Webfram generic model. The arrows show how to estimate the proportion of exposures causing more than a given level of effect; in this case,10% or more species will exceed their toxicity endpoint in 33% of exposures (95% confidence interval 5% - 96%). Clear communication of assessment outputs is one of the key requirements for successful introduction and acceptance of probabilistic methods. To assist in meeting this requirement, a series of investigations was conducted in the EUFRAM project (www.eufram.com) to explore the effectiveness of a range of tabular and graphical formats. Lessons from these studies were used to design and refine the output formats used in the Webfram project, including the graphical formats illustrated in Figures 5-8. To maximise the effectiveness of the graphical displays, the software includes extensive options for user control of the graphical displays, including the following: Plotting distributions in cumulative or complementary cumulative (exceedance) form Option to include or exclude confidence intervals, and to vary confidence level (e.g. 90%, 95% etc) Option to alter minimum or maximum value on each axis, and to plot in natural or logarithmic scale Option to show or hide data points, and labels for points (e.g. species names on SSDs) Control on colours used for points and curves Default settings on all options for quick use Option to download finished graphic in emf, png or fig formats SID 5 (Rev. 3/06) Page 18 of 21 The website contains nearly all of the features identified as potentially desirable in the original project proposal, as shown below (with comments in italics): User interface Home page Information pages, help pages, links to relevant websites (material provided by module projects) Scenario selection screens Data entry screens Output/report option selection screens Option to do model calculations online or in batch Results screens including a range of graphical and tabular outputs, suitable for use in reports/documents, with option to print/download/save online Option to save data/scenarios for later re-use Consistent look and feel for all modules Default options and data for every module (so user can just run or modify a standard example) Information screens on the standard examples (generated by case studies in module projects) List-server to announce changes/additions to registered users (could be added easily if Webfram is endorsed for regulatory use) Option to limit functionality available to different types of users (decided in consultation with PSD to make same functions available to all users) Model inputs Option to vary analysis options (e.g. number of iterations) – but decide during project how much to limit these choices Default distributions for all parameters (from module project case studies) Option to specify own distribution (type and parameters) for each parameter (decided in consultation with PSD to allow changes only for selected parameters, where users are likely to have more relevant data) Option to use default or user’s raw data for bootstrapping? Uncertainty analysis Type(s) of uncertainty propagation to be decided during project and likely to be limited to one or two, but may include: first-order error analysis, 1D and 2D Monte Carlo, Bayesian updating, P-bounds, interval analysis. (2D Monte Carlo selected, see objective 1 in this report). Need to be able to do uncertainty propagation within the software generated by the project, or by running existing packages using macros. Sensitivity analysis (methods to be decided during project) (done for specific models with example datasets in sister projects PS2302 and PS2303) Need to consider what can be done online vs. in batch mode Objective 4. Coordination with sister projects A Committee was established at the start of the work, to coordinate interactions between this project and the other Webfram projects. The Committee comprised: Project manager: Helen Thompson (CSL) (chair) Consortium members: Colin Brown (Cranfield University, subsequently York University), Theo Traas (RIVM), Andy Hart (CSL), Jim Siegmund (Cadmus Group, subsequently freelance consultant) Technical adviser on web software aspects: Matthew Atkinson (CSL) Representatives invited from all the key module projects (Brown and Hart, above, plus Geoff Frampton of Southampton University), and the project on acceptability of effects (Mark Crane, Crane Consultants, subsequently Watts and Crane Associates). Additional members nominated by DEFRA and PSD. The Committee met at the start of the project and subsequently approximately annually, in conjunction with the annual meetings with stakeholders. The Coordinating Committee held four annual stakeholder meetings at CSL (near York). Invitations were sent to a range of invitees including members of the Environmental Panel of the Advisory Committee on Pesticides, and the pesticides industry. Each meeting included presentations on the progress of all the Webfram projects and extensive discussion sessions, and the final meeting also included hands-on practical sessions and break-out groups to provide feedback on the prototype software. The meetings generated a lot of valuable feedback and comment which was very useful in refining the outputs of the projects. SID 5 (Rev. 3/06) Page 19 of 21 Action resulting from this research During the course of the project, the ongoing work and prototype models were presented and demonstrated at a series of events including scientific conferences (SETAC, BCPC, Fresenius, Agchem Forum), workshops (EUFRAM), and annual meetings with invited stakeholders. The models and supporting documentation were also subjected to peer review by independent experts from academia and industry. The feedback from all these activities was used to refine and improve the final versions of both models and software. The final models web-enabled by the project are freely available for use online at www.webfram.com. They are being evaluated by the UK authorities to decide on whether and how they should be used in regulatory risk assessments. Possible future research Although this project has made substantial advances, there are many areas in which further work could be considered. These include: Transfer of the software to a new host server Enhancements and extensions to the existing models Additional models for different scenarios and/or different taxonomic groups Refinements to the user interface Acknowledgements The authors are very grateful to the UK Pesticides Safety Directorate for funding; to Jim Siegmund (formerly of The Cadmus Group Inc., now a freelance consultant at [email protected]) for design and implementation of the Webfram website and programming of some of the models; to Helen Thompson, Colin Brown and other members of the project steering group; to Mark Clook of PSD for advice and feedback; and to the many other individuals who have provided valuable feedback and peer review on the models and software developed by the project. References Cullen AC and Frey HC, 1999. Probabilistic techniques in exposure assessment. A handbook for dealing with variability and uncertainty in models and inputs. Plenum Press, NY. EFSA 2006. Guidance of the Scientific Committee on a request from EFSA related to Uncertainties in Dietary Exposure Assessment. The EFSA Journal, 438, 1-54. European Commission, 2002a. Guidance document on aquatic ecotoxicology in the context of Directive 91/414/EEC. Sanco/3268/2001 rev.4 (final), 17 October 2002, Brussels. European Commission, 2002b. Guidance document on risk assessment for birds and mammals under Council Directive 91/414/EEC. Sanco/4145/2000, 25 September 2002, Brussels. European Commission, 2003. Risk assessment of food borne bacterial pathogens: Quantitative methodology relevant for human exposure assessment. Appendix 3 in: The future of risk assessment in the European Union. The second report on the harmonisation of risk assessment procedures. Scientific Steering Committee. European Commission, Brussels, Belgium. Ferson, S. 2002. RAMAS Risk Calc 4.0 Software: Risk Assessment with Uncertain Numbers. Lewis Publishers, Boca Raton, Florida. Mokhtari A, Frey HC. 2005. Recommended practice regarding selection of sensitivity analysis methods applied to microbial food safety process risk models. Human and Ecological Risk Assessment 11: 591-605. Morgan MG and Henrion M, 1990. Uncertainty. Cambridge University Press. Posthuma L, Suter GW III, Traas TP. 2002. Species sensitivity distributions in ecotoxicology. Lewis Publishers, Boca Raton, FL. Regan HM, Hope BK, Ferson S. 2002. Analysis and portrayal of uncertainty in a food-web exposure model. Human and Ecological Risk Assessment, 8: 1757-1777. Saltelli A, Chan K, EM Scott. 2000. Sensitivity analysis. John Wiley & Sons Ltd, Chichester. SID 5 (Rev. 3/06) Page 20 of 21 US EPA. 1997. Guiding principles for Monte Carlo analysis. US Environmental Protection Agency, Risk Assessment Forum, Washington DC. EPA Document No. EPA/630/R-97/001, March 1997. Available at http://epa.gov/osa/spc/htm/probpol.htm. US EPA. 2001. Risk assessment guidance for Superfund. Volume III – Part A, Process for conducting probabilistic risk assessment. US Environmental Protection Agency, Office of Emergency and Remedial Response, Washington DC. Available at www.epa.gov. Vose, D. 2000. Risk analysis: a quantitative guide. 2 nd edition. John Wiley & Sons Ltd. Zheng JY, Frey HC, 2005. Quantitative analysis of variability and uncertainty with known measurement error: Methodology and case study. Risk Analysis 25: 663-675. References to published material 9. This section should be used to record links (hypertext links where possible) or references to other published material generated by, or relating to this project. The primary output of this project is the modelling software and associated explanatory and supporting information which are freely available for use online at www.webfram.com. The software includes detailed instructions and extensive help screens including more detailed information on the methods, data and assumptions used in the models. SID 5 (Rev. 3/06) Page 21 of 21
© Copyright 2026 Paperzz