ONS Methodology Working Paper Series No 4 Non-probability Survey Sampling in Official Statistics Debbie Cooper and Matt Greenaway June 2015 1. Introduction Non-probability sampling is generally avoided in official statistics, often for good reasons: a lack of selection probabilities makes inference from the sample to the population extremely challenging, quality measures such as standard errors are difficult or impossible to calculate, and so the official statistics context of a wide variety of users, who may use data for different purposes, does not fit well with non-probability methods. However, because of the ever-increasing nonresponse rates, costs associated with probability sampling, and ease of carrying out web surveys some survey researchers have shifted their attention to developing better non-probability sampling and estimation techniques. As there have been numerous developments in the domain of non-probability sampling, this paper endeavours to raise awareness amongst producers of official statistics with regards to challenges and developments relating to non-probability sampling. This paper aims to achieve four main outcomes: i. Provide a concise review of the types of non-probability samples ii. Highlight the key challenges associated with non-probability sampling iii. Increase awareness of techniques available to potentially overcome these challenges iv. Provide guidance to help inform decision-making on whether a non-probability sample is justified In order to achieve these outcomes we first identify the characteristics of non-probability sampling and discuss the growing interest in it. Following this is an overview of various types of non-probability sampling techniques. The key challenges associated with non-probability sampling are then highlighted. Next, techniques developed to overcome some of the challenges associated with non-probability sampling and estimation are discussed. This is followed by a section providing guidance on when the use of non-probability sampling is justified. Finally, a set of recommendations regarding the use of non-probability sampling in official statistics is provided. 2 2. What constitutes non-probability sampling? Non-probability sampling has two distinguishing characteristics: i. one cannot specify the probability of selection for each unit that will be included in the sample ii. it is not possible to ensure that every unit in the population has a nonzero probability of inclusion (Frankfort-Nachmias and Nachmias, 1996) In probability sampling, the ability to calculate selection probabilities allows researchers to create design weights which result in an unbiased estimator. Probability samples also allow for representativeness as each unit in the target population has a nonzero probability of selection, and allow for the estimation of sampling variability – these are crucial advantages. However, non-random nonresponse and undercoverage violate the assumptions of probability sampling, giving them a non-probability element. Various methods have been developed to deal with coverage and nonresponse issues in probability sampling. These include using multiple sampling frames, adjusting the weights for nonresponse and, if relevant, attrition based on sample characteristics and calibrating to target population totals in order to produce more representative estimates. However, the concerns about increasing nonresponse coupled with the high costs associated with traditional probability sampling methods have led some survey researchers to turn their attention to non-probability sampling. This growing interest in non-probability sampling also results from the fact that web data collection (most of which uses non-probability sampling) has become much easier to carry out. It is also much less costly than certain types of probability sampling. Sometimes non-probability sampling is used because there is no other option available to the researcher. This may be caused by the target population being a hidden population (and therefore there is no sampling frame available) or because of limited resource availability. Section 6 will provide guidance with regards to deciding whether use of a non-probability sample is justified. If this is the case, it is essential to bear in mind the challenges associated with non-probability sampling (see Section 4) and attempt to use techniques aimed at overcoming these challenges. Some of these techniques are described in Section 5. 3. Types of non-probability sampling It is extremely difficult to categorise non-probability sampling techniques because there is a lot of inconsistency in literature regarding definitions and applications of the types of nonprobability sampling methods. The blurred boundaries and different interpretations of the types of non-probability sampling should be borne in mind when interpreting the framework below which attempts to identify the main categories of non-probability sampling. 3 Given the multitude of non-probability sampling techniques available, the aim of this section is not to provide a comprehensive review of the types of non-probability sampling methods available but rather to give a flavour for the types of techniques available. This will form the basis for discussing the challenges and limitations of non-probability sampling later on as well as the methods that have been proposed to overcome some of these limitations. A review of literature revealed four common categories for classifying non-probability sampling techniques, these are described below. 3.1 Convenience/accidental sampling According to Baker et al. (2013) “Convenience sampling is a form of non-probability sampling in which the ease with which potential participants can be located or recruited is the primary consideration.” Therefore, no formal sample design is used. Types of convenience sampling techniques include: i. mall-intercept sampling – this is frequently used in market research and involves interviewers attempting to recruit passersby to participate in a survey. ii. volunteer sampling (e.g. some types of online opt-in panels) – this consists of people signing-up to participate in research studies. Volunteer sampling is usually done online whereby volunteers are put on a mailing list and receive e-mail invitations to participate in surveys. 3.2 Purposive sampling This consists of the researcher using their judgement and approaching only those people who they decide are most appropriate to participate in the study e.g. a sample of experts on a particular topic. 3.3 Sample matching This involves selecting a sample that matches a set of population characteristics of interest (rather than bringing the sample and population into alignment after carrying out the survey as is done with post-stratification). The most common type of sample matching is quota sampling (described in further detail in Section 5.1.1). 4 3.4 Chain referral methods These tend to be used for researching rare or hard-to-reach populations. They usually involve obtaining an initial set of respondents (called seeds) from the population of interest and using their links to obtain further respondents from the population of interest. Types of chain referral sampling include: i. snowball sampling - there is a lot of confusion regarding the meaning of snowball sampling. In many texts it is described as a non-probability convenience method used to access hard-to-reach populations whereby respondents from hidden populations are asked to recommend other respondents from the population of interest. However, originally, this method was developed by researchers such as Coleman (1958) and Goodman (1961) to investigate social networks rather than as a means to find participants to interview (Vogt et al., 2012). ii. respondent driven sampling (RDS) - in response to using snowball sampling as a type of convenience sample, survey researchers focused on developing chain referral methods which could be used to produce good estimates. RDS refers to this method (Heckathorn, 2011). As RDS uses a more structured approach to sampling, convenience is not the primary consideration of this type of sampling. RDS is described in further detail in Section 5.1.2. 4. Key challenges associated with non-probability sampling The two main concerns when using non-probability sampling are: i. There is a greater likelihood of selection bias. Consequently, the resulting sample may not be representative of the population ii. It is impossible to utilise unbiased estimators and associated quality measures1 (e.g. variance, standard errors and confidence intervals) These two concerns are described in further detail in Sections 4.1 and 4.2 below. 4.1 Selection bias One of the key challenges when using non-probability sampling is selection bias. Selection bias is “The error introduced when the study population does not represent the target population” (Delgado-Rodriguez and Llorca, 2004). Selection bias occurs during the recruitment and retention of participants and the most effective way of avoiding such bias is by having a well-designed study. 1 5 Some researchers have focused on developing unbiased estimators for use with RDS. However, these require a number of assumptions to be made and should be used with caution. See Section 5.2.1. Selection bias occurs in both probability and non-probability sampling. However, nonprobability sampling is more prone to selection bias. Below is a (non-exhaustive) list of causes of selection bias: i. undercoverage – this occurs when some units in the target population have a zero probability of selection thus making the sample unrepresentative of the population ii. volunteer bias – many non-probability sampling techniques rely on units volunteering to participate in a study and since volunteers may have different characteristics to those who haven’t volunteered, this may result in an unrepresentative sample iii. interviewer/researcher unconscious bias – unconscious biases may influence interviewers/researchers so that they are inclined to select participants with particular characteristics e.g. people who ‘look’ friendly or helpful or people who are more similar to themselves. This is particularly a problem with quota and purposive sampling whereby selection of participants is left to the interviewer/researcher. In all cases above, sampled individuals may differ systematically from non-sampled individuals on variables of interest thus use of the non-probability sample may result in biased estimates. Of the three causes of selection bias listed above, undercoverage may also be an issue in probability sampling. However, this is not usually as extensive/common in probability sampling as it is in non-probability sampling. 4.2 Unbiased estimators and lack of quality measures Standard practice in official statistics, and indeed in most large-scale social surveys, is to utilise probability sampling and design-based estimation, whereby a ‘design weight’ is calculated as the inverse of the selection probability. This produces the Horvitz-Thompson estimator, which is unbiased for any design where all units have a non-zero probability of selection. Frequently, additional auxiliary information is utilised to adjust these design weights, technically making the estimator ‘model-assisted’, although ‘design-based’ is often still used whenever the estimator accounts for survey design. This methodology has a number of advantages – the resulting estimator is unbiased regardless of the purpose for which it is used, and sampling variability can be estimated directly. Since design-based estimation is not suitable for most non-probability samples, these advantages will be lost if a non-probability sample is used. 6 5. Types of sampling and estimation methods developed to overcome issues associated with nonprobability sampling Recently, researchers have focused on developing methods for overcoming the challenges relating to non-probability sampling described above. The methods developed focus on the both the sample selection and estimation stages. Some of these methods are described in Sections 5.1 and 5.2 below. 5.1 Overcoming challenges at the sampling stage When using non-probability sampling, the main challenge at the sampling stage is obtaining a representative sample. Two popular non-probability sampling strategies developed to obtain a more representative sample are sample matching and respondent-driven sampling (RDS). 5.1.1 Sample matching As described in Section 3.3, the most common type of sample matching is quota sampling. In quota sampling, the interviewers are asked to interview a certain number of people (or units) with particular characteristics so that the final sample ‘mirrors’ the target population in terms of these characteristics. In order for this to be successful, good estimates of the population characteristics used for matching need to be available (e.g. the estimates could be obtained from a good quality probability sample or a census). By using quota sampling, researchers hope to achieve a more representative sample (for further details of sample matching see Rivers, 2007; Bethlehem, 2014). However, since the choice of who to interview is still in the hands of the interviewer there may still be a substantial amount of selection bias resulting from interviewers approaching certain people over others (because of the unconscious bias as described earlier). For example, Mosteller et al. (1949) suggest in their review of the 1948 United States election poll results that the unconscious bias of interviewers may have considerably affected the incorrect prediction of the results even though quota sampling was used. Another problem with quota sampling can be undercoverage. For example, a quota sample collected on the High Street will not capture people at home or in work. Consequently, quota sampling alone is not sufficient for obtaining a representative sample. In fact Rubin (1979) recommended using both sample matching and weighting in order to obtain more accurate estimates. Various estimation methods for non-probability sampling are discussed in Section 5.2. 5.1.2 Respondent- driven sampling (RDS) This type of sampling is mainly aimed at sampling hidden (or hard-to-reach) populations. It is a type of chain referral sampling that uses link-tracing to obtain respondents from the target population. It is typically used when a sampling frame is not available. 7 RDS consists of two distinct sampling phases: in the first phase a convenience sample (the seeds at Wave 0) from the target population is chosen. The rest of the sample (Waves 1 onwards) is selected by following the links from previous respondents. This method, developed by Heckathorn (1997) uses an innovative approach for recruiting participants after Wave 0 because respondents are given a fixed number of coupons to hand out to other members of the target population. People who decide to participate in the survey simply take the coupon to the survey centre. Therefore, after Wave 0, each successive wave of the sample consists of population members who are given coupons by members of the previous wave and return those coupons to the survey centre. This process is repeated several times (until the desired sample size is achieved) so that each time respondents from one wave drive the following wave (Gile and Handcock, 2010). Using coupons in this way reduces confidentiality concerns in marginalized populations. Moreover, it enables the researcher to track social networks for use in estimation. Respondents usually receive additional compensation for each successful recruitment. Respondents handing out coupons are asked to report how many coupons they have distributed. This enables the researchers to develop more accurate estimation methods (refer to Heckathorn, 2011 for a description of the various estimators available using RDS). Moreover, the numerous waves in the study reduce the dependence of the final sample on the original convenience sample (Gile and Handcock, 2010). 5.2 Overcoming challenges at the weighting and estimation stage As described in Section 4.2, it is difficult to obtain unbiased estimates and calculate traditional quality measures when using non-probability sampling. In order to overcome these difficulties during the weighting and estimation stage, researchers have developed a number of methods for use with various sampling techniques. This section will focus on a number of these methods. The aim of this section is not to provide instructions regarding how to apply the methods described, but rather, the aim is to make the reader aware of the weighting and estimation methods that exist when using non-probability sampling. This will enable readers to make better informed decisions regarding the type of non-probability sampling method that would best suit their needs. Section 4.2 outlined why non-probability sampling typically rules out the traditional ‘designbased’ approach for estimation. The main alternative is ‘model-based’ estimation, where selection probabilities are not accounted for, and the estimator is based on a (explicit or implicit) model. The precise specification of this estimator will depend on the type of sample and the type and purpose of the estimate required. Some of the methods below propose the use of model-based estimation. In general, it is important to note that a model-based estimator can produce inaccurate or misleading results if the underlying model is incorrect, which is often impossible to verify; and that there are typically a number of models or estimation methods to choose from, each of which may produce different estimates. In contrast, a design-based estimator: will not (usually) produce biased estimates, is appropriate in most situations, and is (broadly) unique 8 – there is only one Horvitz-Thompson estimator for a given design. For this reason, modelbased estimators should be treated with caution, and users should be wary that the results could always be open to dispute. 5.2.1 Available estimators and methods for calculating quality measures when using RDS i. Estimators With reference to non-probability sampling, Salganik (2006, p.i98) stated: “For many years, researchers thought it was impossible to make unbiased estimates from this type of sample. However, it was recently shown that if certain conditions are met and if the appropriate procedures are used, then the prevalence estimates from respondent-driven sampling are asymptotically unbiased.” Heckathorn (2011) provides a description of RDS and the various estimators available when using this sampling approach. Moreover, he specifies the strengths and limitations of these estimators in his paper. It is essential to bear in mind that these estimators require a number of assumptions to be made. Gile and Handcock (2010) caution users of RDS that biased estimates may be produced when these assumptions are not met. In a separate research strand using link-tracing designs Chow and Thompson (2003) proposed a Bayesian approach for estimation. They state that when prior information is available for the characteristics one wants to estimate, then their Bayesian approach should provide better estimators than when no prior information is used. When this prior information is not available, Chow and Thompson (2003) suggest conducting a sensitivity analysis. ii. Quality Measures In terms of quality measures available when using RDS, Salganik (2006) proposes a bootstrap method to construct confidence intervals around estimates produced from RDS samples. Furthermore, following the calculation of design effects for his data, he provides advice regarding the sample sizes required for RDS studies. He recommends that when using RDS, researchers should use a sample size twice as large as that required under simple random sampling. For link-tracing designs using the Bayesian approach described above, Chow and Thompson (2003) describe how, once the estimators are obtained using this approach, not much more effort is required to obtain interval estimates in order to assess the accuracy of estimators. Credibility intervals for use with the Bayesian approach are described in further detail in Section 5.2.3. RDS is not suitable for all research and is generally used for researching hidden populations. When RDS is not a suitable non-probability sampling method other estimators and quality measures are required. The use of Propensity Score Adjustments, described next, is one alternative. 9 5.2.2 Using weighting for estimation We have already outlined how design-based estimation with non-probability sampling is impossible. However, survey researchers have proposed the use of Propensity Score Adjustments (PSA) to approximate a design-based approach. PSA has been largely used and tested on web panel surveys. There are various methods for using PSA. One of these methods, outlined by Valliant and Dever (2011) involves constructing pseudo design weights and using covariates from a reference (probability) survey to adjust these weights for nonresponse. These adjusted weights are then used to construct estimators. In order to construct the pseudo design weight in the first place, Valliant and Dever (2011) propose that if a subsample from a large panel is used, then the pseudo design weights could be calculated as the inverses of the selection probabilities from the panel. For a comparison of the quality of estimators using PSA see Valliant and Dever (2011). Lee (2006) calculated the bias of estimates when using PSA (by comparing PSA weighted and unweighted estimates to the reference survey estimates) as well as the standard errors. He found that although PSA seems to reduce bias resulting from nonresponse, it seems to increase variance. Consequently, this should be borne in mind when using PSA techniques. Lee (2006) also recommends that covariates that are highly related to the study outcomes should be used in the PSA. For further detail on using PSA for weighting and estimation see Lee and Valliant (2009) and Lee (2006). 5.2.3 Additional quality measures proposed for non-probability sampling Some quality measures for estimators calculated from non-probability samples have been discussed in the previous sections. A number of other quality measures for use with nonprobability sampling have been proposed, including credibility intervals and participation rates. These are briefly described below. i. Credibility intervals seem to be gaining popularity when using online opt-in panels. They should be used when a Bayesian approach is adopted. A credibility interval is similar to a confidence interval in that it is used to provide an indication of uncertainty of estimates. In practice it tends to be calculated in exactly the same way as a confidence interval. However, the interpretation of a credibility interval is different from that of a confidence interval (Gill, 2014). Unlike the confidence interval, the credibility interval is directly related to the actual data distribution. Consequently, the interval may or may not include the estimate (e.g. the mean), depending on whether the actual data distribution is skewed. Therefore unlike the confidence interval, the credibility interval is not an interval around the mean (United States, Environmental Protection Agency). Moreover, “a Bayesian credible interval has a precise probabilistic meaning” (United States Environmental Protection Agency), 2003, p33) so that, for example, a credibility interval of 90% would be interpreted as there being a 90% probability that the true value lies within the credibility interval. For further information on interpreting credibility intervals see AAPOR (2012) and United States Environmental Protection Agency (2003). 10 ii. Unweighted probability survey response rates are calculated as: Since the total number of eligible units is generally not known in non-probability samples it is not possible to calculate response rates. Consequently, some researchers have started using the term ‘participation rates’ for non-probability samples. Participation rates (Baker et al., 2013) can be defined as Baker et al. (2013) state that it is essential for researchers to report on the quality of their estimates in order for readers to be able to use their results appropriately. Unfortunately, there is not currently a widely accepted framework for assessing the quality of estimates resulting from non-probability samples as there is for assessing the quality of estimates produced from probability samples. Therefore, Baker et al. (2013) encourage the development of new quality measures for use with non-probability sampling. They also note the importance of using different terminology for quality measures associated with nonprobability sampling in order to differentiate these from the quality measures associated with probability sampling. 6. When justified? is use of non-probability sampling In designing a study one must consider fitness for purpose. This is a well-known concept in probability sampling too as there is always some degree of compromise that needs to be achieved in terms of cost and precision (or minimisation of error). Groves (2004, p10) emphasises the difference between what he refers to as ‘modellers’ and ‘describers’. Modellers are those researchers from psychometric or econometric backgrounds who are mainly interested in relationships between variables. On the other hand, describers are researchers who are mainly concerned with describing the target population e.g. in terms of means and totals. These include producers of official statistics. Groves (2004) highlights the fact that because of their differing research aims, modellers and describers are interested in different types of errors. As a result, modellers and describers tend to use different sampling techniques. This is important in terms of fitness for purpose because there is no single correct survey method or survey sampling technique. Moreover, there is no single correct level of accuracy that should be achieved when carrying out a survey study. These considerations should be made within the context of the study being carried out bearing in mind the aims of the study. For example, modellers tend to be interested in a narrower range of variables than describers (who tend to conduct large surveys with hundreds of variables); therefore certain techniques such as PSA tend to lend themselves better to modellers’ needs. In the case of 11 PSA this is because it was found that unless the covariates included in the analysis are highly related to the study outcomes (i.e. the variables which will be used to produce estimates), the resulting estimates have similar bias and increased variance compared with estimates produced from the reference survey (Lee, 2006). It follows therefore that it is not possible to making sweeping statements regarding the utility of non-probability sampling techniques in official statistics. However, decisions as to whether to use probability or non-probability sampling boil down to what the researcher is hoping to achieve from the survey. Since in official statistics we are mainly (although by no means exclusively) ‘describers’, we need to consider the implications of using non-probability sampling where a large number of variables are collected and used to estimate a fairly wide variety of characteristics of finite populations. As discussed in previous sections, problems such as selection bias and the necessity of model-based estimation make non-probability sampling much less desirable than probability sampling in this context. In addition, the fact that a wide variety of estimates may be produced makes it extremely challenging to ensure that all estimates are fit for purpose. However, sometimes the researcher has no other feasible option – for example, when the researcher is attempting to describe characteristics of a hidden population for which there is no sampling frame. In such a case it is essential to consider the outcomes one hopes to achieve in order to design a study that will achieve the best possible outcomes despite the challenges associated with it. In particular, we advise limiting the uses to which the estimates can be put in order to ensure that all estimates are fit for purpose, although this is challenging in an official statistics context where statistics must be publicly available. At the very least, it is crucial to ensure that any quality issues stemming from the choice of sampling method are communicated clearly to users. One aim of this paper was to make researchers aware of the various non-probability sampling and estimation techniques available. These should be carefully considered if a decision to use a non-probability sample is made. For example, with hidden populations, the use of RDS may be a suitable option as much more effort has been made to develop unbiased estimators and good quality measures for RDS than for other sampling types such as convenience sampling. Whether using probability or non-probability sampling, it is the responsibility of researchers to consider carefully the most suitable option, state clearly the reasoning behind their choice of sampling and estimation techniques, and make every effort to describe clearly the quality of their resulting estimates thus ensuring compliance as far as possible with the UK Statistics Authority (2009) Code of Practice for Official Statistics. 7. Recommendations Following a review of literature on non-probability sampling, there are three main recommendations regarding the use of non-probability sampling in official statistics. These are: i. Fitness for purpose should be used to drive survey design. 12 ii. Non-probability sampling does not necessarily equate to lack of quality and the various methods available should be carefully considered in order to obtain the best quality estimates for the study at hand. iii. It is essential to be transparent regarding the choice of sampling and estimation techniques, describing the quality of resulting estimates as well as their limitations. 13 References AAPOR. 2012. Understanding a “credibility interval” and how it differs from the “margin of sampling error” in a public opinion poll, [online] Available at: https://www.aapor.org/AAPORKentico/AAPOR_Main/media/MainSiteFiles/DetailedAAPORst atementoncredibilityintervals.pdf [Accessed 7th May 2015]. Baker, R., Brick, M.J., Bates, N.A., Battaglia, M., Couper, M.P., Dever, J.A., Gile. K.J., Tourangeau, R. 2013. Report of the AAPOR Task Force on Non-Probability Sampling. [online] Available at: https://www.aapor.org/AAPORKentico/AAPOR_Main/media/MainSiteFiles/NPS_TF_Report_ Final_7_revised_FNL_6_22_13.pdf [Accessed 7th May 2015]. Bethlehem, J. 2014. Solving the nonresponse problem with sample matching? Statistics Netherlands Discussion Paper, [online] Available at: http://libweb.anglia.ac.uk/referencing/harvard.htm. [Accessed 7th May 2015]. Chow, M. and Thompson, S.K. 2003. Estimation with link-tracing sampling designs: a Bayesian approach. Survey Methodology, 29 (2) 197-205. Coleman, J. S. 1958. Relational Analysis: The Study of Social Organizations with Survey Methods. Human Organization. 17(4), pp. 28-36. Delgado-Rodriguez, M. and Llorca, J. 2004. Bias. Journal of Epidemiology & Community Health. 58, pp.635-641 Frankfort-Nachmias, C. and Nachmias, D. 1996. Research Methods in the Social Sciences. New York: Worth Publishers. Gile, K.J. and Handcock, M.S. 2010. Respondent-Driven Sampling: An Assessment of Current Methodology. Sociological Methodology. 40(1), pp. 285–327 Gill, J. 2014. Bayesian Methods: A Social and Behavioral Sciences Approach. CRC Press. Goodman, L. A. 1961. Snowball Sampling. Annals of Mathematical Statistics. 32, pp.148– 170. Groves, R.M. 2004. Survey Errors and Survey Costs. Wiley & Sons: New Jersey Heckathorn, D.D. 1997. Respondent-driven sampling: A new approach to the study of hidden populations. Sociological Problems, 44( 2), pp. 174-199. Heckathorn, D.D. 2011. Snowball versus Respondent-Driven Sampling. Sociological Methodology, 41(1), pp.355–366. Johnston, L.G. and Sabin, K. 2010. Sampling hard-to-reach populations with respondent driven sampling. Methodological Innovations Online, 5(2), pp.38-48 14 Lee, S. 2006. Propensity Score Adjustment as a Weighting Scheme for Volunteer Panel Web Surveys. Journal of Official Statistics, 22(2), pp. 329–349 Mosteller, F., Hyman, H., McCarthy, P., Marks, E. and Truman, D. 1949. The Pre-Election Polls of 1948: Report to the Committee on Analysis of Pre-election Polls and Forecasts. New York: Social Science Research Council. Rivers, D. 2007. Sampling for Web Surveys. [online] Available at: http://www.laits.utexas.edu/txp_media/html/poll/files/Rivers_matching.pdf [Accessed 7th May 2015]. Rubin, D.B.1979. Using Multivariate Matched Sampling and Regression Adjustment to Control Bias in Observational Studies. Journal of the American Statistical Association, 74(366), pp. 318-328. Available at: http://links.jstor.org/sici?sici=01621459%28197906%2974%3A366%3C318%3AUMMSAR%3E2.0.CO%3B2-P [Accessed 7th May 2015]. Salganik, M.J. 2006. Variance Estimation, Design Effects, and Sample Size Calculations for Respondent-Driven Sampling. Journal of Urban Health: Bulletin of the New York Academy of Medicine. 83(7), pp.i98-i112. Salganik, M.J. and Heckathorn, D. D. 2004. Sampling and Estimation in Hidden Populations Using Respondent-Driven Sampling. Sociological Methodology, 34, pp. 193-239. UK Statistics Authority. 2009. Code of Practice for Official Statistics. Edition 1.0. [online] Available at: http://www.statisticsauthority.gov.uk/assessment/code-of-practice/ [Accessed 7th May 2015]. United States, Environmental Protection Agency. 2003. Occurrence Estimation Methodology and Occurrence Findings Report for the Six-Year Review of Existing National Primary Drinking Water Regulations [online] Available at: http://www.epa.gov/ogwdw000/standard/review/pdfs/support_6yr_occurancemethods_final.p df [Accessed 18th June 2015]. Valliant, R. and Dever, J. A. 2011. Estimating Propensity Adjustments for Volunteer Web Surveys. Sociological Methods & Research, 40(1) pp.105–137. Vogt, P.W., Gardner, D.C., and Haeffele, L.M. 2012. When to Use What Research Design. Guilford Press. 15
© Copyright 2026 Paperzz