A Strategy for Prioritising Non-Response Follow

Papers presented at the ICES-III, June 18-21, 2007, Montreal, Quebec, Canada
A Strategy for Prioritising Non-Response Follow-Up
to Reduce Costs Without Reducing Output Quality
Gareth James
UK Office for National Statistics
Abstract
•
In Office for National Statistics (ONS) business
surveys, a considerable amount of resource is spent
following-up sampled businesses that either delay their
response or do not respond at all. There are pressures
to economise by reducing the amount of follow-up but there are unavoidable quality implications.
•
ONS is investigating methods for targeting nonresponders. Current practice in ONS is to aim for an
overall response rate by survey. We aim to develop a
mechanism that will allow us to prioritise follow-up in
terms of pre-defined quality criteria. Whatever strategy
is employed will require a better understanding of, and
treatment for, non-response; performance measurement
will depend on the quality of imputation for nonresponse. Other issues are: treatment of different types
of non-responders - ie new, large, and chronic nonresponders; and how many small businesses need
following up to avoid bias.
Any inherent differences between the respondents
and non-respondents can lead to non-response bias
being introduced to the estimates. The degree of
this depends on whether values are missing at
random (MAR), missing completely at random
(MCAR) or not missing at random (NMAR) as
described by Little and Rubin (1987).
Failure to address non-response from businesses
can lead to the ingraining of bad habits when the
same businesses are selected on future occasions
for that, or other, surveys. In other words,
allowing non-response in one wave of the survey
may make non-response more likely in later
waves.
ONS conducts many sample surveys of business, and
most are statutory (ie businesses that receive
questionnaires have a legal obligation to comply by
providing requested information). However, nonresponse still exists. Continual improvements are made
to ONS surveys, which should result improved
response, and effort is spent on re-contacting
businesses that haven’t responded to try to gain their
co-operation.
Keywords: Non-response, Bias, Business Survey,
Scoring, Follow-up
1. Introduction
The initiatives which are used to develop the survey
and, in turn, enhance response are often focussed on
questionnaire improvements and the sample design.
Examples of changes to the sample design include
altering rotation patterns so that businesses are retained
in the sample for fewer periods, better control of
overlap between samples of different surveys, and
making changes to the sample allocation itself. The
work conducted on data collection instruments (paper
questionnaires are used for most ONS businesses
surveys) can lead to less burden on businesses and
makes a response more likely. In particular, effort is
made to improve businesses’ understanding of
questions and definitions; improve the layout and
routing of questionnaires; and reduce the number of
questions being asked. Additionally, in a few surveys,
some businesses are offered ways to respond other
than completing a paper questionnaire; this includes
completion of a spreadsheet (for some complex
surveys or businesses) or telephone data entry.
Non-response is an issue which affects many, if not
most, sample surveys. Two different forms are often
defined: unit non-response (when a sampled statistical
unit fails to respond at all to the survey), and item nonresponse (when the unit responds to some but not all of
the questions asked in the survey). Either way, nonresponse means that a full sample is not achieved for
the question under consideration.
There are a number of consequences that can arise
from non-response, many of these particularly relevant
to the periodic, repeating business surveys of the
Office for National Statistics (ONS), which comprise
panel and rotating-sample elements in their stratified
designs.
• The effective (returned) sample size is smaller in
the presence of non-response, leading to less
precise estimates. This can be particularly
pronounced in sampling strata that have small
allocated despatch sample sizes.
On a more routine basis, non-responding businesses
are re-contacted during the course of the survey
1408
Papers presented at the ICES-III, June 18-21, 2007, Montreal, Quebec, Canada
processing cycle. It is only through action such as this,
that response rates can be improved once the
questionnaires have been despatched. Further details
on the current strategy for prioritising responsechasing are described in Section 2.1.
register turnover (that being the auxiliary variable
‘turnover’ held on the sampling frame (the Interdepartment Business Register, or IDBR)).
This rest of this paper describes recent work (it is still
in progress) at ONS to better understand non-response
in business surveys and the impact that it has on
estimates. The overall aim is to develop a new strategy
for response-chasing that will be more efficient than
the current one: it should deliver better quality results
for the same inputs. To do this, information gathered
about non-response will be included in a new method
of prioritising non-responding businesses for follow-up
through the development of a scoring system. This
work is not yet complete, but the initial results are
presented here.
A previous analysis of responses to businesses surveys
had revealed a number of patterns which were used in
designing a intensive follow-up exercise (Section 3)
and will also be used in the specification of a scoring
system to prioritise response-chasing.
2.2 Analysis of response-rates
The main results from the analysis were as follows.
• Particular industries were identified across a range
of surveys and time periods as suffering from
lower response-rates. These industries included
Catering and Hotels, and the evidence was
supported by experience from staff who work on
the response-chasing teams. Possible reasons for
lower response include establishments that are not
open for custom during the day, which makes
contacting them more difficult; missing or
incorrect contact details (an issue which is more
prevalent amongst smaller businesses); and
possible language barriers or communication
problems.
• Bigger businesses (in terms of employment)
tended to take longer to respond to ONS surveys.
Conversely, smaller businesses tended to return
questionnaires earlier, which may be reflect the
greater ease with which small businesses can
provide data. However, when reviewed later in the
survey processing cycle, it was apparent that
bigger businesses had the higher response rate,
although that is somewhat due to the current
response-chasing policy of targeting bigger
businesses before smaller ones when responsechasing.
• The relative ordering of industries by responserate remained fairly stable throughout the survey
processing cycle. In other words, those industries
which had relatively high-response rates at the
early stages of the processing cycle, tended to also
have the relatively high rates towards the end of
the period.
2. Current Response-chasing Practice in ONS
Business Surveys
2.1 Overview of current practice at ONS
In most ONS business surveys, reminder letters are
sent to all non-responding businesses, some of which
are then followed-up by telephone calls. The number
of standard reminder letters varies between surveys,
and in some cases the number sent has been cut in
recent years to help meet budget constraints.
Additional action is sometimes taken in the case of
particular non-respondents (especially those that have
persistently failed to respond); further reminder letters
are sent to the chief executive of the business, and
legal enforcement action could be taken as a last resort.
Response rate targets are used by most ONS business
surveys. These are essentially the ratio of return-todespatch sampled numbers of businesses, but adjusted
for any out-of-scope units. However, the targets
usually don’t take account of the quality or usability of
the information returned. Targets are often set overall
for the survey, although in practice a balance is struck
by trying to achieve ‘reasonable’ response rates in all
of the major output groups (eg if the overall target for
the survey is 80% response, then a similar target would
be set for each main output within the survey).
3. Intensive Follow-Up Study
A more recent development has been the introduction
of weighted response rate targets, where the weights
are derived from auxiliary information held about each
business. An example is the Monthly Inquiry into the
Distribution and Services Sector (MIDSS), which
formerly had a response rate target of 80% in terms of
counts of businesses, but has recently started a trial of
using joint targets of 75% by count and 80% by
3.1 Introduction to the study
To learn more about reasons for non-response and to
provide data that could be used to assess whether nonresponse bias might be present in survey estimates, a
small-scale Intensive Follow-Up (IFU) study of
MIDSS turnover data was run during winter 2006/7. In
the exercise, selected businesses that had failed to
1409
Papers presented at the ICES-III, June 18-21, 2007, Montreal, Quebec, Canada
respond were targeted for further action in an attempt
to gain responses that wouldn’t have otherwise been
collected by the survey. Those businesses were
subjected to several more telephone calls to try to elicit
a response. Once additional responses had been
obtained, the survey estimates could be re-worked and
the impact of the extra responses measured, giving an
indication of non-response bias.
3.2 Results of the IFU
3.2.1 Background and summary information
MIDSS is a monthly survey of businesses and collects
only a few variables (often just one, total turnover). At
the time of the IFU, the total monthly sample size was
approximately 30,000. Once each quarter (reference
months March, June, September and December), an
additional question about employment is asked to a
sub-sample of 20,000 businesses. MIDSS is stratified
by industry (according to groups of Standard Industrial
Classification (2003) codes and into employment size
bands, according to register employment). Ratio
estimation is used for monthly total turnover. MIDSS
was selected for the IFU due to its relatively simple
nature and ONS-standard estimation methods.
So as not to interfere with the usual processes of the
survey operation, an additional team of staff was
employed to conduct the IFU telephone calls. To
distinguish more easily those returns that would have
come in anyway, from those that came only as a result
of the IFU, the exercise took place after the closedown of the survey processing cycle for the
compilation of first estimates.
Only part of the MIDSS sample was chosen for the
IFU. Particular industries were targeted, and different
industries were included in the exercises in each of the
three months in which the IFU ran. The choice of
industries was based upon an evaluation of the likely
effect of non-response; all industries in MIDSS were
first stratified into three groups based upon the
variability of returned data and typical response rates.
Table 1 gives summary information about the IFU and
the questionnaires which were returned.
Survey
reference
month
Number of
nonresponding
businesses
selected for
IFU
IFU
businesses
successfully
contacted
(and as %
of selected)
IFU
businesses
which
returned
data (and as
% of
contacted)
Oct 2006
856
590 (69%)
425 (72%)
Nov 2006
783
557 (71%)
416 (75%)
Dec 2006
792
550 (69%)
356 (65%)
All
2431
1697 (70%) 1197 (71%)
Table 1: IFU summary information, relating to MIDSS
Stratum 1: High variability (of data)
Stratum 2: Low variability and low response-rate
Stratum 3: Low variability and high response-rate
Only a small number of industries fell in Stratum 1,
and all of these were selected as any non-response
could have a big impact on results; about half the
industries in Stratum 2 were selected and only a few
from Stratum 3. In Strata 2 and 3 the industries chosen
for the IFU were selected at random. The total number
of industries selected meant that it was possible to
schedule follow-up calls to all the non-responding
businesses within the industry, and a schedule for
calling was planned with reference to information
learned from a similar exercise on a different survey at
ONS a few years ago.
It was not possible to contact all business selected for
the IFU; about 70 per cent of business were
successfully contacted. Though a breakdown of
reasons for failing to achieve 100 per cent contact is
not available, they can be attributed to some businesses
having ‘died’ but still remaining on the register; some
having out-of-date or otherwise incorrect contact
details; or various other reasons.
The response rate from those businesses where a
contact was made varied over the months between 65
and 75 per cent. It was anticipated that more returns
would be achieved, but the relatively low response rate
shows how difficult it can be to obtain responses from
a hard-core element of non-responders. The somewhatlow response rate means that the hoped-for analysis to
determine non-response bias is not possible as that
would require 100 per cent (or close) response;
therefore only an analysis of the relative non-response
bias is possible.
It was planned for businesses to receive up to five
telephone calls as part of the IFU (therefore these calls
were in addition to any they had received as part of the
usual survey process). Successive calls were timed to
be a few days apart. For each successful contact, the
ONS staff member would record the length of call and
the nature of the call (whether positive: the business
contact was courteous, helpful, etc.; neutral; or
negative: aggressive, irritated, unhelpful). In addition,
on the first successful contact, the business would be
asked the reason for its failing to respond originally to
the survey.
1410
Papers presented at the ICES-III, June 18-21, 2007, Montreal, Quebec, Canada
3.2.2 Reasons given for failing to respond
had been completed – this will be referred to as the
‘final’ data set. In the period between these times,
additional, actual returns had been added to the data
file, replacing values which had been imputed
previously. In turn, the imputations for businesses still
outstanding may also have changed (since the values
imputed for non-respondents are based on the set of
returns of responding businesses). It is also possible for
data already received to be amended if errors are
detected or other information becomes available.
Although all businesses contacted in the IFU were
asked the reason for them failing to respond originally,
not all gave one. Table 2 gives a breakdown of the
reasons where they were given and the corresponding
success rate in gaining a survey response:
Reason for not
responding
Frequency
% of businesses
that then returned
data
85%
Figure 1 shows the percentage difference in industry
estimates between first and final data sets. Industries
are defined here as those at the level of sampling, and
only industries not included in the IFU are in Figure 1;
data from all three months have been combined. For
the non-IFU industries, no additional calls were made
after the usual response-chasing had been completed,
so any additional responses received were not as a
result of IFU response-chasing action, and this
distribution provides a control group for comparison
with industries included in the IFU.
Forgot/ missed
573
date
Too busy to
343
77%
respond
Actively decided
56
39%
not to respond
Table 2: Reasons given for not responding to MIDSS
Of most interest is the relatively low proportion of
questionnaires returned from those businesses which
had consciously decided not to return the
questionnaires originally. However, to have turnedaround over one-third of such businesses could also be
considered something of a success.
F
r
e
q
u
e
n
c
y
3.2.3 Other information recorded
The attitude of the business contact was found to be
generally positive (about 80 per cent of calls); the next
biggest group (about 15 per cent) were neutral and the
other five per cent were negative. These proportions
remained fairly constant with respect to each round of
calls (first call, second call, etc.) whereas one might
have expected businesses to become more irritated
after receiving several response-chasing calls.
100
50
0
- 32
- 24
- 16
-8
0
8
16
pc_change_est
Figure 1: Difference in first and final data set estimates for
the non-IFU industries. Summary measures of the
distribution are: mean = 0.44%, median = 0.48%, standard
deviation = 4.16%.
The duration of calls also seemed not to vary much
with the call-round number. Overall, most calls (about
85 per cent) lasted one to two minutes, about 15 per
cent lasted between three and five minutes (duration
was usually recorded to the nearest whole minute), and
five per cent lasted six minutes or more.
The percentage change in estimates due to later
responses of the 36 industries included in the IFU
study over the three months is shown in Figure 2.
F
r
e
q
u
e
n
c
y
3.2.4 Estimation of non-response bias
Two approaches have been used to assess whether any
significant non-response bias might be present in the
estimates of total turnover from MIDSS.
10
5
0
- 32
- 24
- 16
-8
0
8
16
pc_change_est
The first approach compared the estimates of turnover
at two different points in time: the first data set was
that frozen at the time of the close-down of the survey
for compilation of first estimates (ie after the usual
response-chasing has taken place but before the IFU
commenced), and the second some time after the IFU
Figure 2: Difference in first and final data set estimates for
the IFU industries. One outlier (-198%) is not shown – it was
subsequently found to be caused by a data error - with this
point removed, the summary measures of the distribution are:
mean = 0.90%, median = 0.80%, standard deviation = 5.00%.
1411
Papers presented at the ICES-III, June 18-21, 2007, Montreal, Quebec, Canada
There is no significant difference between the
distributions of changes in estimates for the non-IFU
industries and the IFU industries. However, there are
some observations (industries) in the tails of the
distributions (for both IFU and non-IFU industries)
that are notable for the absolute size of the revision,
suggesting that getting a higher response rate, or
perhaps getting a response from particular contributors,
is important for estimates in particular industries.
over-burdening the smallest businesses. However, the
process of selecting ‘key’ businesses is not necessarily
a scientific one: it is left to the survey manager to
identify the key businesses using whatever analysis she
or he wishes, together with expert knowledge and
experience of the survey. This means that the process
is not particularly transparent, varies from survey-tosurvey and can lead to an ever-expanding number of
key businesses, as the markers used to denoted key
businesses tend not to be removed as frequently as they
are applied.
The biggest change in the estimate of total turnover, a
198% decrease, came from an industry in the IFU.
Surprisingly, it was a Stratum 3 (low variability, high
response) industry. Upon investigation, however, it
was discovered that almost all of the difference could
be attributed to error in the data being detected and
corrected; the large change was not caused by
additional responses being received. Similar examples
have been found in other observations too.
The overall aim of scoring businesses is two-fold. The
first is to systematise the process of response-chasing,
thus removing the need for the survey manager to
identify ‘key’ businesses. Even if those businesses
chosen by the survey manager really are the most
important, using scoring would give a more defensible,
transparent system. The second aim is to select
businesses in a more scientific and efficient way, so
that only those businesses whose response is important
(and there are many dimensions to this) are targeted.
The second approach taken to assess possible nonresponse bias was an analysis of the auxiliary
information held on the IDBR of responding and nonresponding businesses.
Once scores are derived for each business (whether for
all units at the time of selection, or just the nonrespondents) they can be used for a number of
purposes. The first is to prioritise the non-responding
businesses (possibly into groups or individually) for
follow-up, and a second is to categorise businesses for
different follow-up treatments, eg phone calls, e-mail
reminders, chief executive letters.
Analysis of the differences in auxiliary information of
those industries included in the IFU was conducted
within each industry-by-size-band sampling stratum
(so that the effect of chasing the biggest businesses
first is removed). The analysis revealed some
significant differences in the auxiliary information of
respondents and non-respondents in some strata, but
there was no discernible pattern in terms of the timing
(ie whether using the first or final data sets), the
variable analysed (employment or turnover) or the
employment size band across industries. The main
notable observation was that, in almost every case
where the difference was significant, the average value
of the auxiliary variables of the respondents was larger
than that of the non-respondents.
3.3.2 Non-response bias in ONS business surveys
This section is written with reference to MIDSS, but
the principles also apply to other ONS business
surveys as many processes are common.
In MIDSS, where a response is not received from a
business selected for the sample, a value is imputed in
its place. Ratio imputation is used to create a value: the
average growth from the previous period to the current
period of businesses in the same stratum that did
respond is applied to the previous-period survey value
of each non-responding business. In cases where the
business is new to the sample, a value is constructed
based upon its auxiliary information held on the IDBR.
The approach of imputing a value in each case of nonresponse (rather than re-weighting) is used as it is
thought to give more accurate estimates with the
auxiliary information available. Use of this approach
means that any error resulting from non-response
manifests itself as imputation error (ie the value
imputed being different from the value that would have
been returned if the business had responded).
3.3 Scoring methods to prioritise response-chasing
3.3.1 Introduction to scoring
The current response-chasing strategy for most ONS
business surveys is to first send written reminders to
businesses that have failed to respond to the survey.
After that, some of the non-responders are telephoned.
The priority order for doing so is first to target
businesses identified as ‘key’, and then to chase other
non-responding businesses by descending order of size
of register employment. This aim of this procedure is
to get responses from the most ‘important’ businesses
(which assumes that, after ‘keys’, the biggest
businesses are the most important ones) while not
1412
Papers presented at the ICES-III, June 18-21, 2007, Montreal, Quebec, Canada
In recent work at ONS, trials have been conducted on
historical MIDSS data sets to test the feasibility of
scoring methods. Although the work is still in
progress, some preliminary results are available to
report here. The survey variable which has been used
in the trials is (monthly) total turnover, and will be
denoted by yi, where i represents the business i, i = 1,
…, n (here n ƒ 30,000, the full, despatched sample
size). An imputed value for business i will be denoted
ŷi.
imputation error cannot be known in advance, and
therefore must be predicted. Using historical data sets
has meant that such scores can be calculated and then
compared against reality to test the effectiveness of the
scoring system.
Once scores have been calculated there are several
ways of proceeding, but the route that ONS is likely to
take is to rank all non-responding business by their
scores and divide them into groups, chasing all
businesses within each group before moving onto the
next group. The reason for this is a practical one; it is
easier to implement it in this way when a team of
several staff members will be responsible for carrying
out the calls.
The estimate of total monthly turnover is given by
tˆy =
∑w y
i
*
i
i∈S
where:
wi represents the survey weight applied to the
response variable of business i, and is
calculated as the product of the design weight
and a calibration weight (since ratio
estimation is used);
An example of this type of method is described by
McKenzie (2000). In his paper he calculated scores
based upon imputation error in previous data returns,
ranked all businesses according to these scores, and
divided them into deciles denoted 0 (smallest scores, ie
lowest imputation error) to 9 (greatest imputation
error). Businesses which were new to the sample, or
were persistent non-responders (ie those that didn’t
have sufficient previous data returns to allow an
imputation score to be calculated) received a score of
10, and became the highest priority businesses for
response-chasing. These businesses were scored as top
priority to encourage good response-behaviour from
them on future occasions.
the response variable, yi*, is given by
⎧ yi if i ∈ S R
yi* = ⎨
⎩ yˆi if i ∈ S NR
where SR and SNR respectively denote the
responding and non-responding parts of the
sample at any given time.
The total potential imputation error is given by
wi ( yi − yˆi )
∑
A slightly modified method was tested on MIDSS at
ONS in 2001-2. However, problems were encountered
with the implementation, and in particular that too
many businesses were being scored ‘10’, and they
often weren’t the businesses that had the biggest
impact on the results. The trial was somewhat shortlived.
i∈S
ie the weighted sum of the differences between the
actual values that would be returned and the values that
would be imputed in the case of non-response. (Note
that in practice this is difficult to calculate or define
precisely as the values of the imputations depend upon
the set of other responses that have been received).
An example of a scoring method that has been
successfully implemented at Statistics Canada is
described by Daoust (2006). Weighted contributions of
businesses form the basis of the scores, from which
businesses are allocated to one of three groups for
follow-up: a group of businesses that must be
followed-up for a response, a group that won’t be
followed-up and a group which may be followed-up
depending on the resource available.
It is possible that the effect of individually large
imputation errors (which should be avoided where
possible) could be cancelled out by differing signs, and
therefore the quantity of interest for the trials was
chosen to be the total absolute potential imputation
error:
wi yi − yˆ i
∑
i∈S
3.3.3 Preliminary results from on-going work at ONS
A good scoring system will be defined as one that
would give the largest scores to the businesses which
are likely to suffer the greatest weighted imputation
error if they fail to respond. Therefore, the greatest
scores would be attributed to businesses with the
largest values of wi yi − yˆi . Of course, the actual
To test the effectiveness of different score functions,
the proposed methods have been applied to historical
MIDSS data sets. For comparison, the ‘actual’ absolute
imputation error has been calculated using true values
that were returned by businesses and by constructing
1413
Papers presented at the ICES-III, June 18-21, 2007, Montreal, Quebec, Canada
the values that would have been imputed otherwise
(some simplifying assumptions were made to avoid the
iterative nature of imputation). The businesses were
then divided into quintiles based on their actual
imputation error and the total of the absolute
imputation errors within each quintile is shown in
Table 3.
to that score. Table 4 shows, for a range of score
functions, the percentage of the actual imputation error
contained within each quintile of the prediction score
function.
Score
quintile
Score
quintile
4
3
2
1
0
Actual
% of total absolute
imputation
error score
88
8
3
1
<< 1
4
3
2
1
0
Table 3: Distribution of imputation error scores (percentages)
Imp.
error
88
8
3
1
<< 1
Weighted
Prediction
Imp.
error
t-1
73
12
10
3
2
Reg.
turn.
Reg.
emp.
Unweighted
prediction
Reg.
emp.
68
15
8
5
4
42
20
11
9
18
40
15
12
18
15
Table 4: Percentages of actual imputation error contained
within each quintile of various prediction score functions,
including register turnover and employment.
The distribution has a positive skew, one-fifth of the
sample accounts for 88 per cent of the error.
Examination of the individual scores revealed that the
majority of the error comes from just a handful of
businesses, highlighting the need to ensure responses
are gained from those particular few businesses. In
practice, this means it is likely that a threshold will be
used to mark the most important units instead of taking
an arbitrary 20 per cent (or any other fixed proportion)
of businesses.
The best score function of those shown is the weighted
imputation error from the previous month, as its
distribution is most similar to the that of the actual
error. However, there is not much difference between
the effectiveness of that function and that of weighted
register turnover. The use of the latter is appealing,
because it is available directly from the IDBR – it does
not have to be computed and it is available for all
businesses (both aspects unlike previous imputation
error). Further, and more detailed investigation into the
properties of these score functions is planned before a
choice is made.
Different score functions have been tested on their
ability to predict the likely imputation error. Functions
included in the trials have included the previous
period’s imputation error (akin to McKenzie’s
methods), weighted register turnover (similar to
Daoust’s) and register employment (similar to the
current ONS strategy of biggest-first). In general, all
the scores functions take the generic form:
It can be seen that register employment is not so good
a predictor of imputation error for the survey variable
total turnover. When looking at these results, however,
one should bear in mind that particular businesses had
been identified as ‘key’ for MIDSS. These units have
not be treated any differently in the scoring trials
reported here – if they had, then the combined use of
‘keys’ and register employment would lead to more
efficient response-chasing. However, this approach
makes the system more complicated, and it appears
that a simpler and more scientific approach than the
current strategy is available through score functions.
scorei,t = f(wi, imp erri ,(t-1), imp erri ,(t-2), …,
reg empi,t, reg turni,t, …)
where the ‘imp err’ is the imputation error calculated at
a previous period, and ‘reg emp’ and ‘reg turn’ refer to
the auxiliary values of employment and turnover held
on the IDBR (sampling frame). A combination of these
variables – average, maximum, etc. – could also be
used.
3.3.4 Future refinements to scoring
Work will continue to refine the score functions and
decide on which is best to use in terms of the variables
and the number of periods of previous data to include.
Inclusion of further refinements to the score will also
be considered, and ideas for these are discussed.
Finally, once testing has concluded using historical
data sets, live testing is planned. This may well give
further insights to the effectiveness of the system to be
In the trials using MIDSS data, a score has been
computed for each business in the sample according to
the given prediction score function. For functions
which use previous imputation error, new-to-sample
businesses or persistent non-respondents have been
assigned a score equal to the stratum average of
businesses that had responded. Once all businesses
have been assigned a score, they are ranked according
1414
Papers presented at the ICES-III, June 18-21, 2007, Montreal, Quebec, Canada
proposed, as the responses contained in the historical
data sets have all been collected under the current
response-chasing strategy.
The future direction and strategy for response-chasing
will build on these results, resulting in a scoring
system to identify the most important businesses to
improve the quality of survey estimates and outputs.
The proposed system should be simple to use and
justify, and will be more transparent and efficient than
the methods currently used. In addition, it will be
designed in conjunction with other survey processes,
such as sample design and editing tools.
A number of other issues and ideas will be considered
when developing the new response-chasing strategy
using score functions.
• Thresholds will be established to determine those
businesses that are much more important than
others.
• There is an option on whether to score the survey
as a whole, or separately within any specified subset. The effects will be tested, but scoring within
each of the main output groups seems likely.
• All businesses selected could be scored or only
those that have failed to respond.
• Additional changes to individual businesses’
scores may be included. An example would be to
increase the score of persistent non-respondents,
making them more likely to be subject to
response-chasing.
• The score function will be extended to cover
multivariate surveys; McKenzie (2000) gives
ideas on this.
• Scores will need to be updated during the survey
processing cycle to reflect changing priorities after
other returns have been received. For example, if
the three non-responding businesses in the same
stratum have the same, high-priority score, after
any two of them have returned data, getting a
response from the third might become much less
important. The frequency with which scores are
updated needs to be considered carefully.
• The system used by the response-chasing team
must be easy to use and clearly show which
businesses need to be chased and in what order. It
must also include a stopping mechanism, which
may be target-based. However, the use of targets
will be reviewed.
Acknowledgements
Many people have been involved with this project,
both in helping with the IFU study and in the analysis
of data. Particular thanks go to Sarah Green, Nicholas
Woodhill, Anthony Szary and Gary Brown of ONS,
and Pedro Silva from the University of Southampton
for their contributions to this paper and the
presentation at the conference.
References
Daoust, P., 2006, ‘Prioritizing follow-up of nonrespondents for the Canadian quarterly survey of
financial statistics for enterprises’, Conference of
European Statisticians
Little, R.J.A. and Rubin D.B., 1987, Statistical
Analysis with Missing Data, New York: John
Wiles & Sons
McKenzie, R., 2000, ‘A framework for priority contact
of non respondents’, Proceedings of the second
International Conference of Establishment
Surveys
4. Conclusions
This paper has described some of the work in progress
on measuring non-response bias in ONS business
surveys and has described some results from the initial
investigations carried out. Overall, it seems that
increasing all response-rate targets would not be an
efficient way to improve quality. Rather, carefully
targeted response-chasing would be a better approach.
It appears that getting responses from particular
businesses can impact heavily on industry estimates,
and that some industries are more prone to this than
others.
1415