Differences across modes coverage error different members of the

Mixed Modes and
Measurement Error
Gerry Nicolaas
Mixed Modes and
Measurement Error
• Funded under the ESRC Survey Design and
Measurement Initiative
• 3-year contract starting 1 Oct 2007
• Collaboration between academics and data
collection organisation
Research Team
• National Centre for Social Research (NatCen)
Gerry Nicolaas
Steven Hope
• Institute for Social & Economic Research (ISER)
Peter Lynn
Annette Jäckle
Alita Nandi
Nayantara Dutt
• Freelance Survey Methods Consultant
Pam Campanelli
Background
Modes of data collection:
• Interviewer-administered
face-to-face interview (PAPI, CAPI)
telephone interview (PAPI, CATI)
• Self-administered
postal questionnaire
self-completion questionnaires with interviewer
present (PAPI, CASI, A-CASI)
IVR and TDE in telephone surveys
web/email
Choice of mode = trade-off
Sampling
error
Coverage
error
Measurement
error
Cost
Quality
Non-response
error
Recording
error
Time
Differences across modes
• coverage error
different members of the population have a zero
chance of inclusion depending on the mode
• non-response error
differential non-response bias
• measurement error
respondents give different answers in different
modes
Increasing use of mixed modes:
• To maximise response rates (unit & item)
offer choice of mode to respondents
use alternative mode among non-respondents
• To reduce measurement error
self-completion modes for sensitive items within
a f2f or tel interview
• To reduce costs
sequentially with cheapest mode first
different waves of a longitudinal study
Mixing modes:
• Different modes to collect different data items from
the same respondents
e.g. CAPI with CASI module
data comparability not affected
data quality may be improved
• Different modes to collect the same data from
different respondents
e.g. tel follow-up among postal non-responders
e.g. panel survey with wave 1=CAPI & wave 2=CATI
potential for mode effects
For an overview of using mixed
modes in surveys:
• Edith D. de Leeuw (2005) “To Mix or Not to Mix
Data Collection Modes in Surveys”, Journal of
Official Statistics, 21(2), pp. 233–255
• One of Edith’s conclusions:
Hardly any theoretical or empirical knowledge is
available on how to design optimal
questionnaires for mixed-mode data collection.
Study
Design
Main Objective
• Practical advice on how to improve portability of
questions across modes
Which mode combinations are likely to produce
comparable responses?
Which types of questions are more susceptible
to mode effects?
Research Design
• A literature review & framework of mixed modes
develop a theoretical framework
identify gaps in evidence base and formulate
hypotheses to address gaps
• Quantitative data analysis
test hypotheses using existing datasets and new
experimental data
• Cognitive interviewing
explore how respondents process questions in
different modes
The Quantitative Data
• Existing datasets, e.g.
1999 Welsh Assembly Election Study
2005 Social Capital Survey
European Social Survey mode experiments
2006 Health Survey for England London boost
• New experimental data
follow-up surveys to BHPS & NatCen Omnibus
focus on f2f, tel and web comparisons
NatCen Omnibus Survey
• Two rounds of face-to-face data collection
Jul/Aug 2008 and Sep/Oct 2008
• Follow-up surveys after 6 months
Omnibus respondents who agreed to follow-up
and web access
Random allocation = 400 f2f, 400 tel, 400 web
• Cognitive interviews after 6 months
Purposively selected sample of 36 respondents
from follow-up surveys
British Household Panel Study
• BHPS Wave 18 (all f2f)
Sep 2008 – Dec 2008
• Follow-up surveys after 6 months
BHPS respondents who agreed to follow-up and
have web access
Random allocation = 400 tel, 400 web
No separate f2f data collection at this stage
• BHPS Wave 19 (all f2f)
Sep 2009 – Dec 2009
NatCen
Omnibus
20 BHPS
questions
F2F interview after 6 months
20 BHPS
questions
Tel interview after 6 months
20 BHPS
questions
Other
modules
Another 40
questions
Cognitive interviews
after 6 months
Another 40
questions
Web q’naire after 6 months
20 BHPS
questions
Another 40
questions
BHPS W19
(12 months after W18)
BHPS
W18
20 BHPS
questions
Other
BHPS
questions
20 BHPS
questions
Tel interview after 6 months
20 BHPS
questions
Another 40
questions
Web q’naire after 6 months
20 BHPS
questions
Another 40
questions
Other BHPS
questions
Key Features of Design
• Repeated measures
enables estimation of mode effects in measures of
change
• Random allocation to modes
• Compare ‘seasoned’ panel members with ‘fresh’
survey sample members
• Cost-efficient design to collect very rich
experimental data
Limitations
• Restricted to respondents with web access
primarily relying on randomisation within the sample
as our basis for inference
relatively broad basis for extrapolation to general
population compared to other mixed mode studies
• BHPS f2f follow-up is 12 months later rather than 6
months
overcome by comparing data from Omnibus and its f2f
follow-up after 6 months with data from BHPS and its
f2f follow-up after 12 months
Literature
Review
Aims of literature review
• A review of the evidence of differences in
measurement due to mode for:
Different types of questions
Different combinations of modes
• Identify gaps (mode pairs and question types) in
the evidence base
• Formulate hypotheses to address these gaps
The literature review
• Initially over 700 papers identified
• Currently screening papers for relevance and
summarising relevant papers
• Criteria for inclusion:
comparison of 2 or more modes
modes of survey data collection
measurement error
Classify existing evidence:
• The question
question type (e.g. attitude, behaviour, other factual)
question format (e.g closed/open, scale, # categories)
task difficulty
sensitivity of question
• The mode comparison
interviewer presence (face-to-face, telephone, none)
delivery of question (visual, aural)
response list (visual, aural)
recording of responses (oral, written)
Continued on next slide
Classify existing evidence:
• The results
Hypotheses tested
Indicators and statistical methods
Results
Synthesis of literature
• Causes of differential measurement error
e.g. interviewer presence, cognitive task
• Nature of differential measurement error
e.g. social desirability bias, survey satisficing
• Magnitude of differential measurement error
Causes of Mode Effects on Measurement (Roberts, Jäckle & Lynn; 2006)
Privacy/legitimacy
Interviewer presence:
Anonymity vs. rapport
Comprehension
Willingness to disclose?
Retrieval
Judgement
Response
Sufficient Effort?
Task difficulty
Stimulus:
Cognitive task
*
R motivation
*
R ability
Interviewer presence:
Pace, non-verbal communication, multitasking
Initial observations from lit review:
• Many experiments are not theory-driven
Focus is on descriptive comparisons of response
distributions across modes
Lack of generalisable inferences about causal
mechanisms
• Many papers provide insufficient information about the
questions and modes being tested
question type & format, sensitivity of question, task
difficulty
interviewer presence, delivery of question and
response options, recording of responses
Next steps
Next steps:
• Complete the literature review and develop
theoretical framework
• Identify gaps in the evidence base
• Design extra 40 questions for mixed modes
experiment
RC-33 Conference in Naples,
1-5 Sept 2008
• Research team is chairing a session on Mixed
Modes and Measurement Error
• Steven Hope presenting a paper on preliminary
results of literature review