Giving stated preference respondents

Giving stated preference respondents “time to think”: results from four countries
Joseph Cook a
Marc Jeuland b
Brian Maskery c
Dale Whittington d,e
a
Evans School of Public Affairs, University of Washington, USA
Sanford School of Public Policy, Duke University, USA
c
International Vaccine Institute, Seoul, Korea
d
Departments of Environmental Sciences & Engineering, and City & Regional
Planning,University of North Carolina at Chapel Hill, USA
e
Manchester Business School UK
b
Abstract
Previous studies have found that contingent valuation (CV) respondents who are given overnight to
reflect on a CV scenario have 30 - 40% lower average willingness-to-pay (WTP) than respondents who
are interviewed in a single session. This “time to think” (TTT) effect could explain much of the gap
between real and hypothetical WTP observed in experimental studies. Yet giving time to think is still rare
in CV or choice experiment studies. We review the literature on increasing survey respondents’
opportunities to reflect on their answers and synthesize results from parallel TTT studies on private
vaccine demand in four countries. Across all four countries, we find robust and consistent evidence from
both raw data and multivariate models for a TTT effect: giving respondents overnight to think reduced the
probability that a respondent said he or she would buy the hypothetical vaccines. Average respondent
WTP fell approximately 40%, and household WTP (to vaccinate all members) fell 30%. Respondents
with time to think were also more certain of their answers, and a majority said they used the opportunity
to consult with their spouse or family. We conclude with a discussion of why researchers might be
hesitant to adopt the TTT methodology.
DRAFT AUGUST 2010
Giving stated preference respondents “time to think”: results from four countries
I.
Introduction
Stated preference (SP) valuation methods--contingent valuation (CV) surveys and choice
experiments (CE) -- are now widely used to study individual and household preferences in various policy
sectors (environment, transportation, health, social welfare programs) and throughout the world. Stated
preference research from its inception has been subject to the criticism that it suffers from hypothetical
and other biases. The charge is that hypothetical responses do not predict actual behavior, and that
welfare estimates arising from these studies overestimate true willingness-to-pay. There have now been
over two decades of careful empirical work that has improved the methodology (more so for contingent
valuation than for choice experiment), including the well-known NOAA panel guidelines, the inclusion of
“cheap talk” scripts, and the recent attention to the incentive compatibility of SP scenarios (Carson and
Groves 2007)1. At virtually every turn, most SP researchers tend to make research choices that will
provide the most conservative estimates of willingness-to-pay (WTP) in order to guard against charges of
inflated welfare estimates (for policy analysis, benefit-cost analysis, etc).
Hypothetical bias probably remains, however (Harrison 2006). The rise of the behavioralists in
economics and the increasing visibility of field studies may soon focus even more criticism on this bias in
existing stated preference studies. We discuss one SP design feature – “time to think” – that has not yet
been widely accepted, yet could reduce much of the gap observed in the real vs. hypothetical field studies.
Many real-world situations, either in the context of voting for referenda or private purchasing decisions,
potentially involve some degree of deliberation. Many consumers and voters will want to think carefully
about the program or referendum with regards to their budget or their annual tax bill before reaching their
decision. They may want to confer with partners, family members or friends. They may seek out
alternative sources of information. Other people may do none of these things. Yet a dominant feature of
nearly all existing in-person stated preference interviews is that respondents are given the scenario and are
asked to quickly make their decisions during a single interview. [This is not true of mail surveys, of
course, although most SP researchers would probably agree that in-person interviews remain the gold
standard for high-quality SP studies because of the ability of a well-trained interviewer to control the
information and flow of the interview]. Beginning in the 1990’s, a handful of split-sample studies gave
some respondents overnight to think about the scenario before reaching a decision. The studies found
clear evidence that respondents who had a chance to deliberate were less likely to say yes to the
1
There is also evidence that contingent valuation studies do in fact predict actual behavior with a reasonable degree
of accuracy (Griffin et al. 1995).
hypothetical scenario, lowering average willingness-to-pay in the sample population. Some respondents
who were initially given no time to think were then given the opportunity to revise their answers
overnight. In many cases, they revised their answers downwards.
Despite this, very few SP studies have incorporated a time-to-think treatment to minimize
hypothetical bias. Researchers may be wary of the logistical and financial difficulties of implementing
TTT (because it requires two interviews). Some may feel that giving time to think does not in fact
replicate real-world situations since many or most voters (in the public good context) may pay little
attention to an issue until they enter the voting booth. Another valid theoretical objection is that giving
respondents time to think introduces the possibility of strategic behavior, especially if survey participants
have opportunities to discuss their answers with each other.
We synthesize results from four parallel TTT studies conducted in India, Mozambique, Vietnam
and Bangladesh carried out from 2002 to 2006. The studies estimated private demand for improved,
“next generation” cholera and typhoid vaccines that are unavailable in these countries. Although the
research designs differed somewhat, all four studies used the same general split-sample approach to
examine the effect of time to think on both contingent valuation (India, Bangladesh, Mozambique) and
choice experiment (Vietnam) responses. Although the full results from these studies have recently
appeared in the literature (Cook et al. 2007, Lucas et al. 2007, Whittington et al. 2008 and Islam et al.
2008), the papers were not focused on the examination of the effect of TTT (with the exception of Cook
et al 2007). Furthermore, because of their public health and development focus, they may not have come
to the attention of many SP practitioners, many of whom are environmental economists. We find
consistent and robust results that a time-to-think treatment lowers average WTP on the order of 30-40%
and increases respondents’ certainty in their answers.
The next section reviews the literature on giving respondents time to think, including related
research on test-retest studies, behavioral economics, group elicitation approaches, and two recent studies
on attention and advanced awareness in the context of choice experiments. The third section describes
the research design and the four sites where we gave respondents time to think about whether they would
buy a hypothetical cholera or typhoid vaccine for themselves or their household members. The fourth
section describes our main time-to-think results. The concluding section offers reflections on the
advantages and disadvantages of giving stated preference respondents time to think in stated preference
surveys.
II.
Literature
Since differences in preferences between respondents with and without time to think could reflect
a lack of stability of preferences, we begin with a brief review of the large number of “test-retest” studies
conducted since the mid-1980’s. Test-retest studies measure the reliability of CV by measuring
willingness to pay (WTP) at two or more points in time – either from a panel of the same respondents or
from two independent samples drawn from the same population. The earliest of these studies were
panels, administering the exact same survey to the same respondents separated by a gap of 1 month
(Jones-Lee et al. 1985) or 2 weeks (Kealy et al. 1988, Kealy et al. 1990) Both studies found no statistical
evidence for differences in responses. Because of concerns that respondents simply anchored on their
previous answers, subsequent studies lengthened the interval between interviews to several months or
years. Loomis (1989,1990) surveyed the same respondents about water quality improvements in Mono
Lake 9 months after the initial survey and found that initial and follow-up WTP were significantly
correlated. In a more recent example, Brouwer et al (2008) found that 80 percent of Bangladeshi
respondents who were contacted six months after the original in-person WTP survey said they would
answer the WTP in the same way. Four different “test-retest” studies were conducted with independent
samples from a single population at two time periods (Reiling et al. 1990,Teisl et al. 1995,Carson et al.
1997,Whitehead and Hoban 1999). All four studies found that WTP was stable over the time periods
tested, ranging from several months to several years2. Many of the test-retest studies, as well as other
early contingent valuation studies, were mail surveys, so that respondents could answer at their leisure
and in effect had time to think. Among test-retest studies, only Carson et al (1997), Jones-Lee et al
(1985), and Brouwer et al (2008) have used in-person interviews.
A series of contingent valuation studies in the early 1990’s first examined the effect of giving
stated preference respondents more “time to think” about their answers to valuation questions.
Whittington et al (1992) examined WTP for public standpipes and private water connections in Anambra
State (Nigeria). Half of respondents in the study heard the CV scenario during the first visit from
interviewers but were given overnight to think about it before responding. The other half responded
during the first interview; these respondents were subsequently given the opportunity to revise their
answers overnight. Elicitation was a double-bounded (bidding game) question, analyzed using four
different approaches. The researchers found that giving respondents time to think decreased bids,
program ratings and likelihood of connection for both public taps and private connections; the results
were robust across the four modeling approaches. Average WTP fell by approximately 37% for public
taps and 32% for private taps. Informal group interviews conducted after the interview uncovered no
2
One exception is Whitehead and Hoban (1999), who found that WTP for a generalized scenario of environmental
protection was lower at t1. They also observed, however, that the percent of the underlying population with
favorable views toward government as well as the percent who ranked environmental concerns highly had decreased
over the same period.
evidence that the extra time induced people to answer strategically. Similarly, Lauria et al (1999) found
that giving respondents time to think reduced bids for improved sanitation services in the Philippines.
They found evidence of a starting point bias in the double-bounded referendum questions (higher starting
bids increased WTP), but found that giving time to think erased that bias. In contrast, Whittington et al
(1993) did not find that time to think had a robust effect on bids for sanitation services in Ghana. Time to
think only reduced WTP for sewer connections within the subset of respondents with water connections.
There has been little additional empirical investigation on the subject since the 1990’s3.
Svedsater (2007) examined hypothetical donations to a environmental program among 111 students in
London. He used a multiple-bounded discrete choice format with explicit levels of uncertainty (i.e. “I am
90% sure I would pay”, Welsh and Poe 1998), and gave one third of students one week to consider their
responses. Svedsater found that TTT reduced respondents’ uncertainty lowered WTP. No split sample
time-to-think studies have been conducted in the general population of an industrialized country,
however, though there have been three recent and parallel lines of research that echo some of the same
findings.
The first line is in experimental economics. Baik et al (1999) explored ways to narrow the gap
between experimentally observed strategies and theoretically-optimal strategies (from a game theoretic
3
Though not a split-sample time-to-think experiment, Holmes et al (1998) had tourists in Brazil complete a conjoint
survey on their WTP for Brazilian forest protection using a laptop. The laptop showed two alternatives with varying
attributes (including entrance fees and lodging costs) and had respondents mark a strength of preference scale
varying from 1 (strongly prefer Alternative A) to 5 (indifferent) to 9 (strongly prefer Alternative B). Because the
computer recorded the time it took for respondents to complete the survey, they included this response time variable
in statistical models as an interaction with all utility parameters. In this design, it is clear that time spent thinking
about the choice is confounded with the degree to which the respondent took the task seriously and devoted time to
answering questions thoughtfully (in fact, the authors call the variable “respondent involvement”). Still, the study
provides some interesting empirical evidence. First, respondents who took longer to respond had stronger
preferences (i.e. fewer “indifferent” or “weakly prefer” responses) than those who took less time. The interactions
between response time and other utility parameters were nearly all significant, though the authors made no attempt
to explain them. Finally, Krinsky-Robb draws on the mean and standard deviations of the estimated utility
parameters were used to calculate WTP as a function of response time, which revealed that WTP fluctuated wildly
for response times below 15 minutes, but began stabilizing and decreasing for response times greater than 15
minutes.
standpoint) in endogenous timing games4. One of their split sample treatments examined the effects of
giving subjects two days to work on their strategy versus 20 minutes. The authors found that giving
respondents more time to think narrowed the gap between observed strategies and the subgame perfect
strategy. In a multivariate model, it was the only statistically significant predictor of whether subjects
used subgame perfect strategies, increasing the probability by 21%. Cherry and Kroll (2003) used
experimental subjects to measure the degree of strategic voting in open primary elections. Because they
felt the task was complex and repeating tasks would lead to learning effects not seen in real-world voting
situations, the authors gave all respondents two days with the primary ballots to consider their responses.
Similarly, many field experiments, which seek to model real behavior in a semi-controlled experiment,
give participants time to think by design (see List and Lucking-Reiley 2002 for one example from a
charitable giving experiment).
A second parallel line of research is the use of groups to elicit WTP. Alternately called the
“market stall” (Macmillan et al. 2002, Lienhoop and Macmillan 2002) or “citizen jury” approach
(Alvarez-Farizo and Hanley 2006), these surveys are typically split sample experiments where half of
respondents complete a CV or choice experiment study in one interview5. The other half of respondents
meet in groups to discuss the scenarios and ask clarifying questions of a moderator. At the end of the
hour-long session they answer a valuation question individually and confidentially. They then attend
another group session one week later after they have had a chance to discuss with family, seek out more
information, consider budget constraints, etc., and answer the same valuation question. MacMillan et al
(2002) found that respondents who were interviewed individually without time to think had mean WTP 2
4
Endogenous timing games involve players that are unevenly matched – there is a “favorite” who has greater than
50% chance of winning and an “underdog”. In such games the subgame perfect equilibrium is for the underdog to
commit to an effort level before the favorite. This allows the favorite to respond efficiently to the underdog’s nowrevealed lower strength. As cited in Baik et al, these games have been applied to defense, environmental disputes,
rent seeking and even track-and-field data from the 1992 Olympics. Subjects in Baik et al played an abstract game
with real payoffs determined by how their strategies fared in a “tournament” with other players’ strategies.
5
Another similar line of studies examined the difference in the percentage of respondents who would agree to
connect to improved water and sanitation services when the elicitation approach was an individual interview versus
a community meeting (Davis and Whittington 1998, Whittington et al. 2000, Whittington et al. 1998). Unlike the
“market stall” approach, however, these studies did not have a second follow-up community meeting to assess
whether preferences had changed, and the authors were unable to match individuals’ socioeconomic data with their
responses in community meetings (i.e. voting was by secret ballot). Furthermore, unlike the other studies mentioned
above, participants in the group meetings were not drawn from a random sample, but were voluntary participants
who had heard about the meetings. Participants in the 1-2 hour long meetings had more time (and probably more
information) to consider their responses than individually-interviewed respondents, yet the authors found that their
responses were roughly the same (Davis and Whittington 1998). They also found that respondents who participated
in the group meetings expressed more uncertainty about their answers (Whittington et al. 2000).
- 4 times higher than those who participated in this “market stall” group approach. Also, 37% of group
respondents changed their answers during the intervening week. These respondents were less certain of
their responses than those interviewed individually. Alvarez-Farizo and Hanley (2006) also found that
WTP fell when respondents answered an identical set of choice experiment tasks as a group (the group
respondents in that study were actually a subset of the individually-interviewed respondents, potentially
introducing selection problems).
The third line of research is the use of a “drop-off” protocol (Subade 2007). In a study valuing
reef protection in the Philippines, Subade randomly assigned half of respondents to complete a standard
in-person CV interview. Interviewers made personal contact with the other half of respondents, but left
the survey with the respondent to complete. They made an appointment to retrieve the completed
questionnaire and answer any questions. The author found that WTP among the sample who completed
the drop-off protocol was approximately 50% lower than WTP estimates from the sample respondents
who completed the standard in-person interview. One drawback of this method is that, like a mail survey,
it requires that respondents be literate, which may be a problem in some populations in developing
countries. Like mail surveys, the drop-off protocol also does not carefully control the flow of the
information in the questionnaire, allowing respondents to jump ahead in the survey instrument in
unobserved ways. Nevertheless, the design shares the common feature of TTT designs in that
respondents have overnight to consider the responses and discuss with family. The dropoff protocol has
been successfully implemented elsewhere in Asia (Labao et al. 2008,Nabangchang 2008).
Two new working papers in the choice experiment literature are also relevant. Bateman et al
(2008) examine the effect of “advanced awareness” of the tasks in a choice experiment. The authors
examine demand for a pure public good (drinking water quality), and split their sample by showing a
subset of respondents all of the attributes, their levels, and all of the choice tasks before they were actually
asked to make their choices on each task. The remainder of the sample completed the choice tasks in the
typical sequential fashion with no advanced awareness. Note, however, that all interviews were still done
in one session, so respondents did not have the opportunity to seek out more information or discuss their
choice with family members. The researchers were primarily interested in testing the incentive
compatibility of choice experiments compared to the standard single binary-choice, referendum question
(i.e. CV). In particular, they were concerned with ordering effects and the possibility that advanced
awareness would induce strategic behavior. Interestingly, they find that advanced awareness respondents
were more likely to agree to the highest bids for both a “large” and “small” water quality improvement,
though differences in acceptance rates were mixed at lower prices. Unlike the comparison subsample,
they found that advanced awareness respondents failed a scope test; mean (and median) WTP for a small
improvement in the public good was not statistically different from WTP for the large improvement.
Advanced awareness also did not seem to reduce the noise in responses (i.e. it had no significant effect on
the error variance of the utility function). They conclude that respondents who are shown all attribute
levels and tasks in advance seem to display a “cost-averaging” decision-making framework where they
assume that the true cost of providing the public good is the average of the costs shown in all tasks. This
type of behavior would not be incentive-compatible.
Cameron and DeShazo (2010) develop a theoretical model for stated preference and revealed
preference data that acknowledges that respondents allocate scarce cognitive resources and time to
attributes and choice tasks in a rational way. They argue that economists may incorrectly infer a low
marginal utility for an attribute that a respondent cares about but chose to pay little attention to because of
time constraints or because of small variations in that attribute’s levels. SP researchers, who often go out
of their way to ensure the salience of the cost variable to respondents, may in fact be directing an
artificially high level of attention to this attribute, lowering the estimated marginal utility of income and
artificially inflating estimated WTP. They hypothesize that respondents compare the time and cognitive
costs of allocating more attention to the marginal attribute with the costs of making a suboptimal choice
due to lack of attention. They then apply an empirical model to a large choice experiment examining
WTP for changes in health risk, and find that respondents do appear to differentially allocate attention
across attributes. In our context, their model suggests that giving respondents time to think about choice
tasks may relax the attention – or “time budget constraint”– leading to a different pattern of attention
allocation across tasks than the single-interview format.
III.
Study Sites and Research Design
As part of the Diseases of the Most Impoverished (DOMI) program, researchers from the
University of North Carolina at Chapel Hill and host-country partners conducted a series of studies on
private demand for cholera and typhoid vaccines in six countries from 2002 to 2006.6 Donors and
policymakers were interested in understanding household demand for “next generation” vaccines which
had been the target of scientific research to improve their efficacy and lower their cost. Although some
respondents had experience with older, less effective versions of these vaccines, the improved “next
generation” cholera or typhoid vaccines were generally unavailable in these countries except in limited
trials designed to evaluate the vaccines’ protective effectiveness, during which they were distributed free
of charge (Jeuland et al. 2009c). Because of this unavailability, we relied on nonmarket valuation
6
The DOMI program is administered by the International Vaccine Institute (IVI) with support from the Bill and
Melinda Gates Foundation.
techniques to measure household demand (Canh et al. 2006,Lucas et al. 2007,Islam et al. 2008,Kim et al.
2008,Whittington et al. 2008). This information was then used to evaluate the economic attractiveness of
investments in these vaccines using both benefit-cost and cost-effectiveness techniques (Cook et al.
2008,Cook et al. 2009b,Jeuland et al. 2009a,Jeuland et al. 2009b). It was also used to examine crosssubsidy schemes (where adults sales cross-subsidize free vaccines for children (Lauria et al. 2009) and
optimal vaccine subsidies in the presence of a herd immunity externality (Cook et al. 2009a). We focus
here on the subset of these studies that provided stated preference respondents time to think about their
answers. These were conducted in Hue (Vietnam), Kolkata (India), Beira (Mozambique) and Matlab
(Bangladesh).
Although the study team had to tailor surveys to local conditions and attitudes (after careful focus
groups and pretesting), the structure and content of the questionnaires in all four countries were similar.
We begin by describing the research design in Kolkata (India), and then discuss how the design in other
sites differed.
India
Stated preference surveys were conducted in 2004 in two areas of Kolkata (formerly Calcutta): a
densely-crowded low income slum (Tiljala), and a neighborhood with more diverse incomes and living
conditions (Beliaghata)7. The split-sample TTT experiment was conducted only in Beliaghata (n=559).
As in all the field sites, interviews were done in person with a team of well-trained local enumerators.
Half of contingent valuation respondents completed the interview in one session (i.e. “no time to think”).
After enumerators described the hypothetical vaccine (respondents were asked about cholera or typhoid
vaccines, but not both) and delivered a standard “cheap talk” script, respondents were asked whether they
would buy it for themselves and/or their household members if the two doses of the vaccine regimen were
available at a local clinic or pharmacy at one of four randomly selected prices. “Time-to-think”
respondents were presented with the same information about the hypothetical vaccine but were not
immediately given the price. Instead, interviewers said the following:
7
More details on the Kolkata field site, the CV design and results can be found in Whittington et al. 2008.
Similarly, details on the Bangladesh, Mozambique and Vietnam studies can be found in Islam et al. 2008, Lucas et
al. 2007, and Cook et al. 2007.
“We are almost at the end of our first interview, and I want to thank you very much for your
time. I would like to return again tomorrow to ask you more questions. I will ask you whether
you would want to buy this vaccine for yourself as well as for other members of your
household if it were sold at a certain price. I would encourage you to think overnight about
how much this new vaccine is worth to you, and the range of prices you might be willing to
pay for this vaccine for yourself and for your household members. You may also want to
discuss these decisions with other members of your household.”
The interviewer also left a “reminder” card with the respondents that summarized information
which had already been discussed. For respondents who were asked about a cholera vaccine, it read:
“In thinking about whether you might want to purchase this vaccine, please keep in mind:
•
Yours and your family’s income and economic status
•
Your risk of getting cholera;
•
We still have other ways to treat cholera such as oral dehydration solution; and
•
The vaccine is 50% effective in preventing cholera for 3 years.”
During the second interview, the interviewer asked the exact same set of contingent valuation
questions as for no time-to-think respondents, including a question which asked respondents how certain
they were of their response. The interviewer also asked several time-to-think debriefing questions,
including how long respondents spend thinking about the decision, who they discussed the decision with,
and other sources of information they consulted. Both subsamples also completed a sliding scale
payment exercise using the analogy of a traffic light where “green” prices were ones that a respondent
was completely sure she would pay, “red” prices were ones that a respondent was completely sure she
would not pay, and “yellow” prices were ones over which the respondent was uncertain (see Figure 1 and
Whittington et al 2008). Table 1 summarizes the research design in Kolkata as well as the three other
sites described below.
Bangladesh
Fieldwork was conducted in Matlab, a rural area southeast of Dhaka, in the summer of 2005. The
contingent valuation survey focused only on cholera vaccines. The time-to-think research design is
nearly identical to the one used in Kolkata. The TTT subsample was also not given a specific price to
consider overnight, but rather asked to think about the value of the vaccine to them, and to discuss the
decision with other family members. The stoplight exercise was only completed for the TTT subsample.
The final sample size was 582 respondents split among six prices.
Mozambique
This study, conducted in the coastal city of Beira in the summer of 2005, also focused on cholera
vaccines and used contingent valuation. As in the other sites, half of Beira respondents completed the
survey in one interview (no time to think), though here the format of the time-to-think experiment was
somewhat different. Unlike in India and Bangladesh, the time-to-think respondents were all given one of
five specific prices during the first interview and were told to think how many household members they
would purchase for at that price. During the second interview, respondents reported how many vaccines
they would buy for the household members. Unlike the other studies, however, the respondent did not
answer a single-price, binary choice question about their demand for themselves. Instead, respondents
completed the same stoplight exercise used in Kolkata to elicit willingness to pay for vaccines for
themselves. Furthermore, half of the time-to-think respondents completed this payment card exercise
during the first interview, were given the vaccine price, and then given overnight to think. The second
subsample of time-to-think respondents completed the payment card exercise in the second interview,
after they answered the household demand question. In addition, the starting point on the payment card
exercise (i.e. whether one starts at a high price first and moves down, or starts at the lowest price first and
moves up) was randomized. A figure summarizing the research design in Beira is provided in the
supplementary appendix (Figure A1, reprinted from Lucas et al 2007); “TL” refers to the sliding scale
“traffic light” exercise. The final sample size was 991 households.
Vietnam
Fieldwork was conducted in Hue in the summers of 2002 (contingent valuation for typhoid) and
2003 (contingent valuation for cholera and choice experiment for both vaccines). All contingent
valuation interviews were conducted during one interview (i.e. no time to think); the time-to-think split
sample experiment was only conducted for choice experiment respondents. Choice experiment
respondents were asked to complete a series of six choice tasks. Each task required the respondent to
compare a hypothetical cholera vaccine and a hypothetical typhoid vaccine. The attributes of the vaccines
were their effectiveness, its duration, and its price. Respondents could choose either vaccine or neither.
Half of these choice experiment respondents were given the set of six choice tasks (in the form of
laminated cards) to examine and mark overnight (Cook et al. 2007). The other half completed the cards
in the first interview, but were given the opportunity to revise their answers overnight if they wished,
following the approach of Whittington et al (1992) in Nigeria. The final sample size was 400 households.
IV.
Results
Nonparametric results
1. Respondents’ demand for vaccines for themselves
We first present raw results for respondent demand on the effect of giving time to think. In
Kolkata and Matlab, respondents were less likely to say they would buy a cholera or typhoid vaccine for
themselves if they were given time to think (Table 2). With the exception of respondents who were
asked about cholera vaccines at the highest price in both sites, time to think reduced the percentage of
respondents who said yes by 4 – 25%.
Figure 2 uses raw responses from the sliding scale “traffic light” exercise in Kolkata to report the
percentage of respondents who were completely certain they would buy the vaccine over a range of
prices. Giving time to think clearly reduced this percentage for typhoid vaccines, though somewhat less
so for cholera vaccines. For typhoid vaccines in Kolkata, the midpoint, lower bound and upper bound
were all lower among TTT respondents (Table 3). The midpoint is 40% lower, and all differences are
statistically significant. The length of the uncertainty interval is also shorter for TTT respondents, a result
to which we return later in the paper. Although the direction of the effect is the same for cholera vaccines
in Kolkata, none of the differences are statistically significant. In Beira, the midpoint, lower bound and
upper bound were again all lower among TTT respondents who did the sliding scale exercise after the
single-price question (Table 3). The midpoint is 30% lower. For respondents who did the sliding scale
exercise during the first interview, however, only the differences in midpoint and upper bound are
statistically significant, and the midpoint is 12% lower.
Finally, Figure 3 presents raw responses from the choice experiment survey in Vietnam. Because
we used the exact same choice tasks in the NTTT and TTT subsample, we are able to directly compare
raw responses. Each data point plotted in the graph reflects the percentage of respondents who said they
would buy neither vaccine at the offered price in the task. The solid line denotes responses where the
percentage of respondents who said “neither” on a given card is exactly the same. Points lying above this
line indicate that a higher percentage of TTT respondents bought “neither” than NTTT respondents. A
higher percentage of TTT respondents opted for “neither vaccine in 16 of 18 choice tasks.
Like the results for respondent demand, our raw response data consistently show that demand for
vaccines for household members is lower when respondents are given time to think. In the three sites
where we conducted a split-sample TTT experiment with household demand, the percentage of household
members vaccinated is lower for TTT respondents at nearly every price level (Table 4).
Multivariate models
Although exact modeling specifications differed somewhat between field sites (see Cook et al.
2007,Lucas et al. 2007,Islam et al. 2008,Whittington et al. 2008 for more details), much of our approach
was similar. We analyzed respondent demand at a single vaccine price (Kolkata and Matlab) using a
multivariate probit model and analyzed information from the stoplight sliding scale exercise using OLS.
We used a random-parameters logit model for the choice experiment data in Hue. We modeled
household demand in Kolkata, Beira and Matlab with a negative binomial model, a variant of a Poisson
count model appropriate for situations where data are over-dispersed (conditional mean and variance of
data are not equal).
Our multivariate results confirm the pattern seen in the raw data. Even after controlling for a
number of covariates for vaccine demand, a dummy variable for whether a respondent was given time to
think was negative and statistically significant in all models, including probit, OLS, and mixed logit
models of respondent demand, and negative binomial models of household demand (see the
supplementary appendix for the full multivariate results). Giving respondents time to think lowered the
average number of vaccines a household said it would purchase by -0.6 vaccines at the mean price and 1.1 vaccines at the lowest price in Beira (average marginal effects with other covariates set to their sample
means). The average number of household vaccines fell by -1.3 vaccines at US$0.50 and -1.1 vaccines at
US$1.00 in Matlab.
Willingness-to-pay
We also find that average respondent willingness-to-pay was on average 39% lower in the TTT
subsamples (Table 5). Estimates from both parametric (probit and mixed logit) models and
nonparametric models (Turnbull, Kristrom, and midpoint of uncertainty range) consistently show these
reductions. The smallest difference in WTP is for cholera respondents in Kolkata using the Turnbull
estimator of mean WTP (4% less). This is largely driven by the fact that the percent of cholera
respondents who said they would buy a vaccine for themselves was very similar at the highest price in the
two subsamples (see left panel of Figure 2). The largest difference is in predictions from the mixed logit
model between cholera subsamples in Hue. In fact, median WTP is predicted to be negative for a cholera
vaccine, leading to a very large difference between the NTTT and TTT subsamples. This negative WTP
estimate is driven by a large coefficient on vaccine effectiveness in the choice experiment model
(respondents were shown vaccines with 50, 70, and 99% effectiveness), leading to a negative predicted
WTP for the 50% effective vaccine. Average household willingness to pay drops on average 29% when
respondents are given time to think (Table 5).
Certainty and Preference Errors
We find evidence that providing time to think increases the certainty that respondents have in
their responses. In Kolkata, 86% of TTT respondents felt “very certain” about their answer about buying
a cholera vaccine for themselves, compared to 68% of NTTT respondents. This difference was
statistically significant at the 1% level. There was, however, no statistical difference in certainty among
respondents who were asked about a typhoid vaccine. In Beira, 75% of the “uncertain” respondents were
in the NTTT subsample. Giving TTT also reduced the length of the uncertainty interval of respondent
WTP (elicited from the sliding scale exercise) in some subsamples in Beira and Kolkata (described
above).
The choice experiment study in Hue also found that giving TTT reduced the number of
respondents who made preference errors (violations of stability, monotonicity, or transitivity). Of the 200
respondents in NTTT subsample, 22 (11%) made some type of preference errors. Fourteen of 200 (7%)
respondents in the TTT subsample made a preference error, a statistically significant difference (onetailed t-test, p=0.05). Furthermore, when NTTT respondents were given the opportunity to revise their
answers overnight, their revised responses showed fewer preference errors and were no longer statistically
different than the TTT responses. A multivariate probit model confirmed that giving TTT reduced the
probability of making a preference error (at the 10% level). When TTT was interacted with education,
the model showed that TTT reduced preference errors in respondents with secondary school education by
35% but only 6% for respondents with a primary school education. All TTT variables and interaction
terms were highly significant in this interaction model.
Debriefing questions: time spent thinking and discussion with others
Most respondents who were given the opportunity to think about the valuation question did so. In
Vietnam, Kolkata, Matlab and Beira, only 2%, 11%, 5%, and 5% of TTT respondents respectively
reported spending no time on the task. Figure 4 shows the distribution of time spent thinking across all
four sites, with a significant fraction reporting either 30 minutes or one hour. The average time spent
across all four sites was 37 minutes, with a median of 20 minutes (Table 6).
A strong majority in all four sites reported discussing the decision with their spouse (Table 6).
Across all four sites, 69% of respondents consulted with their spouse. Fewer consulted with someone
outside the household (23% across all sites), and, in Kolkata and Hue, even fewer reported using other
information (we did not ask questions about use of other information in Beira and Matlab).
We used these debriefing questions to explore which elements of the time-to-think protocol might
be influencing responses. First, we replaced minutes spent thinking (imputing a zero for all NTTT
respondents) for the time-to-think dummy variable in the multivariate models of respondent and
household demand. We did not find a statistically significant relationship between minutes spent on the
task and responses to the single-price respondent demand question (multivariate probit)8. There was,
however, a negative and statistically significant relationship with the midpoint of the uncertainty interval
in the OLS models of respondent demand in the stoplight exercise in Kolkata (more minutes spent
thinking lowered the midpoint). We found a similar statistically-significant relationship in household
demand in Kolkata and Matlab, but not in Beira. Next, we replaced the time-to-think dummy with a
dummy variable tracking whether the respondent had discussed the question with a household member.
This also was not a statistically-significant predictor of respondent demand in the single-price models, but
it was for OLS models of the midpoint of the uncertainty interval in Kolkata. It was also negative and
statistically-significant in household models in Kolkata, Beira and Matlab. Third, we explored whether,
among TTT respondents, those who spent longer on the task reported lower respondent or household
demand. We did not find any statistically significant relationships. Similarly, the TTT respondents who
did not discuss the decision with a household member did not have different respondent or household
demand than the TTT respondents who did.
Finally, we estimated a multivariate probit model of the length of the uncertainty interval from
the stoplight exercise in Kolkata and Beira (these results are shown in the supplementary appendix).
Controlling for other factors such as education, income, the offered vaccine price, and others, we still find
that the TTT dummy is negative and statistically significant in both Kolkata and Beira. Time spent
thinking did not have a statistically-significant effect in either site. A dummy variable indicating that the
respondent discussed the decision did have a negative and statistically-significant effect in Beira but not
in Kolkata.
V.
Discussion
In summary, we find that giving respondents overnight to think about their responses has a large
and consistent effect across all four study sites. This was true for both the contingent valuation and stated
choice methodologies. Time-to-think lowered the percentage of respondents who said they would
purchase a cholera or typhoid vaccine for themselves across a range of bid prices, and lowered the
percentage of household members for whom the respondent said they would purchase a vaccine. It
decreased average willingness-to-pay by 30-40% in both respondent and household demand models. It
reduced the likelihood that a choice experiment respondent made a preference error. We find that, given
8
Results discussed in this paragraph are available from the authors on request.
the opportunity, over half of respondents consulted with their spouse about their vaccine purchase
decision, and spent 20 – 30 minutes thinking about the decision.
In most cases, time-to-think also increased the certainty that respondents had in their answers.
Our result stands in contrast to MacMillan et al (2002), who find that respondents who were given time
and who completed a group elicitation approach were less certain than respondents who completed the
survey in one session individually. Group interactions, however, could have been the driver of increased
uncertainty rather than the effect of time to think. Our finding is similar, however, to Svedsater (2007),
who found that TTT reduced respondents’ uncertainty. Comparing real vs. hypothetical (stated)
purchases of a diabetes management program, Blumenschein et al (2008) find that dropping respondents
who said they were less than “absolutely sure” from the analysis reduced nearly all of the gap between
real and hypothetical choices, though a standard cheap talk script had no significant effect. Giving
respondents time-to-think (and reducing their uncertainty) may therefore provide a way to use more of the
preference data collected while still reducing hypothetical bias.
Why have so few TTT studies been conducted? For many researchers who are aware of the TTT
protocol, the first response to this question will be cost. In-person stated preference surveys are already
expensive endeavors. Most of this expense derives from the fact that skilled interviewers must be
employed and respondents compensated for their time in completing the interview. Although there is a
substantial fixed cost to developing the survey instrument and training enumerators, adding a second
follow-up interview will certainly increase survey expenses considerably. It is also probably no
coincidence that all existing time to think studies have been conducted in developing countries where the
cost of implementing a high-quality stated preference survey is much lower because both interviewer
salaries and any household compensation are much lower than in industrialized countries. Another way
of viewing this dilemma: for the same total cost, a researcher might implement a TTT study but be forced
to reduce the sample size. The researcher might then be concerned with the decrease in statistical
precision surrounding estimation of WTP. This should be weighed against the risk (for policy analysis)
that welfare estimates from conventional single-visit survey methodologies might be 30-40% too high.
Another potential drawback with the TTT method pointed out by Carlsson (2010) is that when
respondents have the ability to consult with other household members or friends it is not clear whose
preferences are being recorded. As he notes, however, this is not necessarily undesirable if householdlevel decisions are being studied (as was the case in our vaccine studies), and in fact may be more
representative of how decisions are made in reality.
Conducting a survey with a time-to-think protocol is also logistically complicated. As described
above, the study in Kolkata interviewed households in one of the poorest slums in Kolkata called Tiljala.
Because of safety concerns (our interviewers were not comfortable interviewing households after dark,
and there was a large riot during our fieldwork), we decided to finish the survey work quickly and did not
give respondents time to think. To avoid the potential for strategic bias and for households to receive
prior (and unobserved) information about the survey from their neighbors, we also completed the sites
where respondents were given no time to think first and then completed interviews with the second
subsample. This sort of logistical problem, however, exists for all split-sample studies, and might be less
of a concern for studies in which all respondents are given time to think (though researchers may still
want to ask households if they have heard about the study and what information they have already
learned).
Another reason why time-to-think designs may still be relatively rare may be the fact that inperson interviews for valuation work in industrialized countries are now the exception rather than the
norm. The need for such a design decreases considerably for more commonly used mail surveys, which
allow respondents to reflect longer on their choices. For other survey modes in which response rates and
selection problems are a major concern, such as telephone surveys, it may be even more difficult to
contact respondents multiple times. Researchers would have difficulty in estimating population WTP if a
sizeable number of respondents completed only the first half of the interview. We did not find this type
of sample selection problem in our four TTT studies. Representative Internet panels (like those run by
Knowledge Networks) are increasingly being used in developed countries, and have the flexibility to
allow respondents time to reflect.
Acknowledgements
We would like to thank the International Vaccine Institute and its director John Clemens, along with a
number of the DOMI project staff, including Jacqueline Deen, Lorenz von Seidlein, R. Leon Ochiai, and
Andrew Nyamete. We are also thankful for the assistance of our partners in each of the four countries,
including: India’s NICED (Dr. Susmita Chatterjee, Dr. S.K. Bhattacharya, and Dr. Dipika Sur),
Vietnam’s NIHE (Dr. Do Gia Cahn), Mozambique’s Center for Environmental Hygiene and Medical
Examination (Dr. Marcelino Lucas, Dr. Avertino Barreto, Dr. Francisco Songane, Arnaldo Cumbane,
Arminda Macamuale, and Nivalda Nazaro) and Bangladesh’s ICDDR-B (Dr. Ziaul Islam). All four
country studies were funded with support from the Bill and Melinda Gates Foundation. We also thank
Sonia Akter for thoughtful comments and suggestions on an earlier draft.
References
Alvarez-Farizo, B. and N. Hanley (2006). "Improving the process of valuing non-market benefits:
combining citizens' juries with choice modeling." manuscript.
Baik, K. H., T. L. Cherry, et al. (1999). "Endogenous timing in a gaming tournament." Theory and
Decision 47: 1-21.
Bateman, I. J., R. T. Carson, et al. (2008). Choice set awareness and ordering effects in discrete choice
experiments. European Association of Environmental and Resource Economists, Gothenburg,
Sweden.
Blumenschein, K., G. C. Blomquist, et al. (2008). "Eliciting Willingness to Pay Without Bias: Evidence
from a Field Experiment*." The Economic Journal 118(525): 114-137.
Brouwer, R., S. Akter, et al. (2008). "Economic valuation of flood risk exposure and reduction in a
severely flood prone developing country." Environment and Development Economics 14: 397417.
Cameron, T. A. and J. R. DeShazo (2010). Differential attention to attributes in utility-theoretic choice
models. Manuscript version Feb 7 2010.
Canh, D. G., D. Whittington, et al. (2006). "Household demand for typhoid fever vaccines in Hue,
Vietnam." Health Policy and Planning: 241-255.
Carlsson, F. (2010). "Design of Stated Preference Surveys: Is There More to Learn from Behavioral
Economics?" Environmental and Resource Economics 46(2): 167-177.
Carson, R. and T. Groves (2007). "Incentive and informational properties of preference questions."
Environmental and Resource Economics 37(1): 1573-1502.
Carson, R. T., W. M. Hanemann, et al. (1997). "Temporal reliability of estimates from contingent
valuation." Land Economics 73(2): 151-163.
Cherry, T. L. and S. Kroll (2003). "Crashing the party: an experimental investigation of strategic voting
in primary elections." Public Choice 114: 387-420.
Cook, J., M. Jeuland, et al. (2009a). "Re-visiting socially-optimal vaccine subsidies: an empirical
application in Kolkata, India." Journal of Policy Analysis and Management 28(1): 6-28.
Cook, J., M. Jeuland, et al. (2008). "The cost-effectiveness of typhoid Vi vaccination programs:
Calculations for four urban sites in four Asian countries." Vaccine 26: 6305-6316.
Cook, J., D. Sur, et al. (2009b). "Evaluating investments in typhoid vaccines in two slums in Kolkata,
India." Journal of Health, Population and Nutrition forthcoming.
Cook, J., D. Whittington, et al. (2007). "The reliability of stated preferences for cholera and typhoid
vaccines with time to think in Hue, Vietnam." Economic Inquiry 45(1): 100-114.
Davis, J. and D. Whittington (1998). ""Participatory" research for development projects: A comparison
of the community meeting and household survey techniques." Economic Development and
Cultural Change: 73-93.
Griffin, C. C., J. Briscoe, et al. (1995). "Contingent valuation and actual behavior: predicting connections
to new water systems in the state of Kerala, India." World Bank Economic Review 9(3): 373-395.
Harrison, G. W. (2006). "Experimental evidence on alternative environmental valuation methods."
Environmental and Resource Economics 34: 125-162.
Holmes, T., K. Alger, et al. (1998). "The effect of response time on conjoint analysis estimates of
rainforest protection values." Journal of Forest Economics 4(1): 7-28.
Islam, Z., B. Maskery, et al. (2008). "Private demand for cholera vaccines in rural Matlab, Bangladesh "
Health Policy 85(2): 184-195.
Jeuland, M., J. Clemens, et al. (2009a). "Incorporating herd protection into cost-effectiveness calculations
of new-generation oral cholera vaccines: a multi-site analysis." Value in Health forthcoming.
Jeuland, M., M. Lucas, et al. (2009b). "A cost-benefit analysis of cholera vaccination programs in Beira,
Mozambique." World Bank Economic Review 23(2): 235-267.
Jeuland, M., M. Lucas, et al. (2009c). "Estimating the private benefits of vaccination against cholera in
Beira, Mozambique: A travel cost approach." Journal of Development Economics 91: 310-322.
Jones-Lee, M. W., M. Hammerton, et al. (1985). "The value of safety: Results of a national sample
survey." The Economic Journal 95(377): 49-72.
Kealy, M. J., J. F. Dovidio, et al. (1988). "Accuracy in valuation is a matter of degree." Land Economics
64(2): 158-173.
Kealy, M. J., M. Montgomery, et al. (1990). "Reliability and predictive validity of contingent values:
Does the nature of the good matter?" Journal of Environmental Economics and Management 19:
244-263.
Kim, D., D. G. Canh, et al. (2008). "Private demand for cholera vaccines in Hue, Vietnam." Value in
Health 11(1): 119-128.
Labao, R., H. Francisco, et al. (2008). "Do Colored Photographs Affect Willingness to Pay Responses for
Endangered Species Conservation?" Environmental and Resource Economics 40(2): 251-264.
Lauria, D. T., B. Maskery, et al. (2009). "An optimization model for reducing typhoid cases in developing
countries without increasing public spending " Vaccine 27: 1609-1621.
Lauria, D. T., D. Whittington, et al. (1999). Household demand for improved sanitation services: a case
study of Calamba, Phillipines. Valuing environmental preferences: theory and practice of the
contingent valuation method. K. Willis and I. Bateman. Oxford, UK, Oxford University Press:
540-584.
Lienhoop, N. and D. Macmillan (2002). "Valuing wilderness in Iceland: New approaches in contingent
valuation." Icelandic Journal of Geography 18-19(1): 29-40.
List, J. A. and D. Lucking-Reiley (2002). "The effects of seed money and refunds on charitable giving:
Experimental evidence from a university capital campaign." Journal of Political Economy 110(1):
215-233.
Loomis, J. B. (1989). "Test-retest reliability of the contingent valuation method: a comparison of general
population and visitor responses." American Journal of Agricultural Economics February: 7684.
Loomis, J. B. (1990). "Comparative reliability of the dichotomous choice and open-ended contingent
valuation techniques." Journal of Environmental Economics and Management 18: 78-85.
Lucas, M., M. Jeuland, et al. (2007). "Private demand for cholera vaccines in Beira, Mozambique."
Vaccine 25(14): 2599-2609.
Macmillan, D. C., L. Philip, et al. (2002). "Valuing the non-market benefits of wild goose conservation: a
comparison of interview and group-based approaches." Ecological Economics 43: 49-59.
Nabangchang, O. (2008). "Mobilizing resources for marine turtle conservation in Asia: A cross-country
perspective." ASEAN Economic Bulletin 25(1): 60-69.
Reiling, S. D., K. J. Boyle, et al. (1990). "Temporal reliability of contingent values." Land Economics
66(2): 128-134.
Subade, R. F. (2007). "Mechanisms to capture economic values of marine biodiversity: The case of
Tubbataha Reefs UNESCO World Heritage Site, Philippines." Marine Policy 31(2): 135-142.
Svedsater, H. (2007). "Ambivalent statements in contingent valuation studies: inclusive response formats
and giving respondents time to think." Australian Journal of Agricultural and Resource
Economics 51(1): 91-107.
Teisl, M. F., K. J. Boyle, et al. (1995). "Test-retest reliability of contingent valuation with independent
sample pretest and posttest control groups." American Journal of Agricultural Economics 77(613619).
Welsh, M. P. and G. L. Poe (1998). "Elicitation Effects in Contingent Valuation: Comparisons to a
Multiple Bounded Discrete Choice Approach." Journal of Environmental Economics and
Management 36(2): 170-185.
Whitehead, J. C. and T. J. Hoban (1999). "Testing for temporal reliability in contingent valuation with
time for changes in factors affecting demand." Land Economics 75(3): 453-465.
Whittington, D., J. Davis, et al. (1998). "Implementing a demand-driven approach to community water
supply planning: A case study of Lugazi, Uganda." Water International 23: 134-145.
Whittington, D., J. Davis, et al. (2000). "Designing a "neighborhood" deal for urban sewers: A case study
of Semarang, Indonesia." Journal of Planning Education and Research 19: 297-308.
Whittington, D., D. T. Lauria, et al. (1993). "Household demand for improved sanitation services in
Kumasi, Ghana: A contingent valuation study." Water Resources Research 29(6): 1539-1560.
Whittington, D., V. K. Smith, et al. (1992). "Giving respondents time to think in contingent valuation
studies: a developing country application." Journal of Environmental Economics and
Management 22: 205-225.
Whittington, D., D. Sur, et al. (2008). "Rethinking cholera and typhoid vaccination policies for the poor:
Private demand in Kolkata, India." World Development 37(2): 399-409.
TABLES AND FIGURES
Figure 1. Diagram explaining the stoplight
analogy to Kolkata respondents for the
sliding scale exercise
Table 1. Summary of time-to-think studies on cholera and typhoid vaccines
Hue,
Vietnam
Cholera, Typhoid
Kolkata,
India
Cholera, Typhoid
Matlab,
Bangladesh
Cholera
Beira,
Mozambique
Cholera
Stated
preference
method in split
sample
experiment
Respondent
preference
information in
split sample
experiment
Choice experiment
CV
CV
CV
Choice experiment:
two vaccine
alternatives, three
attributes (price,
efficacy and
duration)
CV, Sliding scale
payment card,
take if free or pay
anything?
CV,
take if free or pay
anything?
Sliding scale
payment card, take
if free or pay
anything?
Household
demand
information in
split sample
experiment
Given price
overnight?
None
CV
CV
CV
Yes, given choice
tasks overnight
No
No
Yes
Subsample sizes
200 NTTT,
200 TTT
140 Cholera
NTTT
140 Cholera TTT
139 Typhoid
NTTT
140 Typhoid TTT
312 NTTT
279 TTT
499 NTTT
497 TTT
(also split on 6
vaccine prices, n =
37 -58 per price
cell)
(also split on
whether sliding
scale was before or
after household
demand, starting
point of sliding
scale, and 5 vaccine
prices, n∼50 per cell
Vaccine
(6 choice tasks
completed per
respondent)
(also split on 4
vaccine prices,
n∼35 per price
cell)
Table 2. Percent of respondents who said they would purchase a vaccine for themselves, by time to
think.
US$0.15 US$0.37 US$0.74 US$1.10 US$4.5 US$9.0
Matlab NTTT
85%
76%
65%
47%
19%
8%
Matlab TTT
82%
65%
40%
33%
9%
11%
Kolkata Cholera NTTT
Kolkata Cholera TTT
US$0.22 US$0.56 US$1.11 US$11.11
89%
68%
63%
18%
82%
60%
49%
18%
Kolkata Typhoid NTTT
Kolkata Typhoid TTT
91%
88%
86%
69%
63%
52%
20%
9%
Figure 2. Respondents in Kolkata who were certain that they would purchase the vaccine for themselves,
by price and time to think (from sliding scale exercise).
Table 3. The effect of TTT on the stoplight sliding scale exercise for respondent demand (US$)a
Subsample and variable
KOLKATAb
Cholera
Midpoint
Upper bound
Lower bound
Length of range
Typhoid
Midpoint
Upper bound
Lower bound
Length of range
BEIRAc
Sliding scale after
single-price (“TL2”)
Midpoint
Upper bound
Lower bound
Length of range
NTTT
TTT
t-statistic
3.39
5.06
1.71
3.35
n= 136
2.93
4.46
1.40
3.07
134
0.85
0.74
0.81
0.43
3.57
5.21
1.92
3.29
n= 135
2.14
3.07
1.20
1.87
131
2.90***
3.14***
1.75*
2.62***
0.93
1.36
0.49
0.87
n= 239
0.65
0.97
0.32
0.64
226
4.92***
5.06***
3.90***
4.35***
0.75
1.11
0.39
0.73
238
1.81*
2.20***
0.77
2.65***
Sliding scale before
single-price (“TL1”)c
Midpoint
Upper bound
Lower bound
Length of range
n=
a
0.85
1.29
0.42
0.87
234
Notes: * Indicates significance at the 10% level, ** at the 5% level, *** at the 1% level
b
Excludes out of market respondents as well as 17 respondents who said they might be willing to pay US$111.
c
Excludes respondents without a true upper bound (green at all prices). Excludes respondents without a lower bound (green at
all prices or red at all prices). Excludes respondents with an undefined upper or lower bound.
Figure 3. Percent of respondents, with and without time to think, who chose neither vaccine (18 tasks)
(reprinted from Cook et al 2007)
Table 4. Average percent of household members who would be vaccinated at specified price, by time to
think (TTT) in Kolkata, Matlab and Beira
Matlab NTTT
Matlab TTT
US$0.15 US$0.37 US$0.74 US$1.10 US$4.5 US$9.0
76%
75%
60%
53%
16%
4%
63%
53%
31%
27%
4%
10%
Beira NTTT
Beira TTT
US$0.2 US$0.82 US$1.64 US$2.86 US$4.08
74%
53%
38%
26%
18%
60%
30%
17%
16%
12%
Kolkata Cholera NTTT
Kolkata Cholera TTT
US$0.22 US$0.56 US$1.11 US$11.11
88%
58%
57%
33%
76%
57%
43%
27%
Kolkata Typhoid NTTT
Kolkata Typhoid TTT
85%
77%
79%
61%
64%
45%
37%
10%
Table 5. Respondent and household willingness-to-pay estimates (US$) for cholera and typhoid vaccines
Respondent WTP
Kolkata Cholera NTTT
Kolkata Cholera TTT
% reduction
Kolkata Typhoid NTTT
Kolkata Typhoid TTT
% reduction
Beira Cholera NTTT
Beira Cholera TTT
% reduction
Matlab Cholera NTTTd
Matlab Cholera TTT
% reduction
Matlab Cholera NTTTf
Matlab Cholera TTT
% reduction
Hue Cholera NTTTg
Hue Cholera TTT
% reduction
Hue Typhoid NTTT
Hue Typhoid TTT
% reduction
Turnbull
lower-bound
(mean)
2.5
2.4
4%
2.8
1.6
43%
Kristrom
mid-point
(mean)
5.2
4.5
13%
5.5
3.9
29%
Multi-variate probit
(median)
5.1
3
41%
5.2
3.6
31%
b
b
b
b
b
b
1.5
1.2
23%
2.0
1.4
29%
h
h
3.4
2.2
35%
4.8
2.4
51%
h
h
2.3
1.1
53%
2.5
1.1
58%
h
h
h
h
h
h
h
h
Household WTP
Midpoint of
uncertainty range from
sliding scale (mean)
3.39
2.93
14%
3.57
2.14
40%
Mixed logit
(median)
a
a
a
a
0.93c
0.65
30%
e
e
a
a
e
e
a
a
h
h
1.92
-0.09
105%
4.36
1.74
60%
h
h
a
a
Negative binomial
count (mean)
35
27
23%
29
23
21%
13.3
7.7
42%
14.3
10
30%
12.3
9.0
27%
Notes: a No choice experiment study carried out. b Respondent demand in Beira was only elicited through the stoplight sliding scale exercise; there is no respondent binary choice CV data to
c
analyze using the Turnbull or Kristrom estimators or multivariate probit. Estimates shown are for the Beira subsample who did the sliding scale exercise after the single-price household question.
d
For the subsample who did the sliding scale in the first interview the average midpoint was $0.85 in the NTTT subsample and $0.75 in TTT subsample. For the subsample in the government service
f
e
area (see Islam et al 2008 for details) The stoplight sliding scale exercise was only done for the TTT subset in Matlab. For the subsample in the ICDDR,B service area (see Islam et al 2008 for
more details). g The effectiveness and duration of the vaccine offered to respondents in Hue varied. Results here are for the characteristics most similar to the real “next-generation” vaccine: 50%
effective for 3 years for cholera, and 70% effective for 3 years for typhoid. h Although we elicited respondent demand for cholera and typhoid vaccines in Hue (see Canh et al. 2006 and Kim et al.
2008), we conducted the split sample TTT experiment only in the choice experiment study.
0
5
Percent
10
15
20
Figure 4. Amount of time that time-to-think respondents reported spending on the task, all four sites
pooled (n=1449 )
0
50
100
Time spent thinking about tasks (mins)
Note: the rightmost bar collapses all times longer than 150 minutes.
150
Table 6. Debriefing questions for time to think respondents.
Kolkata
Time spent thinking:
Mean (SD)a 28 (37)
Mediana 15
Percent who consulted with a
household member
Percent who consulted with
someone outside the
household
Percent who used other
information
N=
a
Hue
Beira
Matlab
All sites
32(38)
20
52 (52)
30
30 (35)b
20
37 (43)
20
74%
56%
81%
64%
69%
25%
4%
36%
27%
23%
6%
2%
n/a
n/a
3%
272
400
487
279
1438
Drops 11 records where reported time spent thinking was over 5 hours, including one respondent who reported thinking about
the task for 24 hours.
b
This data based on n= 436 for this question in Beira
SUPPLEMENTAL APPENDIX
Figure A1. Time-to-think research design in Beira, Mozambique
(Reproduced from Lucas et al 2007)
Table A1. Multivariate models of respondent and household demand: Kolkata (from Whittington et al 2008) a
Household Demand:
Respondent demand:
Negative Binomial Count
Multivariate probit
Model
Respondent yes/no to
Count of household members
“would buy for
respondent would purchase for
Dependent variable
yourself?”
(including respondent)
Price (Rs)
–4.7e–3(5.4e–4)***
–3.0e–3(4.0e–4)***
Price * Cholera
1.2e–3 (6.8e–4)*
7.8e–4 (5.4e–4)
Time to think
–0.35 (0.13)***
–0.26 (0.06)***
Cholera
–0.29 (0.15)*
–0.12 (0.07)*
Monthly per capita income (000 Rs) ‘
0.17 (0.09)*
0.038 (0.01)***
Male
–0.02 (0.15)
0.04 (0.08)
0.48 (0.22)**
0.38 (0.15)**
Education Low b
Education Mid b
0.82 (0.25)***
0.53 (0.15)***
b
0.54 (0.28)*
0.39 (0.16)**
Education High
Age
–0.01 (0.01)
0.00 (0.01)
Muslim
0.18 (0.76)
–1.15 (0.45)**
No. young children
–0.22 (0.13)*
0.04 (0.08)
No .school-aged children
–0.15 (0.08)*
–0.01 (0.05)
No. adults
0.062 (0.03)*
0.11 (0.02)***
–0.019 (0.01)**
–0.01 (0.01)
Distance to health facility
Likely to get disease – Resp.
–0.05 (0.13)
0.02 (0.08)
Likely to get disease – Child
–0.07 (0.08)
Disease is serious for adults
0.03 (0.13)
0.01 (0.07)
Disease is serious for children
–0.12 (0.08)
0.33 (0.13)**
0.11 (0.06)*
Know person with disease
Never boil drinking water
–0.35 (0.13)***
–0.17 (0.06)***
–0.14 (0.15)
–0.04 (0.08)
HH member had vaccine
Constant
0.85 (0.48)*
0.72 (0.28)***
551
551
N
Pseudo-R2
0.29
0.10
2
158
283
Wald Chi-sq (p < Wald χ )
c
Alpha (std error)
n/a
0.16 (0.05)***
a
* Indicates significance at the 10% level, ** at the 5% level, *** at the 1% level. Excluding nine scenario-rejecters. Robust standard
errors in parenthesis. b Excluded category is no education c Alpha is the overdispersion parameter of the negative binomial model, and
Stata 8.0 empirically tests whether alpha is significantly different from zero. When alpha=0 there is no overdispersion in the data and
the Poisson count model is sufficient. When alpha≠0 then the negative binomial is the appropriate model.
Table A2. Multivariate models of household demand and respondent demand (from stoplight sliding scale): Beira (from Lucas et al 2007)
Table A3. Multivariate (negative binomial) models of household demand: Matlab (from Islam et al 2008)
T-statistic in parentheses; * Indicates significance at the 10% level, ** at the 5% level, *** at the 1% level
Table A4. Multivariate (probit) models of respondent demand: Matlab
Price (Tk)
Time to think
Male
Resident from ICDDRB service area
Age
Education 1-5 years
Education 6-10 years
Education >10 years
Log income per capita
Number of household members
Practice Hinduism
Serious or very serious for children
Serious or very serious for adults
Cholera likely for respondent
Cholera likely for children
Someone in the hh has had cholera
Know someone other than hh member
who has had cholera
Respondent had prior vaccine, satisfied
Respondent had prior vaccine, not
satisfied
Treat water
Constant
Observations
Model 1
-0.00412
-0.34
0.322
-0.085
-0.0209
0.119
0.264
0.296
0.38
0.0947
-0.209
-1.798
583
(0.000501)***
(0.155)**
(0.111)***
-0.188
(0.00715)***
-0.163
-0.216
-0.208
(0.105)***
(0.0379)**
-0.192
(0.812)**
Model 2
-0.00429
-0.42
0.304
-0.134
-0.0208
0.113
0.205
0.283
0.425
0.107
-0.157
-0.289
0.433
0.327
0.099
0.0533
(0.000516)***
(0.158)***
(0.123)**
-0.184
(0.00667)***
-0.162
-0.199
-0.224
(0.101)***
(0.0392)***
-0.175
-0.184
(0.140)***
(0.134)**
-0.151
-0.161
0.0584
0.132
-0.159
-0.0959
-0.785
0.214
-2.309
583
-0.524
-0.135
(0.768)***
Robust standard errors in parentheses; * Indicates significance at the 10% level, ** at the 5% level, *** at the 1% level
Table A6. OLS model of length of uncertainty range in stoplight sliding scale exercise for respondent demand: Kolkataa
Model 1
Price (Rs)
Price * Cholera
Time to think
Cholera
TTT*cholera
Time spend thinking (mins)
1.683
-0.00263
-0.99
0.985
Model 2
(1.49)
(0.00)
(0.432)**
(0.530)*
1.716
-0.00267
-1.574
0.425
1.139
Model 3
(1.49)
(0.00)
(0.610)**
(0.68)
(0.86)
1.591
-0.00255
0.974
-0.00575
Model 4
(1.50)
(0.00)
(0.532)*
1.684
-0.00268
1.025
(1.50)
(0.00)
(0.533)*
(0.01)
Discuss with a household
member? (Yes=1)
-0.488
Monthly per capita income
(000 Rs) ‘
Male
Education Low b
Education Mid b
Education High b
(0.45)
-0.00679 (0.13)
-0.0185 (0.13)
0.00195 (0.13)
0.00479 (0.13)
-0.11 (0.48)
-0.06 (0.48)
-0.0938 (0.49)
-0.112 (0.49)
0.57 (0.77)
0.566 (0.77)
0.443 (0.78)
0.435 (0.77)
1.67 (0.829)**
1.7 (0.829)**
1.547 (0.832)*
1.511 (0.829)*
1.132 (0.93)
1.139 (0.93)
1.013 (0.93)
1.003 (0.93)
Age
-0.00671 (0.03)
-0.00794 (0.03)
-0.00222 (0.03)
-0.00344 (0.03)
Distance to health facility
-0.0586 (0.0324)*
-0.0593 (0.0324)*
-0.052 (0.03)
-0.0542 (0.0324)*
Likely to get disease – Resp.
0.697 (0.45)
0.678 (0.45)
0.711 (0.45)
0.681 (0.45)
Disease is serious for adults
0.216 (0.44)
0.202 (0.44)
0.193 (0.45)
0.19 (0.45)
Know person with disease
-0.176 (0.44)
-0.169 (0.44)
-0.106 (0.44)
-0.13 (0.44)
Muslim
-2.228 (2.95)
-2.13 (2.95)
-2.088 (2.97)
-2.182 (2.97)
Never boil drinking water
-1.172 (0.447)***
-1.165 (0.447)***
-1.172 (0.449)***
-1.182 (0.449)***
HH member had vaccine
-0.181 (0.51)
-0.215 (0.51)
-0.191 (0.52)
-0.183 (0.52)
Constant
3.139 (1.439)**
3.464 (1.458)**
2.593 (1.423)*
2.775 (1.441)*
Observations
536
536
536
536
R-squared
0.06
0.063
0.051
0.052
a
* Indicates significance at the 10% level, ** at the 5% level, *** at the 1% level. Excluding nine scenario-rejecters. Robust standard
errors in parenthesis. b Excluded category is no education Table A7. OLS model of length of uncertainty range in stoplight sliding scale exercise for respondent demand: Beira
Model 1
Price (Mts)
Time to think
Time spent thinking (mins)
0.000905
-0.0996
Model 2
(0.000491)*
(0.0327)***
Model 3
0.000894
(0.000494)*
-2.57E-05
-0.00021
Discuss with a household
member? (Yes=1)
Number of hh members
TL1, start low
TL1, start high
TL2, start high
Urban zone 3
Urban zone 4
Single
Widow
Divor
Male
Muslim
Protestant
Household cholera death
Household cholera
Diarrheal incidence in bairro
Mosquito nets (y/n)
Soap in house
Asset 4 (motor vehicle)
Income quartile 2
Income quartile 3
Income quartile 4
Primary school only
Secondary school
Post-secondary school
Traveled to Esturro
Constant
Observations
R-squared
a
-0.0229
-0.0428
0.0477
0.0873
0.153
0.0311
-0.0177
-0.00844
-0.0436
0.0438
-0.0433
0.00975
-0.088
0.0381
-0.00804
0.00627
-0.00145
0.0696
0.0493
0.0969
0.125
0.143
0.139
-0.0324
0.0582
0.33
899
0.067
(0.00645)***
-0.037
-0.0417
(0.0520)*
(0.0664)**
-0.0513
-0.06
-0.0592
-0.0647
-0.0365
-0.0537
-0.0352
-0.0942
-0.044
-0.00813
-0.0331
-0.0407
-0.0863
-0.0468
(0.0524)*
(0.0695)*
(0.0408)***
(0.0605)**
-0.0943
-0.0417
(0.114)***
-0.0222
-0.0473
0.0463
0.0902
0.154
0.0395
-0.0164
0.000763
-0.0275
0.0464
-0.061
-0.0019
-0.0737
0.0377
-0.0068
0.0199
-0.00091
0.0813
0.0503
0.104
0.134
0.151
0.148
-0.0174
0.067
0.24
899
0.057
(0.00656)***
-0.0375
-0.0416
(0.0523)*
(0.0664)**
-0.0518
-0.0582
-0.0591
-0.0628
-0.0368
-0.0536
-0.0354
-0.0981
-0.0441
-0.00818
-0.0331
-0.0409
-0.0858
-0.0472
(0.0527)**
(0.0703)*
(0.0412)***
(0.0608)**
-0.0939
-0.0415
(0.112)**
0.00092
(0.000492)*
-0.0899
(0.0326)***
-0.0227
-0.0426
0.0463
0.0823
0.148
0.0354
-0.0266
-0.0128
-0.0468
0.0464
-0.045
0.00575
-0.0818
0.0432
-0.0071
0.0134
-0.00108
0.0671
0.0526
0.102
0.135
0.149
0.147
-0.0256
0.0608
0.289
899
0.065
(0.00648)***
-0.0372
-0.0418
-0.052
(0.0661)**
-0.0514
-0.0604
-0.0598
-0.0645
-0.0365
-0.0538
-0.0352
-0.095
-0.044
-0.00811
-0.0332
-0.0409
-0.0865
-0.0469
(0.0525)*
(0.0698)*
(0.0407)***
(0.0603)**
-0.094
-0.0417
(0.111)***
* Indicates significance at the 10% level, ** at the 5% level, *** at the 1% level. Excluding nine scenario-rejecters. Robust standard
errors in parenthesis. b Excluded category is no education