the double-channeled effects of experience in individual decisions

THE DOUBLE-CHANNELED EFFECTS OF EXPERIENCE
IN INDIVIDUAL DECISIONS∗
P EIRAN J IAO†
Abstract
People often make decisions with both descriptive and experiential information available.
For example, investors in the financial markets typically have access to both descriptions from
brochures, financial reports, analysts, etc., as well as own experience. However, little is known
about the role of experience in individual economic decision-making with the presence of descriptive information. This paper investigates the issue by experimentally testing the effects of
experience on individuals using an investment task with choice feedback and varying levels of
descriptive information. In the experiment, participants observed prices of hypothetical assets,
made buying decisions and reported subjective beliefs. We document the double-channeled effects of experience: when elicited beliefs were controlled for, participants significantly relied
on experience regardless of the descriptions, behaving consistently with the law of effect, i.e.
repurchasing more assets on which they gained rather than lost; additionally, beliefs were also
distorted by experience, in that participants were more optimistic about assets from which they
gained, and pessimistic about previously unowned assets. In a calibration exercise, reinforcement learning significantly added predictive power to expected utility models. These results
can inform theories of individual decisions and the design of behavioral interventions.
Keywords: Description, Experience, Law of Effect, Biased Belief, Individual Choice.
JEL Classification Numbers: C91, D03, D83.
Word Count: 7515
∗
The author thanks Peyton Young, Vincent Crawford, Peter Wakker, Olivier l’Haridon, Aurélian Baillon, Joshua
Tasoff, Henrich Nax, John Jensenius III, Chen Li and participants at the CESS colloquia at Nuffield College, Learning,
Games and Networks Seminar in University of Oxford, ETH Zurich, Erasmus School of Economics, Experimental
Finance Conference 2015, 6th Annual Conference of ASFEE for helpful discussions, suggestions and comments on
this project, and Kaixin Bao, Taotao Zhou and Patrick Lu for excellent research assistance during early stages of the
project. This research was supported by the George Webb Medley/OEP Fund and the John Fell OUP Research Fund
under project code CUD09320.
†
Address: Department of Economics, Mannor Road Building, Mannor Road, Oxford, OX1 3UQ, UK; Fax: +44
(0)1865 271094; Telephone: +44 (0)1865 278993; E-mail: [email protected].
1
1
Introduction
Conventional economic models assume that individual decision makers update beliefs follow-
ing the Bayes’ rule, correctly weighting experienced and observed outcomes, as well as descriptive
information. However, evidence in the lab and field documents clear departures. For instance, in
games players tend to repeat the actions that brought experienced gains even when the environment has changed [Camerer and Ho, 1999]; in individual decisions when the Bayesian predictions
clashed with reinforcement learning subjects followed each rule almost half of the time [Charness and Levin, 2005]. Evidence in the financial market suggests that individual investors (but not
professionals) increase their subscription to IPO auctions subsequent to previous successful experience, as if they overweighted personal experience [Kaustia and Knüpfer, 2008; Chiang et al.,
2011]. People seem to have the tendency to overweight experienced outcomes relative to observed outcomes and descriptive information, compared with what a Bayesian agent would do.
The role of experience in individual economic decisions is not yet well-understood. Our goal is to
use an experiment to create clear tension between rational Bayesian strategy and the experiencebased learning rule in individual decision-making under uncertainty, in an investment context. By
comparing actual choices with the clear Bayesian benchmark, we test (1) whether participants
overweighted personal experience, even when such experience was information-free, i.e. bearing
no information value for a Bayesian updater, (2) whether their behavior was consistent with the
law of effect, i.e. the probability of taking an action increasing in its past reinforcement, and (3)
whether experience induced any belief distortion. Our results uncovered positive answers to all
three questions. A better understanding of these issues can inform individual decision theories,
and help design effective tools to debias individuals, such as in financial markets.
Individual choice models in conventional and even main stream behavioral economics do not
distinguish two cases: description-based decisions, where the decision maker possesses complete information about the incentive structure, and experience-based decisions, where the decision maker has no knowledge about the environment but can sample or experience the outcomes.
Psychologists documented discrepancies between description and experience-based decisions, i.e.
2
the description-experience gap:1 for example, Barron and Erev [2003] find subjects exhibited underweighting of rare events in experience-based decisions, instead of overweighting them as in
prospect theory; Jessup et al. [2008] and Lejarraga and Gonzalez [2011] find experience overwhelms descriptive information when both are available. The effect of varying information structures on behavior has been studied in repeated games,2 but these issues were not well-understood
by economists in individual decisions. To the author’s knowledge, attempts to incorporate the
description-experience gap, particularly the effects of experience, in an individual economic or financial decision environment limit to the following: Abdellaoui et al. [2011] estimate prospect theory parameters and find less overweighting of rare events under experience than under descriptionbased context; Ben-Zion et al. [2010] experimentally show that under-diversification can result
from providing feedback on both diversified and non-diversified investment alternatives; Kaufmann et al. [2013] and Bradbury et al. [2014] both use simulated experience to help investors
understand risk.
The reasons for using an investment task in our experiment are threefold. Firstly, investment
decisions, like most real-world problems, are relatively more complex than the clicking paradigm
adopted in most psychological studies.3 Admittedly they have some advantages, but whether their
findings can be generalized to more complex and realistic situations is questionable [see e.g. Hau
et al., 2008]. Secondly, the financial market is a natural environment where decision makers have
access to both descriptive and experiential information. Descriptions are available through investor
education, fund brochures, financial reports, financial analysts, etc., but are often ignored: individual investors lose from under-diversification [Kelly, 1995] and the disposition effect [Shefrin and
Statman, 1985], both clearly against the advices in mainstream investor education. First-hand
experience from own trading history is often overweighted, such as in the IPO market [Kaustia
1
See Rakow and Newell [2010] for a review of this issue in psychology.
In game theoretic experiments, the experimenter either provided players with only descriptions of the game
[Rapoport et al., 2002], only feedback [Mookherjee and Sopher, 1994; Nax et al., 2013; Friedman et al., 2015] or
both; theories have demonstrated the equilibrium property of adaptive learning dynamics under limited information
settings [see e.g. Foster and Young, 2006; Young, 2009].
3
See e.g. Erev and Haruvy [2013] for a review of studies using the clicking paradigm, where participants make
binary choices between two buttons on a screen, each associated with some binary lotteries, known or unknown to the
decision maker.
2
3
and Knüpfer, 2008].4 And thirdly, individual investors’ learning from experience is an intriguing topic. Larger quantity of experience has been shown to induce better performance [Nicolosi
et al., 2009] and reduce behavioral biases [Dhar and Zhu, 2006]. If investors do overweight experience, then good experience can be dangerous, leading to overconfidence[Gervais and Odean,
2001], over-participation such as in 401(k) portfolio allocation [Choi et al., 2009], and unjustifiable
repurchases [Strahilevitz et al., 2011].
In our experiment, participants observed price sequences of 4 hypothetical assets generated
by 4 equally-likely underlying processes with drifts and had two tasks each period: they made
buying decisions, knowing that shares bought would be automatically sold when the new prices
were announced, and they predicted the probabilities of price increases. The price sequences were
designed so that knowledge of the underlying processes and observation of historical prices would
suffice for an agent to adopt the simple Bayesian strategy that relies only on the price change
directions, rendering experienced gains/losses information-free, so that the predictions of Bayes’
rule clashes with that of reinforcement learning in some cases. We find participants fail to sufficiently account for the descriptive information, but instead overweight on their own experience.
Different treatment conditions varied the amount of descriptive information provided regarding
the price generation processes, but this did not alter the effect of experience on decisions, suggesting that descriptions were possibly underweighted or sometimes ignored. More importantly,
we do find behavior consistent with the law of effect: controlling for elicited beliefs, participants
were more likely to repurchase assets from which they gained than those from which they lost.
They also placed considerable weight on the most recent payoffs, discounting past experience
heavily. The experienced gains/losses had additional explanatory power than just observed price
increases/decreases.
Additionally, experience distorted beliefs: participants were more optimistic about the assets
that offered better obtained payoffs, and believed those with positive forgone payoffs would mean
revert. We also tested the effect of two other stimuli, namely, regret and surprise, both believed
4
Another source of experience from observing that of others [see e.g. Kaustia and Knüpfer, 2012] will not be
discussed here.
4
to influence behavior when choice feedbacks are available [see e.g. Zeelenberg et al., 2001; Nevo
and Erev, 2012], but both exerted limited impact on choices. A further evaluation of predictive
performance among models with and without learning reveals that although reinforcement learning alone did not fit the data satisfactorily, adding it to the expected utility models significantly
improved predictive power.
2
The Experiment
2.1
Design
Each session of the experiment contained three rounds, with T = 15 decision periods per
round. In each round, participants were faced with a different set of K = 4 hypothetical assets,
labeled W, X, Y and Z. The price-generating processes were similar to Weber and Camerer [1998]
and Jiao [forthcoming] in their studies of the disposition effect. The price of an asset in any period
was determined in two independent steps. First, the price change direction was determined by one
of 4 equally-likely underlying processes, which differed only in the probability (θ) of price going
up every period, with θ = 65%, 55%, 45%, 35% respectively. The price went down with probability
1 − θ and could not stay unchanged. Then in the second step, the price change magnitude was
randomly drawn from {1, 3, 5}. Prices of all assets in all periods were independently determined.
The underlying process that generated all prices of an asset remained the same throughout the
same round. The participants could see all 4 price sequences evolving on the same chart every
round. The price sequences were predetermined using a computer random number generation
program. For the ease of data pooling and comparison, all participants viewed the same set of
predetermined price sequences but in different random orders. Figure I shows a snapshot of the
experimental interface. Before starting the experiment, all participants needed to correctly answer
all comprehension questions regarding the experimental setup.
There were 4 treatment conditions differing in the tasks after observing the price sequences,
and the information revealed about the price-generating processes. In condition A, participants
5
were clearly instructed about how prices were generated. Then in the beginning of each round,
they received an endowment of 500 ECU.5 Before the start of period 1, they could see 6 periods of
price history, and in each of the 15 subsequent periods, before seeing new prices announced, they
had two tasks: the Prediction task and the Buying task. The Prediction task asked participants to
predict the probability of price going up for each asset, by indicating their guesses on a slider bar
from 0 to 100 percent with increments of 1 percent. The buying task was to purchase one share of
one asset at the last observed price, knowing that it was to be automatically sold when new prices
were announced. The reason for automatic selling is to give everyone equal opportunities to obtain
experience.6
After making decisions in a given period, participants saw new prices of all assets, as well as
their updated balances, and they knew the payoff that could have been obtained if a different asset
had been selected. They also saw their cumulative earnings from each asset. If the payoff from
asset k in period t is denoted by πk,t , the cumulative earning, Rk,t , was calculated as follows:
Rk,t = Rk,t−1 + I(k, k(t))πk,t ,
(1)
where I(k, k(t)) is an indicator function with the value of 1 when asset k is chosen in period t, and
0 otherwise. Once a round was concluded, the balance was not carried over to the next round. The
Prediction task was incentivized using the Quadratic Scoring Rule (QSR) [Offerman et al., 2009]
and elicited beliefs were then corrected for risk attitudes.7 Participants were given an illustrative
table and examples to understand the QSR, and were also told the payoff-maximizing strategy
5
ECU is the monetary unit in the experiment, with an exchange rate of 60 ECU = £1.
If free buying and selling were allowed, a participant could buy and hold an asset without realizing any gain or
loss. Although it may also be interesting to study the difference between experiencing capital versus paper gains and
losses, it is not pursued in this paper.
7
Suppose Belief represents the reported probability, then the prediction reward Q(Belief ) was calculated in the
following way, where m = 100 and n = 100 are parameters that determine the size of rewards:
(
m + 2n × Belief − n × [Belief 2 + (1 − Belief )2 ],
if the actual price increased
Q(Belief ) =
2
2
m + 2n × (1 − Belief ) − n × [Belief + (1 − Belief ) ], if the actual price decreased.
6
The beliefs below were corrected for risk attitudes if not otherwise mentioned. The correction procedure can be
found in Offerman et al. [2009], with the CRRA utility and a probability weighting function of following form:
ω(p) = exp(−(−β(−ln(p)))α ) where we fix α = 0.65 and β = 1.
6
was to always report one’s true beliefs. Once a participant completed the experiment, the ending
balance of a randomly chosen round determined the Buying task reward; and one randomly chosen
prediction determined the Prediction task reward.
The advantage of this experimental method is a simple Bayesian benchmark. With the knowledge of those underlying processes, the Bayesian strategy should be to count the number of ups and
downs in each sequence and to choose the asset with the most number of ups every period, so the
price change magnitudes, experienced or observed, should carry no information value. The three
price change magnitudes create clashes between the Bayesian predictions and reinforcement learning: the asset with the largest price increase does not necessarily coincide with the one with the
most ups. Therefore, departures from the Bayesian benchmark by overweighting one’s experience
can be cleanly detected. This experiment is complementary to empirical studies of reinforcement
learning in finance, because experience is not necessarily information-free in real-world settings,
making clear identification of experience overweighting relative to the Bayesian benchmark difficult.
Compared with Condition A, participants in Condition B Round 3 only had the Prediction
task, but not the Buying task. Two things were different in Condition C. On the one hand, in
Rounds 1 and 2, Condition C participants received reduced descriptive information: they were told
everything except the probabilities of price increase: with regard to price change directions, they
were only told that there were 4 equally-likely processes, HH, H, L and LL, with the following
order of their probability of price increase each period: θHH > θH > 50% > θL > θLL . Even
without specific probabilities, the ordering should be sufficient for a Bayesian updater to follow
the Bayesian strategy. On the other hand, in Condition C Round 3, participants obtained the
same information as in Condition A, but only had the Buying task. In Condition D Rounds 1 and
2, participants received even less information: they were not told how the price change directions
were determined, but only knew prices could increase or decrease and the price change magnitudes
were independently randomly drawn from {1, 3, 5}. This does not provide enough information for
Bayesian updating. Lastly in Condition D Round 3, participants were given full information. In all
7
rounds of Condition D, participants completed both Buying and Prediction tasks. With both buying
and prediction tasks present, the issue of spillover effect between these two decisions emerges. To
address this, we use two approaches: we let subjects completed these tasks in random orders to
mitigate the spillover effect, and we use the variation of tasks across treatments to test for it.8 The
reduced information treatments were for the purpose of testing whether reliance on experience was
sensitive to the completeness of descriptions.
2.2
Procedure
The experiment was conducted at the Centre for Experimental Social Sciences, Nuffield College, University of Oxford. There were a total of 4 sessions and 89 participants (excluding 1
participant who withdrew): 23 in each of Conditions A, B and C, and 20 in Condition D. Among
the recruited participants, 58 were students in the University of Oxford and the rest residents in
Oxford city, with 62% male and an average age of 31. Among the participants, 16 had trading
experience, mostly managing portfolios of less than £10000 for less than a year.
In each session, participants were first consented, and then seated in front of isolated computer
stations. They read instructions of the experiment on the computer screen and were allowed to ask
questions or to withdraw without sacrificing the show-up fee at any time during the session.9 They
were not allowed to use calculators or any computer program, but could use paper and pen.
Upon completing the experiment, participants answered a short questionnaire regarding their
demographic information, economics and mathematics background, real-world trading experience
and risk attitudes. The risk preference questionnaire was adapted from [Weber et al., 2002]. And
we also used a simple incentivized Multiple Price List to elicit risk attitudes. Upon finishing the
survey, they were paid privately one at a time in a separate room. The payment included a show-up
fee of £4, plus the payment to each part of the experiment. The average earning per subject was
8
We find that in Condition C Round 3 where participants only had the Buying task, only 25.22% of the decisions
were consistent with Bayesian, significantly different (p < 0.01) from 31.48%, the proportion of buying decisions
consistent with Bayesian in presence of the Prediction task. This suggests that belief elicitation also drove beliefs
towards Bayesian.
9
The experimental instructions are available upon request.
8
£20.70. Each session took around 2 hours.
3
Theories and Hypotheses
Psychologists have a long history of studying the effect of past experience on decisions. In
the hungry cat experiment of Thorndike [1898], a cat was placed in a box and could free itself by
either pressing a lever or pulling a loop; it initially experimented with many actions but gradually
took less time to obtain freedom in successive trials. This suggests that actions are reinforced
by their consequences: association of a response to a stimulus will be strengthened if followed
by a satisfying state, and weakened if followed by an annoying state. This theory, the law of
effect [see e.g. Herrnstein, 1970], predicts a positive correlation between choice probability and
rewarding past experience, requiring no prior knowledge regarding the environment or formation
of subjective beliefs about the future. This effect has been studied under reinforcement learning in
modern psychology, and found numerous empirical supports [see e.g. Suppes and Atkinson, 1960].
Economists have studied this type of learning mostly in game theoretic interactions. For example, reinforcement learning in games typically assume that reinforcements associated with an
action are updated by obtained payoffs, which is positively correlated with choice propensities, and
experiments demonstrated that these models outperform equilibrium predictions in a wide range
of games [see e.g. Erev and Roth, 1998; Camerer and Ho, 1999]. In the context of individual
decisions, instead of interacting with other players, the decision maker receive reinforcements on
actions by interacting with the environment. Empirical work, such as [Kaustia and Knüpfer, 2008],
has shown the plausibility of such behavior in complex financial contexts. However, empirical
studies cannot determine it underlying mechanism, because (1) investors’ choice sets in the field
cannot be easily determined, i.e. we do not know what assets they monitor and what their forgone
opportunities are, and (2) we cannot calculate the objectively correct weight that should be placed
on experience. In the experiment, if participants had placed too much weight on experience, we
would expect the law of effect to manifest: the probability of choosing asset k should be positively
9
correlated with the reinforcement on that asset; hence relative to the Bayesian benchmark, better
experienced outcomes on asset k increases the chance of it being repurchased. Reinforcement on
asset k in period t is updated as follows,
Rk,t = φRk,t−1 + I(k, k(t))πk,t ,
(2)
where φ, the discount rate of past reinforcements, captures forgetting: when φ = 1 the reinforcement updating converges to equation (1); φ = 0 implies extreme forgetting, so that only the most
recent payoff matters. The value of this parameter may vary across participants. An issue in most
reinforcement learning models is to determine the initial reinforcement value and choice propensity before the decision maker obtains any experience. The experimental setting here evades this
issue because participants observed some price history before starting, which could give them some
initial propensity to choose.
The probability that a participant chooses asset k in period t + 1 is a monotonically increasing
function of its reinforcement. This function can take many forms, including exponential (logit),
power, and normal (probit) functions. Here we use the logistic probability rule:
eλRk,t
,
Pk,t+1 = PK
λRh,t
h=1 e
(3)
where λ regulates the sensitivity of choice probability with respect to reinforcements, and the
decision maker will choose the option with the highest Pk,t+1 . To save a parameter, we assume
λ = 1. The advantage of this functional form over others is its validity even with negative payoffs.
Law of effect implies a win-stay-lose-shift strategy or trial-and-error learning [see e.g. Young,
2009; Callander, 2011]. It generates an asymmetry in the inertia propensity documented in many
experimental studies, i.e. a tendency for positive correlation between recent and current choices
[see e.g. Suppes and Atkinson, 1960; Erev and Haruvy, 2005; Cooper and Kagel, 2008]. In terms
of investment decisions, this means investors are more likely to repurchase assets on which they
gained than those on which they lost [Strahilevitz et al., 2011], although past and future perfor10
mances are not necessarily correlated.
Additionally, Jessup et al. [2008] compare the feedback and no-feedback treatments in the
presence of descriptions and find that feedback overwhelms the effect of descriptions, reducing the
overweighting of rare events and driving subjective beliefs toward objective probabilities. Lejarraga and Gonzalez [2011] further experimentally demonstrate that descriptions are neglected in the
presence of feedback regardless of task complexity. In voluntary contribution games, [Nax et al.,
2013] demonstrate that the asymmetric inertia learning pattern persists in voluntary contribution
games under a black box setting where only feedback on own actions were provided, and in a full
description environment. One possible reason for overweighting experience in real-world financial
decisions is the large quantity and complexity of descriptive information in the market. In our experiment, descriptions about the underlying processes should drive beliefs towards Bayesian, with
perfect descriptions in Conditions A and B. With reduced information in Condition C Rounds 1
and 2, a Bayesian updater still has enough information to follow the Bayesian strategy. In Condition D Rounds 1 and 2, the information setup does not allow a Bayesian updating, but could only
rely on experience. By contrast, if descriptions were neglected, we would expect the law of effect
to be robust under different information settings.
In other words, if following the law of effect is our natural tendency, then reliance on it should
be unconditional, independent from the information value of experience, knowledge about the
environment, or nature of the task. Hence the following hypotheses.
Hypothesis 1 Law of Effect: Pleasant experience with an asset increases the probability of the asset being repurchased subsequently, and unpleasant experience decreases it, even when experience
has no information value.
Hypothesis 2 Descriptions: Pleasant experience with an asset increases the probability of the
asset being repurchased subsequently, and unpleasant experience decreases it, regardless of the
descriptions provided.
The above hypotheses mainly deal with the effect of obtained payoffs. In the experiment,
since feedback on all assets were available, participants could see what payoffs would have been
11
obtained if a different asset had been selected. The effect of forgone payoffs from previously unselected alternatives is another important feature of learning, and also a very realistic consideration.
For example, the belief learning models in game theory, such as fictitious play, take into account
of forgone payoffs [see e.g. Brown, 1951; Fudenberg and Levine, 1998], assuming that players
calculate expected payoffs based on observations of past outcomes of all alternatives and choose
the one with maximum expected payoff. Many reinforcement learning models take into account
forgone payoffs as well: Erev and Roth [1998] model experimentation by adding a parameter for
the weight placed on unchosen alternatives according to their similarity to the successful ones;
the Experience Weighted Attraction (EWA) model [Camerer and Ho, 1999] considers the relative weights on forgone payoffs and obtained payoffs, without similarity judgment, hence the new
reinforcement updating rule:
Rk,t = φRk,t−1 + [δ + (1 − δ)I(k, k(t))]πk,t ,
(4)
where δ is the relative weight placed on forgone payoffs. If δ = 0, no attention is paid to unchosen
alternatives; if δ = 1, all alternatives, chosen or unchosen, are equally-weighted. Together with
the law of effect, this implies that assets with better payoffs are more likely purchased, especially
the gains are obtained. Camerer and Ho [1999] call this a ‘law of simulated effect’, as if decision
makers simulate outcomes of unchosen alternatives in their mind.
Forgone payoffs can influence behavior through the feeling of regret. For example, when the
forgone payoff is higher than the obtained payoff the decision maker might think that she could
have done better had she chosen another option. Psychologists find that experienced regret makes
the decision maker more willing to search for (even irrelevant) information [Shani and Zeelenberg,
2007], and that it provokes the feeling that one should have known better and the willingness to
correct one’s mistake Zeelenberg et al. [1998].10 Hart and Mas-Colell [2000] introduce a ’regret
matching’ adaptive procedure in games, in which the probability of choosing the current strategy
10
Post-decision experienced regret should be distinguished from anticipated regret. Economists have explicitly
modelled the latter, predicting that people make decisions to minimize anticipated regret [see e.g. Loomes and Sugden,
1982].
12
is decreasing in the size of regret, measured using the forgone payoff from unchosen strategies.
If this were true, larger regret should make participants less likely to repurchase, which poses a
limit to the positive recency resulted from law of effect: among assets offering the same obtained
payoff, the one associated with less regret should be more likely repurchased.
Hypothesis 3 Regret: A previously sold asset is more likely repurchased if it was associated with
a smaller regret.
Additionally, Nevo and Erev [2012] document another learning pattern in decision from experience, i.e. a tendency to change action after both a sharp increase and a sharp decrease of payoff,
dubbed surprise-trigger-change. Their study was motivated by [Karpoff, 1988] who find volume
spiked after both sharp price increases and decreases, suggesting increased willingness to sell after
both winning and losing. Nevo and Erev [2012] propose a model, called Inertia, Sampling and
Weighting (I-SAW), where they assume people have an innate tendency for inertia, and surprising
outcomes decrease their probability of remaining in the inertia state. If this were true, it implies
negative recency after a surprising positive obtained payoff, and positive recency, implied by law
of effect, after an unsurprising positive obtained payoff and a negative obtained payoff, hence an
asymmetry. With regard to forgone payoffs, they find negatively surprising forgone payoffs also
reduce inertia. They measure surprise using two gaps: the gap between the obtained payoffs from
action k in period t and in period t − 1, and the gap between the obtained payoff from action k in
period t and the average payoff from that action in all previous periods. In their clicking paradigm,
a surprise is just the rare event of a large positive (or negative) payoff between the two possible outcomes associated with a lottery. In our experiment, however, we define surprise as the difference
between realized payoff and expected payoff calculated from elicited beliefs, hence the following
hypothesis.
Hypothesis 4 Surprise Triggers Change: Positively surprising obtained payoffs and negatively
surprising forgone payoffs decrease the probability of repurchase.
13
4
Results and Discussion
4.1
Law of Effect
Law of effect predicts that participants should be more likely to choose an asset that brought
better payoffs in the past. In order to test this we use two different measures of experience with
an asset: the non-discounted cumulative obtained profits up to period t − 1 within a given round
(CumP rof it), or the most recent profit in period t − 1, i.e. RecP rof it. Out of 3402 total buying
decisions, 2278 (or 66.69%) were inconsistent with the Bayesian strategy.11 Among them, 818
(or 35.91%) were consistent with purchasing assets with the highest CumP rof it, and 1344 (or
59.00%) were consistent with purchasing assets with the highest RecP rof it. Another way to
evaluate this is to see how participants behaved when the prediction of Bayesian updating and law
of effect clashed. There were 261 cases where the Bayesian prediction clashed with buying at the
highest CumP rof it: 169 (or 64.75%) were consistent with Bayesian and 48 (or 18.39%) with
the latter. There were 397 cases where the Bayesian prediction clashed with buying the highest
RecP rof it: 193 (or 48.61%) were consistent with Bayesian and 176 (44.33%) with the latter.12
Thus experienced payoffs mattered, especially the most recent ones.
With each decision on an asset in a period as an observation, among the 13664 observations (not
including first periods and not including Condition B Round 3), given 4390 (3451) opportunities to
purchase at positive (negative) cumulative profit, 1053 or 23.99% (884 or 25.62%) were actually
purchased; given 1639 (1455) opportunities to purchase at positive (negative) recent profit, 906
or 55.27% (572 or 39.31%) were purchased. This comparison is graphically illustrated in Figure
II. The difference between proportions purchased was not significant when cumulative profit is
considered (p = 0.10), but highly significant with recent profit (p < 0.01).
[Insert Figure II about here.]
11
Period 1 decisions were excluded because the law of effect is only applicable once the participant obtained some
experience.
12
This is comparable to the finding in Charness and Levin [2005], that when predictions of Bayesian and reinforcement learning clashed in their experiment of individual decisions, subjects followed each in about half of the
cases.
14
We next test the law of effect in regressions controlling for beliefs and other variables, in order
to see whether experience adds additional explanatory power for buying decisions. At the decision
level, we use conditional fixed-effect logistic regressions grouped by participants, with robust standard errors. The dependent variable Buy is equal to 1 if the participant decided to buy the asset,
and 0 otherwise. The grouping controls for within-subject correlations. The independent variables
include the following: CumP rof itP os (RecP rof itP os) is a dummy variable for positive cumulative profits (recent profits) on an asset; Belief is the elicited probabilistic beliefs about price
going up next period, which is expected to be positively correlated with buying; and two additional
variables control for the observed price patterns. LastU p is equal to 1 if the recent outcome was
a price increase, and 0 otherwise; M oreU p is equal to 1 if there were more ups than downs in
the sequence, and 0 otherwise. LastU p controls for the effect of a recent observed price increase,
rather than experienced; M oreU p is the variable that a Bayesian decision maker cares about. Table
I reports the results.
[Insert Table I about here.]
All regressions clearly show significant positive correlation between belief and the buying decision, meaning that participants in general had little confusion about the task at least in terms of
making buying decisions consistent with their beliefs. Regression (1) in Table I shows that once
beliefs are controlled for, cumulative profits add no explanatory power. Regressions (2) and (4)
show that participants tended to buy an asset that brought them positive recent obtained profit, even
after controlling for the observed price patterns.
Both LastU p and M oreU p are significant in Regression (3), meaning that on the one hand
participants were, to some extent, momentum buyers chasing past returns, so that simply observing
good recent performance, without experiencing it, increased their willingness to buy; and on the
other hand they cared about the number of ups as a Bayesian does. In Regression (4), however,
recent experienced profit is still significant while LastU p is not, suggesting that participants cared
more about the experienced positive payoffs than observed ones. This is supportive evidence for
Hypothesis 1, suggesting the validity of law of effect even after controlling for beliefs, especially
15
when we consider most recent experience, and that direct experience of an outcome influences
decisions more than just observations.
Another piece of evidence for Hypothesis 1 is from asymmetric inertia, i.e. the probability
of maintaining one’s choice from last period (repurchasing an asset) after gains and losses.13 We
define a recent pleasant outcome in two ways. Firstly, it could be a larger absolute recent gain.
Panel A of Figure III illustrates a comparison of repurchases after gains and losses. The height of
the bars reflects the proportions repurchased out of opportunities to repurchase after experiencing
different levels of recent obtained profits, {−5, −3, −1, 1, 3, 5}. The proportion of repurchases
after a gain was generally larger than that after a loss, but the relationship seemed nonlinear in the
domain of losses.
Alternatively, a recent pleasant outcome could be an increase of experienced payoff from period
t − 2 to t − 1.14 There were 1317 opportunities to repurchase when payoff increased: in 716 (or
54.37%) of them the asset was repurchased; only 253 (or 35.34%) of those repurchases were
consistent with Bayesian. There were 1577 opportunities to repurchase when payoff decreased: in
689 (or 43.69%) of them the asset was repurchased; 292 (or 42.38%) of these repurchases were
Bayesian. The differences in proportions repurchased and proportions consistent with Bayesian
after payoff increases versus decreases were both significant (p < 0.01). That is to say participants
were more likely to repurchase (or remain in the inertia state) after a pleasant outcome, and that
they were more Bayesian in repurchase decisions after an unpleasant outcome. These results are
illustrated in Panel B of Figure III. The evidence using both assessors of recent experience suggests
the existence of asymmetric inertia implied by law of effect.
[Insert Figure III about here.]
The results in Table II further supports this. We use the conditional logistic regression grouped
by participants, but with decision in each period (not on each asset) as an observation. We separately estimate for the gain and the loss domains in order to detect asymmetry in behavior. The
13
The analyses above use a decision on each asset in each period as an observation, while here we treat each period
as an observation.
14
Note that the assets chosen in these two periods need not be the same.
16
dependent variable is a binary variable, Repurchaset , that is equal to 1 when in period t the decision maker repurchased an asset that was purchased in period t − 1. Beliefs are controlled for:
Belieft−1 is the period t − 1 elicited belief of price increase on the asset purchased. We create
two dummy variables in each of the gain and loss domains for the price change magnitudes of 3
and 5, so the coefficients represent their effects relative to 1 or −1. Increaset is equal to 1 if the
participant experienced a payoff increase from period t − 2 to period t − 1, and 0 otherwise. The
determinants of repurchases after a gain and a loss were different. Regressions (1) and (3) implies
asymmetric inertia: after a larger gain, participants were more likely to repurchase, and after a
larger loss, they were more likely to shift.15 However, once we consider relative payoff increase in
Regressions (2) and (4), the significance level on recent payoffs decreased in the domain of gains;
in the domain of losses, payoff increase had no effect, possibly because a loss was rarely associated with a payoff increase. Thus participants probably cared about whether there was a relative
increase in payoff, rather than just absolute gains.16 Another asymmetry lies in the effect of cumulative profits: positive cumulative profits increased the likelihood of repurchase after gains, but
had no significant effect after losses.
[Insert Table II about here.]
4.2
The Effect of Descriptions
The above regressions were pooling together treatments with different information structures.
Next we investigate the effect of differences in descriptions. In Treatment A and B with perfect
descriptions, 36.79% of the buying decisions were Bayesian, and in Treatment C with reduced
information, the number was 31.06% (difference: 5.73%, p < 0.01), although both groups of
participants should be able to use the Bayesian strategy with the information given.17 This suggests
that the descriptions were not completely neglected: more complete descriptions drove beliefs
15
The nonlinearity observed in Figure III Panel A manifests in the insignificant coefficient on recent profit −3.
The insignificance of Increase after a loss is not due to insufficient number of observations. There were 325
cases with payoff increases among 1576 cases of negative recent profits.
17
Treatment D is not shown here, because their information did not allow them to use the Bayesian strategy.
16
17
towards Bayesian.
Next we test whether descriptions influenced reliance on law of effect. To do this, we introduce binary variables C and D for Conditions C and D respectively, and interact them with
RecP rof itP os, so that significant coefficients on the interaction term would suggest differences
in the effect of experience in those conditions. The results are reported in Regression (5) of Table I.
The sign on the coefficients are not consistent with the conjecture, but they are insignificant. Being
in the conditions with less information did not significantly influence the extent to which participants relied on experience after beliefs are controlled for. This is consistent with Hypothesis 2 and
with findings in the description-experience gap literature. Therefore in the following analysis we
continue to pool all treatments together. We will show more supportive evidence for Hypothesis 2
in Section 4.4.
4.3
Regret and Surprise
Regrett−1 , the experienced regret from choosing asset k in period t − 1, is measured by the
absolute difference between obtained payoff from asset k in period t − 1 and the maximum realized payoff from unchosen assets in period t − 1. Computed from the 6 possible price change
magnitudes, the size of regret can be {0, 2, 4, 6, 8, 10}. First, if participants were more careful after
a larger regret as proposed by psychologists, we would expect more buying decisions consistent
with Bayesian. However, although there was such a tendency, the differences were not significant
at the 95% level, as shown in Panel A of Figure IV. On the other hand, according to regret theories,
larger Regrett−1 leads to a smaller probability of repurchasing, or a larger probability of shifting
to another asset. We calculated the proportion of repurchases out of opportunities to repurchase
after experiencing each value of Regrett−1 . Figure IV Panel B illustrates the results. Although
zero regret was associated with the highest repurchase rate, significantly different from those of
all other regret values, the trend was nonlinear, and the repurchase rates among all other regret
values were mostly not significantly different. Further tests in Regressions (1) and (4) of Table III
suggest regret had no significant effect on repurchases. This can emerge under two scenarios: (1)
18
based on equation (4), participants placed too little weight on forgone payoffs, or (2) their beliefs
were also distorted, i.e. they were pessimistic about unchosen assets that gave them the regret. The
regression results give some credits to the first possibility, because beliefs were already controlled
for. An analysis of beliefs in Section 4.5 will explore the second possibility.
[Insert Figure IV about here.]
Hypothesis 4 involves two types of surprise:
a positively surprising obtained payoff
(SObtained), defined as the difference between the obtained payoff from asset k in period t − 1
and the expected payoff calculated using elicited beliefs on asset k in period t − 2; a negatively surprising forgone payoff (SF orgone), defined as the absolute value of the largest negative surprise
from an unselected asset in period t−1 based on expectations elicited in period t−2. The surprisetrigger-change theory predicts larger values of both variables should lead to lower probability of
repurchase.
[Insert Table III about here.]
Regressions (2) and (5) of Table III test the effects of surprise. In the domain of gains, a larger
surprise from obtained payoff increased the probability of repurchase, which is contradictory to
Hypothesis 4; and in the domain of losses, the surprise from obtained payoff had insignificant effect. The negative surprise from forgone payoff had significant effect in none of the regressions.
Putting all variables together in Regressions (3) and (6), we find the above results still hold. In
our experiment, regret and surprise did not have the effects we expected, or the effect of reinforcement overwhelmed the effect of surprise. This was probably because the task was slightly more
complicated than in a typical clicking paradigm, involving belief updating and buying decisions.
Belief distortion could still be a plausible explanation: participants were optimistic about assets
from which they gained.
We also tested the effects of demographic variables by interacting them with independent variables in regressions (3) and (6) of Table III, but did not find significant differences. The only exception was that younger male participants had (sometimes marginally) significantly larger propensity
19
to rely on experience after gains.18
4.4
Simulation Test
The foregoing analyses use experience as an explanatory variable to establish correlations between experience and repurchasing decisions. We next conduct simulations to measure the predictive power of reinforcement learning. Six models are evaluated. The first is a baseline model
of risk-neutral expected utility with elicited beliefs, containing no parameter. The second is a
parsimonious reinforcement learning model with only one parameter, φ, the discount rate of past
reinforcements, and reinforcements update according to equation (2). The third is a two-parameter
reinforcement learning model with both φ and δ, whereas the latter is the weight on unchosen
alternatives, and reinforcements update according to equation (4). These two parameters capture
the main findings in the results above, namely the law of effect, asymmetric inertia and the effect
from forgone payoffs. For parsimony, we do not introduce parameters particularly for the effects
of regret or surprise because there was scant evidence for them in the experiment.
In the fourth model, we use CRRA expected utility, together with beliefs corrected for risk
attitudes.19 Before making buying decision for period t, the expected utility from buying asset k in
period t can be calculated according to the following:
1
1
E[Uk,t ] = Beliefk,t × [ (Wt−1 + E[∆P ])θ ] + (1 − Beliefk,t ) × [ (Wt−1 − E[∆P ])θ ],
θ
θ
(5)
where Beliefk,t is the probabilistic belief about the price of asset k going up in period t; 1 − θ is
the coefficient of relative risk aversion; Wt−1 is the ending balance in period t − 1; E[∆P ] = 3
is the expected price change, assuming that participants understood the random draw of price
change magnitudes. Then the decision maker would just choose the asset with the highest expected
18
This points to a plausible mechanism for the finding of Greenwood and Nagel [2009] that younger inexperienced
fund managers were more likely to follow trends during the tech bubble, possibly because they were more likely to
overweight recent personal experience.
19
The correction procedures can be found in Offerman et al. [2009]. Note that in the first model above, beliefs
were not corrected because of the assumed risk neutrality. In the second and third models, there was no need to correct
because beliefs were not included.
20
utility.20
The fifth model is a combination of the CRRA expected utility model and the one-parameter
reinforcement learning model. In this case, one extra parameter was added: the choice probability
of each asset will be a weighted average of the probabilities calculated from each model, with
α being the weight placed on reinforcement learning and (1 − α) on CRRA. Similarly, the last
model combines CRRA with the two-parameter reinforcement learning model, with α regulating
the relative importance of these two components.
The model parameters were estimated using nonlinear least square method. The model fits
were evaluated using the log likelihood (LL) calculated across all sampled periods.21 In order
to compare across models with different numbers of parameters, we also calculated the Akaike
Information Criterion (AIC) and the Bayesian Information Criterion (BIC).22 Table IV reports the
results. The whole sample contained 3094 observations. Reinforcement learning alone did not
generate very good fit, possibly because of its complete reliance on experience neglecting beliefs.
The CRRA model, even with only one parameter, performed much better than the baseline model.
Adding reinforcement learning to CRRA improved the fit. The estimated values of φ were small,
confirming that participants in this experiment discounted past reinforcements heavily, and placed
more weight on recent experience. The value of δ suggests that previously unchosen assets were
weighted less than 50% than experienced outcomes. This can partially explain why they did not
shift to the assets that offered high forgone payoffs.
[Insert Table IV about here.]
20
The risk attitude parameter was estimated from their asset choices, using choices in the MPL, and the questionnaire responses. But poor correlation was found among these different measures of risk attitudes, consistent with
previous evidence of the domain specificity of risk preference elicitation[see e.g. MacCrimmon and Wehrung, 1990].
21
For instance, for the two-parameter reinforcement learning model:
!
T X
N
K
X
X
i
LL(φ, δ) =
ln
(I(k i , k(t)i )Pk,t
) ,
(6)
t=1 i=1
k=1
where T denotes the number of sampled periods; N denotes the number of participants; K = 4 is the number of assets
to choose from; I(k i , k(t)i ) is an indicator function that is equal to 1 when participant i chose asset k in period t, and
i
0 for unchosen assets; Pk,t
is participant i’s probability of choosing asset k in period t calculated from the model.
22
If q represents the number of parameters and M represents the number of observations, then AIC is LL − q, and
BIC is LL − (q/2) × log(M ).
21
We then use two-thirds of the sample to calibrate parameter values and validate the model by
predicting into the rest of the sample, using the Mean Squared Deviation (MSD) to evaluate the
prediction performance.23 Even without elicited beliefs, the reinforcement learning models generated good prediction performance: the one-parameter reinforcement learning model performed
slightly better than the baseline; the two-parameter reinforcement learning model performed better
than the CRRA model. Adding the learning component to expected utility significantly improved
predictive performance. The best fit and predictive power both belong to the CRRA combined with
the two-parameter learning model.
We also conducted an additional test of the effect of descriptions, by comparing the estimated
values of α between Conditions A, B and Conditions C, D. There is no statistically significant
difference, reassuring that the differences in descriptions did not influence the extent to which participants relied on experience. Additionally, although there was considerable variation across individuals in the parameter values, there was no significant difference between demographic groups
based on gender, student status, math and economics background, and trading experience.
4.5
Experience-Induced Belief Distortion
So far our argument has been that experience added explanatory and predictive power even
after beliefs are controlled for, without considering the formation of beliefs per se. There could
be another channel, i.e. belief distortions, through which experience influenced decisions. In
order to test this, we start by investigating whether positive experienced payoffs, either cumulative
(CumP rof itP os) or recent (RecP rof itP os), influenced the participants’ optimism towards the
asset. Table V summarizes elicited beliefs, as well as the benchmark Bayesian beliefs, grouped by
positive and negative obtained payoffs.
23
This is how MSD was calculated:
M SD =
T X
N X
K
i
X
[Pk,t
− I(k i , k(t)i )]2
t=t i=1 k=1
where t is the starting period of the validation sample.
22
(T − t)N K
,
(7)
[Insert Table V about here.]
There were two important patterns. Firstly, participants tended to believe better outcomes for
assets on which they had better experience. This tendency is inconsistent with Bayesian updating. Particularly, we calculated the average elicited beliefs under positive and negative obtained
profits, and compared with Bayesian beliefs. Under both measures of experience, participants had
significantly higher beliefs about price increase (or were more optimistic) for assets on which they
gained than on which they lost. The gap was larger when we considered the most recent experience.
Bayesian beliefs were roughly the same for positive and negative experience groups.
Secondly, participants had a tendency to believe assets that offered positive forgone payoffs
to subsequently mean revert. To show this, we calculated a measure of belief in mean reversion,
called Signed Reaction Measure (SRM ).24 SRM is negative when the participant believed in
mean reversion, or that an asset whose price went up (or down) last period to be more likely to
go down (or up). We found belief in mean reversion on unchosen assets that yielded positive
forgone payoffs, and belief in continuation on assets that yielded negative forgone payoffs (difference=0.0985, p < 0.01). But Bayesian beliefs exhibited no such pattern. To put it another way, on
average, participants tended to weakly believe bad things would happen to unchosen assets. These
results suggest that previous ownership matters when it comes to belief formation.
The following regressions in Table VI confirm the above tendencies. The dependent variable, Belief , is the elicited beliefs corrected for risk attitudes. Experience is captured by two
binary variables for positive cumulative and recent profits respectively. Bayesian is the benchmark Bayesian belief. Elicited beliefs had a significant positive correlation with Bayesian beliefs.
Cumulative profits were not significant in explaining beliefs, but positive recent profits were significantly biasing beliefs upwards. In Regression (3) controlling for observed price increase, a recent
experienced price increase had a drop in its coefficient magnitude but was still highly significant,
suggesting experienced outcomes had additional influence than just observed ones. Regression
24
SRM is equal to the elicited belief minus 0.5, when the previous outcome was an up; and it is equal to the
elicited belief minus 0.5, multiplied by −1 when the previous period outcome was a down. This measure was adapted
from an experimental study of the law of small numbers in Asparouhova et al. [2009].
23
(4) introduced forgone payoffs, F orgoneHigh, a binary variable that is equal to 1 for a foregone
payoff higher than the obtained payoff. The negative sign on this variable indicates that participants tended to guess lower probability of price increase for assets that performed better than their
purchased asset. These results point towards an experience-induced belief distortion, that has not
been well documented or understood in the economics literature. They can potentially partially
reconcile the findings regarding regret and surprise as discussed above.25
[Insert Table VI about here.]
4.6
Discussion
In conventional theories of economics and finance, the past should only be relevant for future
decisions in its informational value, and thus for this purpose no distinction is made between experienced and observed past outcomes. However, empirical and experimental studies document clear
departures from these predictions. The cohort effect offers an example: people who experienced
the great depression exhibited different risk preferences and investment choices compared with
those who only learned from textbooks [Malmendier and Nagel, 2009; Malmendier et al., 2011].
Investors tend to place too much weight on their own experience for various reasons, such as the
difficulty in collecting and processing descriptive information, or simply emotions [Shiller, 2002].
In this paper, we find strong evidence for the law of effect in experimental asset buying tasks
with full descriptive information and complete feedback. This effect can potentially explain some
empirical puzzles in finance, such as investors’ unjustifiably erecting and eradicating certain investment styles [Barberis and Shleifer, 2003]. For another instance, Feng and Seasholes [2005]
find that individual investors’ sophistication and trading experience combined completely eliminated their tendency to hold losers, but only 37% of their tendency to sell winners, for which they
could not find a plausible explanation. This might be driven by naı̈ve reinforcement learning or
25
A comparison of beliefs in Condition B Round 3, where there was no Buying task, with the other rounds revealed
the following: in Condition B Round 3 the elicited beliefs had a correlation coefficient of 0.1297 (p < 0.01) with
Bayesian beliefs; and this correlation coefficient was only 0.0855 (p < 0.01) for the rest of rounds where participants
had both Buying and Prediction tasks. This is another piece of evidence for experienced outcomes to potentially drive
beliefs away Bayesian when participants had stakes involved in the Buying task.
24
asymmetric inertia: selling winners always leads to capital gains, which positively reinforces the
action, while holding on to losers leads to mixed results and the negative outcomes could make
it unattractive due to loss aversion. More empirical and experimental studies should address the
relation between experience-based learning and the role of experience in disposition effect.
We show that in presence of both descriptions and feedback, decision makers may overweight
experience regardless of the descriptive details. The relative importance of descriptions and experience can be exploited to help individual investors understand the risks associated with financial
products [Kaufmann et al., 2013; Bradbury et al., 2014]. Additionally, we find that people attached
considerable weight to recent experience. This can cause the reinforcement strategy to be lossinducing, because personal recent experience is just a small and possibly unrepresentative sample
of the underlying distribution. The management literature argues that reliance on experience may
lead managers to overlook long-term objectives, leading to failure [Levinthal and March, 1993].
Rare disastrous events may not exist in recent experience, leading people to under-predict a crisis,
hence the black swan argument in Taleb [2007], such as during the formation of an asset price
bubble.
Additionally, beliefs were distorted by experience in our experiment. In the literature, belief
distortions can be caused by the desirability of these events [see e.g. Mayraz, 2011], or by the
similarity with past cases in memory [Billot et al., 2005]. However, little has been done with
regard to belief distortion induced by different experienced payoffs associated with states in the
past. We uncovered some evidence that beliefs were more optimistic on assets that offered positive
obtained payoffs but more pessimistic on unchosen assets, and thus ownership mattered. This could
be driving our participants to repurchase assets from which they obtained positively surprising
payoffs, and not to shift to assets that offered high forgone payoffs. Cognitive dissonance can
also play a role in belief formation here. These patterns of beliefs call for further theoretical and
empirical works.
25
5
Conclusion
This paper uses an experimental investment task with full descriptive information and feed-
back to investigate the role of experience in individual investment decisions under uncertainty.
This issue has not been well-understood and has important implications for theories and empirical
studies of individual investor behavior. The experiment creates an environment with clear tension
between Bayesian strategy and the experience-based reinforcement strategy, so that experienced
outcomes has no value to a Bayesian agent with complete knowledge of the incentive structure.
In the experiment, participants made buying decisions and price predictions after observing some
price sequences. We document a double-channeled mechanism through which experience influence decisions. After beliefs are controlled for, participants made decisions consistent with the
law of effect, repurchasing assets on which they had good prior experience, especially placing
more weight on recent outcomes. Experienced outcomes had additional explanatory power than
observed ones. On the other hand, beliefs were also biased by obtained and forgone payoffs.
Additionally, we find that larger regret, positively surprising obtained payoff and negatively
surprising forgone payoff did not significantly decrease the propensity to repurchase in the subsequent period; a positive surprise from obtained payoff actually increased repurchases. It is possible
that in more complex settings, such as in financial context, certain learning patterns may be different from those documented in the clicking paradigm. The biased beliefs can potentially explain
these differences: after an asset had a recent value appreciation, beliefs about future price increase
tended to be biased upward when the gain was obtained, and downward when the gain was forgone. The experimental results in this paper call for more research efforts on learning in individual
decisions, especially in financial contexts, and on experience-induced belief distortions.
D EPARTMENT OF E CONOMICS , U NIVERSITY OF OXFORD
26
References
A BDELLAOUI , M., O. L’H ARIDON AND C. PARASCHIV, “Experienced vs. described uncertainty:
Do we need two prospect theory specifications?,” Management Science 57 (2011), 1879–1895.
A SPAROUHOVA , E., M. H ERTZEL AND M. L EMMON, “Inference from streaks in random outcomes: Experimental evidence on beliefs in regime shifting and the law of small numbers,”
Management Science 55 (2009), 1766–1782.
BARBERIS , N.
161–199.
AND
A. S HLEIFER, “Style investing,” Journal of Financial Economics 68 (2003),
BARRON , G. AND I. E REV, “Small feedback-based decisions and their limited correspondence to
description-based decisions,” Journal of Behavioral Decision Making 16 (2003), 215–233.
B EN -Z ION , U., I. E REV, E. H ARUVY AND T. S HAVIT, “Adaptive behavior leads to underdiversification,” Journal of Economic Psychology 31 (2010), 985–995.
B ILLOT, A., I. G ILBOA , D. S AMET AND D. S CHMEIDLER, “Probabilities as Similarity-Weighted
Frequencies,” Econometrica 73 (2005), 1125–1136.
B RADBURY, M. A., T. H ENS AND S. Z EISBERGER, “Improving investment decisions with simulated experience,” Review of Finance (2014), rfu021.
B ROWN , G. W., “Iterative solution of games by fictitious play,” Activity Analysis of Production
and Allocation 13 (1951), 374–376.
C ALLANDER , S., “Searching and learning by trial and error,” American Economic Review 101
(2011), 2277–2308.
C AMERER , C. AND T. H O, “Experience-weighted attraction learning in normal form games,”
Econometrica 67 (1999), 827–874.
C HARNESS , G. AND D. L EVIN, “When optimal choices feel wrong: A laboratory study of
Bayesian updating, complexity, and affect,” American Economic Review (2005), 1300–1309.
C HIANG , Y.-M., D. H IRSHLEIFER , Y. Q IAN AND A. E. S HERMAN, “Do investors learn from
experience? evidence from frequent ipo investors,” Review of Financial Studies (2011), hhq151.
C HOI , J. J., D. L AIBSON , B. C. M ADRIAN AND A. M ETRICK, “Reinforcement learning and
savings behavior,” The Journal of Finance 64 (2009), 2515–2534.
C OOPER , D. J. AND J. H. K AGEL, “Learning and transfer in signaling games,” Economic Theory
34 (2008), 415–439.
D HAR , R. AND N. Z HU, “Up close and personal: Investor sophistication and the disposition
effect,” Management Science 52 (2006), 726–740.
E REV, I. AND E. H ARUVY, “Generality, repetition, and the role of descriptive learning models,”
Journal of Mathematical Psychology 49 (2005), 357–371.
27
———, “Learning and the economics of small decisions,” in J. H. Kagel and A. E. Roth, eds., The
handbook of experimental economics (2013).
E REV, I. AND A. E. ROTH, “Predicting how people play games: Reinforcement learning in experimental games with unique, mixed strategy equilibria,” American Economic Review (1998),
848–881.
F ENG , L. AND M. S. S EASHOLES, “Do investor sophistication and trading experience eliminate
behavioral biases in financial markets?,” Review of Finance 9 (2005), 305–351.
F OSTER , D. P. AND H. P. YOUNG, “Regret testing: Learning to play Nash equilibrium without
knowing you have an opponent,” Theoretical Economics (2006).
F RIEDMAN , D., S. H UCK , R. O PREA AND S. W EIDENHOLZER, “From imitation to collusion:
Long-run learning in a low-information environment,” Journal of Economic Theory 155 (2015),
185–205.
F UDENBERG , D. AND D. K. L EVINE, The Theory of Learning in Games (Cambridge, MA: The
MIT Press, 1998).
G ERVAIS , S. AND T. O DEAN, “Learning to be overconfident,” Review of Financial studies 14
(2001), 1–27.
G REENWOOD , R. AND S. NAGEL, “Inexperienced investors and bubbles,” Journal of Financial
Economics 93 (2009), 239–258.
H ART, S. AND A. M AS -C OLELL, “A simple adaptive procedure leading to correlated equilibrium,” Econometrica 68 (2000), 1127–1150.
H AU , R., T. J. P LESKAC , J. K IEFER AND R. H ERTWIG, “The description–experience gap in risky
choice: The role of sample size and experienced probabilities,” Journal of Behavioral Decision
Making 21 (2008), 493–518.
H ERRNSTEIN , R. J., “On the Law of Effect1,” Journal of the experimental analysis of behavior
13 (1970), 243–266.
J ESSUP, R. K., A. J. B ISHARA AND J. R. B USEMEYER, “Feedback produces divergence from
prospect theory in descriptive choice,” Psychological Science 19 (2008), 1015–1022.
J IAO , P., “Belief in Mean Reversion and the Disposition Effect: An Experimental Test,” Journal
of Behavioral Finance (forthcoming).
K ARPOFF , J. M., “Costly short sales and the correlation of returns with volume,” Journal of Financial Research 11 (1988), 173–188.
K AUFMANN , C., M. W EBER AND E. H AISLEY, “The role of experience sampling and graphical
displays on one’s investment risk appetite,” Management Science 59 (2013), 323–340.
K AUSTIA , M. AND S. K N ÜPFER, “Do investors overweight personal experience? Evidence from
IPO subscriptions,” The Journal of Finance 63 (2008), 2679–2702.
28
———, “Peer performance and stock market entry,” Journal of Financial Economics 104 (2012),
321–338.
K ELLY, M., “All their eggs in one basket: Portfolio diversification of US households,” Journal of
Economic Behavior &amp; Organization 27 (1995), 87–96.
L EJARRAGA , T. AND C. G ONZALEZ, “Effects of feedback and complexity on repeated decisions
from description,” Organizational Behavior and Human Decision Processes 116 (2011), 286–
295.
L EVINTHAL , D. A. AND J. G. M ARCH, “The myopia of learning,” Strategic management journal
14 (1993), 95–112.
L OOMES , G. AND R. S UGDEN, “Regret theory: An alternative theory of rational choice under
uncertainty,” The economic journal (1982), 805–824.
M AC C RIMMON , K. R. AND D. A. W EHRUNG, “Characteristics of risk taking executives,” Management science 36 (1990), 422–435.
M ALMENDIER , U. AND S. NAGEL, “Depression babies: Do macroeconomic experiences affect
risk-taking?,” Technical Report, National Bureau of Economic Research, 2009.
M ALMENDIER , U., G. TATE AND J. YAN, “Overconfidence and early-life experiences: the effect
of managerial traits on corporate financial policies,” The Journal of finance 66 (2011), 1687–
1733.
M AYRAZ , G., “Wishful Thinking,” Available at SSRN 1955644 (2011).
M OOKHERJEE , D. AND B. S OPHER, “Learning behavior in an experimental matching pennies
game,” Games and Economic Behavior 7 (1994), 62–91.
NAX , H. H., M. N. B URTON -C HELLEW, S. A. W EST AND H. P. YOUNG, “Learning in a black
box,” Department of Economics WP, University of Oxford 653 (2013).
N EVO , I. AND I. E REV, “On surprise, change, and the effect of recent outcomes,” Frontiers in
psychology 3 (2012).
N ICOLOSI , G., L. P ENG AND N. Z HU, “Do individual investors learn from their trading experience?,” Journal of Financial Markets 12 (2009), 317–336.
O FFERMAN , T., J. S ONNEMANS , G. VAN DE K UILEN AND P. P. WAKKER, “A truth serum
for non-bayesians: Correcting proper scoring rules for risk attitudes,” The Review of Economic
Studies 76 (2009), 1461–1489.
R AKOW, T. AND B. R. N EWELL, “Degrees of uncertainty: An overview and framework for future
research on experience-based choice,” Journal of Behavioral Decision Making 23 (2010), 1–14.
R APOPORT, A., D. A. S EALE AND J. E. PARCO, “Coordination in the aggregate without common
knowledge or outcome information,” in Experimental business research (Springer, 2002), 69–
99.
29
S HANI , Y. AND M. Z EELENBERG, “When and why do we want to know? How experienced regret
promotes post-decision information search,” Journal of Behavioral Decision Making 20 (2007),
207–222.
S HEFRIN , H. AND M. S TATMAN, “The disposition to sell winners too early and ride losers too
long: Theory and evidence,” The Journal of finance 40 (1985), 777–790.
S HILLER , R., “Bubbles, human judgment, and expert opinion,” Financial Analysts Journal 58
(2002).
S TRAHILEVITZ , M. A., T. O DEAN AND B. M. BARBER, “Once burned, twice shy: How naı̈ve
learning, counterfactuals, and regret affect the repurchase of stocks previously sold,” Journal of
Marketing Research 48 (2011), S102–S120.
S UPPES , P. AND R. C. ATKINSON, Markov learning models for multiperson interactions, volume 5 (Stanford University Press, 1960).
TALEB , N. N., The Black Swan: The Impact of the Highly Improbable (Random House Group,
2007).
T HORNDIKE , E. L., “Animal intelligence: An experimental study of the associative processes in
animals,” Psychological Monographs: General and Applied 2 (1898), i–109.
W EBER , E. U., A.-R. B LAIS AND N. E. B ETZ, “A domain-specific risk-attitude scale: Measuring
risk perceptions and risk behaviors,” Journal of behavioral decision making 15 (2002), 263–290.
W EBER , M. AND C. F. C AMERER, “The disposition effect in securities trading: An experimental
analysis,” Journal of Economic Behavior and Organization 33 (1998), 167–184.
YOUNG , H. P., “Learning by trial and error,” Games and economic behavior 65 (2009), 626–643.
Z EELENBERG , M., J. J. I NMAN AND R. G. P IETERS, “What we do when decisions go awry: Behavioral consequences of experienced regret,” Conflict and tradeoffs in decision making (2001),
136–155.
Z EELENBERG , M., W. W. VAN D IJK , A. SR M ANSTEAD AND J. DER P LIGT, “The experience
of regret and disappointment,” Cognition and Emotion 12 (1998), 221–230.
30
purchased, and 0 otherwise. The independent variables include the following: Belief is the
elicited belief; CumProfitPos and RecProfitPos are binary variables that are equal to 1 for
positive cumulative profit or recent profit respectively; LastUp is a binary variable that is 1 if the
last outcome was up; MoreUp is a binary variable if the sequence contained more ups than
downs; C (or D) is a binary variable for Condition C (or D) Rounds 1 and 2. Robust standard
errors are in parentheses.
(*** p<0.01, **
* p<0.10)
Table I: D ETERMINANTS
OFp<0.05,
B UYING
D ECISIONS : L AW OF E FFECT
(1)
Belief
8.4898
***
(0.9376)
CumProfitPos
(2)
8.0263
(3)
***
(0.8864)
8.1389
***
(0.8919)
(4)
7.8456
***
(5)
7.8433***
(0.8728)
(0.8720)
0.8253***
0.8687***
(0.1083)
(0.1180)
0.1336
0.1328
(0.0940)
(0.1091)
(0.1098)
0.4443***
0.4397***
0.4381***
(0.1508)
(0.1436)
(0.1435)
0.1376
(0.0915)
0.8850***
RecProfitPos
(0.0987)
LastUp
0.3032
MoreUp
***
C×RecProfitPos
-0.2692
(0.2156)
D×RecProfitPos
-0.0784
(0.2371)
N
LL
Pseudo R
2
12956
12956
12956
12956
12956
-5563.16
-5438.35
-5611.40
-5393.70
-5392.24
0.2117
0.2294
0.2195
0.2357
0.2359
Note: This table reports the results of testing effect of experience and description on buying
decisions, using conditional logistic regression grouped by subjects, and a decision on each asset
in each period as an observation. The dependent variable is Buy, which is equal to 1 if the asset
is purchased, and 0 otherwise. The independent variables include the following: Belief is the
elicited belief; CumP rof itP os and RecP rof itP os are binary variables that are equal to 1 for
positive cumulative profit or recent profit respectively; LastU p is a binary variable that is 1 if the
last outcome was up; M oreU p is a binary variable if the sequence contained more ups than downs;
C (or D) is a binary variable for Condition C (or D) Rounds 1 and 2. Robust standard errors are
in parentheses. (*** p < 0.01, ** p < 0.05, * p < 0.10)
31
Belieft-1 is elicited probabilistic belief on the asset purchased in the previous period; CumProfit is
equal to 1 for positive cumulative profits on an asset up to period t-1, and 0 otherwise; Increaset
is equal to 1 when the payoff increased from period t-2 to t-1; and there are four binary variables,
one for each possible period t-1 profit, 3, 5, -3 and -5. Regressions (1) and (2) are in the domain
of gains; (3) and (4) are in the domain of losses. Robust standard errors are in parentheses. (***
p<0.01, ** p<0.05,
* p<0.10)
Table II:
D ETERMINANTS OF R EPURCHASES : A SYMMETRIC I NERTIA
After Gains
(1)
Belieft-1
0.0287
***
(0.0060)
CumProfitPos
0.9831
***
(0.1779)
RecProfit=3
0.3566
**
(0.1492)
RecProfit=5
0.5022
***
(0.1855)
After Losses
(2)
0.0290
***
(0.0059)
1.0197
***
(0.1811)
0.2809
(3)
0.0259
***
(0.0081)
(0.0082)
-0.2824
-0.2807
(0.2019)
(0.1998)
-0.0826
-0.0803
(0.1522)
0.3330*
(0.1957)
(0.1880)
RecProfit=-5
-0.5125
**
(0.2281)
0.4482
***
LL
Pseudo R
2
(0.1928)
-0.5073**
(0.2511)
0.0134
(0.1291)
N
0.0259***
*
RecProfit=-3
Increaset
(4)
(0.1766)
1540
1540
1108
1108
-756.95
-750.26
-430.90
-430.90
0.1137
0.1215
0.0539
0.0539
Note: This table shows the results of testing the effect of gains and losses on repurchasing,
using conditional logistic regression grouped by subjects, and a decision in each period as an
observation. The dependent variable is Repurchaset , which is equal to 1 if the asset bought in
previous period was repurchased, and 0 otherwise. Independent variables include the following:
Belieft−1 is elicited probabilistic belief on the asset purchased in the previous period; CumP rof it
is equal to 1 for positive cumulative profits on an asset up to period t-1, and 0 otherwise; Increaset
is equal to 1 when the payoff increased from period t−2 to t−1; and there are four binary variables,
one for each possible period t − 1 profit, 3, 5, -3 and -5. Regressions (1) and (2) are in the domain
of gains; (3) and (4) are in the domain of losses. Robust standard errors are in parentheses. (***
p < 0.01, ** p < 0.05, * p < 0.10)
32
previously defined, as well as the following: Regret is the difference between obtained payoff
and the largest forgone payoff in period t-1; SObtained is the difference between the obtained
payoff and the participants’ expected payoff from asset k in period t-1; SForgone is the
difference between the forgone payoff and the expected payoff from an unchosen asset k in
period t-1. Regressions (1) to (3) are in the domain of gains; (4) to (6) are in the domain of losses.
Robust standard
in parentheses.
p<0.01, ** :p<0.05,
* p<0.10)
Tableerrors
III: Dare
ETERMINANTS
OF(***
R EPURCHASES
R EGRET
AND S URPRISE
After Gains
(1)
Belieft-1
0.0597
***
(0.0083)
CumProfitPos
1.0834
***
(0.1878)
After Losses
(2)
0.0599
***
(0.0079)
0.9685
***
(0.1822)
RecProfit=3
(3)
0.0598
***
(0.0079)
0.9651
***
(0.1838)
(4)
0.0791
***
(5)
0.0793
***
(6)
0.0791***
(0.0180)
(0.0178)
(0.0181)
0.0636
0.0652
0.0283
(0.2368)
(0.2316)
(0.2332)
0.1079
(0.1597)
RecProfit=5
0.1163
(0.2026)
RecProfit=-3
-0.2091
(0.2339)
RecProfit=-5
-0.3491
(0.3414)
Increaset
0.3880
***
(0.1406)
Regret
0.6283
***
(0.1444)
-0.0108
(0.0354)
SObtained
0.0971
0.0273
0.0181
(0.1615)
(0.2201)
(0.1983)
(0.2585)
0.0278
0.0128
0.0435
(0.0475)
(0.0571)
(0.0345)
0.1142
SForgone
0.6504
***
***
0.1145
***
-0.0279
-0.0297
(0.0235)
(0.0232)
(0.0284)
(0.0299)
0.0502
0.0511
0.0084
0.0061
(0.0379)
(0.0381)
(0.0346)
(0.0345)
N
1442
1442
1442
917
917
917
LL
-648.55
-629.10
-628.69
-311.25
-310.78
-310.06
0.1841
0.2085
0.2091
0.1726
0.1738
0.1757
Pseudo R
2
Note: This table reports the results of testing hypotheses involving regret and surprise-trigger
change, using conditional logistic regression grouped by subjects, and a decision in each period as
an observation. The dependent variable is Repurchaset , which is equal to 1 if the asset bought
in period t − 1 was repurchased in period t, and 0 otherwise. Independent variables include those
previously defined, as well as the following: Regret is the difference between obtained payoff and
the largest forgone payoff in period t − 1; SObtained is the difference between the obtained payoff
and the participants expected payoff from asset k in period t − 1; SF orgone is the difference
between the forgone payoff and the expected payoff from an unchosen asset k in period t − 1.
Regressions (1) to (3) are in the domain of gains; (4) to (6) are in the domain of losses. Robust
standard errors are in parentheses. (*** p < 0.01, ** p < 0.05, * p < 0.10)
33
weight on unchosen alternatives; θ is the coefficient of relative risk aversion; α is the weight
placed on reinforcement learning. The parameter estimates using the whole sample are reported
first. Then we use two-thirds of the sample to calibrate the models, and predict choices in the rest
of the sample. For comparison of model fit with varying number of parameters, Log-likelihood
(LL), AIC, and BIC are reported. For evaluation of predictive power, Mean Squared Deviation
(MSD) is reported. (* Table
indicates
best fit.)
IV: the
M ODEL
C ALIBRATION AND VALIDATION
Models
(1) Baseline
(2) RL1
(3) RL2
(4) CRRA
(5) CRRRA
(6) CRRA
+RL1
+RL2
Parameter Estimates (Whole Sample, N=3094)
Parameters
0
ϕ
1
2
0.1211
(0.2490)
δ
1
3
4
0.1650
0.1887
0.2123
(0.3080)
(0.3140)
(0.3382)
0.2330
0.4418
(0.2844)
(0.3938)
θ
0.6240
0.5085
0.5882
(0.3965)
(0.4527)
(0.4337)
0.3195
0.3932
(0.2532)
(0.3222)
α
LL
-5116.41
-5867.66
-5621.28
-3799.55
-3542.18
-3245.08
AIC
-5116.41
-5868.66
-5623.28
-3800.55
-3545.18
-3249.08*
BIC
-5116.41
-5871.68
-5629.32
-3803.57
-3554.24
-3261.15*
Calibration (Sample Size=2078)
LL
-3547.12
-3954.82
-3798.36
-2593.74
-2421.84
-2217.76
AIC
-3547.12
-3955.82
-3800.36
-2594.74
-2424.84
-2221.76*
BIC
-3547.12
-3958.64
-3806.00
-2597.56
-2433.30
-2233.04*
Validation (Sample Size=1016)
LL
-1569.29
-1999.78
-1901.29
-1226.17
-1167.93
-1092.09
MSD
0.1982
0.1965
0.1824
0.1627
0.1533
0.1428*
Note: This table reports the results of the performance of six models: (1) baseline model: risk
neutral expected utility with elicited beliefs; (2) reinforcement learning with one parameter; (3)
reinforcement learning with two parameters; (4) CRRA expected utility with beliefs corrected for
risk attitudes; (5) CRRA expected utility with corrected beliefs, and reinforcement learning with
one parameter; (6) CRRA expected utility with corrected beliefs, and reinforcement learning with
two parameters. Parameters: φ is the discount rate of past reinforcements; δ is the relative weight
on unchosen alternatives; θ is the coefficient of relative risk aversion; α is the weight placed on
reinforcement learning. The parameter estimates using the whole sample are reported first. Then
we use two-thirds of the sample to calibrate the models, and predict choices in the rest of the
sample. For comparison of model fit with varying number of parameters, Log-likelihood (LL),
AIC, and BIC are reported. For evaluation of predictive power, Mean Squared Deviation (MSD) is
reported. (* indicates the best fit.)
34
(RecProfit). The table also compares Signed Reaction Measure (SRM) of beliefs on an asset
unchosen in the last period after its positive and negative forgone payoffs. SRM is defined as the
elicited belief minus 0.5, multiplied by -1 if the previous outcome was a price decrease, and not
if the previous outcome was a price increase. Negative SRM indicates belief in mean reversion.
In each case, the comparison results of Bayesian beliefs (Bayesian) are also reported, as a
benchmark. The p-values from t-tests
Standard
errors are in parentheses.
Table are
V: Sreported.
UMMARY
OF B ELIEFS
CumProfit
Positive
Negative
N
3584
2814
Bayesian
0.5044
0.5033
(0.0554)
(0.0561)
0.4530
0.4389
(0.2134)
(0.2172)
Elicited Belief
RecProfit
p-value
p=0.43
p<0.01
Positive
Negative
1954
975
0.5165
0.5153
(0.0555)
(0.0584)
0.5038
0.4562
(0.2326)
(0.2219)
p-value
p=0.59
p<0.01
Forgone Payoff
Positive
Negative
N
3104
4438
Bayesian
0.0154
0.0145
(0.0514)
(0.0521)
SRM of
-0.0429
0.0556
Elicited Belief
(0.2084)
(0.2321)
p=0.46
p<0.01
Note: This table reports the comparison results of elicited beliefs after positive and negative
cumulative profits on an asset (CumP rof it), after positive and negative last period profit on an
asset (RecP rof it). The table also compares Signed Reaction Measure (SRM ) of beliefs on an
asset unchosen in the last period after its positive and negative forgone payoffs. SRM is defined
as the elicited belief minus 0.5, multiplied by -1 if the previous outcome was a price decrease,
and not if the previous outcome was a price increase. Negative SRM indicates belief in mean
reversion. In each case, the comparison results of Bayesian beliefs (Bayesian) are also reported,
as a benchmark. The p-values from t-tests are reported. Standard errors are in parentheses.
35
Independent variables include the following: Bayesian is the benchmark beliefs calculated using
the Bayes’ Rule; CumProfitPos is equal to 1 if the cumulative profit on an asset was positive,
and 0 otherwise; RecProfitPos is equal to 1 if the recent profit on an asset was positive, and 0
otherwise; LastUp is equal to 1 if the last observed outcome was a price increase; ForgoneHigh
is equal to 1 if the forgone payoff from the asset was higher than the obtained payoff last period.
Standard errors clustered by subjects
(*** p<0.01, ** p<0.05, * p<0.10)
Tableare
VI:inBparentheses.
ELIEF D ISTORTION
(1)
Bayesian
0.3059
***
(0.1026)
CumProfitPos
0.0216
(2)
0.2473
***
(0.0920)
*
(0.0119)
RecProfitPos
(3)
0.1713
*
0.1016
(0.0100)
LastUp
0.2047**
(0.0939)
(0.0918)
0.0123
0.0121
(0.0120)
***
(4)
0.0770
***
(0.0093)
0.0519
***
(0.0125)
(0.0119)
0.0671***
(0.0092)
0.0534***
(0.0124)
-0.0246***
ForgoneHigh
(0.0070)
Constant
N
R
2
0.2823
***
0.3054
***
0.3166
***
0.3079***
(0.0517)
(0.0464)
(0.0476)
(0.0466)
13432
13432
13432
13432
0.0085
0.0305
0.0440
0.0465
Note: This table reports the results of testing the determinants of elicited beliefs, using OLS regressions. The dependent variable is Belief , the elicited belief about price going up on each asset.
Independent variables include the following: Bayesian is the benchmark beliefs calculated using
the Bayes Rule; CumP rof itP os is equal to 1 if the cumulative profit on an asset was positive,
and 0 otherwise; RecP rof itP os is equal to 1 if the recent profit on an asset was positive, and 0
otherwise; LastU p is equal to 1 if the last observed outcome was a price increase; F orgoneHigh
is equal to 1 if the forgone payoff from the asset was higher than the obtained payoff last period.
Standard errors clustered by subjects are in parentheses. (*** p < 0.01, ** p < 0.05, * p < 0.10)
36
W X
X Y Z W X
Y Z X Panel A
W X
X Y Z W X
X X
X Y Z Panel B
Figure 1 Experiment Screenshots
Note: This figure shows two Figure
example
from
the experimental computer interface.
I: screenshots
E XPERIMENT
S CREENSHOTS
Panel This
A is the
screen
participants
sawscreenshots
in Period 1from
of Round
1 in Stage 2computer
of the experiment,
Note:
figure
shows
two example
the experimental
interface. when
Panel
they
were
asked
to
make
the
buying
decision.
Panel
B
is
the
screen
that
announced
the
result
of
A is the screen participants saw in Period 1 of Round 1 in Stage 2 of the experiment, when they
Periodasked
1 of to
Round
after
that period
concluded.
were
make1the
buying
decision.
Panel B is the screen that announced the result of Period
1 of Round 1 after that period concluded.
37
.6
Proportions Repurchased
.3
.4
.5
.2
Negative
Positive
Recent Profits
Negative
Positive
Cumulative Profits
Figure 2 Proportions Purchased Given Positive and Negative Experience
Note:
is a barPchart
showing
the proportions
purchased
out ofE XPERIENCE
opportunities to
FigureThis
II: Pfigure
ROPORTIONS
URCHASED
G IVEN
P OSITIVE AND
N EGATIVE
purchase
an asset
at chart
(1) positive
recent
profit, out
andof(2)
positive and
negative
Note:
This figure
is a bar
showingand
the negative
proportions
purchased
opportunities
to purchase
cumulative
profit
in
all
previous
periods
in
a
given
round.
The
95%
confidence
interval
bars
an asset at (1) positive and negative recent profit, and (2) positive and negative cumulative profit
in
were
added.
all previous periods in a given round. The 95% confidence interval bars were added.
38
.7
Proportions Repurchased
.2
.3
.4
.5
.6
.1
0
-5
-3
-1
1
Recent Profit
Actual Values
Fited Values
3
5
95% CI of Actual
95% CI of Fitted
.3
.4
Proportions
.5
.6
Panel A
Payoff Decrease
Payoff Increase
Payoff Decrease
Repurchases/Opportunities
Payoff Increase
Bayesians/Repurchases
Panel B
Figure 3 Asymmetric Inertia Figure III: A SYMMETRIC I NERTIA
Note: Panel
PanelAAis aisbar
a chart
bar chart
showing
the proportion
repurchased
out of opportunities
to
Note:
showing
the proportion
repurchased
out of opportunities
to repurchase
repurchase
after each
possible
payoff
fromperiod,
the previous
period,
with 95%
after
each possible
payoff
from the
previous
{−5, −3,
−1, 1, {-5,-3,-1,1,3,5},
3, 5}, with 95% confidence
confidence
interval
bars;
theissolid
line is afit,quadratic
and the
lines
the 95%
interval
bars;
the solid
line
a quadratic
and the fit,
dashed
linesdashed
indicate
theindicate
95% confidence
interval
of the
fitted values.
shows Panel
two things:
(1) the
repurchased
out of
confidence
interval
of the Panel
fitted Bvalues.
B shows
twoproportions
things: (1)
the proportions
opportunities
to
repurchase
after
a
payoff
increase
and
after
a
payoff
decrease;
(2)
the
proportion
repurchased out of opportunities to repurchase after a payoff increase and after a payoff decrease;
of
decisions
the Bayesian
strategy
repurchases
after a out
payoff
(2)repurchase
the proportion
of consistent
repurchasewith
decisions
consistent
withouttheof Bayesian
strategy
of
increase
and
after
a
payoff
decrease,
with
95%
confidence
interval
bars.
repurchases after a payoff increase and after a payoff decrease, with 95% confidence interval
bars.
39
.7
.6
Proportions Bayesian
.2
.3
.4
.5
.1
0
0
2
4
6
Last Period Regret
Actual Values
Fited Values
8
10
95% CI of Actual
95% CI of Fitted
0
.1
Proportions Repurchased
.2
.3
.4
.5
.6
.7
Panel A
0
2
4
6
Last Period Regret
Actual Values
Fited Values
8
10
95% CI of Actual
95% CI of Fitted
Panel B
Figure 4 The Effect of Regret Figure IV: T HE E FFECT OF R EGRET
Note: Regret
Regret isis defined
definedasasthe
theabsolute
absolutevalue
valueofof
difference
between
obtained
payoff
Note:
thethe
difference
between
thethe
obtained
payoff
and and
the
the
maximum
forgone
payoff
in
the
previous
period.
Potential
magnitudes
of
regret
in
the
maximum forgone payoff in the previous period. Potential magnitudes of regret in the experiment
experiment
{0,2,4,6,8,10}.
A is showing
a bar chart
showing the
proportions
decisions
are
{0, 2, 4, 6,are
8, 10}.
Panel A is Panel
a bar chart
the proportions
decisions
consistent
with
Bayesian
after predictions
different magnitudes
of regret,
with 95%ofconfidence
interval
Panel
consistentpredictions
with Bayesian
after different
magnitudes
regret, with
95% bars.
confidence
B
is a barbars.
chartPanel
showing
of assets the
repurchased
out of opportunities
to repurchase
interval
B isthea proportions
bar chart showing
proportions
assets repurchased
out of
after
different
magnitudes
of
regret,
with
95%
confidence
interval
bars.
In
both
panels,
the
solid
opportunities to repurchase after different magnitudes of regret, with 95% confidence interval
line
a quadratic
fit, and
dashed
thefit,
95%
confidence
interval
the fittedthe
values.
bars.isIn
both panels,
thethe
solid
linelines
is a indicate
quadratic
and
the dashed
linesofindicate
95%
confidence interval of the fitted values.
40