A Leaky Competing Accumulator Model of Lexical Decision

tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
Journal of Experimental Psychology:
Learning, Memory, and Cognition
2012, Vol. ●●, No. ●, 000 – 000
© 2012 American Psychological Association
0278-7393/12/$12.00 DOI: 10.1037/a0026948
OBSERVATION
How to Say “No” to a Nonword:
A Leaky Competing Accumulator Model of Lexical Decision
Stéphane Dufau, Jonathan Grainger, and Johannes C. Ziegler
Centre National de la Recherche Scientifique, Marseille, France, and Aix-Marseille University
We describe a leaky competing accumulator (LCA) model of the lexical decision task that can be used
as a response/decision module for any computational model of word recognition. The LCA model uses
evidence for a word, operationalized as some measure of lexical activity, as input to the YES decision
node. Input to the NO decision node is simply a constant value minus evidence for a word. In this way,
evidence for a nonword is a function of time from stimulus onset (as in standard deadline models)
modulated by lexical activity via the competitive dynamics of the LCA. We propose a simple mechanism
for determining the value of this constant online during the first trials of a lexical decision experiment,
such that the model can rapidly optimize speed and accuracy in discriminating words from nonwords.
Further optimization is achieved via trial-by-trial adjustments in response criteria as a function of task
demands and list context. We show that the LCA model can simulate mean response times and response
distributions for correct and incorrect YES and NO decisions for a number of benchmark experiments that
have been shown to be fatal for deadline models of lexical decision. Finally, using lexical activity
calculated by a computational model of word recognition as input to the LCA decision module, we
provide the first item-level simulation of both word and nonword responses in a large-scale database.
Keywords: lexical decision, leaky competing accumulator, response criteria, item-level simulations,
nonword responses
of data obtained with this task (see Balota & Chumbley, 1984).
One solution to this general problem, proposed by Grainger and
Jacobs (1996), is to develop computational models that incorporate
a key distinction between task-specific and task-independent core
processes. In doing so, Grainger and Jacobs applied the notion of
functional overlap to illustrate how different tasks used to investigate a given phenomenon can each contribute to uncovering the
basic mechanisms underlying the phenomenon of interest, as long
as the contribution of task-specific mechanisms is well understood.
Continuing this tradition, we argue that lexical decision involves
some very important differences with respect to normal, everyday,
word recognition and that understanding these differences is absolutely critical for making progress in this field. However, we
also believe that when performing the lexical decision task, participants apply very general principles of human information processing expressed in terms of optimization of performance (Anderson, 1990; Oaksford & Chater, 1998). In our opinion, these
general principles do not dispense with a more fine-grained analysis of the specific mechanisms involved in performing a particular task, and it is such fine-grained analyses that advance understanding, once the more general principles have been accepted
(e.g., Perry, Ziegler, & Zorzi, 2010).
The lexical decision task, first used by Rubenstein, Garfield, and
Millikan (1970), remains one of the most widely used behavioral
measures of silent word reading. A recent smartphone application
even enables lexical decision data to be collected from thousands
of participants all over the world (Dufau et al., 2011). Participants
in this task are requested to classify as rapidly and accurately as
possible a given string of letters as being a word or not. Variation
in response times and in accuracy to different word stimuli is taken
to reflect variation in the ease with which these words are read.
Performance in the lexical decision task has therefore played a
primary role in developing and testing theories of visual word
recognition. However, as with that for any other measure of word
recognition, lexical decision performance is prone to measurement
biases and response strategies, which complicate the interpretation
Stéphane Dufau, Jonathan Grainger, and Johannes C. Ziegler, Centre
National de la Recherche Scientifique and Laboratoire de Psychologie
Cognitive, Aix-Marseille University, Marseille, France.
This research was supported by European Research Council Advanced
Grant 230313, awarded to Jonathan Grainger. We thank Max Coltheart,
Pablo Gomez, and Eric-Jan Wagenmakers for constructive comments on an
earlier version of this work. Matlab code for the LCA decision module is
available from Stéphane Dufau ([email protected]).
Correspondence concerning this article should be addressed to Jonathan
Grainger, Laboratoire de Psychologie Cognitive, Aix-Marseille Université, 3,
Place Victor Hugo, 13331 Marseille, France. E-mail: [email protected]
Lexical Decision as Evidence Accumulation
The present work follows the tradition of recent accounts of
performance in the lexical decision task (Norris, 2006; Ratcliff,
Gomez, & McKoon, 2004; Wagenmakers, Ratcliff, Gomez, &
1
tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
DUFAU, GRAINGER, AND ZIEGLER
2
Fn1
The Present Approach
a straightforward implementation of what could be conceptualized
as a dynamic deadline account of lexical decision and, therefore,
an alternative to standard deadline models, such as the multiple
read-out model (MROM; Grainger & Jacobs, 1996). One key
aspect of the MROM is that the temporal deadline is adjusted as a
function of lexical activity generated by the stimulus, such that the
deadline is adjusted upward with greater values of lexical activity
and downward with lower values of lexical activity (see also
Coltheart, Davelaar, Jonasson, & Besner, 1977). In this way, the
temporal deadline in the MROM is essentially a mechanism for
combining evidence for a word with elapsed time (time from
stimulus onset) in order to generate a NO response. A NO response
is generated if not enough evidence for a YES response has
accumulated before the deadline has been reached. However, Ratcliff et al. (2004) and, more recently, Wagenmakers et al. (2008)
have outlined a number of deficiencies with this particular mechanism. The deficiencies that were pinpointed in these studies are
summarized in Table 1. All concern either correct response times
(RTs) to nonword stimuli or error RTs to word stimuli. Thus, these
deficiencies are all related to performance driven by the temporal
deadline mechanism. Here we describe a solution for generating
NO responses in the LCA model that is conceptually similar to the
temporal deadline mechanism. We examine whether this dynamic
deadline model overcomes the deficiencies listed by Ratcliff et al.
(2004) and Wagenmakers et al. (2008) with respect to standard
deadline models such as the MROM (Grainger & Jacobs, 1996).
A primary goal in the present work is to understand the mechanism used by participants in a lexical decision task to discriminate
words from nonwords. After all, that is the definition of the lexical
decision task. As described above, deadline models provide a
straightforward solution here. For models like the diffusion model
and the Bayesian Reader, the solution is not so straightforward
(e.g., the virtual nonword mechanism of the Bayesian Reader).
Here we offer an easily implementable solution within the framework of LCA models.
Another important goal in the present study is to investigate the
role of adjustments in response criteria within the LCA framework
as a means of capturing effects of list context in the lexical
decision task. This second goal is a straightforward extension of
the criterion adjustment approach developed by Grainger and
Jacobs (1996) in their account of list-context effects on lexical
decision (see also Perea, Carreiras, & Grainger, 2004). The approach adopted here is not to optimize data fitting via parameter
tuning routines but to understand changes in model performance
(mimicking changes in human performance) by controlled adjustments of a handful of parameters. Put differently, the philosophy
of our approach is not to seek which parameters need to be
adjusted in order to accurately fit data but rather to test whether
theoretically motivated adjustments of parameters allow the model
to account for the data.2
Finally, it is important to note that our aim is not to compare an
LCA model of lexical decision with other eminent approaches,
In the present work we apply an alternative framework for
understanding speeded binary decision making, the leaky competing accumulator (LCA) model of Usher and McClelland (2001).
The specific architecture of the LCA model provides an ideal
framework for conceptualizing the basic mechanisms involved in
making a lexical decision. In particular, the LCA framework offers
1
The lexical decision task is not the only task in this category of binary
decisions, one other prominent example being the standard old–new decision task used to study recognition memory.
2
We are not claiming that our approach is superior in any way, just that
the focus is different.
McKoon, 2008) in applying the basic notion of noisy accumulation
of evidence over time. Input to the decision process is therefore a
measure of evidence in favor of a word response and evidence in
favor of a nonword response. However, although it is fairly
straightforward to specify what might constitute evidence for a
word, defining what might be evidence for a nonword is not so
straightforward. This is one specificity of the lexical decision task,
as opposed to other binary decision tasks, such as deciding whether
a stimulus is moving to the left or to the right, is a square or a
circle, or is red or blue. In a sense, the problem here is that
evidence for a nonword is just less of the same measure that is used
to provide evidence for a word. This is unlike binary decision
tasks, where qualitatively different kinds of evidence are associated with the two responses, and it is this particular characteristic
of the lexical decision task that was the major motivation for
deadline accounts of nonword decision making (Grainger & Jacobs, 1996).1
Ratcliff et al. (2004) described a diffusion model account of the
lexical decision task. They assigned a positive value to word
stimuli, whose strength depended on word frequency, and a negative value to nonword stimuli, whose strength was determined by
how much the nonword resembled real words. Of course, this
approach begs the question of how a lexical processor could
actually transform the evidence extracted from a string of letters
into positive and negative values. One possibility would be that the
system learns, via exposure to words and nonwords during practice, the amounts of activation generated by word and nonword
stimuli and, as a function of that experience, sets a zero point on
the activation (wordness) dimension separating positive values
from negative values. This corresponds to the drift rate criterion in
the diffusion model.
Norris’ (2006) model, the Bayesian Reader, performs lexical
decision by calculating the posterior probability that the stimulus is
a word and comparing this with the probability that it is a nonword.
The posterior probability of a word response for a particular
stimulus is calculated using the probabilities of each individual
word, given the stimulus (something akin to a global lexical
activation value). The posterior probability of a nonword response
is calculated by assuming the existence of a single virtual nonword
that has a frequency approximately equal to the average word
frequency (of words that are encountered in an experiment) and
whose distance from the input is greater than the distance of the
nearest word by a value that varies as a function of the similarity
of words and nonwords in an experiment. This implies that participants in a lexical decision experiment must evaluate the average frequency of the nonword stimuli in the experiment. They
could do so by estimating the amount of lexical activation that
nonwords compared with the word stimuli generate, in a way
similar to the setting of the drift rate criterion in the diffusion
model.
T1
Fn2
tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
LCA MODEL OF LEXICAL DECISION
3
Table 1
The Four Deficiencies of the Multiple Read-Out Model (MROM) of Lexical Decision
1. In MROM with word targets, error RTs cannot be faster than correct RTs without having RTs in the leading edge of the distribution (.1 quantile)
that are too fast.
2. In MROM with word targets, as word frequency increases, correct RTs get faster and accuracy increases but error RTs do not get faster.
3. In MROM with nonword targets, the RT distributions for correct RTs are not skewed like the distributions for correct RTs to words.
4. In MROM with nonword targets, correct RTs cannot be faster than error RTs, and in general MROM cannot generate fast correct responses to
nonwords while keeping response accuracy at an acceptable level.
Note. The first three points are from Ratcliff et al. (2004, p. 178), and the fourth point is from Wagenmakers et al. (2008). Italics highlight where the
problems arise. RTs ⫽ response times.
such as the diffusion model (Ratcliff et al., 2004). This particular
comparison has already been made in the work of Ratcliff and
Smith (2004). We also acknowledge the obvious similarities across
these different approaches and the fact that they can be shown to
be mathematically equivalent under certain ranges of parameter
settings (e.g., Bogacz, Brown, Moehlis, Holmes, & Cohen, 2006).
However, we believe that important differences emerge at a conceptual level and that in this respect the LCA framework offers a
more straightforward implementation of the concepts we wish to
express. Most important, the LCA model can be seen as a module
for lexical decision making that can be appended to any computational model of visual word recognition that outputs a measure of
lexical activation for a given string of letters as input (see Ziegler,
Grainger, & Brysbaert, 2010, for a review of such models). In this
way, lexical input values can be fed into the LCA response/
decision module on a trial-by-trial basis, thus providing the opportunity for performing item-level analyses of simulation results.
An LCA Account of Lexical Decision
F1
Our LCA model of lexical decision is shown in Figure 1.
Applied to the specific case of lexical decision, bottom-up input to
the YES response node is evidence for a word that is derived from
the stimulus. In this respect we follow standard practice. What is
original in our approach is that the input to the NO response node
is simply a constant value minus the evidence for a word. Our
model can therefore be thought of as implementing a dynamic
deadline mechanism for nonword decision making, in that, like the
MROM deadline mechanism (Grainger & Jacobs, 1996), the time
it takes to make a nonword response depends only on lexical
activity (i.e., evidence for a word) and elapsed time.
In our LCA model there are two sources of lexical influences on
activity in the NO response node (see Figure 1). The first influence
comes from the standard practice of having total input to the two
response nodes equal to a constant (Usher & McClelland, 2001).3
This implies that in the absence of any positive evidence for a
nonword (the solution adopted here), the bottom-up input to the
NO response node is equal to the constant total input value minus
the evidence for a word extracted from the stimulus. The use of a
total constant input also differentiates the current LCA model from
previous proposals (Davis, 2010; Grainger, Dufau, & Ziegler,
2009), in which input to the NO decision node was only a function
of elapsed time. The second source of lexical influences in our
LCA model arises via the mutually inhibitory connections that
exist between the two response nodes, such that a rise in activity
in one automatically causes a reduction in activity in the other and
vice versa. Both of these mechanisms are fundamental to the LCA
approach in general. Finally, the other ingredient of LCA models
is the “leak” or decay of information that is associated with the
noisy accumulation of information over time.
Implementation
We implemented a linear version of LCA in which accumulators
are updated according to the following equations:
yYES 共i ⫹ 1兲 ⫽ yYES 共i兲 ⫹ ⌬YES 共i兲
(1)
yNO 共i ⫹ 1兲 ⫽ yNO 共i兲 ⫹ ⌬NO 共i兲
(2)
where yYES and yNO are variables representing the activities of the
YES and NO accumulators (here, the calculation of the ith ⫹ 1
cycle), and deltas are variation terms of the corresponding cycle.
Deltas are calculated with the following equations:
⌬ YES 共i兲 ⫽ cYES 共i兲 䡠 关⫺w 䡠 yNO 共i兲 ⫺ k 䡠 yYES 共i兲 ⫹ IYES ⫹ GYES 共i兲兴
(3)
⌬ NO 共i兲 ⫽ cNO 共i兲 䡠 关⫺w 䡠 yYES 共i兲 ⫺ k 䡠 yNO 共i兲 ⫹ INO ⫹ GNO 共i兲兴
(4)
Figure 1. General architecture of the LCA model of lexical decision.
YES and NO decision nodes receive bottom-up information from their
associated inputs: lexical input for the YES node and constrained input
(constant minus YES input) for the NO node. The decision nodes mutually
inhibit each other. LCA ⫽ leaky competing accumulator.
3
Note that the constant total input is not applied here in order to solve
the scaling problem, as is typically done when applying parameter-fitting
routines to LCA models (Donkin, Brown, & Heathcote, 2009). As shown
in Table 2, a majority of the parameters are fixed in our approach.
Furthermore, in the General Discussion we discuss how this constant input
value might change as a function of experimental conditions, hence avoiding the problems noted by Donkin et al.
Fn3
tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
DUFAU, GRAINGER, AND ZIEGLER
4
Figure 2. An illustration of processing in the LCA model. Panel A: YES and NO activities with three parameter
sets illustrating the noisy accumulation of information and the between-node inhibition (10 trials per simulation;
from left to right: YES input ⫽ .12, .11, .08; NO input ⫽ .08, .09, .12). The first parameter set gives a greater
advantage to YES activity than the second one and therefore inhibits NO activity more quickly. The third
parameter set gives an advantage to NO activity. Panel B: Blowup of the second parameter set simulation (left,
YES activity; right, NO activity). Each activity is associated with a decision criterion (black solid lines), whose
value is drawn from a normal distribution (starting mean ⫽ .5; standard deviation ⫽ .05) and is subject to a
trial-by-trial adjustment. A response is made when one of the two activities crosses its threshold, generating a
distribution of simulated response times expressed in cycles for both the YES and NO activities. In this
simulation (Panel B), the parameter set represents weak evidence for a word, and the two distributions
correspond to correct (YES) and error (NO) responses to word targets. LCA ⫽ leaky competing accumulator.
Fn4
where k and w are the leakage and inhibition factors, IYES and INO
are the input to the YES and NO accumulators, GYES and GNO are
Gaussian noise terms associated with activity in the YES and NO
accumulators, and cYES and cNO are bounding factors ensuring that
yYES and yNO values stay within [0;1].4
IYES ⫹ INO is a constant.
F2
(5)
The model therefore makes decisions on the basis of the accumulation of noisy, leaking, and competing information over time (see
Figure 2, Panel A). Whenever the accumulation of activation in
either the YES or the NO decision node reaches the criterion value
specified for that trial by the setting of response thresholds, a
lexical decision response is made and processing is terminated.
The response corresponds to the decision node whose threshold
was reached first, and the response time corresponds to the time
from stimulus onset to the time at which the threshold was reached.
In our simulations, we also maintained the principle of adding
noise to the decision criteria that was an essential ingredient of
MROM (Figure 2, Panel B; Grainger & Jacobs, 1996; Jacobs &
Grainger, 1992). One key aspect of the model is the implementation of trial-by-trial adjustments of response criteria that occur
within a simulation. For example, in simulating a lexical decision
experiment where the focus is on response speed rather than
4
Following the equations of interactive activation (McClelland & Rumelhart, 1981), activation is bounded between 0 and 1, but removing this
constraint does not affect the performance of our LCA model.
tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
LCA MODEL OF LEXICAL DECISION
T2, AQ:1
accuracy, the YES and NO threshold values decrease by a constant
value on each trial as long as a correct response is given by the
model. When an error is made, thresholds are reset to their initial
values. In experiments where task instructions stress accuracy
rather than speed, a smaller constant value for threshold adjustments is used (see Table 2).
Parameter Settings
T3
Following the parameter setting approach adopted by Grainger
and Jacobs (1996), the core parameters of the LCA model were
hand-tuned to a target set of data (Experiments 1 and 2 from
Ratcliff et al., 2004). These parameter values are given in Table 3.
Once the core parameters of our LCA model were fixed, the only
parameters that were allowed to vary in order to accommodate
specific patterns of results were (a) the values of bottom-up input
to the model (to capture changes in stimulus-driven factors) and
(b) the values of response criteria, which were modified on a
trial-by-trial basis (to capture influences of task and list context).
Bottom-Up Input
The different word frequency conditions were simulated by a
gradient of lexical input values ranging from .11 (very lowfrequency words) to .12 (high-frequency words; see Table 2).
Lexical input differed for the two classes of nonword stimuli tested
in the simulations. Orthographically regular, pronounceable pseudowords were assigned a lexical input value of .0875, whereas
nonwords formed of random strings of letters were assigned a
lexical input value of .0825. Input to the NO decision node was
simply a constant value (.2) minus the lexical input. For example,
a lexical input of .11 would generate a NO input of .09 (word
stimulus), and one of .09 would generate a NO input of .11
(nonword stimulus).
Response Criteria Adjustments
As noted above, one key feature of the present theoretical work
is the implementation of trial-by-trial adjustments in response
criteria. The initial settings for YES and NO response criteria were
drawn independently from a normal distribution of mean .5 and
standard deviation .01. The initial value of the mean and the value
5
Table 3
Values of the Core (Nonmodifiable) Parameters of the LCA
Model of Lexical Decision
Parameter
Value
Activity curves
w (inhibition)
k (leakage)
IYES ⫹ INO (total input)
GYES and GNO mean (Gaussian noise)
GYES and GNO SD (Gaussian noise)
yYES and yNO intervals
Measurements
Response criteria SD
No. trials per run
No. runs per simulation
Note.
.15
.05
.2
0
.05
[0;1]
.01
1,500
50
LCA ⫽ leaky competing accumulator; SD ⫽ standard deviation.
of the standard deviation did not change across simulations. At the
end of every trial, the mean values of both response criteria were
adjusted as a function of response accuracy, such that if a correct
response was made these values were decreased by a small constant. Whenever an error occurred, the mean values of response
thresholds were reset to their initial values. The threshold adjustment constant varied as a function of task instructions, type of
nonwords, and percentage of word versus nonword stimuli (see
Table 2).
Simulation Studies
Our LCA model was first evaluated against the results of Experiments 1 and 2 of Ratcliff et al. (2004) in Simulation Study 1.
This seemed the most appropriate set of benchmark phenomena for
an initial evaluation, given that (a) these were the data used to
falsify the MROM and (b) these data were used as the unique
benchmarks in a recent evaluation of the Bayesian Reader model
(Norris, 2009). The model was also evaluated against the experiments of Wagenmakers et al. (2008) in Simulation Study 2 and the
English Lexicon Project (ELP) lexical decision data (Balota et al.,
2007) in Simulation Study 3. The results of Wagenmakers et al.
provide another important benchmark for models of the lexical
Table 2
Bottom-Up Lexical Input Values (Evidence for a Word), Initial Threshold Values, and (in Parentheses) the Corresponding
Trial-by-Trial Adjustment Values Used in the Simulations
Lexical input values
Simulation
1
2
3
Stimuli/conditions
Words
Words
Words
Words
Words
Words
Words
and
and
and
and
and
and
and
pseudowords
random strings
pseudowords (accuracy stressed)
pseudowords (speed stressed)
pseudowords (75% words)
pseudowords (75% pseudowords)
pseudowords
HF
LF
VLF
NW
Threshold values
YES / NO (adjustment)
.12
.12
.12
.12
.12
.12
BIAM
.115
.115
.115
.115
.115
.115
BIAM
.11
.11
.11
.11
.11
.11
BIAM
.0875
.0825
.0875
.0875
.0875
.0875
BIAM
.5 (.003)/.5 (.003)
.5 (.02)/.5 (.03)
.5 (.003)/.5 (.003)
.5 (.05)/.5 (.05)
.5 (.02)/.5 (.003)
.5 (.003)/.5 (.02)
.5 (.001)/.5 (.001)
Note. HF ⫽ high-frequency; LF ⫽ low-frequency; VLF ⫽ very low-frequency; NW ⫽ nonwords; BIAM ⫽ bimodal interactive activation model
(Diependale et al., 2010).
tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
6
DUFAU, GRAINGER, AND ZIEGLER
decision task, and the ELP simulation provided an initial test of the
possible integration of the LCA decision mechanism within the
larger framework of models of word recognition and lexical decision (Coltheart, Rastle, Perry, Langdon, & Ziegler, 2001; Grainger
& Jacobs, 1996).
In the experiments of Ratcliff et al. (2004), high-frequency
(HF), low-frequency (LF), and very low-frequency (VLF) words
were intermixed with regular pronounceable nonwords (pseudowords) in Experiment 1 and random letter strings in Experiment 2.
Wagenmakers et al. (2008) tested HF, LF, and VLF words intermixed with pseudowords while manipulating speed versus accuracy instructions in their Experiment 1. In these experiments,
correct responses to words were faster and accuracy was higher
when nonwords were random strings than when they were pseudowords, and responses were faster but accuracy was lower when
speed was stressed in the instructions to participants. Wagenmakers et al. (2008) reported the results of a second experiment in
which they manipulated the proportion of word versus nonword
stimuli, with one condition having 75% words and another condition having 75% nonwords. Overall, their results showed that RTs
were faster and accuracy was higher for the stimulus category
(word vs. nonword) that occurred the most frequently in a given
condition.
The simulations reported here were structured like a behavioral
experiment. Each simulation study consisted of 50 consecutive
runs, equivalent to 50 virtual participants. Each run was composed
of 750 words and 750 nonwords randomly ordered, except for the
simulations in which word/nonword proportion was manipulated,
where there were 1,125 words and 375 nonwords in the 75% word
simulation and 375 words and 1,125 nonwords in the 75% non-
word simulation. Number of cycles to reach either the YES or the
NO decision threshold was recorded on each trial, as well as the
winning response (YES or NO).
Simulation Study 1
Simulation Study 1 used the conditions tested in Experiments 1
and 2 of Ratcliff et al. (2004), in which HF, LF, and VLF words
were either intermixed with orthographically regular and pronounceable nonwords (pseudowords, Experiment 1) or with random letter strings (random strings, Experiment 2). The effects of
this manipulation of type of nonword were simulated by a change
in the rate of the trial-by-trial adjustments of response criteria (see
Table 2). The rationale behind this particular simulation is discussed below. Figure 3 shows the scatterplots of simulated and
experimental mean RTs for the different conditions tested in these
experiments. The 16 mean RTs are derived from the combination
of stimulus type (HF, LF, and VLF words, and nonwords), correctness of response (correct and error), and type of nonword
(pseudowords and random strings). The results of these simulations clearly show that our LCA model provides a good fit to the
experimental data. It should be noted that the LCA model does not
exhibit the same problem as does the Bayesian Reader model,
which systematically underestimated error RTs in these experiments (see Norris, 2009, Tables 1 and 2).
Most important, as can be seen in Figure 3, is that the LCA
model accurately captures the pattern of RTs generated by the NO
decision node (correct responses to nonwords and errors to words).
This is exactly where the MROM (Grainger & Jacobs, 1996)
failed, which demonstrates that a dynamic deadline account of
Figure 3. Performance of the LCA model in simulating the response time (RT) condition means from Ratcliff
et al. (2004, Experiments 1 and 2). R2 is the percentage of explained variance. Correct responses are in dark gray,
errors are in light gray. PW ⫽ pseudowords; HF ⫽ high frequency; LF ⫽ low frequency; VLF ⫽ very low
frequency; RS ⫽ random strings; LCA ⫽ leaky competing accumulator.
F3
tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
LCA MODEL OF LEXICAL DECISION
lexical decision, as implemented in our LCA model, can overcome
the fundamental limitations of standard deadline models. We return to this point in greater detail in the General Discussion.
Simulation Study 2
F4
Simulation Study 2 tested the LCA against the data from
Wagenmakers et al. (2008) in which HF, LF, and VLF words were
intermixed with orthographically regular, pronounceable pseudowords. The same parameter set as in Simulation Study 1 was
maintained in these simulations. In Experiment 1 of Wagenmakers
et al. (2008), participants were given either instructions that
stressed accuracy of responding or instructions that stressed speed
of responding. The effects of this manipulation were simulated by
changes in the rate of trial-by-trial adjustments in response criteria.
In the speed condition, adjustment of both response criteria was
greater than in the accuracy condition (see Table 2). In Experiment
2, Wagenmakers et al. manipulated the percentage of words versus
pseudowords in the experiment (75% words in one condition and
25% in the other condition). Once again, these manipulations were
simulated in LCA by trial-by-trial adjustments of response criteria.
In the condition with a high proportion of words, adjustment of the
YES decision node was greater than adjustment of the NO decision
node, and in the condition with a high proportion of nonwords, the
opposite adjustment was implemented (see Table 2).
Figure 4 presents the results of simulations run on the LCA
model. The core parameters of the LCA model were the same as
those used in Simulation Study 1. Quite remarkably, by simply
modifying the rate of adjustment of response criteria, the LCA
7
model captured practically the same amount of variance in mean
RT (82%) as in Simulation Study 1 that involved not only a
different experimental manipulation (type of nonword), but also
different stimuli and participants. Once again, the LCA model
captures key patterns in the data that the MROM could not capture.
First, in the speed instructions condition tested by Wagenmakers et
al. (2008), the LCA model correctly simulates the faster RTs
obtained to error responses than to correct responses (diamond
shapes in Figure 4), and one can also see a clear frequency effect
in both the correct and the error RTs that is well captured by the
LCA model. Figure 5 (left panel) shows the distribution of simulated correct and error RTs to word targets in the speed instructions
condition, and it therefore demonstrates that RTs can be faster on
errors than on correct responses without RTs in the leading edge of
the distribution being too fast. Second, correct RTs to nonword
stimuli can be faster than error RTs to nonword stimuli, as was the
case in the accuracy instructions condition of Wagenmakers et al.
and as captured by LCA (circle shapes in Figure 4). This is further
demonstrated in Figure 5 (right panel), showing the distribution of
simulated RTs for correct and error responses to nonwords in the
accuracy instructions condition of Wagenmakers et al. Here we
clearly see a shift in the complete RT distribution, with correct RTs
being faster than error RTs for nonword stimuli. Furthermore,
Wagenmakers et al. demonstrated that this pattern of correct and
error RTs to nonword stimuli is exaggerated in the 75% nonword
condition they tested, but the opposite pattern is seen in the 75%
word condition. Figure 4 shows that LCA perfectly captures this
pattern. Finally, Figure 6 shows the mean accuracy per condition
Figure 4. Performance of the LCA model in capturing the pattern of mean response times (RTs) per condition
in Wagenmakers et al. (2008). R2 is the percentage of explained variance. Correct responses are in dark gray,
errors are in light gray. PW ⫽ pseudowords; HF ⫽ high frequency; LF ⫽ low frequency; VLF ⫽ very low
frequency; acc ⫽ accuracy instructions; speed ⫽ speed instructions; 75W ⫽ 75% word stimuli; 75NW ⫽ 75%
nonword stimuli; LCA ⫽ leaky competing accumulator.
F5
F6
tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
8
DUFAU, GRAINGER, AND ZIEGLER
Figure 5. Left panel: Distributions (quantile means) of simulated correct and error response times (RTs) in the
LCA model for high-frequency (HF), low-frequency (LF), and very low-frequency (VLF) words under the speed
instruction conditions of Wagenmakers et al. (2008). Right panel: Distributions (probability density) of
simulated correct and error RTs to nonwords under the accuracy instruction conditions of Wagenmakers et al.
(2008). LCA ⫽ leaky competing accumulator.
in Simulation Studies 1 and 2, in order to demonstrate that the
LCA model’s success in simulating mean RT is not achieved at the
cost of accuracy.
Summarizing the results of Simulations Studies 1 and 2, we
conclude that the LCA model provides an excellent fit to the major
data patterns presented by Ratcliff et al. (2004) and Wagenmakers
et al. (2008). Most important, with the same set of core (nonmodifiable) parameters, our model generated excellent fits to the data of
Wagenmakers et al., via theoretically motivated, trial-by-trial adjustments in the values of response criteria.
Simulation Study 3
Figure 6. Mean accuracy in the six experiments and simulations of the
present study. In black, pseudowords; in dark gray, very low-frequency
words; in light gray, low-frequency words; in white, high-frequency words.
PW ⫽ pseudowords (Ratcliff et al., 2004, Experiment 2); RS ⫽ random
strings (Ratcliff et al., 2004, Experiment 1); Accuracy ⫽ accuracy stressed
(Wagenmakers et al., 2008, Experiment 1); Speed ⫽ speed stressed
(Wagenmakers et al., 2008, Experiment 2); 75W ⫽ 75% words (Wagenmakers et al., 2008, Experiment 3); 75NW ⫽ 75% nonwords (Wagenmakers et al., 2008, Experiment 4).
Simulation Study 3 tested the LCA model against the data from
the English Lexicon Project (ELP; Balota et al., 2007) in which
words were intermixed with orthographically regular, pronounceable pseudowords. From the 40,000⫹ words in the ELP database,
we extracted the 5-letter and 6-letter words (842 words) that had
been tested in Ratcliff et al. (2004). We also randomly selected 842
5-letter and 6-letter nonwords from the ELP database. These words
and nonwords were then submitted as input to the bimodal version
of the interactive activation model (Diependaele, Ziegler, &
Grainger, 2010). For each stimulus, we took the activation level of
the most activated word unit at the 15th cycle (i.e., before asymptote). These activation values were then linearly transformed to fit
the range of input values used in Simulation Studies 1 and 2,
and these values were used as input to the YES response node in
LCA. In this simulation, the same core parameter values as in
Simulation Studies 1 and 2 were maintained (see Tables 2 and 3).
The value of the threshold adjustment parameter was set to a small
value for both thresholds, given the large number of trials in these
simulations.
Figure 7 shows the simulation run on LCA. Although the
percentage of explained variance (37%) is lower than in the
previous simulations, it must be remembered that variance is
explained here at the item level, as opposed to variance in mean
RT in Simulation Studies 1 and 2. One should also keep in mind
that the amount of reproducible variance in such large-scale data-
F7
tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
LCA MODEL OF LEXICAL DECISION
AQ: 2
9
Figure 7. LCA simulation of lexical decision response times (RTs) extracted from the English Lexicon Project
database (Balota et al., 2007) corresponding to the 842 words tested by Ratcliff et al. (2004) and 842 randomly
selected nonwords (left panel, behavioral RT distributions; bottom panel, simulated RT distributions; central
panel, scatterplot of mean RT per item). R2 is the percentage of explained variance. NW ⫽ nonwords; HF ⫽
high-frequency words; LF ⫽ low-frequency words; VLF: very low-frequency words. LCA ⫽ leaky competing
accumulator.
bases is probably no more than 40% (see Rey, Courrieu, SchmidtWeigand, & Jacobs, 2009). Furthermore, this result is quite encouraging, given that no parameter tuning was performed in order
to increase the amount of variance that could be explained. Figure 7 also shows that LCA generates RT distributions that change
as a function of word frequency and lexical status (word vs.
nonword), much in the same way as the empirical distributions. To
our knowledge, this is the first simulation of large-scale lexical
decision data that includes a simulation of nonword responses as
well as word responses. It is also the first simulation of large-scale
lexical decision data involving trial-by-trial adjustments in response criteria (see Perea et al., 2004, for similar simulations of a
standard lexical decision experiment). The relative success of this
simulation opens up a whole new perspective for analyzing sequential dependencies in such large-scale data and examining the
consequences of adjustments in response criteria on the evolution
of correct and error RTs as well as error rates in such data.
General Discussion
The lexical decision task is probably the most popular task used
to study silent single-word reading. Yet, we still have no clear
understanding of how participants in a lexical decision experiment
perform this task. In particular, it is not at all clear how participants
respond negatively when the stimulus is not a word. One simple
solution to this problem, the temporal deadline mechanism (implemented in the MROM; Grainger & Jacobs, 1996), was rejected
on the basis of recent analyses of the patterns of RTs obtained to
correct and error responses in the lexical decision task (Ratcliff et
al., 2004; Wagenmakers et al., 2008; see Table 1). In the present
work, we tested a dynamic deadline account of lexical decision
implemented within the framework of an LCA (Usher & McClelland, 2001). This model has two decision nodes, one for a YES
decision and one for a NO decision. Input to the YES decision node
is a function of lexical activity generated by the stimulus, whereas
input to the NO decision node is equal to a constant value minus
tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
10
DUFAU, GRAINGER, AND ZIEGLER
the input to the YES decision node. The activation dynamics of the
two decision nodes follow the standard equations of LCAs (Usher
& McClelland, 2001), with decay over time (leak) and lateral
inhibition between nodes (competition).
Simulation Studies 1 and 2 provided a test of LCA on two
independent sets of experimental data. Simulation Study 1 was
used to hand-tune the core parameters of LCA. Having fixed these
core parameter values, LCA was able to capture the effects of type
of nonword (pseudowords vs. random strings) by adjusting the
lexical input generated by these two types of stimuli and by
adjusting the rate of adjustment of response criteria. LCA captured
83% of the variance across the 16 mean RTs reported by Ratcliff
et al. (2004, Experiments 1 and 2).
The same core parameters whose values were adjusted to fit the
empirical data in Simulation Study 1 were used in the second
simulation study. The only parameter that changed between these
two simulation studies was the value of the trial-by-trial adjustment of response criteria. Despite the absence of any systematic
parameter tuning, LCA accounted for 82% of the variance across
the 32 mean RTs reported in Wagenmakers et al. (2008). More
precisely, the model provided an accurate account of variations in
mean RT and error rate to word and nonword stimuli as a function
of changes in speed versus accuracy instructions and the percentage of words versus nonwords in the experiment. Overall, the
results of Simulation Studies 1 and 2 showed that the LCA model
easily passed the tests that had been fatal for the MROM (see
Table 1) and that proved problematic for Norris’ (2009) Bayesian
Reader model (i.e., incorrect predictions concerning error RTs).
We analyze this particular aspect of the simulation results in more
detail in the following section.
Simulation Study 3 tested LCA against a subset of words and
nonwords extracted from the English Lexicon Project database of
item-level lexical decision latencies. Input to the YES response was
word activation level as calculated by the bimodal interactive
activation model (Diependaele et al., 2010), which provides a
measure of word activity for word and nonword stimuli of varying
length. The core parameter values used in Simulation Studies 1 and
2 remained unchanged. The model provided a good fit to the
item-level data, suggesting that it can be usefully appended to
word recognition models in order to account for performance in
the lexical decision task.
In the following sections, we first analyze precisely how the
LCA model overcomes the deficiencies of its predecessor: the
multiple read-out model (MROM) of lexical decision (Grainger &
Jacobs, 1996). Then we discuss the main contributions of the
present approach in increasing our understanding of how participants actually perform the lexical decision task. In particular, we
describe the LCA solution to the hard question in modeling lexical
decision, that is, what constitutes evidence for a nonword? Finally,
we discuss a key novel aspect to the current approach, that is, the
implementation of trial-by-trial adjustments in response criteria.
The Four Deficiencies of MROM
In Table 1 we listed the four deficiencies of the MROM
(Grainger & Jacobs, 1996), as identified in the work of Ratcliff et
al. (2004) and Wagenmakers et al. (2008). Here, we analyze the
success of the LCA model in passing these tests.
Considering first of all Points 1 and 2, both of which concern
RTs in response to word stimuli. MROM was found to be incapable of producing fast error RTs to word targets without generating excessively fast RTs in the first decile of the RT distribution
(Point 1), and error RTs to word targets in MROM did not show
a word frequency effect that is seen in the empirical data (Point 2).
The LCA simulation of the speed instruction condition tested by
Wagenmakers et al. (2008) revealed faster error RTs than correct
RTs to word targets (see Figure 4) without distorting the RT
distribution (see Figure 5, left panel). Also, the results of Simulation Studies 1 and 2 both revealed a word frequency effect in the
error RTs to word stimuli (see Figures 3 and 4).
Next, let us consider Points 3 and 4, both of which concern RTs
in response to nonword stimuli. MROM was found to generate RT
distributions for correct responses to nonword stimuli that did not
have the same rightward skew as the distributions for correct
responses to words (Point 3), and MROM could not generate fast
correct RTs to nonwords (and never faster than error RTs to
nowords) without generating too many errors (Point 4). Figure 5
(right panel) reveals that the RT distributions for nonword stimuli
in the LCA model show the distinctive rightward skew. The
simulation results reported in Figure 3 clearly demonstrate that the
LCA model can generate RTs for correct responses to nonword
targets that are faster than the error RTs to the same targets.
Finally, accuracy in the LCA model compares favorably with
accuracy in the experiments, as shown in Figure 6.
Input to the NO Response
Having abandoned a deadline account of generating NO responses in lexical decision, as implemented in the MROM
(Grainger & Jacobs, 1996), we note that the hard question for
alternative accounts of lexical decision, such as the diffusion
model (Ratcliff et al., 2004), the Bayesian Reader (Norris, 2009),
or the present LCA model, is to define what constitutes evidence
for a nonword. Defining evidence for a word (i.e., a YES response)
is relatively straightforward. In the first two simulation studies of
the present work, we followed Ratcliff et al. in using word frequency to determine these input values. In the third simulation
study, we used a measure of lexical activity derived from a
computational model of visual word recognition (Diependaele et
al., 2010). The main contribution of the present work concerns our
proposal for how to calculate evidence for a nonword.
In the LCA model, input to the NO decision node (i.e., evidence
for a nonword) is equal to a constant value minus the input to the
YES decision node (i.e., a measure of lexical activity generated by
the stimulus). The model therefore has to know only two things in
order to calculate input to the NO decision node: the amount of
lexical activity generated by the stimulus and the total input value
that optimizes speed and accuracy of responses to words and
nonwords.
So, a key remaining question here concerns how the total input
value could be computed online during the first trials of a lexical
decision experiment? One simple solution is to monitor the lexical
activity generated by word and nonword stimuli on each trial
(these correspond to the lexical input values given in Table 3, and
to compute the sum of these two values. This sum can be used to
provide an initial estimate of the total input value after having seen
only one word and one nonword in a lexical decision experiment.
tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
LCA MODEL OF LEXICAL DECISION
This initial value can then be adjusted in order to optimize performance, much in the same manner as response criteria are
adjusted on a trial-by-trial basis in our model. Increasing the total
input value will generate faster RTs to nonwords but will also
generate more errors to word stimuli, as input to the NO response
will eventually be greater than input to the YES response. Errordriven adjustments of this value can therefore be used to rapidly
home in on an optimal setting. A similar mechanism could be
applied to the diffusion model (Ratcliff et al., 2004), where evidence for a word and evidence for a nonword would be averaged,
rather than summed, in order to provide an initial estimate of the
drift rate criterion.
It is important to note here that our proposed mechanism for
computing the optimal value of the total constant input in LCA is
analogous to the way in which response criteria are adjusted in our
model. This implies that the modifiable parameters in our LCA
model all involve either stimulus-driven changes in lexical input
values or trial-by-trial adjustments of the mean values of the YES
and NO decision thresholds and the total constant input in order to
optimize performance.
Online Adjustments of Response Criteria
One of the key features of the LCA model, in line with the
general approach to modeling adopted in the MROM (Grainger &
Jacobs, 1996), is that effects of list context and task instructions
are simulated via adjustments of response criteria. In the present
study, we demonstrated that such adjustments in the thresholds of
the YES and NO decision nodes provided an accurate account of
how different types of nonword affect responding to word stimuli
(an effect of list context), how responses to word and nonword
stimuli are affected by the proportion of words versus nonwords in
the experiment (another effect of list context), and how responses
to word and nonword stimuli are affected by stressing speed over
accuracy or vice versa (an effect of task instructions). The general
principle that was applied in this endeavor is that participants in a
lexical decision experiment are trying to optimize their performance in line with the instructions they receive and do so by
monitoring their performance on a trial-by-trial basis. This critical
information provided by such monitoring is whether or not a
mistake was made, and it is this information about accuracy of
responding that was used to adjust response criteria in the present
work.
In implementing such trial-by-trial adjustments of response criteria, we follow in the steps of a number of prior attempts to
describe such adjustments in the light of empirical data concerning
post-error responding (e.g., Laming, 1968; Rabbitt, 1966; Rabbitt
& Rodgers, 1977). In particular, conflict monitoring theory (Botvinick, Braver, Barch, Carter, & Cohen, 2001) accounts for errorrelated changes in performance by postulating the occurrence of
conflict in association with errors. Conflict is used to change the
activation level of output units, which is of course functionally
equivalent to modifying response thresholds, as in the present
approach. This allowed Botvinick et al. to account for the standard
pattern of slower RTs and higher levels of accuracy following an
error (but see Notebaert et al., 2009, for different patterns of
post-error performance).
We implemented a relatively simple version of this general
“monitor and adjust” principle, which despite its simplicity pro-
11
vided an accurate account of the behavioral data. The mean values
of the thresholds of YES and NO decision nodes were lowered on
every correct trial by a small constant value that was determined
by task instructions and list context, and the rate of change could
be different for the two decision nodes. Every time there was an
error, the values of both thresholds were reset to their initial
values.
To simulate the influence of type of nonword on responses to
both nonword and word targets, we increased the rate of adjustment of response thresholds in the easy nonword condition (i.e.,
random strings), and this increase was slightly greater for the NO
threshold than the YES threshold. The NO threshold can be lowered to a certain extent without an increase in false negative errors,
as long as the YES threshold is lowered at the same time. The YES
threshold can be lowered to a certain extent without an increase in
false positive errors because random string nonword stimuli have
high input activation to the NO response node. Similar adjustments
of the rate of change of response thresholds were used to simulate
the effects of the proportion of word versus nonword stimuli in an
experiment. With more words than nonwords, the YES response
threshold was lowered by a greater amount on each correct trial,
and with more nonwords than words it was the NO response
threshold that was adjusted by a greater amount. The lowering of
response thresholds as a function of the proportion of words and
nonwords enables optimization of responding for a majority of
trials in the experiment. Finally, with speed instructions, the trialby-trial threshold adjustment was increased by the same amount
for both decision nodes, thus simulating participants’ attempts to
respond faster to both word and nonword stimuli. This generated
more false negative and false positive errors, and the resulting
error RTs were faster due to the fact that they are mostly generated
with the lowest settings of response criteria. Because errors immediately generate a resetting of response criteria to their initial
values, correct responses in these conditions are therefore slower
on average than error responses.
Thus, overall our simple implementation of trial-by-trial adjustments in response criteria provided an accurate simulation of the
effects of list context and task instructions as tested in the studies
of Ratcliff et al. (2004) and Wagenmakers et al. (2008). Although
this study is clearly only a first approximation with respect to how
participants in a lexical decision experiment adjust to such contextual factors, the relative success of this approach encourages
further tests of more sophisticated adjustments in response criteria
with the aim to capture trial-by-trial changes in human performance.
Conclusions
In 1996, Grainger and Jacobs presented an influential account of
the lexical decision task (the multiple read-out model; MROM), in
which time from stimulus onset was used to trigger NO responses
via a temporal deadline. This account has been rejected on the
basis of results showing fast NO responses in certain conditions in
lexical decision experiments (Ratcliff et al., 2004; Wagenmakers
et al., 2008). Here we implemented and tested a dynamic deadline
model of the lexical decision task using the framework of leaky
competing accumulators (Usher & McClelland, 2001). The LCA
model was shown to overcome the deficiencies associated with the
MROM, and overall it provided an accurate account of variations
tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z xppws S⫽1 12/29/11 1:20 Art: 2011-2553
DUFAU, GRAINGER, AND ZIEGLER
12
in RTs for correct and error responses to word and nonword
stimuli as a function of the type of nonword, the proportion of
words versus nonwords in the experiment, and whether or not task
instructions stressed speed or accuracy of responding. Finally, by
using lexical activity calculated by a computational model of
visual word recognition as lexical input to the LCA decision
module, we were able to provide the first item-level simulation of
both word and nonword responses in a large-scale database.
References
Anderson, J. R. (1990). The adaptive character of thought. Hillsdale, NJ:
Erlbaum.
Balota, D. A., & Chumbley, J. I. (1984). Are lexical decisions a good
measure of lexical access? The role of word frequency in the neglected
decision stage. Journal of Experimental Psychology: Human Perception
and Performance, 10, 340 –357. doi:10.1037/0096-1523.10.3.340
Balota, D. A., Yap, M. J., Cortese, M. J., Hutchison, K. A., Kessler, B.,
Loftis, B., . . . Treiman, R. (2007). The English Lexicon Project. Behavior Research Methods, 39, 445– 459. doi:10.3758/BF03193014
Bogacz, R., Brown, E., Moehlis, J., Holmes, P., & Cohen, J. D. (2006). The
physics of optimal decision making: A formal analysis of models of
performance in two-alternative forced choice tasks. Psychological Review, 113, 700 –765. doi:10.1037/0033-295X.113.4.700
Botvinick, M. M., Braver, T. S., Barch, D. M., Carter, C. S., & Cohen, J. D.
(2001). Conflict monitoring and cognitive control. Psychological Review, 108, 624 – 652. doi:10.1037/0033-295X.108.3.624
Coltheart, M., Davelaar, E., Jonasson, J. T., & Besner, D. (1977). Access
to the internal lexicon. In S. Dornic (Ed.), Attention and Performance VI
(pp. 535–555). Hillsdale, NJ: Erlbaum.
Coltheart, M., Rastle, K., Perry, C., Langdon, R., & Ziegler, J. C. (2001).
DRC: A dual route cascaded model of visual word recognition and
reading aloud. Psychological Review, 108, 204 –256. doi:10.1037/0033295X.108.1.204
Davis, C. J. (2010). The spatial coding model of visual word identification.
Psychological Review, 117, 713–758. doi:10.1037/a0019738
Diependaele, K., Ziegler, J. C., & Grainger, J. (2010). Fast phonology and
the bimodal interactive activation model. European Journal of Cognitive
Psychology, 22, 764 –778. doi:10.1080/09541440902834782
Donkin, C., Brown, S. D., & Heathcote, A. (2009). The overconstraint of
response time models: Rethinking the scaling problem. Psychonomic
Bulletin & Review, 16, 1129 –1135. doi:10.3758/PBR.16.6.1129
Dufau, S., Duñabeitia, J. A., Moret-Tatay, C., McGonigal, A., Peeters, D.,
Alario, F.-X., . . . Grainger, J. (2011). Smart phone, smart science: How
the use of smartphones can revolutionize research in cognitive science.
PLoS ONE, 6, e24974. doi:10.1371/journal.pone.0024974
Grainger, J., Dufau, S., & Ziegler, J. C. (2009). A new model of lexical
decision: MROM⫹. Abstracts of Psychonomics Society, 14, 27. Retrieved from http://www.psychonomic.org/pdfs/Abstracts09.pdf
Grainger, J., & Jacobs, A. M. (1996). Orthographic processing in visual
word recognition: A multiple read-out model. Psychological Review,
103, 518 –565. doi:10.1037/0033-295X.103.3.518
Jacobs, A. M., & Grainger, J. (1992). Testing a semistochastic variant of
the interactive activation model in different word recognition experiments. Journal of Experimental Psychology: Human Perception and
Performance, 18, 1174 –1188. doi:10.1037/0096-1523.18.4.1174
Laming, D. R. (1968). Information theory of choice-reaction times. London, England: Academic Press.
McClelland, J. L., & Rumelhart, D. E. (1981). An interactive activation
model of context effects in letter perception: Part 1. An account of basic
findings. Psychological Review, 88, 375– 407. doi:10.1037/0033295X.88.5.375
Norris, D. (2006). The Bayesian reader: Explaining word recognition as an
optimal Bayesian decision process. Psychological Review, 113, 327–
357. doi:10.1037/0033-295X.113.2.327
Norris, D. (2009). Putting it all together: A unified account of word
recognition and reaction-time distributions. Psychological Review, 116,
207–219. doi:10.1037/a0014259
Notebaert, W., Houtman, F., Van Opstal, F., Gevers, W., Fias, W., &
Verguts, T. (2009). Post-error slowing: An orienting account. Cognition,
111, 275–279. doi:10.1016/j.cognition.2009.02.002
Oaksford, M., & Chater, N. (1998). Rational models of cognition. New
York, NY: Oxford University Press.
Perea, M., Carreiras, M., & Grainger, J. (2004). Blocking by word frequency and neighborhood density in visual word recognition: A taskspecific response criteria account. Memory & Cognition, 32, 1090 –
1102. doi:10.3758/BF03196884
Perry, C., Ziegler, J. C., & Zorzi, M. (2010). Beyond single syllables:
Large-scale modeling of reading aloud with the Connectionist Dual
Process (CDP⫹⫹) model. Cognitive Psychology, 61, 106 –151. doi:
10.1016/j.cogpsych.2010.04.001
Rabbitt, P. M. A. (1966). Errors and error correction in choice reaction
tasks. Journal of Experimental Psychology, 71, 264 –272.
Rabbitt, P. M., & Rodgers, B. (1977). What does a man do after he makes an
error? An analysis of response programming. Quarterly Journal of Experimental Psychology, 29, 727–743. doi:10.1080/14640747708400645
Ratcliff, R., Gomez, P., & McKoon, G. (2004). A diffusion model account
of the lexical decision task. Psychological Review, 111, 159 –182. doi:
10.1037/0033-295X.111.1.159
Ratcliff, R., & Smith, P. L. (2004). A comparison of sequential sampling
models for two-choice reaction time. Psychological Review, 111, 333–
367. doi:10.1037/0033-295X.111.2.333
Rey, A., Courrieu, P., Schmidt-Weigand, F., & Jacobs, A. M. (2009). Item
performance in visual word recognition. Psychonomic Bulletin & Review, 16, 600 – 608. doi:10.3758/PBR.16.3.600
Rubenstein, H., Garfield, L., & Millikan, J. A. (1970). Homographic
entries in the internal lexicon. Journal of Verbal Learning and Verbal
Behavior, 9, 487– 494. doi:10.1016/S0022-5371(70)80091-3
Usher, M., & McClelland, J. L. (2001). The time course of perceptual
choice: The leaky, competing accumulator model. Psychological Review, 108, 550 –592. doi:10.1037/0033-295X.108.3.550
Wagenmakers, E. J., Ratcliff, R., Gomez, P., & McKoon, G. (2008). A
diffusion model account of criterion shifts in the lexical decision task.
Journal of Memory and Language, 58, 140 –159. doi:10.1016/
j.jml.2007.04.006
Ziegler, J. C., Grainger, J., & Brysbaert, M. (2010). Modelling word
recognition and reading aloud. European Journal of Cognitive Psychology, 22, 641– 649. doi:10.1080/09541446.2010.496263
Received February 18, 2011
Revision received December 1, 2011
Accepted December 5, 2011 䡲
JOBNAME: AUTHOR QUERIES PAGE: 1 SESS: 1 OUTPUT: Thu Dec 29 01:20:29 2011
/tapraid5/zfv-xlm/zfv-xlm/zfv00212/zfv2713d12z
AUTHOR QUERIES
AUTHOR PLEASE ANSWER ALL QUERIES
AQ1: Author. Table 3 was cited before Table 2, so they were switched.
AQ2: Author. Figure 7. The caption could make a clearer distinction between the panels.
1