Interactions between Prediction, Perception and Episodic Memory

Interactions between Prediction, Perception and Episodic Memory
DISSERTATION
Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy
in the Graduate School of The Ohio State University
By
Adam E. Hasinski
Graduate Program in Psychology
The Ohio State University
2015
Dissertation Committee:
Per B. Sederberg, Ph.D., Advisor
Trisha Van Zandt, Ph.D.
Julie Golomb, Ph.D.
© Copyrighted by
Adam E. Hasinski
2015
Abstract
For a variety of reasons, cognitive scientists tend to divide the study of the brain into
separate domains. Some of the most studied domains are episodic memory, visual
perception, and prediction. However, the underlying processes that give rise to one’s
memories, perceptions, and predictions do not operate independently. Instead each of
these processes influences the others. Previous research has demonstrated that the
memory system is critical for the generation of predictions (Buckner, 2010; Corbit &
Balleine, 2000; Johnson, Meer, & Redish, 2007; Lisman & Redish, 2009). Other
researchers have argued that portions of the memory system are actually a part of the
visual perception system (Baxter, 2009; Bussey & Saksida, 2007). Here we present
research that furthers our understanding of how these processes interact. First we
demonstrate, through a series of behavioral experiments, that predictions influence
memory. Specifically, in high-predictability environments, we find enhanced memory for
items that could have been predicted to recur, but did not. We then present a
computational model of recognition memory that attempts to formally explain this pattern
using a prediction-based learning mechanism. Finally, we present neural evidence that
the information content of the perceptual system is contingent on subsequent memory.
Taken together, these results demonstrate how the perception of present stimuli and the
prediction of future stimuli interact with the memory system.
ii
Dedicated to Dave Qualkinbush, Jim Petitto, Harry Finkbone, and Joe Panaccione,
who taught me how to focus, persevere, and how to get up after being knocked down.
iii
Acknowledgments
I would like to thank my advisor for his help and instruction over the years. I
would also like to thank the committee members for their time and feedback, both during
and before the dissertation process.
I would also like to thank my family, especially my parents, for their endless
support and patience. I did not fully appreciate their devotion until I came to graduate
school.
Finally, I would like to thank the friends I met in Columbus for their
encouragement and camaraderie over the years.
iv
Vita
June 2004 .......................................................Westlake High School
2004-2008 ......................................................Presidential Scholar, Ohio University
2008 ...............................................................B.A. Psychology & History, Ohio
University
2008-2009 .....................................................Graduate Fellow, Department of
Psychology, The Ohio State University
2009-2011 .....................................................Graduate Teaching Associate, Department
of Psychology, The Ohio State University
2010 ...............................................................M.A, Psychology, The Ohio State
University
2011-2015 .....................................................Graduate Research Associate, Department
of Psychology, The Ohio State University
Publications
T.A. Smith, A.E. Hasinski, P.B. Sederberg, “The context repetition effect: Predicted
events are remembered better, even when they don’t happen”. Journal of Experimental
Psychology: General, 142(4):1298, 2013.
v
M.A. Martens, A.E. Hasinski, R.R. Andridge, W.A. Cunningham, “Continuous cognitive
dynamics of the evaluation of trustworthiness in Williams syndrome”. Frontiers in
Psychology, 3, 2012.
J.J. Ratcliff, G.D. Lassiter, V.M. Jager, M.J. Lindberg, J.K. Elek, A.E. Hasinski, “The
hidden consequences of racial salience in videotaped interrogations and confessions”.
Psychology, Public Policy & Law, 16(2):200-218, 2010.
Fields of Study
Major Field: Psychology
vi
Table of Contents
Abstract ............................................................................................................................... ii
Acknowledgments.............................................................................................................. iv
Vita...................................................................................................................................... v
Table of Contents .............................................................................................................. vii
List of Tables ..................................................................................................................... xi
List of Figures ................................................................................................................... xii
Chapter 1: Introduction ...................................................................................................... 1
Context: The Cornerstone of Episodic Memory ............................................................. 2
Memory does not Operate in Isolation ............................................................................ 4
Memory and Perception............................................................................................... 5
Memory and Prediction ............................................................................................... 7
Chapter 2: Repeating an Item’s Temporal Context Enhances its Memorability .............. 10
Overview of Experiments .......................................................................................... 14
Experiment 1: A Predictable Environment with Scenes ............................................... 16
Method ....................................................................................................................... 16
vii
Results ....................................................................................................................... 20
Preliminary Discussion .............................................................................................. 22
Experiment 2: A Predictable Environment with Words................................................ 24
Results ....................................................................................................................... 25
Preliminary Discussion .............................................................................................. 29
Experiment 3: An Unpredictable Environment with Images ........................................ 30
Method ....................................................................................................................... 31
Results ....................................................................................................................... 31
Preliminary Discussion .............................................................................................. 33
Results ....................................................................................................................... 34
Preliminary Discussion .............................................................................................. 35
Discussion ..................................................................................................................... 35
Implications for the Study of Memory ...................................................................... 36
Open Questions and Future Directions ...................................................................... 39
Conclusions ............................................................................................................... 41
Chapter 3: A Formal Account of the Context Repetition Effect ...................................... 43
The Context Repetition Effect ................................................................................... 44
A Formal TD-SR Account of the Context Repetition Effect .................................... 47
TCM-SR: TCM with a TD-SR Learning Rule .............................................................. 49
viii
Simulation and Results .................................................................................................. 55
Hierarchical Bayesian Framework ............................................................................ 59
Results ....................................................................................................................... 64
Discussion ..................................................................................................................... 72
Other Models and the Context Repetition Effect ...................................................... 73
Differences from TD-SR ........................................................................................... 74
Limitations and Future Work .................................................................................... 76
Conclusions ............................................................................................................... 80
Chapter 4: Face-specific Information in the Fusiform Face Area Depends on Memory.. 81
Method .......................................................................................................................... 87
Subjects ...................................................................................................................... 87
Stimuli ....................................................................................................................... 87
Procedure ................................................................................................................... 88
MRI Data Acquisition ............................................................................................... 92
Calculating Face-specific Information ...................................................................... 98
Mixed-effects Regression ........................................................................................ 100
Results ......................................................................................................................... 103
Behavioral Results ................................................................................................... 103
Neuroimaging Results ............................................................................................. 105
ix
Discussion ................................................................................................................... 109
Conclusions ................................................................................................................. 113
Chapter 5: General Discussion........................................................................................ 114
Memory and Prediction ........................................................................................... 114
Memory and Perception........................................................................................... 117
Bridging Domains.................................................................................................... 118
Limitations and Future Directions............................................................................... 118
Recognition vs. Recall ............................................................................................. 118
The Neural Correlates of the Context Repetition Effect ......................................... 120
A Model of Memory, Perception, and Prediction ................................................... 122
The Role of Attention .............................................................................................. 123
Conclusions ................................................................................................................. 124
Bibliography ................................................................................................................... 125
x
List of Tables
Table 1. Free and fixed parameters. .................................................................................. 58
Table 2. Essential equations. ............................................................................................. 63
xi
List of Figures
Figure 1. Structure of encoding lists. Each color represents an example triplet-pair for
one of the four main experimental conditions. The leftmost condition, RCRT, does not
appear in Experiments 3 and 4. First and second triplets appear on upper and lower rows,
respectively. The lag between actual triplet pairs was randomized (see Experiment 1
methods). ........................................................................................................................... 18
Figure 2. Memory performance across conditions. Bars show the mean hit rate for all
target conditions and the false alarm rate for lures. Panels A, B, C, and D correspond to
Experiments 1, 2, 3, and 4, respectively. Error bars indicate standard errors of the mean.
........................................................................................................................................... 28
Figure 3. Context repetition effects. Bars show the effect of contextual repetition: Within
each panel, the left bar indicates the mean difference in hit rate between once-presented
targets whose contexts later repeat and control targets (RCNT1 vs. NCNT1), while the
right bar indicates the mean difference in hit rate between once-presented targets that
follow repeated contexts and control targets (RCNT2 vs. NCNT2). Panels A, B, C, and D
correspond to Experiments 1, 2, 3, and 4, respectively. Error bars indicate standard errors
of the mean. ∗ 𝑝 < 0.05. ................................................................................................... 29
Figure 4. Hierarchical structure. Each model variant was fit to all participants i from
Experiment 1 (High Predictability) and all participants j from Experiment 3 (Low
xii
Predictability). The same hyperpriors were used across both experiments for all
parameters except 𝛾. ......................................................................................................... 62
Figure 5. Model fits: Memory performance across conditions. Dark blue bars show the
mean hit rates across conditions, and false-alarm rates for lures. Light blue bars indicate
mean model behavior for each condition using MAP estimates. Panel rows correspond to
different experiments. ....................................................................................................... 65
Figure 6. Model fits: Context repetition effects. Dark blue bars show the behavioral
context repetition effect. Light blue bars indicate model behavior using MAP estimates.
Within each panel, the left bars indicate the mean difference in hit rate between oncepresented targets whose contexts later repeat and control targets (RCNT1 vs. NCNT1).
The right bars indicate the mean difference in hit rate between once-presented targets that
follow repeated contexts and control targets (RCNT2 vs. NCNT2). Panel rows
correspond to different model variants. Panel columns correspond to different
experiments. ...................................................................................................................... 66
Figure 7. Model fits: Individual differences. Scatterplots show the relationship between
each participant’s context repetition effect and the simulated context repetition effect for
their corresponding sub-model. Best fitting regression lines and 𝑅2 values are included.
Panel rows correspond to different model variants. Panel columns correspond to different
experiments and presentations. ......................................................................................... 67
Figure 8. Model comparison: Variants 1 and 2. The histogram shows the frequency of
within-participant differences in BPIC scores between Variants 1 and 2. Greater values
indicate a stronger preference for Variant 2, with 0 indicating no preference.................. 70
xiii
Figure 9. Model comparison: Variants 2 and 3. The histogram shows the frequency of
within-participant differences in BPIC scores between Variants 2 and 3. Greater values
indicate a stronger preference for Variant 3, with 0 indicating no preference.................. 72
Figure 10. Overview of experimental design. (A) Faces were presented one at a time, for
1.2 seconds, with a randomly varied interstimulus interval. Each target face occurred
twice. (B) Each presentation of a given target face occurred in either the same context
(low encoding variability) or in different contexts (high encoding variability). Same-face
similarity was computed between a first-presentation target (red square) and its identical
second presentation (dark-blue square). Different-face similarities were computed
between the first-presentations and non-identical second presentations with matching
genders and encoding variability (light-blue square)........................................................ 91
Figure 11. Functional and anatomical ROIs. Superposition of every subject’s FFA (red),
OFA (green) and PPA (blue), along with aIT (violet) in MNI standard space. ................ 97
Figure 12. Subsequent memory performance as a function of encoding condition. Bars
show mean proportion of faces recognized during the memory test for faces that occurred
under high encoding variability (dark green) and low encoding variability (light green).
For a reference, the mean proportion of once-presented items that were subsequently
remembered is shown in grey. Error bars reflect standard errors. The solid red line
indicates the mean proportion of lures that were incorrectly judged as having been seen
previously, with dashed red lines indicating standard error. .......................................... 105
Figure 13. Average similarity between face repetitions. Bars show similarity between
first presentation targets and either (dark blue) same or (light blue) different second
xiv
presentations, split by memory performance. (A) FFA shows significant face-specific
information–greater similarity between identical faces than between different faces of the
same gender–only for faces that are subsequently remembered. The difference in facespecific similarity between remembered and forgotten faces is also significant. (B) aIT
shows the same pattern, although the difference in face-specific information is no longer
significant after correcting for multiple comparisons. † 𝑝 < 0.05 uncorrected, ∗ 𝑝 < 0.05
corrected, ∗∗ 𝑝 < 0.01 corrected. ................................................................................... 108
xv
Chapter 1: Introduction
Research into human episodic memory has come a long way since Ebbinghaus
shuffled cards with three-character nonsense syllables, presented them to himself to the
beat of a metronome, and then tested his own memory (Ebbinghaus, 1913). Modern
episodic memory researchers replace shuffled cards with randomized study lists,
presented with precise timing via computer. The content of these lists can be anything:
printed or spoken words, images of faces, places, or even images collected from
participants’ own lives. As memory scientists adjust various factors of these lists, they
look for effects: reliable patterns of behavior from the participants they study. These
patterns provide insight into the mechanisms underlying memory. Some of the most wellknown effects include: primacy, enhanced memory for the first few items on a study list;
recency, enhanced memory for the last few list items; contiguity, the tendency to recall
items that were studied near a previously recalled item; and the von-Restorff effect,
enhanced memory for items that stand out from the rest of the list (Kahana, 2012). To
make sense of the dozens of memory effects and paradigms, memory theorists develop
computational models. These models are explanations of memory phenomena, written in
the language of math to avoid the ambiguities of informal languages. To date, no unified
theory of episodic memory has been proposed that can account for all of these
phenomena. Instead different research groups propose various models to explain handfuls
1
of effects or particular paradigms. While the mechanisms of these models vary widely, a
nearly ubiquitous feature is the use of context.
Context: The Cornerstone of Episodic Memory
Context is perhaps the single most important construct in the study of memory.
Most models of memory posit that effective retrieval from long-term memory depends
upon either associating studied items with context (Dennis & Humphreys, 2001; Gillund
& Shiffrin, 1984; Howard & Kahana, 2002; Polyn, Norman, & Kahana, 2009;
Raaijmakers & Shiffrin, 1981; Sederberg, Howard, & Kahana, 2008) or encoding an
item’s context, along with the item itself, as a memory trace (Hintzman, 1988). Specific
memories can be retrieved by cueing the memory system with a particular context. In the
real world this is presumed to happen spontaneously all the time. For instance, when you
enter a store—a specific context—you may be reminded of a friend you ran into the last
time you were in that store. In the laboratory, this often takes the form of asking
participants to recall items that occurred on specific study lists. Alternatively, the degree
of association between an item and a particular context can be used to judge whether or
not that item previously occurred in that context. As an example, an eyewitness may be
shown a lineup and asked if any of the suspects were at the scene of the witnessed crime.
In the lab, item recognition tests ask participants to judge whether test items occurred on
earlier study lists. Thus, context not only enables one to retrieve information from
memory, but also to identify when and where that information was encoded. Perhaps in
part because context is such an important and enduring idea in memory research, context
has been defined in many different ways.
2
In the most general sense, context refers to the relatively stable aspects of one’s
experience. “Relatively stable” certainly does not mean fixed or constant. In fact many
models of memory define context as something that changes (Estes, 1955; Howard &
Kahana, 2002; Shankar & Howard, 2012). Rather, context is something that is stable
relative to the more fleeting stimuli that make up an episode of experience. The actual
information that comprises context varies from theory to theory. Rather than thinking
about one or another theory as having the “correct” definition of context, it is useful to
see different theories and paradigms as focusing on different aspects of context.
A common definition of context is the spatial environment (e.g. a store or a
psychology lab). Changing the spatial context between study and test phases has been
shown to diminish memory performance (Falkenberg, 1972). For instance, one classic
study had divers learn and recall lists of words underwater and on land. Memory
performance was degraded when participants had to recall words in an environment that
was different from the encoding environment (Godden & Baddeley, 1975). Aspects of
one’s internal state can also be viewed as a type of context. Indeed changing various
aspects of one’s current physiology, including the amount of caffeine consumed
(Mystkowski, Mineka, Vernon, & Zinbarg, 2003) or degree of inebriation (Weingartner,
Adefris, Eich, & Murphy, 1976) has been shown to impede memory performance. Even
one’s mood can be considered a context, with studies again showing better memory for
congruent study-test moods (Ucros, 1989).
One commonality across the above-mentioned research is that context was treated
as discrete. Participants learned items in one stable context, and then were tested in either
3
that same context or a different one. Alternatively, context can be defined as evolving
continuously over time (Estes, 1955; Howard & Kahana, 2002; Shankar & Howard,
2012). Early conceptualizations of a continuously evolving context envisioned it as
fluctuating randomly (Estes, 1955). More recent accounts define context as a recencyweighted summation of temporally recent information (Howard & Kahana, 2002;
Shankar & Howard, 2012). In other words, recent experience serves as the context for
one’s current experience at this particular moment. As this moment passes, it becomes a
part of the context of the next moment. As time progresses and experiences drift farther
away from the current point in time, these experiences decay out of context. The rate of
this decay is often assumed to be constant (Howard & Kahana, 2002; Polyn et al., 2009;
Sederberg et al., 2008). However, recent work integrating theory from cognitive science
with empirical findings from animal learning strongly suggests that this decay may be
nonlinear, such that elements of context decay rapidly at first, but then remain active at
low levels for an extended period of time (Shankar & Howard, 2012).
As a final note, we emphasize a distinction between the actual information that
makes up one’s experience—be it environmental information or internally generated
information—and the memory system’s representation of that information. For brevity, in
this document the term “context” will be used to describe experiential information and its
representation in the brain.
Memory does not Operate in Isolation
Of course, episodic memory is just one of many cognitive processes, all of which
work together to produce human behavior and subjective experience. To completely
4
understand memory, researchers will need theories that describe the interactions between
episodic memory and these other processes. In order to develop integrative theories,
experimental research is needed to explore how manipulations of memory affect other
processes and vice versa. In the present research, we contribute to this endeavor by
exploring how memory interacts with prediction and perception. Specifically, we
investigate the relationship between perceptual representation and the episodic memory
system, as well as how the memory system is influenced by prediction.
We have chosen these three processes—memory, perception, and prediction—
because they represent stimuli and events at different points in time: the past, present, and
hypothetical future, respectively. At encoding, the memory system forms enduring traces
of the information that is being currently processed by the perceptual system. Later, these
traces can be retrieved in order to reconstruct previously perceived information.
Information in the memory system can also be used to generate predictions about what
events or stimuli are likely to be perceived in the near future. Thus, while these three
processes are likely fully interconnected, the relationships between memory and both
prediction and perception may be the most critical. We focus on these relationships in the
present work.
Memory and Perception
Memory and perception are typically viewed as distinct processes that correspond
to different neural substrates, the medial temporal lobe memory system (MTL) and the
ventral visual pathway, respectively. However, this view has been challenged in recent
years (Suzuki & Baxter, 2009). Some have argued that portions of the MTL, the
5
perirhinal cortex in particular, may also be involved in object perception (Baxter, 2009;
Bussey & Saksida, 2007). In essence, the MTL is viewed as an extension of the visual
pathway. Other researchers have contested this interpretation of the existing evidence
(Suzuki, 2009; Suzuki & Baxter, 2009). Additionally, recent neuroimaging work has
shown that some of the same parietal regions are active during both memory and
perceptual tasks (Cabeza et al., 2011). The researchers argue that these parietal regions
may be fulfilling similar attentional roles for both memory and perception. This claim is
consistent with the recent distinction made between external attention, the selection of a
subset the perceivable environment, and internal attention, the selection of certain stored
memories (Chun, Golomb, & Turk-Browne, 2011). Thus, the idea that memory and
perception overlap is not new; however, their relationship is far from understood.
In the present research, we further explore the connection between memory and
perception. We were specifically interested in the perception of, and memory for, human
faces. We placed participants inside an MRI scanner and used fMRI to measure their
brain activity while they viewed a series of unfamiliar faces. After participants exited the
scanner, they completed a surprise recognition test for the faces they saw. We used
multivariate neuroimaging techniques to compare the distributed neural representations
evoked by different and identical faces. Our investigation focused on the ventral visual
processing pathway, an interconnected set of regions responsible for representing what
we see. Our results showed evidence that face-specific information is present in ventral
visual regions. However, this face specific information was only found for faces that were
subsequently remembered during the recognition test. In other words, later memory for a
6
face is correlated with the amount of face-specific information present in the online
neural representation of that face. Although several interpretations of these findings are
possible, our results further establish the link between perception and memory.
Memory and Prediction
A great deal of research has shown that prediction is one of the core functions the
human brain (Bar, 2009). This is not surprising given the advantages offered by
prediction. As generating predictions requires a prior knowledge base, it is not surprising
that the episodic memory system—which encodes, maintains, and retrieves information
about experiences (Davachi, 2006; Eichenbaum, 2000, 2004)—has been also been
implicated as a source of predictions (Buckner, 2010; Johnson et al., 2007; Levy, 1989;
Levy, Hocking, & Wu, 2005; Lisman & Redish, 2009; Schacter, Addis, & Buckner,
2007). It may be that the same reconstructive processes that give rise to memory retrieval
also enable prediction and mental simulation more generally (Maguire & Hassabis, 2011;
Schacter & Addis, 2009).
In the present research, we argue that prediction can also influence memory. This
assertion stems from a framework (Gershman, Moore, Todd, Norman, & Sederberg,
2012) that merges reinforcement learning rules (Dayan, 1993; White, 1995), with models
of episodic memory (Howard & Kahana, 2002; Polyn et al., 2009; Sederberg et al.,
2008). Based on this framework, we developed a computational model that learns via
prediction and is fully capable of reproducing recognition memory behaviors. Because of
this prediction-based learning, the model posits a very particular effect: Once-presented
items will be better remembered if they are predicted to recur. However, this hypothesis
7
cannot be tested using conventional memory paradigms. The reason for this is that most
paradigms never repeat items. Even in cases where repetitions do occur, list
randomization eliminates any signal that would suggest repetitions may be coming. In
other words, nothing about the experimental setting enables the participant to effectively
generate predictions. To test this new model, we needed a new paradigm that enables the
memory system to make use of prediction-based learning.
We developed such a paradigm and conducted a series of behavioral experiments
to test whether prediction-based learning actually occurs when applicable. For some
study list items, a repetition of that item was preceded by a repetition of its original
context. Using context, participants could predict when these item repetitions would
occur. We focused on subsequent memory performance for other items that did not
repeat, even though their contexts did. Across two studies we demonstrated a context
repetition effect: Later repetition of an item’s context enhanced memory for that item. We
used different stimulus types across the two experiments, images and words, to ensure
that the results are not unique to one stimulus category. We did not find this effect in two
additional experiments where we eliminated the ability to predict item repetitions based
on context. These results suggest that the human memory system is capable of learning
based on prediction; however, it may only do so when predictions are perceived as
credible.
We begin in the next chapter by describing our behavioral paradigm and results in
more detail. For clarity, we keep our theoretical explanation pithy while discussing the
empirical findings. In the following chapter, we fully describe our theory. This
8
description is accompanied by a computational implementation, along with qualitative
and quantitative assessments of how well the model fits our behavioral data. We then use
neuroimaging to explore the interplay between memory and perception. Finally, we
present some considerations and future directions for the research presented here.
9
Chapter 2: Repeating an Item’s Temporal Context Enhances its Memorability1
The ability to predict the occurrence of future stimuli and environments confers
obvious advantages in terms survival and well-being. Indeed, this utility may have
secured prediction’s place as one of the fundamental functions of the brain, across
domains and levels of analysis (Bar, 2009). Prediction plays a potential role in everything
from perception (Stokes, Atherton, Patai, & Nobre, 2012) and stimulus–response learning
(Rescorla & Wagner, 1972) to the experience of emotions (Kirkland & Cunningham,
2011) and the navigation of social interactions (Mitchell, 2009). In particular, the medial
temporal lobe (MTL) has been implicated as a system for producing such predictions
(Buckner, 2010; Johnson et al., 2007; Levy, 1989; Levy et al., 2005; Lisman & Redish,
2009; Schacter et al., 2007).
The MTL has long been known to be critical for storing and retrieving episodic
memories (Eichenbaum, 2000, 2004), which it accomplishes by binding neural
representations of stimuli to representations of the context in which they are experienced
(Davachi, 2006). Context can refer to any element of one’s experience, including spatial
information (Mizumori, Ragozzino, Cooper, & Leutgeb, 1999), environmental
configuration (Fanselow, 2000), and internally generated mental information (Kennedy &
1
This chapter presents the research previously published in an article to The Journal of Experimental
Psychology: General (Smith, Hasinski, & Sederberg, 2013). The dissertation author was an author on that
publication.
10
Shapiro, 2004). As mentioned previously, when we refer to context, we are referring
specifically to temporal context: a neural representation of recently experienced past
states (Howard & Kahana, 2002). Previous research has utilized the concept of temporal
context to account for many aspects memory, including human behavioral effects
(Howard & Kahana, 2002; Polyn et al., 2009; Sederberg et al., 2008) and non-human
MTL function (Howard, Fotedar, Datey, & Hasselmo, 2005).
Given that the MTL is critical for the formation and retrieval of memories
(Davachi, 2006; Eichenbaum, 2000, 2004), as well as the use of those memories to
generate predictions (Buckner, 2010; Lisman & Redish, 2009; Schacter et al., 2007), a
logical question is whether or not these predictions can also subsequently influence
memory processes. The answer to this question depends on the learning mechanism used
by the MTL. Several models of episodic memory (Howard & Kahana, 2002; Polyn et al.,
2009; Sederberg et al., 2008) posit that the MTL uses Hebbian learning (Hebb, 1949),
whereby associations are formed between context and the to-be-learned stimulus. Other
memory models similarly learn associations between stimuli that occur close in time, as
well as between those stimuli and a list context (Gillund & Shiffrin, 1984; Raaijmakers &
Shiffrin, 1981). No variant of this simple associative learning involves prediction. If the
memory system utilizes a simple Hebbian or associative learning rule, as these models
suggest, then predictions should not influence memory.
Alternatively, “prediction-based” learning rules also modify associations between
context and predicted stimuli. These predictions are generated using information the
memory system has already learned. Variants of prediction-based learning have been
11
successfully used in models of stimulus-response learning (Rescorla & Wagner, 1972)
and semantic memory (Shankar, Jagadisan, & Howard, 2009). Prediction-based learning
is also widely used in the field of reinforcement learning (Sutton & Barto, 1998) to
develop agents that maximize rewards by optimally navigating from state to state (Dayan,
1993; White, 1995). Although prior theoretical work has proposed applying a predictionbased learning rule to the task of learning episodes (Gershman et al., 2012), to our
knowledge the behavioral consequences of using such rule have not been tested
empirically.
Comparing prediction-based learning to simple Hebbian learning is not
meaningful using typical memory paradigms where encoding lists are carefully
randomized to avoid regularities or where each stimulus is presented only once. When
stimuli are encountered only once, predictions cannot be effectively made. As a result,
prediction-based and Hebbian learning rules would behave the same way: merely binding
each stimulus to context. When sequences of stimuli are repeated, however, predictions
about what stimuli will be experienced next can be made based on prior experience.
These predictions can be used to enhance learning beyond what is possible with
traditional Hebbian learning. Under Hebbian learning, repetitions will lead to
strengthened memory only for the repeated stimuli. In contrast, prediction-based learning
rules can also modify memory based on what is predicted to recur. Thus, comparing
memory models using Hebbian vs. prediction-based learning requires an experimental
design where study lists occasionally repeat sequences, thereby introducing predictability
into the experimental environment.
12
In such an environment, prediction-based learning can deviate from Hebbian
learning in two ways: Firstly, predictions about what the current stimulus should be—
based on prior learning—can curb additional learning for the current stimulus.
Specifically, stimuli that were accurately predicted to recur do not require much
additional strengthening. This is analogous to error-driven rules, used in animal models
of stimulus-response learning (Rescorla & Wagner, 1972), where only the predictionerror is learned. In this way, prediction-based learning can avoid catastrophic strength
growth, which can plague Hebbian algorithms (Goodhill & Barrow, 1994).
Secondly, by learning predictions about what stimulus should appear in the near
future—again based on previous experience—memory can also be strengthened for
stimuli that were predicted to recur but never do. For example, if a sequence of three
stimuli is experienced and later the first two stimuli from that sequence are reexperienced, one may then predict to see the third stimulus again. A prediction-based
learning system could learn this prediction, strengthening memory for the third stimulus
even if it does not actually recur. Simple Hebbian learning will not learn this prediction.
We will define that third stimulus as a “target” because memory strength for it will allow
us to discriminate between Hebbian and prediction-based accounts of episodic learning.
Consistent with prior memory research (Howard & Kahana, 2002; Polyn et al., 2009;
Sederberg et al., 2008), we will define the two stimuli that preceded the target as its
“context.” Concretely, prediction-based learning rules predict that a once-presented
stimulus whose temporal context is later repeated should be better remembered relative to
other once-presented stimuli whose contexts do not repeat. Hebbian learning rules predict
13
no difference in subsequent memory. We will refer to enhanced memory for oncepresented targets whose temporal contexts later repeat as a “context repetition effect.”
In summary, prediction-based learning offers advantages over Hebbian learning.
By making use of prediction, a memory system could both bound the growth of memory
strengths and strengthen memory for items that are predicted to reappear but do not.
Whether the human memory system actually employs prediction-based learning is an
open empirical question. Because Hebbian and prediction-based learning produce
divergent predictions only when sequences of stimuli are repeated, one cannot
discriminate between these frameworks using conventional memory paradigms where
stimulus order is randomized to avoid regularities. Hence, we developed a paradigm with
repeated stimulus sequences designed to elicit a context repetition effect. The presence or
absence of such an effect would be evidence that the brain uses prediction-based or
Hebbian learning, respectively. We conducted a series of experiments to answer this
question.
Overview of Experiments
Here we present four experiments that tested for the existence of prediction-based
learning as indicated by the presence or absence of a context repetition effect.
Specifically, we created a controlled environment that featured repetitions of sequences
of stimuli. In Experiments 1 and 2, the list sequences were designed such that the
recurrence of certain target stimuli could be predicted by the repetition of context stimuli
that preceded these targets’ first presentations. Experiments 3 and 4 omitted this
regularity. We used two different stimulus classes—images for Experiments 1 and 3 and
14
words for Experiments 2 and 4. These classes were used to investigate whether any
effects found could be influenced by stimulus imageability or rehearsability. The
resulting set of four experiments comprise a 2 (environmental predictability: high
vs. low) X 2 (stimulus type: images vs. words) factorial design.
The predictability factor investigates whether the presence or magnitude of a
context repetition effect is contingent on the existence of predictable repetitions in the
current environment. Because prediction-based learning will be most useful in
circumstances where accurate prediction is possible, we posit that the memory system
modulates its generation or use of predictions based on environmental predictability. The
result should be a down-weighting of prediction-based learning in unpredictable
environments. Therefore, we hypothesize that a context repetition effect will occur only
in the high-predictability experiments (1 and 2) and not—or at least to a lesser degree—in
the low-predictability experiments (3 and 4).
The stimulus type factor investigates the generality of a context repetition effect
across stimulus classes that differ in terms of imageability and rehearsability. If context
repetition effects are the result of stimulus vividness or rehearsal (Hasher & Zacks,
1979), then such an effect should only appear in Experiment 1 or 3, respectively. We do
not have strong a priori hypotheses regarding this factor.
15
Experiment 1: A Predictable Environment with Scenes
Method
Participants. Participants were 40 Ohio State introductory psychology students.
Each participant provided informed consent and participated in exchange for partial
credit in the course.
Design. Experiments 1 and 2 utilized an identical, within-subject design. The
experiment was divided into 3 parts: encoding, distraction, and test.
During encoding, participants experienced study lists as streams of stimuli,
presented one at a time. Unbeknownst to participants, these stimuli were arranged as
pairs of triplets. For each triplet, the first two stimuli were defined as the (temporal)
context of the third stimulus: the target. These triplets were paired such that each triplet
had a corresponding triplet that occurred either earlier or later in the same list. The lag
between these corresponding triplets varied randomly from six to 21 stimuli (i.e. two to
seven intervening triplets). Figure 1 provides a visualization of the triplet-pair structure
and experimental conditions.
The different experimental conditions were determined based on differences
between triplets of each pair. Experiments 1 and 2 utilized a 2 (context: same
vs. different) X 2 (target: same vs. different) factorial design. This led to four primary
conditions: Firstly, a repeated-context, repeated-target condition (RCRT) featured
identical context and target stimuli for both triplets within each triplet-pair (i.e. 2
identical triplets for each pair). Critically, this condition creates a high-predictability
environment where the recurrence of targets can be predicted when a context is repeated.
16
Secondly, a repeated-context, novel-target condition (RCNT) features identical contexts
but different targets for both triplets within each triplet-pair (i.e. the first and second
elements of both triplets in each pair are identical). This condition provides an
opportunity to elicit a context repetition effect: Later repetition of a target’s context may
enhance memory for that target even though it is never repeated. Together these two
conditions create an experimental environment where the repetition of a particular
context indicates a 50% chance that the next stimulus will be the target that previously
followed this context.
Thirdly, a novel-context, repeated-target condition (NCRT) features identical
targets but different contexts for both triplets in each triplet-pair (i.e. the third element of
both triplets in each pair are identical). This condition allows us to estimate memory
performance for twice-presented stimuli in the absence of contextual repetitions.
Fourthly, in a novel-context, novel-target condition (NCNT) neither the context stimuli
nor the target stimuli are identical for paired triplets (i.e. no identical elements between
triplets of a given pair). This condition serves as a control, allowing us to estimate
baseline memory performance for once-presented stimuli.
Note that this design produces four nested sub-conditions within the two noveltarget conditions: RCNT1 or NCNT1, the first triplet within each triplet-pair, and RCNT2
or NCNT2, the second triplet within each triplet-pair. To look for effects of contextual
repetition on memory for once-presented targets, we contrasted the memory performance
for each repeated-context novel-target sub-condition with its corresponding novel-context
novel-target sub-condition (RCNT1 vs. NCNT1 and RCNT2 vs. NCNT2, respectively).
17
The first contrast lets us test for a context repetition effect for memory. The second
contrast allows us to distinguish between associative and predictive accounts of this
context repetition effect.
Figure 1. Structure of encoding lists. Each color represents an example triplet-pair for
one of the four main experimental conditions. The leftmost condition, RCRT, does not
appear in Experiments 3 and 4. First and second triplets appear on upper and lower rows,
respectively. The lag between actual triplet pairs was randomized (see Experiment 1
methods).
Materials. We chose images as stimuli because we wanted a stimulus set that
would not lend itself to verbal rehearsal. Specifically, we used a pool of 645 color
photographs of indoor (336) and outdoor (309) scenes that had been used in previous
research (Turk-Browne, Simon, & Sederberg, 2012). Stimuli for each participant were
drawn randomly from this pool. Each study list consisted of 96 presentations, but only 72
unique stimuli, because some presentations were repetitions. We presented five unique
encoding lists to each participant. For each encoding list, we created a corresponding 96item test list. Each test list contained all 24 targets from the corresponding encoding list
along with 24 randomly selected context items and 48 lures drawn at random from the
stimulus pool. We used the Python Experiment Programming Library (Geller, Schleifer,
18
Sederberg, Jacobs, & Kahana, 2007) to present stimuli to participants and record their
responses.
Procedure. Participants first completed the encoding phase, during which they
were exposed to five lists of stimuli. They were not told that memory would be assessed
later, in an effort to capture incidental learning. Each stimulus was presented for 1200
ms, centered on a monitor, with a 600- to 800-ms jittered interstimulus interval. To keep
participants attending to the stimuli, we asked them to complete a task during the
encoding phase. Participants judged each stimulus to be an image of either an indoor or
outdoor scene by pressing one of two keys.
Participants then completed the distraction phase, which consisted of an n-back
task for 5 minutes. The n-back task is divided into a series of blocks. During each block,
participants are shown a stream of numbers, one at a time. During each presentation,
participants judge whether the current number is the same as the number that occurred n
presentations earlier. The value of n varies across blocks, from 1 to 4, with higher
numbers presenting greater challenge.
Finally, they completed a surprise item-recognition test, which assessed memory
for all target stimuli, a subset of the context stimuli, and an equal number of non-studied
lures. Test probes were presented one at a time for 1,200 ms, with a 600- to 800-ms
jittered interstimulus interval. On each test trial, participants judged a probe to be “old” or
“new” by pressing one of two keys. The order of test trials was randomized and, unlike
the encoding phase, did not feature any repetition or regularity.
19
Results
We quantified memory performance in terms of within-participant hit rates and
false-alarm rates: the number of tested stimuli that a participant judged as “old” divided
by the total number of stimuli for which they made judgments. Hit rates were calculated
separately for targets and context stimuli in each condition and sub-condition and false
alarm rates were calculated for lures (see Figure 2). Considering all the stimuli at once, a
2 (context repetition: novel vs. repeated) X 2 (target repetition: novel vs. repeated) X 2
(stimulus type: context vs. target) repeated measures analysis of variance (ANOVA)
revealed main effects of context repetition, F(1, 39) = 45.90, MSE = 0.0091, p < .0001,
and target repetition, F(1, 39) = 12.15, MSE = 0.0119, p = .0012, but no significant main
effect of stimulus type, F(1, 39) = 2.08, MSE = 0.0080, p = .1570. We also found
significant interactions between stimulus type and both context repetition, F(1, 39) =
26.36, MSE = 0.0084, p < .0001, and target repetition, F(1, 39) = 20.32, MSE = 0.0063, p
< .0001. Given that we found significant interactions with stimulus type, we proceeded to
separately analyze target and context stimuli.
Memory for Targets. We first looked for differences in memory performance
between the four main conditions. A 2 (context repetition: novel vs. repeated) X 2 (target
repetition: novel vs. repeated) repeated measures ANOVA on target stimuli demonstrated
a main effect of target repetition, F(1, 39) = 36.38, MSE = 0.0075, p < .0001. Repeated
targets (M = 0.55, SEM = 0.017) were more likely to be remembered than once-presented
targets (M = 0.47, SEM = 0.016). We found no significant main effect of repeating
context, F(1, 39) = 1.75, MSE = 0.0087, p = .1940. Additionally, we found no interaction
20
between context and target repetition (F < 1). In other words, only the repetition of
targets improved memory performance for targets.
We then examined the nested sub-conditions for the once-presented stimuli,
looking for evidence of a context repetition effect via planned contrasts. We found that
once-presented targets were more likely to be recognized if their contexts were later
repeated (RCNT1) relative to other once-presented targets whose contexts were never
2
repeated (NCNT1), t(39) = 2.46, p = .0185, 𝜔𝑝𝑎𝑟𝑡𝑖𝑎𝑙
= 0.06. In contrast, we found no
difference in memory performance for once-presented targets whose contexts had already
been presented (RCNT2) relative to corresponding once-presented targets with novel
contexts (NCNT2), t(39) = 0.11, p = .9 (see Figure 3, panel A).
Memory for Contexts. Again, we first looked for differences in memory
performance between the four main conditions. A 2 (context repetition: novel
vs. repeated) X 2 (target repetition: novel vs. repeated) repeated measures ANOVA on
context stimuli demonstrated a main effect of context repetition. Specifically, context
stimuli were more likely to be remembered F(1, 39) = 71.51, MSE = 0.0087, p < .0001,
when they were repeated (M = 0.59, SEM = 0.018) than if they were presented only once
(M = 0.47, SEM = 0.015). This finding parallels the effect of target repetition on memory
for targets, mentioned above. The ANOVA revealed no significant main effect of target
repetition (F < 1) or the interaction between context and target repetition (F < 1).
Similar to our analysis of the target stimuli, we next examined the nested subconditions, looking for an effect of target repetition on their corresponding context
stimuli. Planned contrasts revealed no effect of later target repetition on memory for
21
once-presented novel context stimuli. This null result was found both for the first triplets
from each triplet-pair (NCRT1 vs. NCNT1), t(39) = 0.1638, p = .80, as well as the
second triplets from each triplet pair (NCRT2 vs. NCNT2), t(39) = 0.
Preliminary Discussion
Experiment 1 tested for evidence of a context repetition effect in a moderately
predictable environment. Specifically, we looked for changes in recognition memory for
once-presented images due to repetition of their temporal contexts at a later point in time.
We found that repeating the context of a once-presented target (RCNT1) increased the
likelihood of subsequently recognizing that target. However, contextual repetition had no
effect on memory for the new target (RCNT2) that followed a context’s second
presentation.
Not surprisingly, we also observed item repetition effects for both context and
target stimuli: Twice-presented images were more likely to be subsequently recognized
than once-presented images. Additionally, we found no evidence of a bidirectional effect
of repetition. That is, repetition of a target did not alter the probability of subsequently
remembering the context that preceded either its first (NCRT1) or second (NCRT2)
presentation.
This counterintuitive pattern of effects cannot be explained by traditional theories
of memory. Strength-based theories of recognition memory (Murdock, 1993) have no
mechanism for improving memory performance for an item that is not repeated. Other
theories, using associative learning mechanisms to explain recognition memory (Dennis
& Humphreys, 2001; Gillund & Shiffrin, 1984; Howard & Kahana, 2002; Shiffrin &
22
Steyvers, 1997), also have problems explaining the full pattern of results. Generally,
these models say that the likelihood of recognizing a target depends on the strength of the
association between that target and context. Some models could predict that repeating a
context would increase that context’s memory trace. This could then enhance memory for
all targets associated with this context, making them seem more familiar at test. This
could explain half of the context repetition effect: enhanced memory for targets whose
contexts are later repeated. However, this mechanism would also predict enhanced
memory for the novel targets that follow repetitions of context. In contrast, the context
repetition effect in Experiment 1 shows enhanced memory only for targets that follow the
first presentation of a twice-presented context. Some associative theories of memory
would also predict a target repetition effect: enhanced memory for context items that
precede twice-presented targets. However, our results do not bear out this prediction.
A theory of memory that includes a prediction-based learning mechanism
naturally explains the full pattern of results. When a context is experienced for the first
time, the memory system has no basis for making a prediction about what target could
appear next. However, the context does become associated with the actual subsequent
stimulus via a Hebbian-like learning process. When this context occurs again at a later
time, these context-target associations enable the memory system to predict that the same
target may also be repeated. If this prediction reinstates the target’s representation then
the target’s associations to context will be further strengthened. Thus, strengthening can
occur even if the target does not actually repeat. In this case, the novel target that follows
the repeated context (RCNT2) would become associated to that context just like any
23
other once-presented target (e.g. NCNT2). If the target actually does repeat as predicted,
the amount of additional Hebbian learning that occurs during the repetition will only be
as large as the prediction error (Rescorla & Wagner, 1972). Thus, the more a target is
strengthened in memory due to prediction, the less additional strengthening will occur as
a result of its actual repetition. Thus, repeating the context of a target could enhance
memory for it almost as much as repeating the target itself.
Experiment 2: A Predictable Environment with Words
Our second experiment had two purposes: Firstly, we attempted to replicate the
pattern of results found in Experiment 1. Secondly, we used a different stimulus class to
ensure that the results of Experiment 1 are not specific to scenes or highly imageable
stimuli more generally. Thus, Experiment 2 featured an identical design, but used words
as stimuli instead of scenes.
Method
Participants. Participants were 31 Ohio State introductory psychology students.
Each participant provided informed consent and participated in exchange for partial
credit in the course.
Design, Materials, and Procedure. We employed the same experimental design
used in Experiment 1. The stimulus pool consisted of 630 medium- to high-frequency
words, which were selected from the University of Southern Florida free-association
database (Nelson, McEvoy, & Schreiber, 2004). For each participant, a random
subsample from this pool was selected to create unique lists, paralleling Experiment 1.
The only difference in procedure from Experiment 1 was a change to the task that
24
participants performed during the encoding phase. In Experiment 2, participants judged
whether each word represented a living or nonliving thing. Although there is no correct
answer for many of the words in the pool, a prior internal pilot study demonstrated that
approximately the same numbers of words were judged to be living and nonliving.
Results
Considering all stimuli at once, a 2 (context repetition: novel vs. repeated) X 2
(target repetition: novel vs. repeated) X 2 (stimulus type: context vs. target) repeated
measures ANOVA indicated a main effect of context repetition, F(1, 30) = 17.16, MSE =
0.0069, p = .0003; along with a marginal main effect of target repetition, F(1, 30) = 4.12,
MSE = 0.0114, p = .0514. However, we found but no main effect of stimulus type (F <
1). As with Experiment 1, we also found a significant interaction between target
repetition and stimulus type, F(1, 30) = 8.79, MSE = 0.0123, p = .0059. However, unlike
Experiment 1, the interaction between context repetition and stimulus type was not
significant, F(1, 30) = 1.31, MSE = 0.0112, p = .20. Due to the significant target
repetition by stimulus type interaction, we next analyzed context and target items
separately, paralleling the analyses of Experiment 1.
Memory for Targets. We first looked for differences in memory performance
between the four main conditions. A 2 (context repetition: novel vs. repeated) X 2 (target
repetition: novel vs. repeated) repeated measures ANOVA on target items revealed a
main effect of target repetition, F(1, 30) = 13.09, MSE = 0.0115, p = .0011. Specifically,
twice-presented targets (M = 0.67, SEM = 0.022) were more likely to be subsequently
remembered than once-presented targets (M = 0.60, SEM = 0.018). We also found a
25
marginal main effect of context repetition, F(1, 30) = 4.08, MSE = 0.0060, p = .0525.
Targets presented with repeated contexts (M = 0.65, SEM = 0.022) were remembered
marginally better than targets presented with novel contexts (M = 0.62, SEM = 0.019).
We did not find a significant interaction between target and context repetition (F < 1).
We then examined the nested sub-conditions for the once-presented stimuli,
looking for evidence of a context repetition effect via planned contrasts. These contrasts
replicate the results from Experiment 1. Once-presented targets whose contexts were later
repeated (RCNT1) were more likely to be subsequently recognized, relative to other
once-presented targets whose contexts were never repeated (NCNT1), t(30) = 2.24, p =
2
.0327, 𝜔𝑝𝑎𝑟𝑡𝑖𝑎𝑙
= 0.06. Also consistent with Experiment 1, we found no difference in
memory performance between once-presented targets that followed a repetition of
context (RCNT2) and control targets (NCNT2), t(30) = -0.63, p < .5 (see Figure 3, panel
B).
Memory for Contexts. Shifting to the context stimuli, we first looked for
differences in memory performance between the four main conditions. A 2 (context
repetition: novel vs. repeated) X 2 (target repetition: novel vs. repeated) repeated
measures ANOVA on context stimuli revealed a main effect of context repetition, F(1,
30) = 8.94, MSE = 0.0120, p = .0055, consistent with Experiment 1. Twice-presented
context stimuli were better remembered (M = 0.67, SEM = 0.023) than once-presented
contexts (M = 0.61, SEM = 0.019). We found no significant main effect of target
26
repetition on memory for context stimuli (F < 1). Similarly we found no significant
interaction between context and target repetition, F(1, 30) = 2.95, MSE = 0.0138,
p = .0964.
Planned contrasts, testing for an effect of target repetition on memory for the
preceding context stimuli, once again showed no effect of target repetition on memory
performance for once-presented contexts. This was true whether a context preceded the
first presentation of a target (NCRT1 vs. NCNT1), t(30) = 0.55, p = .50, or the second
presentation (NCRT2 vs. NCNT2), t(30) = 0.71, p = .40.
27
Figure 2. Memory performance across conditions. Bars show the mean hit rate for all
target conditions and the false alarm rate for lures. Panels A, B, C, and D correspond to
Experiments 1, 2, 3, and 4, respectively. Error bars indicate standard errors of the mean.
28
Figure 3. Context repetition effects. Bars show the effect of contextual repetition: Within
each panel, the left bar indicates the mean difference in hit rate between once-presented
targets whose contexts later repeat and control targets (RCNT1 vs. NCNT1), while the
right bar indicates the mean difference in hit rate between once-presented targets that
follow repeated contexts and control targets (RCNT2 vs. NCNT2). Panels A, B, C, and D
correspond to Experiments 1, 2, 3, and 4, respectively. Error bars indicate standard errors
of the mean. ∗ 𝑝 < 0.05.
Preliminary Discussion
Experiment 2 replicated the main results from Experiment 1 while using a
different stimulus category, words instead of images. In addition to expected item
29
repetition effects for both context and target stimuli, we once again found a context
repetition effect: enhanced recognition memory performance for once-presented items
whose temporal context is later repeated, but no enhancement for once-presented images
that follow a repetition of context. Together with the findings from Experiment 1, these
results provide strong evidence for the context repetition effect and suggest that it may
occur for a wide variety of stimulus categories.
Experiment 3: An Unpredictable Environment with Images
We describe the encoding phases of Experiments 1 and 2 as “predictable
environments” because the participants saw sequences of three stimuli (context and
target) fully repeated at a later time (the RCRT condition) as well as sequences of two
stimuli (context only) repeated (the RCNT condition). Thus, when a context recurred, it
was possible to predict what target would follow, albeit not perfectly. The RCRT
condition may have indicated to participants that this type of prediction had utility in the
experimental environment. In other words the predictable environment may have
prompted participants’ memory systems to generate or make use of these predictions,
which ultimately enhanced memory.
We test this hypothesis in Experiments 3 and 4 by removing the RCRT condition,
thereby reducing the predictability of the experimental environment. Specifically, target
repetitions can no longer be predicted from repetitions of context. We predict that the
reduction in predictability will reduce a memory system’s use of prediction. This should
eliminate, or at least attenuate, the context repetition effect.
30
Method
Participants. Participants were 52 Ohio State introductory psychology students.
Each participant provided informed consent and participated in exchange for partial
credit in the course.
Design, Materials, and Procedure. Experiment 3 was identical to Experiment 1,
except that the RCRT condition was omitted. This reduced the predictability of the
environment by ensuring that participants never saw repeated sequences of more than two
stimuli. The resulting design featured single main condition with three-levels: NCRT,
RCNT, and NCNT. Consistent with the previous studies, the novel target conditions
contained nested sub-conditions: RCNT1, RCNT2, NCNT1, and NCNT2. The materials
and procedure were also identical to Experiment 1, except that the number of triplet pairs
in each condition was increased slightly to keep the duration and stimulus exposure
comparable to Experiment 1. Each participant saw a total of 450 stimulus presentations
(375 unique pictures), across five blocks. As with Experiments 1 and 2, test lists were
constructed to correspond to each encoding list. Each test list featured all 25 target items
and 20 randomly chosen context items from the relevant encoding list, along with 45
lures.
Results
Because the context and target repetition factors no longer create a 2 X 2 factorial
design (as they did in Experiments 1 and 2), the three remaining conditions were merged
into a single factor. The result is a 3 (condition: NCRT, RCNT, and NCNT) X 2
(stimulus type: context vs. target) repeated measures ANOVA. Paralleling Experiments 1
31
and 2, we found a significant main effect of condition, F(2, 102) = 11.08, MSE = 0.0078,
p < .0001, and a significant interaction between condition and stimulus type, F(2, 102) =
42.49, MSE = 0.0072, p < .0001. Also paralleling Experiments 1 and 2, we found no
main effect of stimulus type (F < 1). We then used planned pairwise t tests to test for
differences between the three experimental conditions separately for targets and contexts.
Memory for Targets. Planned comparisons revealed better memory for repeated
targets (NCRT) than once-presented targets with either once-presented contexts (NCNT),
t(51) = 6.45, p < .0001, or twice-presented contexts (RCNT), t(51) = 7.04, p = .0001.
Additionally, whether context was twice-presented (RCNT) or not (NCNT) had no
significant overall effect on memory for once-presented targets, t(51) = 1.06, p = .20.
Thus, consistent with Experiments 1 and 2, target repetition, but not context repetition
was associated with better memory for target items.
We then tested the novel-target sub-conditions for the context repetition effect.
Importantly, we found no such effect for once-presented targets whose contexts were
later repeated (RCNT1 vs. NCNT1), t(51) = -0.42, p = .60, or for once-presented targets
that followed a repetition of context (RCNT2 vs. NCNT2), t(51) = -0.89, p = .30 (see
Figure 3, panel C).
Memory for Contexts. Memory for context stimuli displayed a pattern consistent
with our previous experiments. Specifically, we found an item repetition effect: better
subsequent memory for twice-presented contexts (RCNT) relative to once-presented
contexts that preceded either once-presented targets (NCNT), t(51) = 4.81, p < .0001, or
twice-presented targets (NCRT), t(51) = 5.45, p < .0001.
32
We also found no evidence that repeating a target influenced memory for the
context items that preceded either the target’s first presentation (NCRT1 vs. NCNT1),
t(51) = 0.34, p = .70, second presentation (NCRT2 vs. NCNT2), t(51) = 0.03, p = .90, or
both presentations taken together, t(51) = 0.22, p = .80.
Preliminary Discussion
Many of Experiment 3’s results are similar to findings from Experiment 1.
However, unlike Experiment 1, we found no evidence of a context repetition effect. The
only difference between these experiments was the removal of the RCRT condition for
Experiment 3. This suggests that predictable environments are necessary to produce a
context repetition effect. To replicate this null result, we ran a second experiment without
the RCRT condition.
Experiment 4: An Unpredictable Environment with Words
Method
Participants. Participants were 71 Ohio State introductory psychology students.
Each participant provided informed consent and participated in exchange for partial
credit in the course.
Design, Materials, and Procedure. Experiment 4’s design was identical to the
design used in Experiment 3; however, the stimulus pool was the same set of words used
in Experiment 2. Accordingly, the encoding task was the same living-nonliving judgment
used in Experiment 2.
33
Results
Experiment 4’s results were similar to the results of Experiment 3. The 3
(condition: NCRT, RCNT, and NCNT) X 2 (stimulus type: context vs. target) repeated
measures ANOVA detected a significant main effect of condition, F(2, 140) = 16.63,
MSE = 0.0065, p < .0001, a significant interaction between condition and stimulus type,
F(2, 140) = 43.56, MSE = 0.0070, p < .0001, but no significant main effect of stimulus
type (F <1).
Memory for Targets. Consistent with the previous studies, we found an item
repetition effect for targets: Twice-presented targets (NCRT) were more likely to be
remembered than once-presented targets with novel contexts (NCNT)), t(70) = 7.24, p <
.0001, or repeated contexts (RCNT), t(70) = 8.09, p < .0001. Additionally, subsequent
memory performance did not differ for novel targets with novel vs. repeated contexts,
t(70) = 0.53, p < .50.
We again found no evidence of a context repetition effect: Recognition memory
was not improved for once-presented targets whose context was later repeated (RCNT1
vs. NCNT1), t(70) = -0.50, p = .60, or for once-presented targets that followed a
repetition of context (RCNT2 vs. NCNT2), t(51) = -0.19, p = .80 (see Figure 3, panel D).
34
Memory for Contexts. Once again, repeated contexts were more likely to be
remembered than the novel contexts that preceded either novel targets, t(70) = 5.49, p <
.0001, or repeated targets, t(70) = 5.51, p < .0001. Further, target repetition did not affect
subsequent memory for contexts (NCRT vs. NCNT), t(70) = 0.69, p = .40. Null effects of
target repetition on context were also found when we separately tested first presentations
(NCRT1 vs. NCNT1), t(70) = 1.11, p = .20, and second presentations (RCNT2 vs.
NCNT2), t(70) = 0.05, p = .90.
Preliminary Discussion
The absence of a detectable context repetition effect in Experiments 3 and 4 is
evidence supporting our claim that only predictable environments enable or cue the
memory system to use prediction-based learning. In environments where meaningful
predictions from context are not possible, the repetition of a target’s temporal context will
not improve memory for that target. Experiments 3 and 4 suggest that human memory
processes are sensitive to the utility of predictions in the current environment and that use
of prediction-based learning is modulated by this utility.
Discussion
In two experiments, we found that repeating the temporal contexts of oncepresented stimuli enhanced their memorability relative to once-presented control stimuli
whose contexts were never repeated. However, we found no difference in subsequent
memory for the similarly once-presented stimuli that followed a repetition of context
relative to once-presented control stimuli. We termed this combined pattern of behavior
the context repetition effect. We expected this pattern if the episodic memory system
35
learned using predictions (Gershman et al., 2012) rather than simply learning contextstimulus associations (Howard & Kahana, 2002). In order to elicit the effect, we
constructed encoding lists where sequences of stimuli were occasionally repeated after
some delay. Thus, sometimes the repetition of a target stimulus could be predicted after a
repetition of the contextual stimuli it originally followed.
A very different pattern emerged when we reduced the predictability of the
experimental environment. In two additional experiments, where full context-target
pairings were never repeated, we found no effect of contextual repetition for the oncepresented targets that followed either presentation of a twice-presented context. These
four experiments, taken together, are strong preliminary evidence that the human episodic
memory system employs a type of prediction-based learning when—and perhaps only
when—the current environment provides regularity. That is, the memory system can
encode information that is predicted to appear in the environment, even when that
information does not, in fact, appear. However, this will only occur when these
predictions have a reasonable chance of being accurate. We found this context repetition
effect for multiple stimulus classes: pictures of scenes and words.
Implications for the Study of Memory
The existence of a context repetition effect means some sort of prediction-based
learning mechanism must be incorporated into any theory that intends to be a truly
comprehensive explanation of human memory. Indeed, while many models of memory
that utilize Hebbian learning (Howard & Kahana, 2002; Polyn et al., 2009; Sederberg et
al., 2008) or similar mechanisms (Gillund & Shiffrin, 1984; Raaijmakers & Shiffrin,
36
1981) are sufficiently powerful to account for many episodic memory effects, they would
fail to reproduce the context repetition effect. No model currently offers anything close to
complete account of memory. Why, then, should models of memory undergo revision to
account for this particular effect, especially when it only manifests under peculiar
experimental circumstances involving repetitions of sequences of stimuli (i.e. contexts
and targets)?
Such regularity is intentionally avoided at all cost in typical, tightly-controlled
memory experiments. Thorough randomization procedures are used to ensure that item or
inter-item idiosyncrasies do not give rise to false positives or mask real effects. In fact,
the memory lists used in most paradigms avoid any stimulus repetition whatsoever.
Studies of the spacing effect (Glenberg, 1979) and item repetition effects (Ratcliff,
Hockley, & McKoon, 1985) are exceptions to this norm only because they, by definition,
require repetition of individual items. Even in these cases, however, list randomization is
still paramount, albeit more complicated, and predictable patterns are still avoided.
Consequently, a context repetition effect could never appear in a typical memory
experiment because the repetition of sequences required to elicit the effect would never
occur.
However, repetitions and regularities are ubiquitous outside the controlled
environment of the laboratory. The route one takes to any frequently visited destination
provides a prime example of repetition of sequences in the real world. Even in cases
where events are not guaranteed to occur—such as whether or not one will meet a
neighbor while walking down the street—potential outcomes can often be predicted.
37
These scenarios are analogous to the experimental environment created for the present
research. Given the amount of regularity—and hence predictability—in real-world
experience, the effects of prediction-based learning may be prevalent in everyday life.
Therefore, rather than viewing the context repetition effect as a peculiar phenomenon,
elicited by contrived circumstances, we view it as an important effect that provides
information both about how the MTL system learns and about the types of memory
phenomena that are being ignored using traditional encoding lists.
Fortunately, the present research is part of a growing literature that emphasizes
the importance of regularity, prediction, and the use of naturalistic stimuli. For instance,
spaced presentations of semantically related words lead to enhanced memory for the first
presentation. This occurs because the presentation of the second, related word reminds
the participant of the first word (Tullis, Benjamin, & Ross, 2014). Prediction is also
important for segmenting one’s stream of experience into separate episodic events
(Zacks, Speer, Swallow, Braver, & Reynolds, 2007). The predictability of the
environment also influences neural responses underlying the perception of visual stimuli
(Summerfield, Trittschuh, Monti, Mesulam, & Egner, 2008; Turk-Browne et al., 2012).
Still other researchers use advancements in experience sampling techniques to investigate
memory for stimuli that actually occur in participants’ everyday lives (Milton et al.,
2011; Nielson, Smith, Sreekumar, & Sederberg, submitted). Experience sampling
methods potentially offer the best chance to capture the regularities present in everyday
life.
38
Open Questions and Future Directions
Although the effects of prediction-based learning and environmental regularities
may be commonplace, important boundary conditions may exist. The current work used
two different stimulus classes: images of scenes, which are both highly imageable and
difficult to rehearse, as well as medium to high frequency words, which are not as
imageable but much easier to rehearse. We found the context repetition effect for both
stimulus types. However, different stimuli may not produce these effects. For instance,
images of fractal patterns, which are imageable but relatively difficult to discriminate,
may not produce a context repetition effect. Another question is the degree of
predictability required to elicit a context repetition effect. In Experiments 1 and 2, which
produced context repetition effects, a repetition of context preceded the repetition of a
target 50% of the time. In Experiments 3 and 4, which did not produce a discernable
context repetition effect, a contextual repetition never signaled the repetition of a target
(i.e. 0% of the time). More work is needed to determine the predictability threshold at
which a context repetition effect will appear, as well as how the magnitude of the effect
relates to further increases in predictability (e.g. 75%). Finally, the effects of additional
repetitions of context are currently unknown. How a third repetition of context, with a
different novel target following it (i.e. RCNT3), influences memory for the targets that
followed the first (RCNT1) and second (RCNT2) presentations of context is an open
question.
Additionally, the neurological mechanisms that give rise to the context repetition
effect specifically, and prediction-based learning more generally are not fully understood.
39
The MTL is believed to play a critical role in the generation of predictions from episodic
memory (Buckner, 2010; Johnson et al., 2007; Levy, 1989; Levy et al., 2005; Lisman &
Redish, 2009; Schacter et al., 2007). However, other brain regions also seem to respond
to manipulations of predictability (Summerfield et al., 2008; Turk-Browne et al., 2012).
Much of this work centers on regions within the ventral visual processing stream, and
looks for changes in neural repetition attenuation: a reduction in neural responsiveness to
subsequent presentations of a particular stimulus (Grill-Spector, Henson, & Martin,
2006). For instance, the fusiform face area, a region of the inferior temporal lobe that
preferentially responds to faces (Kanwisher & Yovel, 2006; Kanwisher, McDermott, &
Chun, 1997), shows greater repetition attenuation during periods (i.e. experimental
blocks) when repetition is more likely to occur (Summerfield et al., 2008). Additionally,
the parahippocampal place area, a region that preferentially responds to scenes (Epstein
& Kanwisher, 1998), produces greater repetition attenuation when a repetition of a target
scene follows a repetition of its initial temporal context (Turk-Browne et al., 2012).
Future work will be required to determine where the neural correlates of the context
repetition effect lie. These correlates may be found in, and perhaps only in, the MTL.
Alternatively, the context repetition effect may have neural correlates in other regions,
perhaps depending on the stimuli used. For instance, scenes and words may produce
neural context repetition effects in the parahippocampal place area (Epstein &
Kanwisher, 1998) and word sensitive regions (Dehaene, Le Clec’H, Poline, Le Bihan, &
Cohen, 2002; McCandliss, Cohen, & Dehaene, 2003), respectively. Our results, along
40
with the existing findings of prediction-related effects in different brain areas, strongly
support the claim that prediction is a fundamental function of the brain (Bar, 2009).
Finally, although the prediction-based learning account is based on mechanisms
from reinforcement learning (Dayan, 1993; Gershman et al., 2012; White, 1995), it has
never been implemented in a full recognition memory model. Even up until this point, we
have only presented it as a verbal theory. This enabled us to focus on the empirical
findings, namely the context repetition effect. However, now that we have established
this effect, we must investigate the extent to which prediction-based learning actually can
account for memory enhancements due to contextual repetitions. We accomplish this in
the next chapter where we describe and test a computational implementation. In that
chapter, we will briefly review the experimental design and main findings before fitting
our model to some of the data described in this chapter. It turns out that capturing both
the context repetition effect in high-predictability environments and the null context
repetition effect in low-predictability environments will require a few important
deviations from conventional prediction-based learning rules.
Conclusions
The painstaking efforts memory researchers go through to “properly” randomize
their study lists may actually be causing them to miss important memory phenomena
caused by environmental regularities. Given that the environment outside the laboratory
is full of regularity and periodicity, it is not surprising that the brain, especially the MTL
makes use of it. We have shown that, in predictable environments, memory for a oncepresented stimulus can be improved by repeating the temporal context of that stimulus.
41
To our knowledge, this behavior can only be explained through a prediction-based
learning mechanism. The results presented here are part of a recent push to examine the
role that such regularities have on memory, perception, and the function of the brain more
generally (Bar, 2009).
42
Chapter 3: A Formal Account of the Context Repetition Effect2
As mentioned in the previous chapter, a considerable literature has developed to
explain the relationship between memory and prediction. Researchers studying both
humans and animals have found converging evidence that the medial temporal lobe
system (MTL), a collection of regions known to be critical for storing and retrieving
memories (Davachi, 2006; Eichenbaum, 2000, 2004), is also involved in making
predictions (Buckner, 2010; Johnson et al., 2007; Levy, 1989; Levy et al., 2005; Lisman
& Redish, 2009; Schacter et al., 2007). Sequences of events are critical for the generation
of predictions (Hawkins, George, & Niemasik, 2009), and the MTL seems especially
suited for learning sequential information (Lisman & Redish, 2009; Skaggs,
McNaughton, Wilson, & Barnes, 1996). Although the neocortex may also learn
sequences and use them to generate predictions (Hawkins et al., 2009), some have argued
that the MTL is required for all but the most basic predictions (Corbit & Balleine, 2000).
It seems then that prediction-making may actually be a primary function of the MTL and
a prevalent process in the brain overall (Bar, 2009; Schacter et al., 2007).
To date, most research at the intersection of memory and prediction has focused
on how memories or the memory system can enable and influence predictions. Less
2
This chapter presents research that is being written up in as a manuscript to be submitted to Psychonomic
Bulletin and Review. The authors will be Hasinski and Sederberg.
43
research has investigated how predictions may influence memory. One exception to this
trend is the set of experimental findings from the preceding chapter (Smith et al., 2013),
which demonstrated a context repetition effect. That is, participants were more likely to
subsequently recognize a once-presented item as occurring in a study list if the context of
that item was repeated later in the study phase. However, no boost in memory
performance was found for once-presented items that followed a repeated context. This
result cannot be readily explained by existing models of episodic memory. We present a
new computational model of memory that formally explains and reproduces the context
repetition effect. We first summarize the prior experimental findings, before explaining
our computational model and procedure. We devote to the rest of this article to assessing
model fit, comparing model variants, and discussing implications and future directions.
The Context Repetition Effect
Deviating from typical episodic memory paradigms that strive to avoid
predictable repetitions of stimuli, we designed a paradigm that intentionally repeated
sequences of stimuli (Smith et al., 2013). More precisely, sequences of three stimuli were
occasionally repeated in the same order at later points during the same encoding list. We
denoted the third item in each triplet as a “target” of interest and defined the preceding
two items as the “context” of the target. The term context is appropriate because the first
two items comprise the most recent experience, or temporal context (Howard & Kahana,
2002), of the target. We named this condition the repeated-context repeated-target
condition (RCRT). The experiment also featured a condition where targets were
presented twice, but the two context stimuli that preceded each presentation of a
44
particular target were unique for each target presentation. We referred to this as the
novel-context repeated-target condition (NCRT). Of particular importance was a third
repeated-context novel-target condition (RCNT), where context stimuli were presented
twice, but the target that followed each presentation of context was unique. As a result,
targets in this condition can be broken down into two sub-conditions: RCNT1 and
RCNT2, corresponding to the targets that followed first or second presentations of
particular contexts, respectively. Finally a novel-context novel-target condition (NCNT)
served as a control, its triplets consisting of only once-presented stimuli. NCNT1 and
NCNT2 sub-conditions correspond to the RCNT1 and RCNT2 sub-conditions. The entire
paradigm consisted of five encoding lists, then a short distractor task, and finally five
corresponding test lists. Readers are referred to the previous chapter for more details
about design and procedure.
Looking at the recognition performance across participants, we found two effects
of repetition. Firstly, we found an effect of item repetition: Twice-presented items were
more likely to be subsequently recognized as occurring during study than once-presented
items. This effect is not surprising as the relationship between additional presentations
and memory performance has been demonstrated (Ratcliff et al., 1985). More
importantly, we also found an effect of contextual repetition on memory for the targets
that followed: Participants were more likely to remember targets whose temporal
contexts were repeated at a later point in time (RCNT1) than other once-presented targets
whose contexts were never repeated (NCNT1). Furthermore, this effect was asymmetric:
Once-presented targets that followed the repetition of a context (RCNT2) were no more
45
likely to be subsequently recognized than control targets (NCNT2). These results were
consistent across two experiments that used very different stimulus categories: images of
scenes and words.
However, a very different pattern emerged when we removed the repeatedcontext repeated-target condition from the experimental design. In two additional
experiments, we found no evidence of elevated recognition memory for once-presented
targets whose contexts were later repeated relative to once-presented control targets
(i.e. RCNT1 vs. NCNT1). Memory performance for once-presented targets that followed
a repeated context did not differ from similar once-presented control targets (RCNT2
vs. NCNT2), consistent with the first two studies. In other words, we found a null context
repetition effect. Additionally, memory for twice-presented targets was significantly
better than memory for once-presented targets, replicating the traditional item repetition
effect that was also found in the first two studies.
To explain the context repetition effect, we previously proposed a verbal theory
centered on prediction-based learning. We argue that the episodic memory system, which
is believed to generate predictions (Bar, 2009; Buckner, 2010; Lisman & Redish, 2009),
may also be influenced by these predictions. In particular, when the memory system
predicts that a particular stimulus or event will appear in the near future, this may
reactivate the representation of that stimulus. This reactivation can then strengthen
memory for the stimulus, perhaps nearly as much as if the item had actually been
repeated. Furthermore, we argue that repetition of item sequences is the critical indicator
that predictions and should be made. As a result, predictions should only occur in
46
predictable environments where accurate forecasting is possible. Experiments 1 and 2
qualify as predictable environments because the presence of the RCRT condition lets
participants learn that occasionally sequences of three stimuli will be repeated. Thus,
when a context is repeated in an RCNT2 triplet, a participant may make a prediction—
perhaps nonconsciously—that the RCNT1 target will reappear. The result would be a
reactivation and strengthening of the RCNT1 target in memory, even though it is never
actually repeated. However, in Experiments 3 and 4, the RCRT condition is missing,
preventing participants from ever developing an expectation that three items in a row
would be repeated. Consequently, predictions are not utilized, and the RCNT1 targets
receive no additional learning.
Although this verbal theory sounds reasonable, it is far from a complete
explanation. It is missing necessary details about how the memory system strengthens
items and turns those memory strengths into recognition judgments. Further, we have not
yet put forth evidence that such a theory can account for the data. We next outline a
formalization of this verbal theory. We present a computational model that inherits
principles both from previous memory models and the reinforcement learning literature.
This model specifies all of the components required to reproduce the recognition
behavior found previously (Smith et al., 2013). Following this outline, we demonstrate
that the model can account for item and context repetition effects.
A Formal TD-SR Account of the Context Repetition Effect
Accounting for the context repetition effect described above requires a memory
model to produce higher likelihoods of recognizing once-presented items whose temporal
47
contexts are later repeated (RCNT1) relative to the likelihoods of recognizing oncepresented control items (NCNT1). However, the model must not produce elevated
recognition likelihoods for once-presented items that follow a repetition of context
(RCNT2). This turns out to be a difficult pattern of behavior to replicate. Most models of
memory would require enhanced or additional processing of an item (Gillund & Shiffrin,
1984; Hintzman, 1988; Murdock, 1993) in order to increase its memory strength. This
could come from many factors, such as elevated attention, additional presentations, or
longer presentation duration. However, none of these factors can explain the context
repetition effect because the RCNT1 targets are presented only once and for the same
duration as any other stimulus. Further, the fact that their contexts are not repeated until
later means there is no cue to make these targets more salient to participants. Some
associative models could explain the enhanced memory for RCNT1 targets: Contextual
repetition would strengthen the representation of context items in memory. Strengthened
context items could then increase the likelihood of recognizing any targets that were
associated with them. However, this would also boost memory for RCNT2 targets, which
is inconsistent with the context repetition effect. It seems, then, that a new mechanism is
required.
One plausible mechanism is a learning rule that makes use of prediction, known
in the reinforcement learning literature as the temporal difference-successor
representation (TD-SR) (Dayan, 1993; White, 1995). As a reinforcement learning
algorithm, TD-SR provides a mechanism for an agent to successfully navigate from one
state to another so as to maximize long-term rewards. Previous research has shown that
48
the TD-SR framework parallels the temporal context model (TCM), a computational
model of memory (Howard & Kahana, 2002). In fact, under certain conditions, as when
the environment avoids repetitions, a TD-SR model simplifies to a Hebbian learning
model that is equivalent to TCM (Gershman et al., 2012). However when regularities
occur, such that sequences of states repeat, TD-SR makes use of past experience to
generate predictions which enhance learning. We propose that a TCM-like memory
model with a TD-SR learning rule (Gershman et al., 2012), which we will refer to as
TCM-SR, can qualitatively account for both aspects of the context repetition effect. This
mechanism is consistent with the verbal theory that was put forth previously to explain
the context repetition effect (Smith et al., 2013). However, the current investigation
formally defines a principled mechanism and implements a complete model that can be
quantitatively evaluated. The remainder of this article focuses on explaining the
computational model and then testing it.
TCM-SR: TCM with a TD-SR Learning Rule
TCM-SR inherits its basic architecture, a two-layer associative network, from
TCM (Howard & Kahana, 2002). An item layer, 𝑓, represents the currently experienced
stimulus3 as a vector of features. The context layer, 𝑡, is the hallmark of TCM-like
models and represents the temporal context of the currently experienced stimulus.
Specifically, the context vector represents recent experience—information that was
recently represented in the item layer—such that the more recently something was
experienced, the stronger its representation will be in context. The context vector can be
3
This can include internally generated information; however, for the present purposes, we will restrict the
item layer to representing the features of the current experimental stimulus.
49
thought of as a graded version of a short-term memory store (Gillund & Shiffrin, 1984).
These two layers are linked via an associative matrix, 𝑀, such that each element of the
context vector is connected to every element of the item layer. The strengths or weights
of these connections are adjusted through learning. The weight of each connection
corresponds to an element value within 𝑀. 𝑀 represents long-term memory and is
queried at test to determine an item’s memory strength.
A new stimulus is presented to the model by activating that stimulus’
representation in item layer, 𝑓. The model then learns by increasing the connection
weights between active elements of 𝑓 and active elements of the context vector, 𝑡. In
most variants of TCM, this learning rule is a simple associative or Hebbian learning
process (Hebb, 1949):
𝑀 = 𝑀 + 𝛼𝑓1 𝑡0𝑇 , (1)
where 𝑀 is a matrix of connection weights, 𝑓1 is a column vector representing the new
stimulus features, 𝑡0 is a column vector representing the current temporal context, and T
represents the transpose operation, turning a column vector into a row vector. The 𝛼
parameter governs the rate of learning. The larger 𝛼 is, the greater the increase to the
connection weights (i.e. the more the model will learn). For the purposes of the present
simulations, we fixed 𝛼 to .5. Importantly this learning rule makes no use of prediction
and cannot reproduce the context repetition effect.
In contrast, the TD-SR learning rule from reinforcement learning makes use of
previously acquired information to generate predictions. It then incorporates these
predictions into the learning mechanism,
50
𝑀 = 𝑀 + 𝛼[𝑓1 − 𝑀𝑓0 + 𝛾𝑀𝑓1 ]𝑡0𝑇 , (2)
where 𝑀𝑓0 is a query of long-term memory using the features of the previous item, 𝑓0 .
The query uses what has already been learned—weights in 𝑀—to determine what has
followed the previous item, 𝑓0 , in the past and then predict that those features should have
reappeared during the current item presentation. Similarly, 𝑀𝑓1 is a query to 𝑀 using the
new item features to generate a prediction about what features are expected to appear
next. The 𝛾 parameter takes values on the interval [0, 1] and weights these predictions,
with larger values of 𝛾 producing stronger predictions. Hence 𝛾 will be critical in
explaining the context repetition effect. The term 𝑓1 − 𝑀𝑓0 is equivalent to the prediction
error term in error-driven learning (Rescorla & Wagner, 1972; Sutton & Barto, 1998).
𝑀𝑓0 serves two functions: Firstly, it keeps connection weights from growing without
bound, because if the prediction was perfect then it will cancel out 𝑓1 . Unbounded weight
growth can be problematic for purely Hebbian models (Goodhill & Barrow, 1994).
Secondly, 𝑀𝑓0 lets the model unlearn predictions that are not borne out. While this is a
useful feature for a reinforcement learning algorithm trying to maximize rewards, it is
less than ideal for a memory system trying to retain episodic information.
To adapt TD-SR to episodic memory, we introduce an important change for the
TCM-SR learning rule:
𝑀 = 𝑀 + 𝛼⟨𝑓1 − 𝑀𝑓0 + 𝛾𝑀𝑓1 ⟩+ 𝑡0𝑇 , (3)
where ⟨•⟩+ represents a rectifying operation such that any elements within • that are less
than zero are set to zero. This modification prevents the model from unlearning features
that were predicted but did not recur. Although this deviates from the TD-SR rule used in
51
reinforcement learning (Dayan, 1993), it makes intuitive sense for an episodic memory
system. A memory system that unlearns past events just because they are not repeated
when expected would be problematic.
When a new stimulus is encountered, both item and context layers are updated.
For item layer, 𝑓, the update is straightforward: Vector elements are activated to the
degree that they represent the stimulus, and no activation for the previous stimulus
remains. That is 𝑓 simply changes from 𝑓1 to 𝑓2 . The updating of context layer, 𝑡, is more
involved. First, the input to 𝑡 is computed. The input, 𝑡 𝐼𝑁 , is a weighted combination of
the previous item’s features and any context information previously bound to those
features:
𝑡 𝐼𝑁 = ∥𝛽𝑓1 + (1 − 𝛽)𝑓1 𝑀∥∥, (4)
where 𝑡 𝐼𝑁 is a vector representing the input to context, 𝑓1 is a vector representing the
previous item features, and 𝑓1 𝑀 represents a query to long-term memory to reinstate any
context information that was previously associated with 𝑓1 . The result of this query is
another vector representing reinstated context. This reinstated context vector would be a
vector of zeros if 𝑓1 has never been experienced before. The 𝛽 parameter takes values on
the interval [0, 1] and governs the tradeoff between how much of 𝑓1 vs. 𝑓1 ’s reinstated
context contributes to 𝑡 𝐼𝑁 . Thus smaller values of 𝛽 mean that input to context will be
dominated by the previous item’s reinstated context, provided that item has been bound
to a prior context. Finally, ∥•∥ represents a normalizing operation to ensure that • is unit
length.
52
Once input to 𝑡 has been calculated, context itself evolves according to the
following rule:
𝑡1 = ∥𝜌𝑡0 + 𝑡 𝐼𝑁 ∥, (5)
where 𝑡1 and 𝑡0 are the new and old context vectors, respectively, and 𝜌 is a parameter on
the interval [0, 1] that governs the rate at which context drifts. Larger values of 𝜌 lead to
a context vector that evolves more slowly. In contrast, a value of 𝜌 at or close to 0 would
produce context vectors with little to no representation of less recent experience.
Finally, we set the value of a specific element of 𝑡, which we denote as
representing the context of the learning phase (i.e. a list context), to
𝑡1𝑙𝑖𝑛𝑑 = 𝜆, (6)
where 𝑙 𝑖𝑛𝑑 denotes the element of 𝑡 representing this list context and 𝜆 is another
parameter on the interval [0, 1], that denotes the value of the list context. Thus the
activation of the list context element will be constant throughout a simulation.
Importantly, the normalizing operation in Equations 4 and 5 do not include the list
context element in the normalization process. That is, vectors 𝑡 𝐼𝑁 and 𝑡 each have a
length of one, ignoring the list context element. The list context element will prove to be
critical for recognition.
During the test phase, test probes are presented to the model just like study items.
Context continues to evolve with each test probe. However, we allow the value of 𝛽 to be
different at test than it was at encoding, so that
𝑡 𝐼𝑁 = ∥𝛽𝑅 𝑝 + (1 − 𝛽)𝑝𝑀∥∥, (7)
53
where 𝑝 is the test probe, and 𝛽𝑅 is a free parameter, on the interval [0, 1], that governs
the degree of context reinstatement during retrieval. This allows the model to have
different degrees of context reinstatement at encoding and retrieval. This is important
because successful recognition will depend on the values of the reinstated context vector,
𝑡 𝐼𝑁 .
The presentation of each test probe first leads to a new input to context, according
to equation 7. The probe’s memory strength, 𝑠, is then computed as a function of the
similarity between this new input to context and the previous context:
𝑠 = 𝑡 𝐼𝑁 ⋅ 𝑡0 𝜉 + 𝜈, (8)
where 𝑠 is the resulting memory strength, 𝑡0 denotes the previous context, and ⋅ denotes
the dot product operation. The more similar 𝑡 𝐼𝑁 and 𝑡0 are, the larger the dot product will
be. The 𝜉 parameter takes values on the interval [0, ∞) and multiplicatively scales this
dot product while 𝜈, a parameter on the interval (−∞, ∞), is an additive increment. When
the probe is a lure, the dot product will be 0. Thus 𝜈 represents the minimum amount of
strength probes will have during the test phase, regardless of evidence coming from longterm memory. Both 𝜈 and 𝜉 are critical for fitting the model’s performance to actual
human recognition behavior.
To compare TCM-SR to behavioral data, which consists of old/new recognition
judgments for every test probe, we must convert each probe’s memory strength, 𝑠, to a
likelihood of responding “old.” To do this, we use a signal detection theory (SDT) (Green
& Swets, 1966) likelihood ratio method (Glanzer, Hilford, & Maloney, 2009).
Specifically, the model makes a standard SDT assumption that memory strength values
54
can come from one of two normal distributions: a noise distribution, if a probe is a lure
that was not studied, or a signal distribution, if a probe was a studied target. We compute
the likelihood, 𝑜, that the model judged each probe to be “old” by taking the probability
density of the strength under the signal distribution over the summed probability densities
of the strength under both the signal and lure distributions, such that
𝑜 = 𝑑𝑛 (𝑠, 𝜇𝑆 , 𝜎𝑆 )/(𝑑𝑛 (𝑠, 𝜇𝑆 , 𝜎𝑆 ) + 𝑑𝑛 (𝑠, 𝜇𝑁 , 𝜎𝑁 )), (9)
where 𝑜 is the resulting likelihood of judging a probe as old, 𝑠 is the memory strength, 𝜇𝑆
and 𝜇𝑁 are parameters for the means of the signal and noise distributions, respectively,
and 𝜎𝑆 and 𝜎𝑁 are standard deviation parameters for the signal and noise distributions
,respectively. Finally, 𝑑𝑛 (𝑥, 𝜇, 𝜎) calculates the probability density of a normal
distribution, with mean, 𝜇, and standard deviation, 𝜎, at value 𝑥. We fix 𝜇𝑁 to be 0 and
set 𝜇𝑆 as a free parameter on the interval [0, ∞). We also make an equal variance
assumption, fixing both 𝜎𝑆 and 𝜎𝑁 to 1. Given these simplifications, the likelihood of
judging a test probe as old becomes:
𝑜 = 𝑑𝑛 (𝑠, 𝜇𝑆 , 1)/(𝑑𝑛 (𝑠, 𝜇𝑆 , 1) + 𝑑𝑛 (𝑠, 0,1)). (10)
Combining a TCM framework with a modified TD-SR learning rule and an SDT
decision rule, we have all the components necessary to reproduce recognition memory
results. In the next section, we explain how we fit this model to the data.
Simulation and Results
We chose to fit our model to Experiments 1 and 3 from the previously described
behavioral research (Smith et al., 2013). Both experiments used images as stimuli.
Experiment 1 created a predictable environment that produced a context repetition effect
55
across participants. Experiment 3 removed the condition that created this predictability.
As a consequence, no context repetition effect was found for Experiment 3. TCM-SR
should reproduce both patterns of behavior. We predict that the only significant change to
the fitted parameters between these two experiments will be the value of 𝛾. Specifically,
we predict higher values of 𝛾 for Experiment 1, relative to Experiment 3.
Experiments 1 and 3 had 40 and 52 participants, respectively. All participants
completing five encoding lists and then five corresponding test lists, with a single
distraction phase separating the encoding and test phases. Because stimuli did not occur
on multiple lists for a given participant, we did not have to worry about inter-list
interactions. Therefore, we trained and tested TCM-SR on one list at a time.
To model the encoding phase, we first initialized the model with an empty longterm memory matrix, 𝑀. We then presented each list item to the model, by activating
specific item features in item layer, 𝑓. For simplicity, we chose a unit-length, localist
representation for each item. Specifically, each item representation was a vector of zeros
except for a single element set to 1. Each item representation was orthogonal to every
other item representation. A localist representation scheme reflects the sparse, largely
non-overlapping representations of the hippocampus (Norman & O’Reilly, 2003) and has
been used successfully in similar models of episodic memory (Howard & Kahana, 2002;
Polyn et al., 2009; Sederberg et al., 2008). However, this representation scheme does not
allow us to realistically simulate perceptual processes. For each presentation, the model
learned associations according to Equation 3 and then updated context according to
Equations 4 and 5.
56
The presence of a distraction phase between encoding and test lists means that
elements of context that were active at the end of the encoding phase (i.e. elements
representing the last several items to occur during encoding) should no longer be active at
the start of the test phase. This is a consequence of the continual evolution of context
according to Equation 5. To simulate this drift in the model, we simply evolved context
once using a more extreme drift parameter:
𝑡1 = ∥𝜌𝑑𝑖𝑠𝑡 𝑡0 + 𝑓𝑑𝑖𝑠𝑡 ∥, (11)
where 𝜌𝑑𝑖𝑠𝑡 was set to .01 in order to clear previous context, 𝑡0 , and 𝑓𝑑𝑖𝑠𝑡 is a feature
vector, orthogonal to all items, representing the distraction task.
To model the test phase, we presented each test probe to the model. This included
lure items that were represented in the same localist fashion as studied target items. For
each probe, memory strength was assessed according to Equation 8, and then the
probability of judging the probe as a studied item was calculated according to Equation
10.
We opted to fit our model to each trial rather than to summary statistics. By fitting
to the trial-level data, we need not concern ourselves with determining a sufficient set of
summary statistics (Turner & Sederberg, 2013). Because our memory model returns a
likelihood for each test trial, fitting the model to all trials simply involves the
multiplication of likelihoods:
𝑛
𝑙 = ∏ 𝑟 (𝑜𝑖 ) + (1 − 𝑟)(1 − 𝑜𝑖 ), (12)
𝑖=1
57
where 𝑙 is the likelihood of the model responding as the participant did on all test probes,
𝑟 represents the participant’s actual response to each probe such that 𝑟 is 0 and 1 for
“new” and “old” responses, respectively, and 𝑜 is the model’s likelihood of responding
“old” to each probe.
For computational efficiency, we first log-transform the trial-level likelihoods,
and sum these log-likelihoods across all trials:
𝑙𝑙 = ∑𝑟log 𝑒 (𝑜) + (1 − 𝑟)log 𝑒 (1 − 𝑜), (13)
where 𝑙𝑙 is the resulting log-likelihood value. We fit our model to the data using these
log-likelihood values.
Table 1. Free and fixed parameters.
Parameter
𝛾
𝛽
𝛽𝑅
𝜌
𝜆
𝜉
𝜈
𝜇𝑆
𝜂
𝛼
𝜇𝑁
𝜎𝑆
𝜎𝑁
𝜌𝑑𝑖𝑠𝑡
Purpose
prediction coefficient
context reinstatement
context reinstatement at test
contextual drift
list context activation
probe strength coefficient
baseline memory strength
signal mean
negative scaling
learning rate
noise mean
signal variance
noise variance
probe strength coefficient
58
Domain
[0,1]
[0,1]
[0,1]
[0,1]
[0,1]
[0,∞)
(−∞,∞)
(−∞, ∞)
[0,1]
.5
0
1
1
.01
Equations
2,3,14,15
4
7
5
6
8
8
9,10
15
1,2,3,14,15
9
9
9
11
Hierarchical Bayesian Framework
We fit our model using a hierarchical Bayesian framework. This framework
requires us to specify a prior distribution for each parameter. The benefit of the Bayesian
approach is that it allows us to go beyond merely finding best-fitting parameter sets as
well as posterior distributions.
The hierarchical nature of our framework lets us capture individual differences
across participant, while constraining participant-level parameters via hyperpriors. The
posterior distribution of these hyperparameters can capture across-participant patterns
within the data. We favor this approach for two reasons: Firstly, the context repetition
effect exhibited considerable variation across participants, even within a given
experiment. The estimation of different parameters for each participant may allow our
model to capture these individual differences. However, each participant only
experienced 20 trials of each condition type. Thus, the amount of data per participant is
quite low, and participant-level parameter estimations may be unreliable. Our hyperpriors
serve to constrain participant-level estimations by making use of what is known about the
dataset as a whole. In other words, our hierarchical Bayesian approach offers us the
chance to look at participant-level differences while still using the entire dataset to
produce stable parameter estimates (Busemeyer & Diederich, 2010).
We chose mildly informative priors for each parameter. For parameters that exist
on the interval [0, 1], including: 𝛽, 𝛽𝑅 , 𝜌, and 𝜆, we defined the functional form of their
priors as normal distributions, which we transformed to the interval [0, 1] using an
inverse logit transform. For each of these parameters, we defined 92 participant-level
59
priors, one for each participant. The parameters for those participant-level priors
depended on mean and standard deviation hyperparameters. Mean hyperparameters were
drawn from normally distributed hyperpriors, 𝛽 𝑀 , 𝛽𝑅𝑀 , 𝜌𝑀 , and 𝜆𝑀 , each defined as a
normal distribution, 𝑁(𝜇 = 0, 𝜎 = 1). Standard deviation hyperparameters were drawn
from hyperpriors, 𝛽 𝑆 , 𝛽𝑅𝑆 , 𝜌 𝑆 , and 𝜆𝑆 , each defined as an inverse gamma distribution,
𝐼𝐺(𝛼 = 3, 𝛽 = 1).
Participant-level priors for 𝛾 were defined in exactly the same way, with one
important exception. The hyperparameters for participant-level gamma priors were drawn
from different hyperpriors, depending on experiment. Specifically, mean
hyperparameters were drawn from normally distributed hyperpriors, 𝛾 𝑀1 or 𝛾 𝑀1 , for
participants from Experiments 1 and 3, respectively. Both hyperpriors were defined as
normal distributions, 𝑁(𝜇 = 0, 𝜎 = 2). Standard deviation hyperparameters were drawn
from hyperpriors, 𝛾 𝑆1 or 𝛾 𝑆3 , for participants from Experiments 1 and 3, respectively.
These hyperpriors were defined as inverse gamma distributions, 𝐼𝐺(𝛼 = 3, 𝛽 = 1). Thus,
although the two 𝛾 hyperpriors are identically defined, they are free to change
independently as the model fits both experiments simultaneously.
Two parameters, 𝜉 and 𝜇𝑆 , were defined as truncated normal distributions on the
interval [0, ∞). Once again, we defined participant-level priors. The lower and upper
limits of each participant-level truncated normal prior were fixed at 0 and 𝑖𝑛𝑓,
respectively. The mean and standard deviations of each participant-level prior were
hyperparameters. Mean hyperparameters were drawn from normally distributed
hyperpriors, 𝜉 𝑀 and 𝜇𝑆𝑀 , each defined as a normal distribution, 𝑁(𝜇 = 0, 𝜎 = 1).
60
Standard deviation hyperparameters were drawn from hyperpriors, 𝜉 𝑆 and 𝜇𝑆𝑆 , each
defined as an inverse gamma distribution, 𝐼𝐺(𝛼 = 3, 𝛽 = 1).
Finally, the parameter 𝜈, was defined on the interval (−∞, ∞) and took the form
of a normal distribution. Again, participant-level hyperparameters were drawn from mean
and standard deviation hyperpriors, 𝜈 𝑀 and 𝜈 𝑆 , respectively. The former was defined as a
normal distribution, 𝑁(𝜇 = 0, 𝜎 = 1), while the latter was defined as an inverse gamma
distribution, 𝐼𝐺(𝛼 = 3, 𝛽 = 1).
61
Figure 4. Hierarchical structure. Each model variant was fit to all participants i from
Experiment 1 (High Predictability) and all participants j from Experiment 3 (Low
Predictability). The same hyperpriors were used across both experiments for all
parameters except 𝛾.
62
We used a custom Markov-chain Monte-Carlo (MCMC) algorithm to find best
fitting parameters and compute a joint posterior distribution. The algorithm searches the
parameter spaces of each participant-level sub-model and each hyperprior, testing 50
parameter sets at a time (i.e. 50 chains). For each participant-level sub-model, the MCMC
algorithm proposes 50 new parameter sets. The algorithm then performs a MetropolisHastings step to calculate the likelihood of each parameter set and determine which
parameter sets will be retained (i.e. added to a parameter chain). After doing this process
on each participant’s sub-model, the algorithm performs an analogous procedure on each
hyperprior. The process then begins again with the next iteration. We performed 3,200
iterations, discarding the first 1,500 as a burn-in period.
Table 2. Essential equations.
Equation
Function
Number
𝑀 = 𝑀 + 𝛼⟨𝑓1 − 𝑀𝑓0 + 𝛾𝑀𝑓1 ⟩+ 𝑡0𝑇
learning rule, Variant 1
3
𝑀 = 𝑀 + 𝛼⟨𝑓1 − 𝛾𝑀𝑓0 + 𝛾𝑀𝑓1 ⟩+ 𝑡0𝑇
learning rule, Variant 2
14
𝑀 = 𝑀 + 𝛼⟨𝑓1 − 𝛾𝑀𝑓0 + 𝛾𝑀𝑓1 ⟩𝜂 𝑡0𝑇
learning rule, Variant 3
15
𝑡 𝐼𝑁 = ∥𝛽𝑓1 + (1 − 𝛽)𝑓1 𝑀∥∥
context reinstatement
4
𝑡1 = ∥𝜌𝑡0 + 𝑡 𝐼𝑁 ∥
context evolution
5
𝑡1𝑙𝑖𝑛𝑑 = 𝜆
list context activation
6
𝑡 𝐼𝑁 = ∥𝛽𝑅 𝑝 + (1 − 𝛽)𝑝𝑀∥∥
test context reinstatement
7
𝑠 = 𝑡 𝐼𝑁 ⋅ 𝑡0 𝜉 + 𝜈
probe strength
8
𝑜 = 𝑑𝑛 (𝑠, 𝜇𝑆 , 1)/(𝑑𝑛 (𝑠, 𝜇𝑆 , 1) + 𝑑𝑛 (𝑠, 0,1))
prob. of responding “old”
10
𝑡1 = ∥𝜌𝑑𝑖𝑠𝑡 𝑡0 + 𝑓𝑑𝑖𝑠𝑡 ∥
context clearing
11
𝑙𝑙 = ∑𝑟𝑙𝑜𝑔𝑒 (𝑜) + (1 − 𝑟)𝑙𝑜𝑔𝑒 (1 − 𝑜)
log likelihood function
13
63
Results
Our model fit the behavioral data quite well. Taking the maximum a posteriori
(MAP) parameter set from each participant sub-model, we were able to produce
simulated hit rates that closely tracked mean behavioral performance across all conditions
and in both experiments (see Figure 5, top panels}. Notably, the model produced higher
hit rates for twice-presented targets than once-presented targets. False-alarm rates for
lures were also appropriately lower than hit rates for targets.
Furthermore, the model also qualitatively captured both components of the
context repetition effect for participants in high-predictability Experiment 1 (see Figure 6,
panel A). Specifically, the model produced higher hit rates for once-presented targets
whose contexts were later repeated relative to control targets (RCNT1 - NCNT1);
however, there was no difference in hit rates between once-presented targets that
followed repeated contexts and control targets (RCNT2 - NCNT2). Additionally, only a
slight advantage for RCNT1 targets over NCNT1 targets was produced in Experiment 3
(see Figure 6, panel B). We return to this point in the discussion.
Going beyond mean model performance, we next checked to see if the model
captured individual differences in the context repetition effect. For first presentations
(RCNT1 - NCNT1), the performances of sub-models appeared to track individual
variability in the context repetition effect (see Figure 7, panels A and B}). However this
correlation only reached statistical significance in the low-predictability experiment,
Pearson’s r = 0.37, n = 52, p = .007.
64
Figure 5. Model fits: Memory performance across conditions. Dark blue bars show the
mean hit rates across conditions, and false-alarm rates for lures. Light blue bars indicate
mean model behavior for each condition using MAP estimates. Panel rows correspond to
different experiments.
65
Figure 6. Model fits: Context repetition effects. Dark blue bars show the behavioral
context repetition effect. Light blue bars indicate model behavior using MAP estimates.
Within each panel, the left bars indicate the mean difference in hit rate between oncepresented targets whose contexts later repeat and control targets (RCNT1 vs. NCNT1).
The right bars indicate the mean difference in hit rate between once-presented targets that
follow repeated contexts and control targets (RCNT2 vs. NCNT2). Panel rows
correspond to different model variants. Panel columns correspond to different
experiments.
66
Figure 7. Model fits: Individual differences. Scatterplots show the relationship between
each participant’s context repetition effect and the simulated context repetition effect for
their corresponding sub-model. Best fitting regression lines and 𝑅 2 values are included.
Panel rows correspond to different model variants. Panel columns correspond to different
experiments and presentations.
67
We then simulated a variant of TCM-SR, which utilized parameter 𝛾 in a slightly
different way. Instead of making 𝛾 a coefficient of predictions of future stimuli, as
originally used (see Equation 3), we now also make 𝛾 a coefficient of predictions for the
current stimulus:
𝑀 = 𝑀 + 𝛼⟨𝑓1 − 𝛾𝑀𝑓0 + 𝛾𝑀𝑓1 ⟩+ 𝑡0𝑇 . (14)
Thus, rather than modulating the learning of predictions of future items, the 𝛾 parameter
in Variant 2 modulates the use of all prediction-based learning. Note that this alteration
does not change the number of free parameters in the model. In all other respects Variant
2 was identical to Variant 1, as was the simulation procedure.
Variant 2 performed similarly well in terms of fitting overall hit rates across
conditions (see Figure 5, middle panels). This variant did not quite capture the full
magnitude of the context repetition effect found in the high-predictability experiment (see
Figure 6, panel C). However, it still qualitatively reproduced the effect. Paralleling the
previous simulation, Variant 2 also produced a minor context repetition effect in the lowpredictability experiment (see Figure 6, panel D).
Variant 2 tracked individual differences in the context repetition effect better than
Variant 1 (see Figure 7, panels E and F}). Individual’s context repetition effects (for first
presentations) were significantly correlated with the model’s behavior both in the highpredictability, r = .48, n = 40, p = .002, and low-predictability experiments, r = .34, n =
52, p = .014.
We then performed a model comparison between Variants 1 and 2 using the
Bayesian predictive information criterion (BPIC) (Ando, 2007). BPIC assesses a model’s
68
ability to predict data using the whole posterior distribution, not just a MAP estimate.
BPIC is also ideally suited for evaluating models where the posteriors were estimated
using MCMC. For both of our model variants, we computed BPIC scores for each
participant sub-model. This allowed us to see if each variant best captured the behavior of
some of the participants.
Our results indicate that Variant 2 outperformed Variant 1 on 72 of our 92
participants (see Figure 8. This seems counterintuitive at first, given that Variant 1
seemed to reproduce the context repetition effect slightly better. However, that effect—
much like hit rates for each condition—is a set of summary statistics. Those summary
statistics partially describe the data, but are not sufficient to fully reproduce it (Turner &
Sederberg, 2013). Because of this, we fit our model variants to each participant’s triallevel data. Thus, our model comparison indicates that Variant 2 provides a somewhat
better explanation of the data overall.
69
Figure 8. Model comparison: Variants 1 and 2. The histogram shows the frequency of
within-participant differences in BPIC scores between Variants 1 and 2. Greater values
indicate a stronger preference for Variant 2, with 0 indicating no preference.
We tested one additional variant of TCM-SR, which built upon Variant 2 by
relaxing the rectification operation in the learning rule (see Equation 14). Instead of
simply setting negative features to 0, Variant 3 scaled down these negative features, so
that
𝑀 = 𝑀 + 𝛼⟨𝑓1 − 𝛾𝑀𝑓0 + 𝛾𝑀𝑓1 ⟩𝜂 𝑡0𝑇 . (15)
The 𝜂 parameter takes values on the interval [0, 1] and multiplicatively scales any
negative features before they become associated with context. Hence, the value of 𝜂
70
controls the rate of unlearning. If 𝜂 is 0, then negative features get similarly set to 0 and
the model behaves identically to Variant 2. If 𝜂 is 1, then the model will unlearn at the
same rate as traditional TD-SR models. We chose normally distributed participant-level
priors and hyperpriors that were consistent with our other parameters. Specifically,
participant-level hyperparameters were drawn from mean and standard deviation
hyperpriors, 𝜂𝑀 and 𝜂 𝑆 , respectively. The former was defined as a normal distribution,
𝑁(𝜇 = 0, 𝜎 = 1), while the latter was defined as an inverse gamma distribution, 𝐼𝐺(𝛼 =
3, 𝛽 = 1).
Qualitatively, Variant 3 also captured the mean hit rates for each target condition
and the false-alarm rate for lures (see Figure 5, lower panels). However, it only partially
reproduced the context repetition effect. In the high-predictability experiment, the
likelihood of recognizing once-presented items whose contexts were later repeated was
only slightly greater than the recognition likelihood for control items (see Figure 6, panel
E). Furthermore, a marginal context repetition effect was also found in the lowpredictability experiment (see Figure 6, panel F), consistent with previous variants.
Variant 2’s sub-models did capture some of the individual differences in the context
repetition effect for first presentations (RCNT1 - NCNT1), both in high-predictability, r =
.39, n = 40, p = .012, and low-predictability, r = .35, n = 52, p = .011, experiments (see
Figure 7, panels I and J}). A quantitative comparison between Variants 2 and 3 using
BPIC revealed Variant 2 to be the better fitting model for all 92 participants (see Figure
9). Thus, Variant 3 was qualitatively and quantitatively inferior to Variant 2, a more
parsimonious model.
71
Figure 9. Model comparison: Variants 2 and 3. The histogram shows the frequency of
within-participant differences in BPIC scores between Variants 2 and 3. Greater values
indicate a stronger preference for Variant 3, with 0 indicating no preference.
Discussion
We fit multiple variants of our TCM-SR model to the experimental data presented
in the previous chapter (Smith et al., 2013). All variants were able to reproduce hit rates
for twice presented stimuli and once-presented control stimuli, as well as false alarm rates
for lures. Importantly, two of these model variants were also able to reproduce the
context repetition effect: elevated recognition likelihoods for once-presented items whose
context is later repeated, but not for the similarly once-presented items that followed
72
repetitions of context. Our computational model is consistent with the verbal theory
proposed previously (T. A. Smith et al., 2013), which argued for prediction-based
learning. However, in the process of formally implementing this framework, we made
critical changes to the learning rule and added the necessary storage, retrieval, and
decision mechanisms in order to adequately simulate memory performance.
The importance of prediction in accounting for the context repetition effect is
plainly evident based on the current simulations. Without predictions of the future, the
model would not be able to produce elevated recognition probabilities for once-presented
targets whose contexts later repeat (RCNT1). The prediction coefficient parameter, 𝛾, is
critical to enabling and regulating this behavior. Without the ability to tune this value, the
model would not be able to capture the null context repetition effect of Experiment 3.
Other Models and the Context Repetition Effect
We know of no existing episodic memory model that, in its current form, can
predict the context repetition effect. Formal models of memory rely on a variety of
encoding mechanisms, including—but not limited to—rehearsal buffers (Gillund &
Shiffrin, 1984; Raaijmakers & Shiffrin, 1981), convolution (Murdock, 1993), exemplar
encoding (Hintzman, 1988) and Hebbian learning (Howard & Kahana, 2002). None of
these learning mechanisms involve learning predictions of the future. Without this
component, current models of memory are incapable of explaining the current findings.
That is not to say that these established models could not be adapted, as we have
adapted TCM, to accommodate our new data. Most episodic memory models could be
modified to learn based on prediction. All that is required is that a given model query the
73
contents of its long-term memory storage with a cue—the currently perceived stimulus—
and then learn the output based on whatever learning rule the model uses. As an example,
consider the search for associative memory model (SAM) (Gillund & Shiffrin, 1984),
which learns associations via rehearsal of items in a short-term buffer. If SAM were
modified such that it 1) generated free-recall outputs during encoding and 2) put those
outputs into the buffer for rehearsal, then the model would begin associating recent items
with items predicted by current cues. To put it in Marr’s (Marr, 1982) terminology, TDSR is not the only algorithm that can solve the problem posed by the context repetition
effect; however, it does characterize the computational steps required for a solution.
Another potential way of accounting for the context repetition effect may come
from the study of semantic memory. The predictive temporal context model (pTCM)
(Shankar et al., 2009) updates semantic representations based, in part, on predictions
generated from current representations. Nevertheless, pTCM is a model of semantic
memory and uses predictions to learn semantic relationships between words, which is
quite different from the goal of recognizing specific episodes. Given that semantic
learning generally requires many encounters with a concept, pTCM as currently
implemented is probably ill-equipped to rapidly learn episodes of experience.
Consequently, pTCM would have to undergo modification before it could function as an
episodic store at all, let alone account for the context repetition effect.
Differences from TD-SR
The learning mechanism in TCM-SR does have subtle, yet important, differences
from the typical TD-SR learning rule. Firstly, our second variant uses the prediction
74
parameter 𝛾 as a coefficient for both the prediction of future and current items. TD-SR,
along with our first variant, places 𝛾 only next to the prediction of future states. This
difference changes the interpretation of 𝛾 and has important implications for the model.
For our second variant, 𝛾 is regulating the learning of any prediction. If 𝛾 is set to zero,
then the model simplifies to Hebbian learning (Hebb, 1949). In our first variant, as well
as the standard TD-SR rule, 𝛾 simply weights how much predictions of the future are
learned. When 𝛾 is set to zero, the learning rule reduces to error-driven learning (Rescorla
& Wagner, 1972). Although Variant 2 outperformed Variant 1 overall, both variants
provided reasonably good fits to the data with 𝛾 set above zero.
Additionally, TCM-SR rectifies the vector of features that gets bound to context
(see Equation 3). Any features with negative values are set to zero; thus no unlearning
occurs in the model. This stands in contrast to conventional TD-SR which unlearns
suboptimal transitions as it gains experience with the environment (Dayan, 1993).
Rectification is a logical adjustment given that the presumed goal of episodic memory is
to retain information. Nevertheless, it could be the case that some degree of unlearning is
tolerable, or even desirable, in TCM-SR. However, our third variant replaced the
rectifying function with a negative-scaling parameter. This parameter scaled down the
magnitude of negative feature values instead of setting them to zero. Interestingly, this
added flexibility did not improve performance and in fact led to a worse fitting model. In
the next section we highlight several other modifications that could be incorporated into
TCM-SR.
75
Limitations and Future Work
Although the present incarnation of TCM-SR qualitatively reproduces the context
repetition effect (Smith et al., 2013), the model does not perfectly capture the data. In
particular, all variants produced a slight context repetition effect in the low-predictability
experiment, which is not present in the behavioral data. Furthermore, sub-models do not
perfectly correlate with individual differences in the effect. One especially noticeable
shortcoming is the fact that TCM-SR does not produce appreciable negative context
repetition effects, despite the fact that some participants exhibit this behavior (see Figure
7). Variants 1 and 2 lack a mechanism (i.e. unlearning) to generate such an effect. In
principle, Variant 3 could unlearn via 𝜂, however this behavior is not produced by the
best fitting sub-models. Below we discuss several potential additions or modifications to
TCM-SR that could improve its explanatory power.
The use of a constant list context unit appears to be an odd choice given that that
basis of our model is a continually evolving context vector, based on TCM (Howard &
Kahana, 2002). However, recent work in neurophysiology and cognitive science
(Shankar & Howard, 2012) suggests that context does not decay at a uniform rate (i.e. in
accordance with drift parameter 𝜌) as suggested by TCM. Rather, different elements of
context are proposed to decay nonlinearly. Briefly, newly activated context elements are
thought to decay rapidly at first; however, their rate of decay diminishes as time passes.
This results in less recent information being represented in context at low but relatively
stable levels of activation. In this framework, effectively recognizing probes would hinge
upon the strength of associations between item features and these relatively stable context
76
elements. We made use of a single list-context element as a computationally simple first
approximation of this more elaborate framework. Future work could replace the list
context unit in TCM-SR with a more advanced implementation of context.
One could frame the context repetition effect in terms of enhanced internal and
external attention (Chun et al., 2011). Encoding into memory is a function of available
attention (Chun & Turk-Browne, 2007). Hence, to the extent that participants have been
paying greater attention to the stimuli during a particular RCNT1 sequence, they should
have formed stronger associations between the context and target stimuli. This would
enable their memory system to make stronger predictions during the RCNT2 sequence.
Future research could treat attention as an individual difference, inferring level of
attention from behavioral (e.g. response latencies during an encoding task, where shorter
latencies correspond to greater attention) or physiological (e.g. pupillometry or eye
tracking) measures. Participants’ attention at encoding should correlate positively with
their memory performance across conditions at test. If the measure of attention were
more fine-grained, such that list- or even item-level indices of attention could be reliably
measured, then those indices should similarly correlate with memory for said lists or
items. The learning rate, 𝛼, could conceivably be scaled by the attention index. This
would allow the model to more accurately predict which stimuli get strongly encoded and
control for variance in memory that results from fluctuations in attention.
Internal attention—attention to internally generated information (Chun et al.,
2011), such as predictions—may help account for the context repetition effect. Under this
interpretation, the memory system is always making predictions if possible. When the
77
environment is less predictable (i.e. Experiment 3) then these predictions receive less
attention. In that sense the prediction coefficient 𝛾 could be thought of as representing the
degree of internal attention directed at predictions. If an appropriate index of internal
attention (e.g. a neural or other physiological measure) could be attained, then prediction
coefficient 𝛾 could be modulated by this index. This could potentially explain why oncepresented items that follow a repetition of context in Experiment 1 but not 3 are better
remembered without requiring separate 𝛾 hyperpriors for each experiment. Of course,
this approach would not explain how attention changes across trials or between
experiments, but would merely account for those changes.
One way of mechanistically accounting for trial-level changes to 𝛾 would be to
add a mechanism to the model that tracks the accuracy of the model’s predictions over
time. A recency-weighted running average of prediction accuracy could be used to scale
𝛾 (or potentially even replace 𝛾). In high-predictability environments, the model would
find that the stimuli it predicts have a relatively high probability of actually occurring.
Consequently, 𝛾 would be scaled up and predictions would be strongly encoding into
memory. However when entering an unpredictable environment, predictions will no
longer be borne out. Hence, the running average of prediction accuracy will diminish and
the value of 𝛾 along with it. As a result, predictions will not be strongly encoded into
memory. This type of model would not only readily fit both high- and low-predictability
environments without separate parameters, but would also make predictions about the
memory performance for stimuli that occur during transitions from predictable to
78
unpredictable environments or vice versa. Future empirical work will need to modify our
behavioral paradigm (Smith et al., 2013) in order to test these predictions.
In some cases, modeling the episodic memory system may not be enough to
capture human behavior. Semantic memory is known to influence episodic memory in
many ways, including the order in which recalls are made during free recall (Bousfield,
1953; Bousfield & Sedgewick, 1944), and even whether or not participants will recall
items that were not studied (Deese, 1959; Roediger & McDermott, 1995). Previous
computational models have incorporated semantic memory strengths to better account for
episodic memory performance for word lists (Kimball, Smith, & Kahana, 2007; Polyn et
al., 2009). Similar mechanisms could be incorporated into a future version of our model.
As such mechanisms are most useful when modeling memory for words, they do not
have much relevance for the present simulations.
Finally, future iterations of this model should strive for greater biological
plausibility. One aspect of the current model that is lacking in biological realism is our
representation of the stimuli. The item layer, 𝑓, uses orthogonal localist representations
for each item. That is, each stimulus is represented by a unit vector with a single non-zero
element. This is a reasonable simplification of representation in the hippocampus, which
likely encodes episodes using mostly non-overlapping activation patterns. However,
other brain regions, such as those that actually represent visual stimuli, use overlapping
distributed representations (Norman & O’Reilly, 2003). To completely understand how
the transient perceptions of current experience lead to enduring and accessible memory
traces, we will need to understand how and where the perceptual system represents
79
stimuli and how those representations interact with the memory system. In the next
chapter, we take a step towards this goal.
Conclusions
Our previous research demonstrated that repeating a sequence of events can
improve memory for information that followed the original presentation of that sequence
(Smith et al., 2013). Importantly, this context repetition effect may only occur in
environments featuring sequential regularities that enable accurate forecasting. Given that
real world experience is full of such regularities, the mechanism that gives rise to this
effect may be employed with considerable frequency. In fact, such a mechanism may be a
fundamental component of the memory-prediction system. We have already argued that
prediction-based learning can account for the context repetition effect. Our formal
account of memory, featuring prediction-based learning, qualitatively captures the effect.
We believe this model provides a principled starting point for further explorations into
the interaction between memory and prediction.
80
Chapter 4: Face-specific Information in the Fusiform Face Area Depends on Memory4
As we mentioned in the previous chapter, our computational model of memory
makes no effort to account for how the perceptual system represents stimuli. We view
this omission as a serious limitation given that these perceptual representations ultimately
provide input to the medial temporal lobe memory system (MTL) (Davachi, 2006;
Eichenbaum, 2000, 2004). In fact, some researchers have argued that these two systems
are not distinct. Instead, portions of the MTL can be considered part of the ventral visual
pathway (Baxter, 2009; Bussey & Saksida, 2007), which is critical for the perception of
objects of all kinds (Martin, 2007). Different portions of the ventral pathway are sensitive
to particular categories of stimuli, such as scenes, words, limbs, and faces (Martin, 2007).
However, what these category-selective regions actually represent is a matter of debate.
In this chapter we focus on representations in face-selective regions and their relationship
with the memory system.
How the human brain represents face information is a core question in cognitive
neuroscience, with implications ranging from machine vision (Tistarelli, Bicego, &
Grosso, 2009) to social biases (Van Bavel, Packer, & Cunningham, 2008). Converging
evidence from human neuroimaging (Collins & Olson, 2014; Haxby, Hoffman, &
4
This chapter presents research that is discussed in a manuscript currently being revised for a resubmission
to NeuroImage. The authors are Hasinski and Sederberg.
81
Gobbini, 2000; Kanwisher & Yovel, 2006; Kanwisher et al., 1997) and cellular
recordings of nonhuman primates (Freiwald, Tsao, & Livingstone, 2009; Tsao, Freiwald,
Tootell, & Livingstone, 2006) supports the existence of specialized regions in the primate
brain that preferentially process faces. It has been proposed that these regions form a
face-processing network, with different regions assumed to play unique but potentially
overlapping roles in the processing of face stimuli (Collins & Olson, 2014; Haxby et al.,
2000; Nestor, Plaut, & Behrmann, 2011). However, what exactly each region
represents—and how they come to represent it—is not yet understood.
Several key face-processing regions lie in the ventral visual processing pathway.
The most studied of these, the fusiform face area (FFA) is a portion of the fusiform gyrus
that responds preferentially to face stimuli relative to many other categories, such as
scenes or objects (Kanwisher & Yovel, 2006; Kanwisher et al., 1997). FFA responds to
constituent face features, but it is also sensitive to complete face configurations,
suggesting it may contain holistic representations of faces (Harris & Aguirre, 2010; Liu,
Harris, & Kanwisher, 2009). In contrast, the more posterior occipital face area (OFA)
does not appear sensitive to intact face configurations, suggesting it represents only the
components of faces (Haxby et al., 2000; Liu et al., 2009; Pitcher, Walsh, & Duchaine,
2011; Pitcher, Walsh, Yovel, & Duchaine, 2007). Additionally, the anterior inferior
temporal region (aIT)5 is strongly implicated in both the perception and memory of faces
(Collins & Olson, 2014). Indeed, lesions of aIT produce person-memory deficits
5
We refrain from describing aIT as a “face area” per se because we reserve that phrase for regions that are
typically functionally defined as more responsive to faces than other stimulus categories. Due known issues
in imaging the anterior temporal lobe with fMRI (Devlin et al., 2000; Visser, Jefferies, & Lambon Ralph,
2009), functional localization of a face-sensitive ROI in anterior inferior temporal lobe is less common.
82
(Gainotti & Marra, 2011; Olson, McCoy, Klobusicky, & Ross, 2013; Ross & Olson,
2010) as well as deficits in face discrimination (Busigny et al., 2014; Olson, Ezzyat,
Plotzker, & Chatterjee, 2014). In contrast, the evidence for FFA’s involvement in facememory is mixed, and OFA is believed to be uninvolved (Collins & Olson, 2014).
The precise function of these regions is still an open question. For instance, the
FFA has been viewed as performing face detection (Kanwisher, Tong, & Nakayama,
1998), face identification (Gauthier et al., 2000), or both (Grill-Spector, Knouf, &
Kanwisher, 2004). Taking advantage of advancements in multivariate analysis of fMRI,
recent research has shown that face-sensitive regions contain information that
differentiates individuals (Anzellotti & Caramazza, 2014). An early study presented one
male and one female to participants many times during scanning. Multivariate analyses
revealed that information in the aIT, but not the FFA, was distinct for each face
(Kriegeskorte, Formisano, Sorger, & Goebel, 2007). Although this result could reflect the
presence of gender information, more recent work also found information in faceselective voxels in the FFA and anterior temporal lobe that individuated different faces of
the same gender (Nestor et al., 2011). More recently, the OFA, FFA, and aIT were shown
to contain information identifying individuals even when some aspect of individuals’
faces changed across presentations. For instance, a pattern classifier trained on faces from
several viewpoints was used to effectively discriminate between those faces when tested
on novel viewpoints (Anzellotti, Fairhall, & Caramazza, 2013). Other researchers
extracted individuating information from faces that persisted across changes in facial
expression (Nestor et al., 2011). Indeed, converging neuroimaging and lesion data
83
suggest that the integration of information across these face areas may be essential for
face identification in humans (Anzellotti & Caramazza, 2014; Collins & Olson, 2014;
Haxby et al., 2001; Natu et al., 2009).
One commonality across these studies is that they all present the same faces
multiple times during scanner sessions, usually with additional presentations during a
pre-scanning training phase. Consequently, participants inevitably become familiar with
the stimuli used. Often, this is training or familiarization process is an intentional, and
perhaps critical, element of the experiment. Because familiarization is a constant feature
of these studies, it leads to a question about when and how this differentiating
information arises in face regions. One possibility is that these regions contain
individuating information regardless of familiarity. Face regions may give rise to
information for individual faces (i.e. face-specific information) through the bottom-up
processing of face stimuli and the representation of each face’s unique constellation of
features. Under this hypothesis, patterns in face-sensitive regions should individuate faces
without any prior exposure. Consequently, face-specific information should be present in
both OFA and FFA, which are believed to represent face components (Harris & Aguirre,
2010; Liu et al., 2009). OFA, which appears to more specifically represent lower-level
features (Haxby et al., 2000; Liu et al., 2009; Pitcher et al., 2011; Pitcher et al., 2007),
may show greater face-specific information than FFA. Finally, face-specific information
may or may not be present in aIT, which is typically associated with more abstracted or
view-invariant representations of identities (Anzellotti et al., 2013; Collins & Olson,
2014; Freiwald & Tsao, 2010).
84
An alternate hypothesis is that familiarization of a face is what leads to this
individuating information. According to this latter hypothesis, face regions should only
individuate faces that are familiar to participants. Additionally, the face regions more
strongly associated with face or person memory, FFA (Grill-Spector et al., 2004) and
especially aIT (Collins & Olson, 2014; Gainotti & Marra, 2011; Olson et al., 2013),
should contain this face-specific information, whereas OFA may not contain information
for individual faces.
The current study directly compares these hypotheses. To do this, we forego
extensive training and repeated presentations of faces during scanning. Instead, we use a
single repetition of target faces during scanning, and then test memory for those faces
after scanning. We operationalize familiarity as subsequent memory for these target
faces. That is, remembered faces will retrospectively indicate whether faces were familiar
to a participant at the initial repetition and forgotten faces will have remained unfamiliar
during study. According to the familiarity hypothesis, only faces that are subsequently
remembered should show significant pattern similarity between presentations.
Furthermore, we would expect this similarity to be present in the more downstream
regions: FFA and especially aIT. However, under the alternate hypothesis, faces should
show significant pattern similarity between presentations regardless of subsequent
memory performance, and this similarity should be stronger in more upstream regions:
FFA and especially OFA.
Potential support for the familiarity hypothesis comes from recent research that
found greater pattern similarity between repetitions of words that were subsequently
85
remembered relative to the similarity observed between repetitions of forgotten words
(Xue et al., 2010)6. The researchers further argued that their finding provided evidence
against the encoding variability hypothesis, which states that repeatedly experienced
items are better remembered if the repetitions occur in varied contexts (i.e., with different
preceding items), as opposed to all repetitions occurring in the same context (Bray,
Robbins, & Witcher, 1976; Martin, 1968; Melton, 1970). Although they did not explicitly
manipulate the context in which spaced items were presented, Xue and colleagues
equated greater pattern similarity across repetitions with decreased encoding variability.
As such, when they observed that items with lower pattern similarity exhibited worse
subsequent memory performance, they concluded this contradicted the encoding
variability hypothesis.
Given that the context across spaced repetitions was never the same in the Xue et
al. study, the relationship between neural pattern similarity, memory, and context has yet
to be fully explored. To that end, we added a secondary question to the current
experiment: Does degree of encoding variability influence subsequent memory via
altering the degree of similarity between presentations? We manipulated encoding
variability by controlling the temporal contexts (Howard & Kahana, 2002; Sederberg et
al., 2008; Turk-Browne et al., 2012) in which faces appeared across repetitions. This
manipulation allowed us to test directly whether encoding variability in the traditional
sense is related to subsequent memory. If we found a significant relationship, we would
6
In fact, Xue et al. (2010) did show greater pattern similarity for remembered faces in a separate
experiment. However, the experiment using face stimuli did not feature trial-level analyses, so they could
not assess whether or not representations for individual faces were more similar across repetitions when
subsequently remembered.
86
then test whether the effect is mediated by the degree of pattern similarity in facesensitive regions. However, if we found context to be unrelated to memory behaviorally,
then we would then drop the mediation analysis and focus instead on our primary
question: What is the relationship between subsequent memory and face-specific
information, irrespective of context?
Method
Subjects
Twenty-seven Ohio State community members (11 female, all right-handed, mean
age 22.3 years) participated in the main experiment and a functional localizer task. All
subjects had normal or corrected-to-normal visual acuity, provided informed consent, and
received monetary compensation. The study protocol was approved by the Institutional
Review Board for Human Subjects at the Ohio State University.
Stimuli
Face stimuli consisted of color photographs of nonfamous male and female faces,
facing forward, from the shoulders up, with neutral expression and various hair styles
(See Figure 1). All faces appeared before a white background. A separate set of faces–
along with sets of scenes, objects, and scrambled objects–were also used in the functional
localizer task. Using separate pools of faces for each task prevented any additional
exposures to our target stimuli before the surprise memory test (see below). A Christie
digital projector displayed images at 60 Hz with a resolution of 1280 x 1024 onto a
screen behind the scanner bore. Participants viewed images with a mirror attached to the
head coil.
87
Procedure
Anatomical Scan. After obtaining informed consent and providing instructions,
we began scanning with a high-resolution T1-weighted anatomical scan, lasting 3
minutes and 25 s, was collected from each subject.
Functional Localizer. Participants then completed a functional localizer task
based on previous work (Epstein & Kanwisher, 1998). This task, which consisted of three
4 minute and 24 s runs, enabled us to functionally define ROIs, independently of our
main experiment. Consistent with similar tasks used by others (Epstein & Kanwisher,
1998), each localizer run alternated between blocks of scenes, faces, objects and
scrambled objects. Each block contained 16 stimuli, all from the same category. Each
stimulus was presented for 500 ms, followed by an ISI of 500 ms. Blocks were arranged
in a pseudorandom fashion. A block of each category would appear once, in random
order, followed by 12 s of fixation. This occurred 3 times per run, with the category
blocks arranged in a different order each time. To keep participants engaged during the
localizer, they were instructed to detect immediate repetitions of stimuli. In each block,
two stimuli were randomly chosen to be repeated immediately. Participants indicated
when they detected an immediate repetition via button press. All responses in the scanner
were made using a Current Designs fiber-optic response pad.
Main Task. The main scanner task consisted of four runs, each lasting 8 minutes
and 48 s. Each run consisted of 96 faces, presented one at a time for 1.2 s, with a
randomly jittered inter-stimulus interval (ISI) between 2.3 s and 5.3 s (see Figure 10A).
88
To keep participants engaged in the task, they were instructed to indicate whether each
face was male or female. Participants indicated their response via the response pad.
The main task featured a structure identical to the paradigm discussed in previous
chapters (Smith et al., 2013), although it used different stimuli and timing sequences.
Because the target stimuli in this experiment are different from our behavioral research,
we will explain the design differently. Each target face appeared twice, with between 5
and 20 other faces occurring in between the first and second presentations (see Figure
10B). Consistent with previous research (Smith et al., 2013; Turk-Browne et al., 2012)
and theoretical work on temporal context (Howard & Kahana, 2002; Sederberg et al.,
2008), we denoted the two faces that preceded each target as the “context” in which each
target presentation occurred. Faces making up the high encoding variability (high-EV)
condition were preceded by different context faces on their first and second presentations.
That is, each face occurred in a unique temporal context on both presentations. In
contrast, faces making up the low encoding variability (low-EV) condition were preceded
by the same context faces on their first and second presentations. However, different lowEV face pairs appeared in different contexts.
Across all runs, each participant viewed 16 targets in the high-EV condition and
16 targets in the low-EV condition, with each target being presented twice. Importantly,
these stimuli were completely novel to the participants. They had no training or previous
exposure to them prior to the study and only minimal familiarization–two presentations–
during the scanner task.
89
Participants were not informed of the organization of the stimulus lists, nor were
they informed that some faces were “target” stimuli. They were simply told that they
would see a sequence of faces, that some of them may be repeated, and that they should
indicate the gender of each face as quickly and accurately as possible. Additional filler
faces were interspersed throughout each run, consistent with similar paradigms used
previously (Smith et al., 2013). These faces were either presented once or twice to
obscure the organizational structure of the target stimuli, but they had no bearing on the
hypotheses in question here.
90
Figure 10. Overview of experimental design. (A) Faces were presented one at a time, for
1.2 seconds, with a randomly varied interstimulus interval. Each target face occurred
twice. (B) Each presentation of a given target face occurred in either the same context
(low encoding variability) or in different contexts (high encoding variability). Same-face
similarity was computed between a first-presentation target (red square) and its identical
second presentation (dark-blue square). Different-face similarities were computed
between the first-presentations and non-identical second presentations with matching
genders and encoding variability (light-blue square).
Recognition Memory Task. Following the last structural scan, participants exited
the scanner and completed a recognition memory task on a laptop in an adjacent testing
room. They were shown all of the target and comparison faces for all four conditions (96)
and 20% of the context faces (38), along with an equal number (134) of lures and asked
91
whether or not each face appeared in the study task. Again, faces were presented one by
one for 1.2 s and subjects had the same amount of time to make a decision. Subjects were
given four response options: “sure old,” “old,” “new,” or “sure new”, with each
corresponding to a different keyboard key. Participants indicated their decisions with a
single keypress.
MRI Data Acquisition
MRI Data were acquired with a 3T Siemens Trio scanner and a 12-channel matrix
head coil at the Center for Cognitive and Behavioral Brain Imaging (CCBBI) at the Ohio
State University. For the main encoding task, functional images were obtained with a T2weighted EPI sequence: TR = 2200 ms, TE = 26 ms, field of view = 250 x 220 mm [read
x phase], flip angle = 75°, thickness = 2.7 mm (2.5 x 2.5 x 2.7 mm voxels). 41 oblique
axial slices were collected in interleaved order. 237 volumes were acquired.
For the functional localizer scans, T2-weighted images were obtained with the
following parameters: TR = 3000 ms, TE = 28 ms, field of view = 250 x 220 mm [read x
phase]; flip angle = 80°; thickness = 2.5 mm (2.5 x 2.5 x 2.5 mm voxels). 47 oblique
axial slices were collected in interleaved order. 85 volumes were acquired.
For each subject, we used the high-resolution T1-weighted anatomical scan to
align scans align all functional scans. 160 sagittal slices were collected, with a thickness
of 1 mm (1 x 1 x 1 mm voxels).
Data Processing
Preprocessing. Processing of both functional and anatomical MRI images was
carried out with a combination of AFNI (Cox, 1996) and FSL (Jenkinson, Beckmann,
92
Behrens, Woolrich, & Smith, 2012; Smith et al., 2004). FSL’s brain extraction tool
(Jenkinson, Pechaud, & Smith, 2005) was applied to the T1-weighted structural images to
isolate voxels containing brain tissue.
For functional scans, we began by discarding the first two volumes of each
functional run. Functional runs were preprocessed and analyzed using AFNI (Cox, 1996).
Processing then followed standard AFNI protocols (Cox, 2012) and included despiking
the time series, aligning the runs together and correcting for head motion. The functional
localizer data were then smoothed using a 4 mm FWHM Gaussian kernel. Because of our
primary analyses involved patterns of activation, we did not smooth the main task data.
Finally, all functional data were scaled to have a mean of 100 as part of AFNI’s standard
data analysis (Cox, 2012).
Modeling the BOLD Response. For both functional tasks, we modeled the neural
response to each stimulus by fitting canonical hemodynamic response functions (HRFs),
convolved with a square function equal to the stimulus duration, in a general linear model
(3dDeconvolve in AFNI) (Ward, 2002) for each participant. Each HRF was estimated by
fitting its amplitude (i.e. one regressor per HRF) to the data. The result of normalizing
our data to have a mean of 100 is that the regression coefficients are equivalent to
estimates of percent signal change (Ward, 2002).
For the functional localizer, separate canonical HRF regressors were entered for
every stimulus category (face, scene, object and scrambled object), convolved to a .5 s
square function, representing stimulus duration. The resulting statistical maps were
similar if we used a single 16 s regressor for each category block.
93
For the main task, all hemodynamic responses were convolved to a 1.2 s square
function, representing stimulus duration. We modeled the hemodynamic response to each
target face presentation with its own regressor, using the least-squares all method
(Rissman, Gazzaley, & D’Esposito, 2004). This produced 64 single-trial regressors (2
presentations for each of 16 targets in both the high-EV and low-EV conditions).
Extracting estimated BOLD response for these individual trials allowed us to find
patterns of activation associated with each individual presentation.
Conventional condition regressors were used to model all non-target stimuli. We
estimated a single hemodynamic response for each type of context and filler face. This
resulted in 20 additional regressors (16 for context faces and 4 for filler faces). This was
done to improve the stability of the estimates of the 64 trials of interest. We also included
6 motion parameters, which were estimated during preprocessing. Finally, we included 4
polynomial regressors to remove slow changes in the BOLD signal (i.e. due to drift in the
magnetic field). This procedure resulted in 94 total regressors for the main encoding task.
Note that even though our task is a rapid-presentation, event-related design, the
spacing between our target stimuli is considerable (mean: 76.6 s, SD: 21.7 s). Thus, we
effectively have a slow-presentation, event-related design where our regressors of interest
are concerned, with BOLD responses due to non-target stimuli accounted for with stable
condition regressors.
Participants’ resulting coefficient maps were transformed to their own highresolution T1-weighted anatomicals. Registration between functional and anatomical
maps was performed using FSL’s linear volumetric registration tool (Jenkinson & Smith,
94
2001; Jenkinson, Bannister, Brady, & Smith, 2002). Coefficient maps in anatomical
space were used for all ROI analyses.
Regions of Interest. We used the functional localizer to define regions of interest
(ROIs) for our main analysis. Consistent with previous research (Anzellotti et al., 2013),
we used a faces - scenes contrast to obtain a statistical map of face selectivity. Left and
right fusiform face areas (FFAs) were identified separately for each subject. Contrasts
were clusterized, using 20 voxels as the minimum cluster size. This ensured that all ROIs
had at least 20 voxels in functional space, providing ROIs of sufficient size to extract
reliable signal from our task data.
The process for selecting ROIs involved three steps: First, for each hemisphere,
we focused on a broad anatomical region that coincides with established FFA location
(e.g. the fusiform gyrus). Then we set an initial threshold equivalent to p = .0001. If no
clusters survived this initial threshold we did not define an ROI for that subject and
region. Finally, if a cluster was found, we decreased the p-value until the cluster shrank to
20 voxels, or as close to 20 as possible without breaking up. All ROIs were between 20
and 40 2.5x2.5x2.7mm voxels in size.
ROI were then transformed into each subject’s high-resolution anatomical space.
We then created a mask out of the 200 most significant 1x1x1mm voxels in anatomical
space. This allowed us to reliably identify anatomical landmarks and ensured that ROIs
would align with our task data, which were also transformed to anatomical space. Finally,
we combined–when possible–left and right hemispheres into a single mask for each ROI.
95
This was done because we did not have specific hypotheses regarding hemisphere and we
wanted to keep the number of statistical comparisons to a minimum.
This process enabled us to create OFA and FFA cluster in 26 of 27 participants.
We were also able to create a parahippocampal place area (PPA) (Epstein & Kanwisher,
1998) ROI in 26 of 27 participants by following the same process but reversing the
statistical contrast to be scenes > faces. We use PPA as a control region in the following
analyses to ensure we don’t get face-specific activation in regions not known to be
sensitive to faces.
We were unable to functionally localize an anterior temporal face area in more
than a small fraction of our participants. This was likely due to signal loss in the ventral
anterior temporal lobe, a known issue (Devlin et al., 2000; Visser et al., 2009). As an
alternative, we created an anatomically defined 400-voxel (1 mm^3 voxels) anterior
inferior temporal lobe ROI (aIT) masks for each participant, using the Harvard-Oxford
atlas distributed by FSL (Mark Jenkinson et al., 2012; Smith et al., 2004) as a reference.
We treat this analogue as a substitute for a functionally defined aIT.
The use of participant-specific ROIs obviated the need for transforming any data
to a standard, across-participant space. Therefore, we perform this second transformation
on our ROI masks only for the purpose of aggregating ROIs across participants for
graphical purposes (see Figure 11). This anatomical-to-standard space transformation was
accomplished with FSL nonlinear volumetric registration tool (Andersson, Jenkinson, &
Smith, 2007). ROIs are aligned to the Montreal Neurological Institute (MNI-152) atlas
provided by FSL (Jenkinson et al., 2012; Smith et al., 2004).
96
Figure 11. Functional and anatomical ROIs. Superposition of every subject’s FFA (red),
OFA (green) and PPA (blue), along with aIT (violet) in MNI standard space.
97
Calculating Face-specific Information
To investigate the presence of face-specific information, we used each
participant’s masks to extract a pattern of neural activity at each ROI for each target
presentation. We then used a trial-level variant of correlation analysis (Haxby et al.,
2001; Xue et al., 2010). Specifically, for each subject we correlated activity for each firstpresentation target with activity for every second-presentation target. This produced two
types of correlations: “same-face” correlations between each first presentation and the
identical second presentation, and “different-face” correlations between each first
presentation and all non-identical second presentations. We then applied a Fisher ztransformation to all correlations, converting them to similarity values on the interval
(−∞, ∞).
The same-face similarities reflect the dependency or overlapping information
between the neural representations of first and second presentations of identical faces.
The different-face similarities are important because collectively they quantify the
amount of overlapping information between neural representations of first-presentation
faces to other non-identical faces on average. In other words, different-face similarities
let us estimate the amount of overlapping information that is due to the stimulus category
of faces. We use these different-face similarities to estimate the amount of information in
the same-face similarity values that can be attributed to representing the same exact face
rather than the same stimulus category. Specifically, for each first presentation, we
98
created a vector of face-specific similarity values: the same-face similarity minus each
different-face similarity:
𝐟𝐢 = [𝑠𝑖 − 𝑑𝑖1 , 𝑠𝑖 − 𝑑𝑖2 , . . . , 𝑠𝑖 − 𝑑𝑖𝑛 ], (16)
where 𝑠𝑖 is the scalar same-face similarity between 𝑖 𝑡ℎ first presentation and its identical
second presentation, din is the scalar different-face similarity between the 𝑖 𝑡ℎ first
presentation 𝑛𝑡ℎ different-face second presentation, and 𝐟𝐢 is the resulting vector of facespecific similarity values.
The different-face similarity values are meant to control for any pattern similarity
that is not due to specific faces. Therefore, we avoid using different-face similarities that
may be low due to systematic differences between first and second presentations. Past
research has shown that full multivariate techniques can be used to distinguish
representations of male and female faces (Kriegeskorte et al., 2007). In searching for
face-specific information in face-sensitive regions, we did not want our results to be due
to gender differences between faces. Similarly, we only use different-face similarity
values when the same-face and different-face targets have the same gender. Additionally,
we only use different-face similarity values when the different-face target occur in the
same encoding variability condition as the same-face target. Finally, we only use
different-face similarity values when the subsequent memory performances of the sameface targets and different-face targets match.
The eligible different-face similarities were used to create face-specific
similarities according to Equation 16. We use these resulting values as the dependent
variable in our neural analyses.
99
Mixed-effects Regression
To test the competing hypotheses that face-specific information is either
dependent on subsequent memory or not, we fit our data with a mixed-effects regression
model. We performed a separate analysis for each of our four ROIs. Our predictor
variable of interest is an indicator variable for whether or not each target face was
recognized during the memory test. Additionally, we wanted to rule out the possibility
that any effect, or lack of effect, could actually be due to the spacing between faces or
overall activation within ROIs. We included two nuisance regressors to control for these
issues.
First, we needed to control for the variation in lag between the first and second
presentations. Autocorrelation is a known feature of fMRI time-series data. In general,
when two presentations are closer to one another in time, the neural patterns they evoke
maybe more similar regardless of the actual evoked neural activity. Thus, the lags
between the first and second presentations are a potential source of variance we must
account for. More specifically, because our face-specific similarity values are the
difference between two values–same-face similarity and different-face similarity–we
needed to account for the relationship between same-face lags and different-face lags.
To accomplish this we created a lag-log-ratio term, which we will explain further:
We began by calculating the same-face lags, the spacing between first presentations and
their identical second presentations, and the different-face lags, the spacing between first
presentations and non-identical second presentations. We then log-transformed both types
of lags. Finally we subtracted the different-face log-lags from the same -face log-lags:
100
𝐥𝐫𝐢
𝑙𝑖𝑠
𝑙𝑖𝑠
𝑙𝑖𝑠
= [log 𝑒 ( 𝑑 ), log 𝑒 ( 𝑑 ), . . . , log 𝑒 ( 𝑑 )] , (17)
𝑙𝑖1
𝑙𝑖2
𝑙𝑖𝑛
where 𝑙𝑖𝑠 is the lag between the first and second presentations of the 𝑖 𝑡ℎ face, ldin is the
scalar lag between the 𝑖 𝑡ℎ first presentation and and the 𝑛𝑡ℎ different-face second
presentation, and 𝐥𝐫𝐢 is the resulting vector of lag-log-ratios. Note that these values are the
log of the ratio between same-face lags over different-face lags.
The resulting lag-log-ratios have three useful properties. Firstly, they combine
both same-face and different-face lags into a single value, keeping the complexity of our
inferential models to a minimum. Secondly the ratio preserves the relative difference in
spacing between same-face pairs and different-face pairs. Finally, the log transformation
provides an intuitive and interpretable zero point. That is, when the same-face lag equals
the different-face lag, the resulting lag-log-ratio value is zero. Thus, this lag-ratio variable
quantifies uneven lag between same- and different-face similarity values. As an
additional safeguard against autocorrelation, we discarded pairings where the lag between
the first and second presentation was less than 30 seconds.
We also wanted to ensure that any face-specific pattern similarity was not due to,
or obscured by, changes in overall level of activation from first to second presentations.
This repetition modulation (RM) has already been shown to covary with familiarity
(Eger, Schweinberger, Dolan, & Henson, 2005), so we created a nuisance regressor to
partial out any covariation between subsequent memory and face-specific similarity that
is shared with RM. We followed a similar approach to our calculation of lag-log-ratios
above. First we calculated the mean difference in activation between first and second
101
presentations (RM) for same-face and different-face pairings. We then log-transformed
these RM values. Finally, we took the difference between all same-face log-RMs and all
the corresponding different-face RMs:
𝐫
𝐦𝐢
𝑠
= [log 𝑒 (
𝑚𝑖
𝑑
𝑚𝑖1
𝑠
), log 𝑒 (
𝑚𝑖
𝑑
𝑚𝑖2
𝑠
), . . . , log 𝑒 (
𝑚𝑖
𝑑
𝑚𝑖𝑛
)] , (18)
𝑠
d
where 𝑚𝑖 is the mean RM between the first and second presentations of the 𝑖 𝑡ℎ face, min
is the scalar mean RM between the 𝑖 𝑡ℎ first presentation and the 𝑛𝑡ℎ different-face second
𝐫
presentation, and 𝐦𝐢 is the resulting vector of RM-log-ratios. These values quantify the
relative difference in RM between same- and different-face pairings, such that the values
would be zero when the relative difference was zero.
We included both lag-log-ratio and RM-log-ratio as nuisance regressors in our
statistical model. However, we had no hypotheses concerning these variables. Therefore,
we avoid making inferences about them, and do not report statistics or significance in this
paper.
Finally, our model included three random effects terms: Firstly, individual brains
may differ in the degree to which they exhibit face-specific similarity. Therefore, we
treated participant as a random effect. This allowed us to account for any variation in
face-specific similarity that may be due to the particular participants in our sample.
Additionally, some faces may tend to evoke more distinct neural responses than others.
To control for any differences in face-specific similarity due to stimulus distinctiveness,
we treated each same-face target and each different-face target as random effects.
102
Determining degrees of freedom for mixed effects regression models is not
straightforward. Instead, parametric statistical inference requires an estimation of the
degrees of freedom (Kenward & Roger, 1997). To circumvent this estimation, we
performed non-parametric permutation tests to ascertain the p-values for each of our
comparisons of interest (Ernst, 2004). For each ROI analysis, we performed 10,000
permutations of the dependent data by randomly flipping the sign of the similarity
difference. We then calculated two-tailed p-values based on the null distribution of tvalues for each regression term of interest. Finally, to correct for multiple comparisons
we used a Benjamini-Hochberg correction (Genovese, Lazar, & Nichols, 2002) to keep
the false discovery rate at 𝛼 = 0.05 across all tests. For significant results, we provide the
raw p-values, followed by corrected p-values in parentheses.
Results
Behavioral Results
Participants responded correctly on 93.1% of all trials of the gender categorization
task during encoding. We defined incorrect responses as those where the participant
provided the wrong gender or failed to make a response within the 1.2 s the face was on
the screen. Only face pairs where the participant responded correctly during both
encoding presentations were included in subsequent neural similarity analyses. This
ensured that differences in similarity were not due to a lack of attention to the faces.
The mean reaction time for completed trials was 647.1 ms (SD=158.1 ms). We
looked for evidence that the first presentation of a face would facilitate the gender
103
judgement response for the second presentation of the same face. We did not find
evidence of repetition priming (mean=4.03 ms, SD=41.48 ms), t(26)=0.49, p=0.62.
Turning to the recognition test, we calculated hit rates and false-alarm rates for
each subject. The mean hit rate for all target items was 51.6% (SD=18.8%), with a
corresponding mean false-alarm rate of 35.9% (SD=14.3%). As expected, given the use
of unfamiliar faces, performance on the memory task was poor, yet, on average,
participants performed above chance. For a comparison, the mean hit rate for non-target,
once-presented faces was 41.6% (SD=13.8%). As would be expected, a paired samples ttest revealed a significant boost in memory performance for twice-presented items,
t(26)=2.36, p=0.026.
Figure 12 shows the memory performance for target trials split by condition. We
find a mean hit rate of 50.9% (SD=21.6%) for faces presented twice in the same context
(low-EV), and a mean hit rate of 52.3% (SD=19.8%) for those presented in different
contexts (high-EV). A paired-samples t-test did not detect a significant difference
between conditions, t(26)=-0.32, p=0.75. Thus, at least in this experiment, variability in
encoding context does not appear to be related to memory performance. Because
encoding variability is unrelated to memory performance, we forego investigating
whether this (null) relationship is mediated by neural similarity. We collapse across
encoding variability conditions for the remainder of our analyses, focusing instead on a
potential relationship between similarity and subsequent memory.
104
Figure 12. Subsequent memory performance as a function of encoding condition. Bars
show mean proportion of faces recognized during the memory test for faces that occurred
under high encoding variability (dark green) and low encoding variability (light green).
For a reference, the mean proportion of once-presented items that were subsequently
remembered is shown in grey. Error bars reflect standard errors. The solid red line
indicates the mean proportion of lures that were incorrectly judged as having been seen
previously, with dashed red lines indicating standard error.
Neuroimaging Results
For each of our four ROIs, we tested whether face-specific similarity was present
for subsequently remembered and subsequently forgotten faces. We also tested for a
difference in face-specific similarity between forgotten and remembered faces, for a total
of 12 tests across ROIs. We conducted these tests using a linear mixed-effects regression
105
procedure, which let us control for nuisance regressors (lag between presentations and
repetition modulation), as well as participant-specific and stimulus-specific effects. We
controlled for multiple comparisons using a Benjamini-Hochberg correction (Genovese et
al., 2002) to keep the false discovery rate at 𝛼 = 0.05. Corrected p-values are shown in
parentheses for significant results.
In FFA, face-specific similarity was greater for subsequently remembered faces,
t=3.29, p=0.001 (.007). Furthermore, significant face-specific similarity was found for
remembered faces, t=2.57, p=0.012 (0.046), but not for forgotten faces, t=0.11, p>0.91
(see Figure 13A).
In aIT, face-specific similarity was also greater for remembered than forgotten
items; however, this finding did not survive correction for multiple comparisons t=1.99, p
= 0.045 (0.135). Once again, we found significant face-specific similarity in for
remembered faces, t=2.39, p=0.007 (0.043), but not for forgotten faces, t=0.85, p>.05
(see Figure 13B).
In OFA, an upstream face-sensitive region that is generally believed to process
face components, we found no effect of memory on face-specific similarity, t=-1.16,
p>0.23. Furthermore, we found no evidence of face-specific similarity for either
remembered, t=-0.33, p>0.74, or forgotten faces, t=0.58, p>0.57.
As expected, our control region—the scene-selective PPA—similarly showed no
effect of memory on face-specific similarity, t=0.89, p>.36. PPA also showed no facespecific similarity for remembered, t=0.33, p>0.74, or forgotten faces, t=-0.37, p>0.69.
106
In support of the familiarization hypothesis, we found that memory performance
was a significant predictor of face-specific similarity in down-stream face-sensitive
regions. We found no evidence of face-specific information in OFA, which is consistent
with the idea that subsequent memory for faces—not their individual features—is an
important moderator.
107
Figure 13. Average similarity between face repetitions. Bars show similarity between
first presentation targets and either (dark blue) same or (light blue) different second
presentations, split by memory performance. (A) FFA shows significant face-specific
information–greater similarity between identical faces than between different faces of the
same gender–only for faces that are subsequently remembered. The difference in facespecific similarity between remembered and forgotten faces is also significant. (B) aIT
shows the same pattern, although the difference in face-specific information is no longer
significant after correcting for multiple comparisons. † 𝑝 < 0.05 uncorrected, ∗ 𝑝 < 0.05
corrected, ∗∗ 𝑝 < 0.01 corrected.
108
Discussion
The present results demonstrate that downstream face-processing regions, such as
the FFA and aIT, contain unique face-specific information after minimal familiarization.
We find no evidence of such face-specific information for faces that participants are
unable to subsequently recognize. This was unlikely to be due to participants not
attending to those stimuli because we only included faces in our analyses that the
participants successfully categorized as male or female within 1.2 seconds. We also find
no evidence of face-specific information in early face-selective regions (OFA) or in a
control region that is not specifically sensitive to faces (PPA). We used subsequent
memory as an index of the strength of encoding into memory across both presentations of
a face; however, the mechanism that drives this encoding–similarity relationship has not
yet been explained. We next outline some potential mechanisms.
One possibility is the reactivation account proposed previously (Xue et al., 2010).
According to this account, memory is most improved when the same trace of a stimulus
is reactivated on subsequent presentations. Thus, greater face-specific similarity actually
reflects this stable reactivation, which gives rise to stronger memory encoding during
each presentation of a given face. Importantly, the current findings suggest this
reactivation refers to the representation of the face itself, rather than the face and its
context as it has been previously discussed (Xue et al., 2010).
A second possibility is a feedback account where patterns in face areas reflect
online perceptual processing that are associated with enduring memory representations in
other regions, such as the hippocampus (Buzsáki & Moser, 2013; Eichenbaum, 2000,
109
2004) or regions that have been implicated in person memory, such as the anterior
paracingulate cortex or amygdala (Gobbini & Haxby, 2007). These memory
representations may feed back to and alter the perceptual representations in upstream
areas such as the FFA during the second presentation of a face. Thus, the stronger a
memory representation is after a first presentation, the more that memory representation
may influence a subsequent perceptual representation to be like the first.
An alternate explanation for the current results is that attention, not memory
encoding per se, is the mechanism that produces face-specific similarity. Specifically,
faces observed under greater attention may be encoded with greater fidelity. This may
also produce more distinctive representations of those faces in face-sensitive regions.
Greater attention during encoding also facilitates subsequent memory (Chun & TurkBrowne, 2007; Turk-Browne, Golomb, & Chun, 2013). While we are agnostic with
regard to potential attention-based mechanisms at this time, future research could use
eye-tracking and pupillometry to provide a quantitative indices of attentional focus and
cortical arousal to help shed light on this hypothesis.
A more extreme characterization of this mechanism is that the lack of facespecific similarity for forgotten items is due to a complete lack of attention during
forgotten trials. Specifically, face areas will not represent stimuli if a participant’s eyes
were closed or looking away. This would also hamper memory performance for faces,
producing a correlation. However, our neural analyses only used face pairs that were
correctly categorized as by gender during both presentations. Performing this task
110
accurately requires at least some degree of attention. Thus, the current results are likely
not driven by an absence of attention during subsequently forgotten trials.
Furthermore, the current results may suggest that attention alone may not be
sufficient to produce face-specific information. If attention were the sole mechanism at
work, we would expect face-specific similarity in OFA precisely because we used the
same image for both presentations of our target faces. The OFA principally represents
lower-level features of faces (Haxby et al., 2000; Liu et al., 2009; Pitcher et al., 2011;
Pitcher et al., 2007), so greater attention should also yield greater representational fidelity
of those features in FFA. That is, the low-level features are the same, so attention to those
features could have produced “features-specific” information in OFA. We found no
evidence of face-specific similarity for familiar faces in OFA. This is consistent with
earlier research that found no effect of familiarity in OFA when using univariate methods
(Davies-Thompson, Gouws, & Andrews, 2009). Our null result in OFA suggests
something more than attention is necessary to produce face-specific information in FFA
and aIT.
Another potential explanation is that both face-specific similarity and memory are
driven by the features of our stimuli. That is, some faces may be more visually distinctive
than others. Greater distinctiveness could produce both greater face-specific similarity
and better memory performance. By including face labels as a random factor in our
statistical analyses, we were able to control for variation in similarity that was attributable
to the varying distinctiveness of different faces. However, this procedure does not control
for within-participant variation in face distinctiveness. That is, a particular face may not
111
be seen as distinctive by the entire sample, but could be particularly distinctive for a
particular participant. This distinctiveness could still affect that participant’s memory and
neural response pattern.
The present results are limited by the difficulty in imaging the anterior temporal
lobes (Devlin et al., 2000; Visser et al., 2009). If we had been able to reliably extract
functional-localized anterior temporal face areas from our participants, we may have
found stronger or different effects in that region. Nonetheless, the fact that our atlasbased aIT results followed the same pattern as the FFA results is evidence that familiarity
may serve the same function in both regions. fMRI protocols optimized to image the
anterior temporal lobes may provide better functional localization for future work.
Although the current work relied on repetitions of identical images of the target
faces, future research may benefit from including alternate images of the same face
(e.g. from different viewpoints) for first and second presentations. Previous work has
demonstrated that the FFA and aIT contain patterns that discriminate highly familiar
identities, even if alternate images are used for the same face (Anzellotti & Caramazza,
2014). This image-invariant representation of identity has been shown with respect to
variations in viewing angle (Anzellotti et al., 2013; Natu et al., 2009) and emotional
expression (Nestor et al., 2011). However, each of these studies relied on identities that
were highly familiar to participants due to training or previous experience. Future work
should investigate the extent to which image variation (e.g. head rotation, morphing
along a dimension, or different expressions) affect representations for untrained
identities.
112
Conclusions
These results begin to elucidate how the face-processing network comes to
represent individual faces. Face-specific information is detectable in FFA and aIT, but
not OFA, on a trial level. Extensive training of the stimuli before or during scanning is
not required to elicit this face-specific information. However, we only detect facespecific information for faces that were subsequently recognized. Thus, these results
implicate the development of familiarity as being directly related to the emergence of
face-specific information in ventral-stream face regions.
These results also have important implications for our memory model. In TCMSR, the item layer’s representation is noiseless and stable throughout the presentation of a
given stimulus. Our results here suggest that this implementation is inadequate. If our
feedback hypothesis is correct, then the item layer representation must evolve
dynamically based on signals coming from both the stimulus and long-term memory. If
the reactivation account is correct, then noise must be added to the item layer
representations. Implementing and comparing these model variants, which could be
constrained using both the neural and behavioral data (Turner et al., 2013), could give
support to one account over the other. Either of these models may, in turn, offer new
predictions regarding memory function, prediction generation, or perceptual processing.
113
Chapter 5: General Discussion
The present work investigated the relationship between episodic memory and both
prediction and perception. To accomplish the former, we developed a novel experimental
paradigm that tests for effects of prediction on memory. These effects are not detectable
using conventional memory paradigms; however, our new design found consistent
evidence that prediction influences memory. We also presented preliminary results from
a formal model of episodic memory that employs prediction-based learning. Our results
demonstrate that our model can account for the behavioral effects we discovered. Finally,
we used fMRI to investigate how neural representations of visual stimuli relate to
subsequent memory. In particular, we found that face-sensitive brain regions represent
information about specific faces only when those faces will be subsequently remembered.
We first review these findings in more detail and discuss their implications. We conclude
with a discussion of future directions.
Memory and Prediction
Adopting a common definition of context as a continuously evolving construct
(Howard & Kahana, 2002; Polyn et al., 2009; Sederberg et al., 2008), in which recent
experience makes up the context of the current experience, we were able to create an
experimental environment in which the repetition of an item’s context sometimes
preceded a repetition of the item itself. This created an experimental setting with
114
predictable regularities. In two experiments, this setting consistently produced enhanced
memory for once-presented items whose contexts were later repeated. The enhancement
was almost as large as the difference between once- and twice-presented items. However,
memory was not enhanced for items that followed a repetition of context. This
asymmetrical context repetition effect is difficult to explain using simple associative
learning mechanisms (Hebb, 1949). However, a learning rule that includes the learning of
predictions (Dayan, 1993; Gershman et al., 2012; White, 1995) has the potential to
account for both aspects of this effect.
Complicating matters is the fact this context repetition effect does not occur in
environments where the repetition of an item’s temporal context never precedes the
repetition of an item (i.e. environments lacking predictability). We interpret this null
context repetition effect, found in two additional behavioral experiments, as evidence that
the episodic memory system is sensitive to the overall predictability of the current
environment. In environments where predictions have little to no chance of being
accurate, the memory system may disregard those predictions instead of learning them.
This sensitivity to the presence of environmental regularities has obvious value.
However, it complicates explanations of how and when the brain produces predictionbased learning.
As a first step towards an adequate explanation, we developed a computational
model of recognition memory that uses a prediction-based learning mechanism adapted
from reinforcement learning (Dayan, 1993; White, 1995). This algorithm produces
behavior equivalent to Hebbian learning in standard memory paradigms. However, in
115
environments where prediction is possible it deviates from Hebbian models (Gershman et
al., 2012). We fit our model to two of our behavioral experiments: one that featured a
high-predictability environment and a significant context repetition effect, and another
that featured a low-predictability environment and a null context repetition effect. The
magnitude of the model’s context repetition effect is determined by a free parameter, 𝛾,
which scales the degree to which predictions are learned by the memory system.
Using a hierarchical Bayesian framework and a custom sampler, we estimated
posterior parameter distributions for each participant and a set of hyperpriors. To account
for behavioral findings in both high- and low-predictability experiments, we included two
hyperpriors for the prediction coefficient, 𝛾. Hyperparameters for a given participant’s 𝛾
parameter were drawn from one of these two hyperpriors, depending whether the
participant experienced the high- or low-predictability environment. Two of our three
model variants qualitatively reproduced the context repetition effect in a highpredictability environment and mostly reproduced the null context repetition effect in a
low-predictability environment. They also accurately recreated hit rates for once and
twice-presented targets, as well as false-alarm rates for lures. Finally, our model was also
able to capture individual differences in the context repetition effect.
Collectively, the empirical and computational modeling results presented here
provide strong evidence that the episodic memory system learns based on prediction.
This is not implausible given that the medial temporal lobe system is both the locus of
episodic memory (Davachi, 2006; Eichenbaum, 2000, 2004) and known to be critical for
generating predictions (Buckner, 2010; Johnson et al., 2007; Levy, 1989; Levy et al.,
116
2005; Lisman & Redish, 2009; Schacter et al., 2007). However, no other episodic
memory model currently makes use of prediction. Given that our experimental paradigm
reflects the regularity and moderate predictability of many real-world scenarios, we argue
that prediction-based learning may be an important—and overlooked—feature of human
memory. Future empirical work is needed to better define the boundary conditions of the
context repetition effect, as well as the pervasiveness of prediction-based learning more
generally. Additionally, more advanced computational models will be needed to explain
how the memory system calibrates 𝛾 to the current environment. Finally, future versions
of our memory model should include perceptual processes, in an effort to elucidate the
interactions between memory and perception.
Memory and Perception
Turning to the relationship between episodic memory and perception, we used
fMRI to investigate the degree to which face-selective brain regions within the ventral
processing stream encode information related to specific faces, rather than faces as a
category. Using multivariate techniques (Haxby et al., 2001; Xue et al., 2010), we found
evidence of face-specific information in two regions—the fusiform face area (Kanwisher
& Yovel, 2006; Kanwisher et al., 1997) and the anterior inferior temporal lobe (Collins &
Olson, 2014). Both regions are believed to process faces holistically (Collins & Olson,
2014; Harris & Aguirre, 2010; Liu et al., 2009). However, this face-specific information
was only found for faces that were recognized in a subsequent memory test. Additionally,
no face-specific information was found in regions that are not sensitive to faces or only
process the low-level features of faces.
117
To our knowledge, this experiment is the first demonstration that face-selective
regions in the ventral pathway can represent information about individual faces that have
never been seen before. This research also contributes to the growing literature that
explores the relationship between perception and memory. Our findings suggest that
researchers who study the neural pathways of perception—especially, though not
exclusively, regarding faces—should consider incorporating a subsequent memory task
into their research. In doing so, they may be able to control for variance in neural or
behavioral data that is due to memory.
Bridging Domains
The present research focused on three popular topics in the cognitive sciences:
memory, perception, and prediction. Our results demonstrate that the cognitive and neural
processes discussed in each of these domains do not operate in isolation. Rather, the
processes interact and influence one another. This statement no doubt holds true for
domains not discussed here, including: attention, decision-making, language, and affect.
Thus, completely understanding one process requires an understanding of the processes
that interact with it. The present research shows the importance of integrating different
domains. We are certainly not the first to take this approach, and we hope the coming
years will see an even greater emphasis placed on such efforts.
Limitations and Future Directions
Recognition vs. Recall
Every experiment in the present research used a typical item recognition test to
assess memory performance. An alternative testing method would have asked
118
participants to freely recall list items in any order. Free recall enables researchers to
investigate which items a participant would retrieve from memory without any
prompting, as well as the order in which these items would come to mind. Free recall
paradigms have been critical to our understanding of memory search and how the
memory system moves from one episodic retrieval to another (Kahana, 2012). However,
recognition tests afford researchers more control because they allow us to assess memory
for every item of interest, and in the order we choose. Recognition tests are also the only
feasible method of testing memory for unfamiliar faces and scenes. Finally, because
recalling items is much more difficult than recognizing them, free recall tests are
typically performed after each study list. Thus participants quickly learn that their
memory will be tested and adopt rehearsal strategies. Consequently, it is not feasible to
investigate incidental learning using free recall. Nevertheless, using words as stimuli, the
present research could be altered to test participants’ recall.
Testing for the presence of a context repetition effect during free recall would
require significant modifications to the paradigm. The number of items in each study list
would need to be reduced. This would then affect the number of conditions that would
occur in each list. Additionally, participants’ recalls would likely focus on the twicepresented items, perhaps to the point of ignoring the more difficult—but quite critical—
once-presented items. A potential solution may be to direct participants to recall the
relevant words by presenting them in one color and the less relevant words in another
color. The recall phase would then begin by asking participants to recall the relevantly
colored words. Significant pilot testing would be required to refine the paradigm.
119
However, the end result would enable us to explore how likely participants are to
transition from repeated contexts to the first and second novel targets (RCNT1 and
RCNT2, respectively). The results could either lend converging support to our current
prediction-based memory model or force us to modify it.
Recall could also tell us more about how memory and perception interact. For
instance, we might find that the amount of stimulus-specific information present in the
perceptual pathway correlates with the recall position of recalled items. Specifically,
study items with greater stimulus-specific representations may have higher probabilities
of being recalled first. Once again, testing recall would require significant modifications
to our experimental design. List length would need to be shortened, and words would
need to be used as stimuli. Fortunately previous research has already identified faceselective regions in the ventral processing stream (Dehaene et al., 2002; McCandliss et
al., 2003). Focusing on those regions, we may find an analogous relationship between
word-specific representations and subsequent memory.
The Neural Correlates of the Context Repetition Effect
The neuroimaging findings presented here focused on the perceptual
representation of faces. In principle, because that experiment contained the same
conditions used in our behavioral analysis of the context repetition effect, those data can
also be used to investigate the neural correlates of the behavioral context repetition effect.
However, this type of analysis presents several challenges. The first issue is
timing: When exactly prediction-based learning occurs is not entirely clear. When a
context is repeated, a prediction about the next target will follow. However, our current
120
model does not specify exact timing. In fact, the prediction-generation and predictionlearning processes may occur right up to, and potentially during, the onset of the actual
target. Furthermore, given the sluggish nature of the BOLD response, any signal
representing the prediction would overlap with BOLD signals coming from both the
repeated context stimuli and the subsequent novel target. The uncertainty about when
prediction-based learning occurs coupled with the overlapping BOLD responses presents
a serious challenge.
Recent advancements in multivariate imaging analysis (Nielson & Sederberg,
submitted) may enable us to detect these prediction signals. The basic goal is to find
specific patterns of neural activation, similar to our prior analyses on the perception of
faces. We may find patterns indicative of predictions in face-selective regions, suggesting
that predictions generated in the MTL affect activation in perceptual layers. This would
be consistent with the feedback account of our current neural findings. However,
information may also be detected in the MTL itself. Additionally, we might find a BOLD
response corresponding to prediction error when the unexpected novel target (RCNT2)
actually follows the repeated context. Such a signal would be indicative of predictiongeneration and would be expected in regions where similar prediction errors have been
reported previously, such as the lateral prefrontal cortex and intraparietal sulcus
(Gläscher, Daw, Dayan, & O’Doherty, 2010)7.
7
Although previous research has also implicated the dopaminergic system (e.g. ventral tegmental area,
substantia nigra, and ventral striatum) as an additional source of prediction error (D’Ardenne, McClure,
Nystrom, & Cohen, 2008; S. M. McClure, Berns, & Montague, 2003), signals found in these regions
correspond to reward prediction errors. The memory paradigm used in the present research does not feature
explicit rewards. Consequently, we have listed the lateral prefrontal cortex and intraparietal sulcus, as
BOLD responses in these regions are known to correspond to state prediction errors (Gläscher et al., 2010).
121
A Model of Memory, Perception, and Prediction
As mentioned previously, our current computational model is lacking any realistic
perceptual mechanism. Consequently, it cannot help us discriminate between multiple
explanations that could account for memory-dependent, face-specific information in the
visual pathway. A reactivation account posits that greater reactivation of the first neural
pattern during the second presentation leads to better memory strength for that face. A
feedback account posits that the second presentation cues the memory system, which
feeds back to the perceptual layer, constraining and biasing the new perceptual
representation. However, these verbal explanations are imprecise and can be difficult to
distinguish from one another. In contrast, formal mathematical explanations can avoid
ambiguities and offer precise predictions, which can then be tested.
Such a model would have key differences from the memory model proposed
earlier in this document. To begin with, our memory model featured a localist
representation of stimulus features. That is, the vector representing each stimulus
included only one nonzero element, which was always set to one to make each vector unit
length. However, distributed representations of stimuli will be critical in the modeling of
our perceptual findings. We may retain a localist representation for a layer symbolizing
the hippocampus, which encodes episodes using non-overlapping patterns (Norman &
O’Reilly, 2003). However, input to that hippocampal layer would need to come from a
perceptual layer, utilizing distributed, overlapping representations. Furthermore, this
perceptual layer would receive feedback from the hippocampal layer, enabling us to test
122
the feedback account. The context layer and associative matrix, as well as the learning
and decision rules, would likely remain in their current form.
The resulting model of perceptual representation, memory function, and
prediction generation would formalize and clarify our competing hypotheses regarding
our neural data, hopefully helping us to compare them. A model with distributed
perceptual representations may also lead to new questions regarding the relationship
between memory and prediction. Finally, this model may allow us to develop novel
hypotheses about the relationship between prediction and perception, a topic we have not
investigated so far. Thus, in an era where memory researchers tend to develop models to
account for specific tasks and effects, our model would be able to do something unique:
explain the interactions between memory, perception, and prediction.
The Role of Attention
Attention is key process in all of the research presented here. In the context
repetition behavioral paradigm, a participant must attend to the overall experimental
environment to learn that this environment is predictable. Similarly, attention to specific
stimuli is essential to successful encoding and the generation of predictions. In the
imaging experiment, fluctuations in attention could be at least partially responsible for
variations in both face-specific information and subsequent memory. Additionally, lapses
in attention almost certainly account for some of the noise in both our behavioral and
fMRI data. However, there is a difference between acknowledging the importance of
attention and actually incorporating attentional processes into one’s explanation.
123
Properly incorporating attention requires two additions to our program of
research. First, we must operationalize and quantify participants’ attention over time.
Eye-tracking could be used to track the focus of participants’ eye gaze, providing
information about their overt attention. Furthermore, pupillometry could be used to
provide an index of participants’ cognitive alertness over the course of an experiment.
Secondly, our computational accounts of the cognitive and neural phenomena would need
to be modified to include attentional mechanisms. Previous research argues that
perception and memory are functions of different types of attention, external and internal,
respectively (Chun et al., 2011). This suggests that incorporating attentional processes
into our model may make the interactions between the perceptual, memory, and
prediction components more cohesive and complete. The resulting model may also make
predictions about attention itself.
Conclusions
The present research explored the relationship between episodic memory and
prediction and perception. Using a novel experimental design and a computational
model, we demonstrated that in relatively predictable environments—not unlike realworld scenarios—humans use predictions to enhance their memory. Using advanced
neuroimaging techniques, we also demonstrated a link between the neural representation
of faces and memory for those faces. This research contributes to our understanding of
cognition and advances the methodologies used in computational modeling and fMRI
data analysis. Perhaps most importantly, this work demonstrates the utility of spanning
levels of analysis and bridging the domains of the cognitive sciences.
124
Bibliography
Andersson, J., Jenkinson, M., & Smith, S. (2007). Non-linear registration, aka spatial
normalisation FMRIB technical report TR07JA2 (Technical Report No. TR07JA2).
FMRIB Analysis Group of the University of Oxford.
Ando, T. (2007). Bayesian predictive information criterion for the evaluation of hierarchical
bayesian and empirical bayes models. Biometrika, 94(2), 443–458.
Anzellotti, S., & Caramazza, A. (2014). The neural mechanisms for the recognition of face
identity in humans. Frontiers in Psychology, 5.
Anzellotti, S., Fairhall, S. L., & Caramazza, A. (2013). Decoding representations of face identity
that are tolerant to rotation. Cerebral Cortex, bht046.
Bar, M. (2009). Predictions: A universal principle in the operation of the human brain.
Philosophical Transactions of the Royal Society of London B: Biological Sciences,
364(1521), 1181–1182.
Baxter, M. G. (2009). Involvement of medial temporal lobe structures in memory and perception.
Neuron, 61(5), 667–677.
Bousfield, W. A. (1953). The occurrence of clustering in the recall of randomly arranged
associates. The Journal of General Psychology, 49(2), 229–240.
Bousfield, W. A., & Sedgewick, C. H. W. (1944). An analysis of sequences of restricted
associative responses. The Journal of General Psychology, 30(2), 149–165.
Bray, J. F., Robbins, D., & Witcher, W. B. (1976). Encoding variability theory and the spacing
effect in associate learning. Memory & Cognition, 4(5), 548–552.
Buckner, R. L. (2010). The role of the hippocampus in prediction and imagination. Annual
Review of Psychology, 61(1), 27–48.
125
Busemeyer, J. R., & Diederich, A. (2010). Cognitive modeling. SAGE.
Busigny, T., Van Belle, G., Jemel, B., Hosein, A., Joubert, S., & Rossion, B. (2014). Facespecific impairment in holistic perception following focal lesion of the right anterior
temporal lobe. Neuropsychologia, 56, 312–333.
Bussey, T., & Saksida, L. (2007). Memory, perception, and the ventral visual-perirhinalhippocampal stream: Thinking outside of the boxes. Hippocampus, 17(9), 898–908.
Buzsáki, G., & Moser, E. I. (2013). Memory, navigation and theta rhythm in the hippocampalentorhinal system. Nature Neuroscience, 16(2), 130–138.
Cabeza, R., Mazuz, Y. S., Stokes, J., Kragel, J. E., Woldorff, M. G., Ciaramelli, E., …
Moscovitch, M. (2011). Overlapping parietal activity in memory and perception:
Evidence for the attention to memory model. Journal of Cognitive Neuroscience, 23(11),
3209–3217.
Chun, M. M., & Turk-Browne, N. B. (2007). Interactions between attention and memory.
Current Opinion in Neurobiology, 17(2), 177–184.
Chun, M. M., Golomb, J. D., & Turk-Browne, N. B. (2011). A taxonomy of external and internal
attention. Annual Review of Psychology, 62(1), 73–101.
Collins, J. A., & Olson, I. R. (2014). Beyond the FFA: The role of the ventral anterior temporal
lobes in face processing. Neuropsychologia, 61, 65–79.
Corbit, L. H., & Balleine, B. W. (2000). The role of the hippocampus in instrumental
conditioning. The Journal of Neuroscience, 20(11), 4233–4239.
Cox, R. W. (1996). AFNI: Software for analysis and visualization of functional magnetic
resonance neuroimages. Computers and Biomedical Research, 29(3), 162–173.
Cox, R. W. (2012). AFNI: What a long strange trip it’s been. Neuroimage, 62(2), 743–747.
D’Ardenne, K., McClure, S. M., Nystrom, L. E., & Cohen, J. D. (2008). BOLD responses
reflecting dopaminergic signals in the human ventral tegmental area. Science, 319(5867),
1264–1267.
Davachi, L. (2006). Item, context and relational episodic encoding in humans. Current Opinion
in Neurobiology, 16(6), 693–700.
126
Davies-Thompson, J., Gouws, A., & Andrews, T. J. (2009). An image-dependent representation
of familiar and unfamiliar faces in the human ventral stream. Neuropsychologia, 47(6),
1627–1635.
Dayan, P. (1993). Improving generalization for temporal difference learning: The successor
representation. Neural Computation, 5(4), 613–624.
Deese, J. (1959). On the prediction of occurrence of particular verbal intrusions in immediate
recall. Journal of Experimental Psychology, 58(1), 17–22.
Dehaene, S., Le Clec’H, G., Poline, J.-B., Le Bihan, D., & Cohen, L. (2002). The visual word
form area: A prelexical representation of visual words in the fusiform gyrus.
Neuroreport, 13(3), 321–325.
Dennis, S., & Humphreys, M. S. (2001). A context noise model of episodic word recognition.
Psychological Review, 108(2), 452–478.
Devlin, J. T., Russell, R. P., Davis, M. H., Price, C. J., Wilson, J., Moss, H. E., … Tyler, L. K.
(2000). Susceptibility-induced loss of signal: Comparing PET and fMRI on a semantic
task. NeuroImage, 11(6), 589–600.
Ebbinghaus, H. (1913). Memory: A contribution to experimental psychology. University
Microfilms.
Eger, E., Schweinberger, S. R., Dolan, R. J., & Henson, R. N. (2005). Familiarity enhances
invariance of face representations in human ventral visual cortex: fMRI evidence.
NeuroImage, 26(4), 1128–1139.
Eichenbaum, H. (2000). A corticalhippocampal system for declarative memory. Nature Reviews
Neuroscience, 1(1), 41–50.
Eichenbaum, H. (2004). Hippocampus: Cognitive processes and neural representations that
underlie declarative memory. Neuron, 44(1), 109–120.
Epstein, R., & Kanwisher, N. (1998). A cortical representation of the local visual environment.
Nature, 392(6676), 598–601.
Ernst, M. D. (2004). Permutation methods: A basis for exact inference. Statistical Science, 19(4),
676–685.
127
Estes, W. (1955). Statistical theory of spontaneous recovery and regression. Psychological
Review, 62(3), 145–154.
Falkenberg, P. R. (1972). Recall improves in short-term memory the more recall context
resembles learning context. Journal of Experimental Psychology, 95(1), 39–47.
Fanselow, M. S. (2000). Contextual fear, gestalt memories, and the hippocampus. Behavioural
Brain Research, 110(12), 73–81.
Freiwald, W. A., & Tsao, D. Y. (2010). Functional compartmentalization and viewpoint
generalization within the macaque face-processing system. Science (New York, N.Y.),
330(6005), 845–851.
Freiwald, W. A., Tsao, D. Y., & Livingstone, M. S. (2009). A face feature space in the macaque
temporal lobe. Nature Neuroscience, 12(9), 1187–1196.
Gainotti, G., & Marra, C. (2011). Differential contribution of right and left temporo-occipital and
anterior temporal lesions to face recognition disorders. Frontiers in Human
Neuroscience, 5, 55.
Gauthier, I., Tarr, M. J., Moylan, J., Skudlarski, P., Gore, J. C., & Anderson, A. W. (2000). The
fusiform “face area” is part of a network that processes faces at the individual level.
Journal of Cognitive Neuroscience, 12(3), 495–504.
Geller, A. S., Schleifer, I. K., Sederberg, P. B., Jacobs, J., & Kahana, M. J. (2007). PyEPL: A
cross-platform experiment-programming library. Behavior Research Methods, 39(4),
950–958.
Genovese, C. R., Lazar, N. A., & Nichols, T. (2002). Thresholding of statistical maps in
functional neuroimaging using the false discovery rate. NeuroImage, 15(4), 870–878.
Gershman, S. J., Moore, C. D., Todd, M. T., Norman, K. A., & Sederberg, P. B. (2012). The
successor representation and temporal context. Neural Computation, 24(6), 1553–1568.
Gillund, G., & Shiffrin, R. M. (1984). A retrieval model for both recognition and recall.
Psychological Review, 91(1), 1–67.
Glanzer, M., Hilford, A., & Maloney, L. T. (2009). Likelihood ratio decisions in memory: Three
implied regularities. Psychonomic Bulletin & Review, 16(3), 431–455.
128
Gläscher, J., Daw, N., Dayan, P., & O’Doherty, J. P. (2010). States versus rewards: Dissociable
neural prediction error signals underlying model-based and model-free reinforcement
learning. Neuron, 66(4), 585–595.
Glenberg, A. M. (1979). Component-levels theory of the effects of spacing of repetitions on
recall and recognition. Memory & Cognition, 7(2), 95–112.
Gobbini, M. I., & Haxby, J. V. (2007). Neural systems for recognition of familiar faces.
Neuropsychologia, 45(1), 32–41.
Godden, D. R., & Baddeley, A. D. (1975). Context-dependent memory in two natural
environments: On land and underwater. British Journal of Psychology, 66(3), 325–331.
Goodhill, G. J., & Barrow, H. G. (1994). The role of weight normalization in competitive
learning. Neural Computation, 6(2), 255–269.
Green, D. M., & Swets, J. A. (1966). Signal detection theory and psychophysicsJohn wiley. New
York.
Grill-Spector, K., Henson, R., & Martin, A. (2006). Repetition and the brain: Neural models of
stimulus-specific effects. Trends in Cognitive Sciences, 10(1), 14–23.
Grill-Spector, K., Knouf, N., & Kanwisher, N. (2004). The fusiform face area subserves face
perception, not generic within-category identification. Nature Neuroscience, 7(5), 555–
562.
Harris, A., & Aguirre, G. K. (2010). Neural tuning for face wholes and parts in human fusiform
gyrus revealed by fMRI adaptation. Journal of Neurophysiology, 104(1), 336–345.
Hasher, L., & Zacks, R. T. (1979). Automatic and effortful processes in memory. Journal of
Experimental Psychology: General, 108(3), 356–388.
Hawkins, J., George, D., & Niemasik, J. (2009). Sequence memory for prediction, inference and
behaviour. Philosophical Transactions of the Royal Society of London B: Biological
Sciences, 364(1521), 1203–1209.
Haxby, J. V., Gobbini, M. I., Furey, M. L., Ishai, A., Schouten, J. L., & Pietrini, P. (2001).
Distributed and overlapping representations of faces and objects in ventral temporal
cortex. Science, 293(5539), 2425–2430.
129
Haxby, J. V., Hoffman, E. A., & Gobbini, M. I. (2000). The distributed human neural system for
face perception. Trends in Cognitive Sciences, 4(6), 223–233.
Hebb, D. O. (1949). The organization of behavior (Vol. 50). New York: Wiley.
Hintzman, D. L. (1988). Judgments of frequency and recognition memory in a multiple-trace
memory model. Psychological Review, 95(4), 528–551.
Howard, M. W., & Kahana, M. J. (2002). A distributed representation of temporal context.
Journal of Mathematical Psychology, 46(3), 269–299.
Howard, M. W., Fotedar, M. S., Datey, A. V., & Hasselmo, M. E. (2005). The temporal context
model in spatial navigation and relational learning: Toward a common explanation of
medial temporal lobe function across domains. Psychological Review, 112(1), 75–116.
Jenkinson, M., & Smith, S. (2001). A global optimisation method for robust affine registration of
brain images. Medical Image Analysis, 5(2), 143–156.
Jenkinson, M., Bannister, P., Brady, M., & Smith, S. (2002). Improved optimization for the
robust and accurate linear registration and motion correction of brain images.
NeuroImage, 17(2), 825–841.
Jenkinson, M., Beckmann, C. F., Behrens, T. E. J., Woolrich, M. W., & Smith, S. M. (2012).
FSL. NeuroImage, 62(2), 782–790.
Jenkinson, M., Pechaud, M., & Smith, S. (2005, June). BET2: MR-based estimation of brain,
skull and scalp surfaces.
Johnson, A., Meer, M. A. van der, & Redish, A. D. (2007). Integrating hippocampus and
striatum in decision-making. Current Opinion in Neurobiology, 17(6), 692–697.
Kahana, M. J. (2012). Foundations of human memory. Oxford University Press.
Kanwisher, N., & Yovel, G. (2006). The fusiform face area: A cortical region specialized for the
perception of faces. Philosophical Transactions of the Royal Society B: Biological
Sciences, 361(1476), 2109–2128.
Kanwisher, N., McDermott, J., & Chun, M. M. (1997). The fusiform face area: A module in
human extrastriate cortex specialized for face perception. The Journal of Neuroscience,
17(11), 4302–4311.
130
Kanwisher, N., Tong, F., & Nakayama, K. (1998). The effect of face inversion on the human
fusiform face area. Cognition, 68(1), B1–B11.
Kennedy, P. J., & Shapiro, M. L. (2004). Retrieving memories via internal context requires the
hippocampus. The Journal of Neuroscience, 24(31), 6979–6985.
Kenward, M. G., & Roger, J. H. (1997). Small sample inference for fixed effects from restricted
maximum likelihood. Biometrics, 53(3), 983–997.
Kimball, D. R., Smith, T. A., & Kahana, M. J. (2007). The fSAM model of false recall.
Psychological Review, 114(4), 954–993.
Kirkland, T., & Cunningham, W. A. (2011). Neural basis of affect and emotion. Wiley
Interdisciplinary Reviews: Cognitive Science, 2(6), 656–665.
Kriegeskorte, N., Formisano, E., Sorger, B., & Goebel, R. (2007). Individual faces elicit distinct
response patterns in human anterior temporal cortex. Proceedings of the National
Academy of Sciences, 104(51), 20600–20605.
Levy, W. B. (1989). A computational approach to hippocampal function. Computational Models
of Learning in Simple Neural Systems, 23, 243–305.
Levy, W. B., Hocking, A. B., & Wu, X. (2005). Interpreting hippocampal function as recoding
and forecasting. Neural Networks, 18(9), 1242–1264.
Lisman, J., & Redish, A. D. (2009). Prediction, sequences and the hippocampus. Philosophical
Transactions of the Royal Society of London B: Biological Sciences, 364(1521), 1193–
1201.
Liu, J., Harris, A., & Kanwisher, N. (2009). Perception of face parts and face configurations: An
fMRI study. Journal of Cognitive Neuroscience, 22(1), 203–211.
Maguire, E. A., & Hassabis, D. (2011). Role of the hippocampus in imagination and future
thinking. Proceedings of the National Academy of Sciences, 108(11), E39–E39.
Marr, D. (1982). Vision: A computational investigation into the human representation and
processing of visual information. new york, NY, USA: Henry holt and co. Inc. June.
Martin, A. (2007). The representation of object concepts in the brain. Annual Review of
Psychology, 58(1), 25–45.
131
Martin, E. (1968). Stimulus meaningfulness and paired-associate transfer: An encoding
variability hypothesis. Psychological Review, 75(5), 421–441.
McCandliss, B. D., Cohen, L., & Dehaene, S. (2003). The visual word form area: Expertise for
reading in the fusiform gyrus. Trends in Cognitive Sciences, 7(7), 293–299.
McClure, S. M., Berns, G. S., & Montague, P. R. (2003). Temporal prediction errors in a passive
learning task activate human striatum. Neuron, 38(2), 339–346.
Melton, A. W. (1970). The situation with respect to the spacing of repetitions and memory.
Journal of Verbal Learning and Verbal Behavior, 9(5), 596–606.
Milton, F., Muhlert, N., Butler, C. R., Smith, A., Benattayallah, A., & Zeman, A. Z. (2011). An
fMRI study of long-term everyday memory using SenseCam. Memory, 19(7), 733–744.
Mitchell, J. P. (2009). Inferences about mental states. Philosophical Transactions of the Royal
Society of London B: Biological Sciences, 364(1521), 1309–1316.
Mizumori, S. J., Ragozzino, K. E., Cooper, B. G., & Leutgeb, S. (1999). Hippocampal
representational organization and spatial context. Hippocampus, 9(4), 444–451.
Murdock, B. B. (1993). TODAM2: A model for the storage and retrieval of item, associative,
and serial-order information. Psychological Review, 100(2), 183–203.
Mystkowski, J. L., Mineka, S., Vernon, L. L., & Zinbarg, R. E. (2003). Changes in caffeine
states enhance return of fear in spider phobia. Journal of Consulting and Clinical
Psychology, 71(2), 243–250.
Natu, V. S., Jiang, F., Narvekar, A., Keshvari, S., Blanz, V., & O’Toole, A. J. (2009).
Dissociable neural patterns of facial identity across changes in viewpoint. Journal of
Cognitive Neuroscience, 22(7), 1570–1582.
Nelson, D. L., McEvoy, C. L., & Schreiber, T. A. (2004). The university of south florida free
association, rhyme, and word fragment norms. Behavior Research Methods, Instruments,
& Computers, 36(3), 402–407.
Nestor, A., Plaut, D. C., & Behrmann, M. (2011). Unraveling the distributed neural code of
facial identity through spatiotemporal pattern analysis. Proceedings of the National
Academy of Sciences, 108(24), 9998–10003.
132
Nielson, D., & Sederberg, P. B. (submitted). MELD paper. Submitted Manuscript.
Nielson, D., Smith, T. A., Sreekumar, V., & Sederberg, P. B. (submitted). Experience sampling
paper. Submitted Manuscript.
Norman, K. A., & O’Reilly, R. C. (2003). Modeling hippocampal and neocortical contributions
to recognition memory: A complementary-learning-systems approach. Psychological
Review, 110(4), 611–646.
Olson, I. R., Ezzyat, Y., Plotzker, A., & Chatterjee, A. (2014). The end point of the ventral visual
stream: Face and non-face perceptual deficits following unilateral anterior temporal lobe
damage. Neurocase, 0(0), 1–9.
Olson, I. R., McCoy, D., Klobusicky, E., & Ross, L. A. (2013). Social cognition and the anterior
temporal lobes: A review and theoretical framework. Social Cognitive and Affective
Neuroscience, 8(2), 123–133.
Pitcher, D., Walsh, V., & Duchaine, B. (2011). The role of the occipital face area in the cortical
face perception network. Experimental Brain Research, 209(4), 481–493.
Pitcher, D., Walsh, V., Yovel, G., & Duchaine, B. (2007). TMS evidence for the involvement of
the right occipital face area in early face processing. Current Biology, 17(18), 1568–
1573.
Polyn, S. M., Norman, K. A., & Kahana, M. J. (2009). A context maintenance and retrieval
model of organizational processes in free recall. Psychological Review, 116(1), 129–156.
Raaijmakers, J. G., & Shiffrin, R. M. (1981). Search of associative memory. Psychological
Review, 88(2), 93–134.
Ratcliff, R., Hockley, W., & McKoon, G. (1985). Components of activation: Repetition and
priming effects in lexical decision and recognition. Journal of Experimental Psychology:
General, 114(4), 435–450.
Rescorla, R. A., & Wagner, A. R. (1972). A theory of pavlovian conditioning: Variations in the
effectiveness of reinforcement and nonreinforcement. In Classical conditioning: Current
research and theory.
Rissman, J., Gazzaley, A., & D’Esposito, M. (2004). Measuring functional connectivity during
distinct stages of a cognitive task. NeuroImage, 23(2), 752–763.
133
Roediger, H. L., & McDermott, K. B. (1995). Creating false memories: Remembering words not
presented in lists. Journal of Experimental Psychology: Learning, Memory, and
Cognition, 21(4), 803–814.
Ross, L. A., & Olson, I. R. (2010). Social cognition and the anterior temporal lobes.
NeuroImage, 49(4), 3452–3462.
Schacter, D. L., & Addis, D. R. (2009). On the nature of medial temporal lobe contributions to
the constructive simulation of future events. Philosophical Transactions of the Royal
Society of London B: Biological Sciences, 364(1521), 1245–1253.
Schacter, D. L., Addis, D. R., & Buckner, R. L. (2007). Remembering the past to imagine the
future: The prospective brain. Nature Reviews Neuroscience, 8(9), 657–661.
Sederberg, P. B., Howard, M. W., & Kahana, M. J. (2008). A context-based theory of recency
and contiguity in free recall. Psychological Review, 115(4), 893–912.
Shankar, K. H., & Howard, M. W. (2012). A scale-invariant internal representation of time.
Neural Computation, 24(1), 134–193.
Shankar, K. H., Jagadisan, U. K. K., & Howard, M. W. (2009). Sequential learning using
temporal context. Journal of Mathematical Psychology, 53(6), 474–485.
Shiffrin, R. M., & Steyvers, M. (1997). A model for recognition memory: REMretrieving
effectively from memory. Psychonomic Bulletin & Review, 4(2), 145–166.
Skaggs, W. E., McNaughton, B. L., Wilson, M., & Barnes, C. (1996). Theta phase precession in
hippocampal. Hippocampus, 6, 149–172.
Smith, S. M., Jenkinson, M., Woolrich, M. W., Beckmann, C. F., Behrens, T. E. J., JohansenBerg, H., … Matthews, P. M. (2004). Advances in functional and structural MR image
analysis and implementation as FSL. NeuroImage, 23 Suppl 1, S208–219.
Smith, T. A., Hasinski, A. E., & Sederberg, P. B. (2013). The context repetition effect: Predicted
events are remembered better, even when they don’t happen. Journal of Experimental
Psychology: General, 142(4), 1298–1308.
Stokes, M. G., Atherton, K., Patai, E. Z., & Nobre, A. C. (2012). Long-term memory prepares
neural activity for perception. Proceedings of the National Academy of Sciences, 109(6),
E360–E367.
134
Summerfield, C., Trittschuh, E. H., Monti, J. M., Mesulam, M.-M., & Egner, T. (2008). Neural
repetition suppression reflects fulfilled perceptual expectations. Nature Neuroscience,
11(9), 1004–1006.
Sutton, R. S., & Barto, A. G. (1998). Reinforcement learning, a bradford book. MIT Press,
Cambridge, MA.
Suzuki, W. A. (2009). Perception and the medial temporal lobe: Evaluating the current evidence.
Neuron, 61(5), 657–666.
Suzuki, W. A., & Baxter, M. G. (2009). Memory, perception, and the medial temporal lobe: A
synthesis of opinions. Neuron, 61(5), 678–679.
Tistarelli, M., Bicego, M., & Grosso, E. (2009). Dynamic face recognition: From human to
machine vision. Image and Vision Computing, 27(3), 222–232.
Tsao, D. Y., Freiwald, W. A., Tootell, R. B. H., & Livingstone, M. S. (2006). A cortical region
consisting entirely of face-selective cells. Science, 311(5761), 670–674.
Tullis, J. G., Benjamin, A. S., & Ross, B. H. (2014). The reminding effect: Presentation of
associates enhances memory for related words in a list. Journal of Experimental
Psychology: General, 143(4), 1526–1540.
Turk-Browne, N. B., Golomb, J. D., & Chun, M. M. (2013). Complementary attentional
components of successful memory encoding. NeuroImage, 66, 553–562.
Turk-Browne, N. B., Simon, M. G., & Sederberg, P. B. (2012). Scene representations in
parahippocampal cortex depend on temporal context. The Journal of Neuroscience,
32(21), 7202–7207.
Turner, B. M., & Sederberg, P. B. (2013). A generalized, likelihood-free method for posterior
estimation. Psychonomic Bulletin & Review, 21(2), 227–250.
Turner, B. M., Forstmann, B. U., Wagenmakers, E.-J., Brown, S. D., Sederberg, P. B., &
Steyvers, M. (2013). A bayesian framework for simultaneously modeling neural and
behavioral data. NeuroImage, 72, 193–206.
Ucros, C. G. (1989). Mood state-dependent memory: A meta-analysis. Cognition and Emotion,
3(2), 139–169.
135
Van Bavel, J. J., Packer, D. J., & Cunningham, W. A. (2008). The neural substrates of in-group
bias a functional magnetic resonance imaging investigation. Psychological Science,
19(11), 1131–1139.
Visser, M., Jefferies, E., & Lambon Ralph, M. A. (2009). Semantic processing in the anterior
temporal lobes: A meta-analysis of the functional neuroimaging literature. Journal of
Cognitive Neuroscience, 22(6), 1083–1094.
Ward, B. (2002). Deconvolution analysis of fMRI time series data. Milwaukee, WI: Biophysics
Research Institute, Medical College of Wisconsin.
Weingartner, H., Adefris, W., Eich, J. E., & Murphy, D. L. (1976). Encoding-imagery specificity
in alcohol state-dependent learning. Journal of Experimental Psychology: Human
Learning and Memory, 2(1), 83–87.
White, L. (1995). Temporal difference learning: Eligibility traces and the successor
representation for actions. Unpublished master’s thesis, De- partment of Computer
Science, University of Toronto.
Xue, G., Dong, Q., Chen, C., Lu, Z., Mumford, J. A., & Poldrack, R. A. (2010). Greater neural
pattern similarity across repetitions is associated with better memory. Science, 330(6000),
97–101.
Zacks, J. M., Speer, N. K., Swallow, K. M., Braver, T. S., & Reynolds, J. R. (2007). Event
perception: A mind-brain perspective. Psychological Bulletin, 133(2), 273–293.
136