Prefrontal Cortex and the Flexibility of Cognitive Control: Rules

Prefrontal Cortex and the Flexibility of Cognitive Control:
Rules Without Symbols
Nicolas P. Rougier
David C. Noelle
Todd S. Braver
Dept of Psychology, CU Boulder
Department of Computer Science
Department of Psychology
INRIA Lorraine, France
Vanderbilt University
Washington University
Jonathan D. Cohen†
Randall C. O’Reilly†
Department of Psychology
Department of Psychology
Princeton University
University of Colorado Boulder
Abstract:
A hallmark of human behavior is the flexibility of cognitive control, which is believed to depend on prefrontal cortex (PFC). Previous work has suggested that specific PFC mechanisms, including active maintenance of rule or goal representations, adaptive gating of these representations, and the top-down biasing
of processing in other cortical areas, are critical for flexible cognitive control. Neural network models have
demonstrated that these mechanisms are sufficient to accurately simulate flexible cognitive control in a
variety of tasks. However, these models have not addressed a fundamental question: How is information
represented in PFC and, critically, how does this develop? Here, we show that PFC-specific mechanisms
interact with the breadth of training experience to produce abstract, rule-like representations that support
generalization of performance in novel task circumstances. We also show that these rule-like representations support patterns of performance characteristic of neurologically intact and frontally-damaged people
on benchmark tasks of cognitive control (Stroop and WCST). Although the rule-like representations that
developed in our model of PFC support flexible cognitive control, they do so in a way that is fundamentally different from symbolic representations characteristic of more traditional unified theories of cognition.
Therefore, these results bear on both the organization and development of PFC at the neurobiogical level,
as well as debates regarding the nature of cognitive flexibility and rule-like behavior at the psychological
level.
† corresponding authors: [email protected],
345 UCB, Boulder, CO 80309-0345 or [email protected],
Green Hall, Princeton, NJ 08544. Supported by ONR grants
N00014-00-1-0246 and N00014-03-1-0428, and NIH grants
MH64445. Last authorship reflects equal contribution; order was determined by a flip of coin. We thank Tim Curran,
Michael Frank, Tom Hazy, Dave Jilk, Ken Norman, Yuko Munakata and members of the CCN lab for helpful comments.
One of the fundamental faculties that distinguishes human cognition and behavior from other species is
our capacity for cognitive control: the ability to behave in accord with rules, goals, or intentions, even when
this runs counter to reflexive or otherwise highly compelling competing responses (e.g., the ability to keep
typing rather than scratch a mosquito bite). A hallmark of the human capacity for cognitive control is its
remarkable flexibility. This flexibility is exhibited in many ways, including the ability to perform a familiar
task in novel environments (generalization; e.g., playing cards with a novel type of deck with new players in
a new location), and the ability to rapidly configure the behavior needed to perform arbitrary new tasks with
relatively little or no prior training (generativity; e.g., playing a new card game for the first time). These
abilities pose a major challenge to theories of cognitive control: How can they be explained in terms of selforganizing mechanisms (that develop on their own, without recourse to unexplained sources of influence or
intelligence; i.e., a “homunculus”), and extensive but not infinite computational resources (1)?
Classic unified theories of cognition, such as ACT-R and SOAR (2), have addressed this issue by assuming that human cognition has the capacity for symbol processing — that is, the ability to execute procedures
for any values of the relevant variables, and to arbitrarily combine these procedures in new ways. While
such theories provide a powerful account of human behavior, they fail to answer several important questions. First, most models based on these theories pre-specify the basic set of procedures that are available,
without explaining how these are acquired. Thus, the modeler is required to imbue such models with much
of the intelligence that is to be explained. Second, one of the great mysteries of neuroscience is how coordinated, goal directed behavior can arise from the highly distributed activity of billions of neurons. Symbolic
models have not addressed this question.
In this paper, we explore an alternative explanation for the origin of cognitive flexibility that addresses
both of these questions: how neural mechanisms can support flexibility of cognitive control, and how this
can develop through experience. Our approach builds on the idea that a primary function of prefrontal cortex
(PFC) is to support rule-like representations that guide behavior (3). Unlike symbolic models, these rule-like
representations do not support arbitrary symbol binding operations, but rather are more like standard neural
network representations, in that they derive their meaning from specific connections between neurons in
PFC that support these representations and neurons in other parts of the brain responsible for carrying out
behavior. By virtue of these connections, sustained patterns of activity in PFC can modulate processing in
other parts of the brain, and thereby govern rule-like behavior. Flexibility of control derives from having a
suitably general set of representations in PFC, and the ability to activate these in task-appropriate contexts.
Extensive neurobiological and theoretical work supports this view. They indicate that PFC exhibits:
1. Robust maintenance of patterns of neural activity over time and against interference from distracting
inputs, so that currently relevant information can be actively held in working memory (4, 5, 6).
2. Adaptive gating of these activity patterns, allowing them to be rapidly and flexibly updated by dynamically switching between robust maintenance and rapid updating (7, 8, 9).
3. Extensive interconnectivity with other cortical areas (e.g., in posterior cortex) (4) allowing patterns of
activity in PFC to modulate processing in pathways responsible for task execution (10, 11).
Neural network models that simulate these PFC mechanisms can explain performance in a wide range
of tasks that rely on cognitive control (10, 12 13, 14), the ability to flexibly shift task sets (15, 16, 17, 14), as
well as behavioral deficits in such tasks associated with disturbances in PFC (18, 9, 14). A critical feature of
these models is the incorporation of rule-like representations in PFC that allow the PFC to exercise control
over processing, and to rapidly shift between tasks. This is consistent with accumulating neurobiological
data indicating that PFC neurons can represent the goals of a task, or the rules needed to perform it (3).
A fundamental limitation of existing models, however, is their use of prespecified representations in PFC,
that are designed “by hand” to perform the particular tasks of interest. Thus, like symbolic models, these
models do not address the question of how these representations develop, and acquire the properties needed
to support the flexible control of behavior. We propose that the development of representations in PFC relies
critically on an interaction between the functional specializations that are characteristic of PFC, as reviewed
above, and exposure to a broad range of tasks during learning. This is consistent with the protracted course
of development of cognitive control, and the human PFC itself (19).
In this article, we describe modeling results that support this proposal, by demonstrating that when a neural network with mechanisms specific to PFC is trained on several different tasks, it develops abstract, rulelike representations that support performance in novel task contexts. Networks without a PFC, or with more
limited training, fail to exhibit this form of flexibility. We present the results of two simulation experiments
a)
b)
No
Cross−Task Generalization
AG
pon
se
PF
(25 C Con
unit text
s)
A1 A2 A3 A4
E1 E2 E3 E4
D1 D2 D3 D4
C1 C2 C3 C4
B1 B2 B3 B4
A1 A2 A3 A4
Left Stimulus
Cu
(16 e Hidd
unit en
s)
Hid
den
Tas
(16 k Hidd
unit en
s)
(83
uni
ts)
Res
C1 C2 C3 C4
B1 B2 B3 B4
E1 E2 E3 E4
D1 D2 D3 D4
C1 C2 C3 C4
B1 B2 B3 B4
A1 A2 A3 A4
Right Stimulus
NF MF SF LF
Task
A
B
C
D
Dimension Cue
E
Generalization % Correct
E1 E2 E3 E4
D1 D2 D3 D4
75
Training Regimen
Task Pairs
All Tasks
50
25
r
c
e
l
f
rio +Re Sel SRN PFC Gat Ful
e
t
− o
s
P
P+
Po
RN N
No PFC S
PFC
Network Configuration
Figure 1: a) The model with the complete PFC system. Stimuli are presented in two possible locations (left, right).
Rows represent different dimensions, labeled A-E, and columns represent different features (1-4). Other inputs include
a task input indicating current task to perform (NF = name feature, MF = match features, SF = smaller feature, LF =
larger feature), and, for the explicit cue condition, a cue as to the currently relevant dimension. Output responses are
generated over the response layer. The AG unit is the adaptive gating unit, providing a temporal-differences (22) based
dynamic gating signal to the PFC context layer. To evaluate the features of this architecture, the following variants
were tested: Posterior: a single large hidden unit layer between inputs and response — a simple model of posterior
cortex without any special active maintenance abilities; P + Rec: posterior + full recurrent connectivity among hidden
units — allows hidden to maintain information over time via attractors; P + Self: posterior + self recurrent connections
from hidden units to themselves — allows individual units to maintain activations over time; SRN: simple recurrent
network, with a context layer that is a copy of the hidden layer on the prior step — widely used form of temporal
maintenance; SRN-PFC: an SRN context layer applied to the PFC layer in the full model (identical to the full PFC
model except for this difference) — tests for role of separated hidden layers; NoGate: the full PFC model without
the AG adaptive gating unit. b) Cross-task generalization results (% correct on test set) for the full PFC network
and a variety of control networks, with either only two tasks (Task Pairs) or all four tasks (All Tasks) used during
training. Overall, the full PFC model generalizes substantially better than the other models, and this interacts with the
level of training such that performance on the All Tasks condition is substantially better than the Task Pairs condition
(with no differences in numbers of training trials or training stimuli). With one feature left out of training for each of
4 dimensions, training represented only 31.6% (324) of the total possible stimulus inputs (1024); The roughly 85%
generalization performance on the remaining test items therefore represents good productive abilities.
using the model. The first tests whether the model’s mechanisms are sufficient to support generalization of
task performance to novel environments, and explores the relationship between this generalization ability
and the factors that lead to the development of rule-like task representations. The second tests whether these
learned PFC representations can account for human performance in tasks known to rely on PFC.
Simulation Study 1
Our model includes mechanisms for actively maintaining representations in a PFC layer, an adaptive
gating mechanism for updating these, and connections that allow activity patterns in PFC to influence processing in other layers of the model (we will refer to these other layers as “posterior cortex”, although in
reality this can include subcortical areas). Active maintenance in our model (Figure 1b) is subserved by
complementary mechanisms at the cellular and circuit levels, which allow individual units to be “latched,”
(i.e., robustly maintained via intrinsic bistability) as well as recurrent connectivity that supports the formation of stable attractor states (6, 20, 17). Adaptive gating is achieved through a mechanism based on phasic
dopamine signals, which allows units in PFC to be transiently responsive to input from posterior structures,
and to latch their state once the input has subsided. In the absence of dopamine signals, representations in
PFC are insensitive to exogenous input, and thereby resistant to interference. The concurrent effects of the
dopamine signal on reinforcement learning (21, 22) allow the system to associate appropriate stimuli with
the gating signal, and thus learn how and when to update representations in PFC (9, 8, 16, 17). Feedback
connections from the PFC to units in posterior areas in our model support the ability of the PFC to influence
processing in pathways responsible for task execution.
We trained a model having all of these mechanisms (the Full PFC model), as well as versions of it
lacking these mechanisms by varying degree, on either 2 or 4 different tasks designed to simulate simple
stimulus processing and active maintenance as demanded by standard laboratory tasks (e.g., the Stroop and
WCST). Each task involved processing multidimensional stimuli (e.g., varying along dimensions such as
size, shape, color, etc) in different ways, including naming a stimulus feature value in a given dimension
(e.g., “blue” for the color dimension), matching features of two stimuli in a given dimension, and ordinal
comparisons of stimulus features along a given dimension (23). All of these tasks required the network to
process stimulus features in one dimension while ignoring features in the other dimensions, maintain this
task set (i.e., knowledge of the relevant dimension) across sequences of trials, and flexibly update it when
the relevant dimension or the task itself changed. To measure generalization, we trained each network on a
subset of stimuli in each task, and then tested it on stimuli that it had not previously seen in that task. Thus,
the network had to generalize what it had learned about these stimuli across the different tasks. To determine
the contribution of the breadth of training experience, we manipulated the range of tasks using two training
regimens (pairs of tasks vs. the full set of four tasks)
Figure 1b shows the effects of network configuration and training regimen on generalization. First, note
that most networks failed to surpass 50% generalization, irrespective of training regimen. This included
networks with full recurrence (and therefore the capacity for active maintenance) throughout the network, as
well as networks with a PFC-like layer including recurrence but lacking an adaptive gating mechanism. Only
the network with a PFC layer that possessed both recurrent connectivity and an adaptive gating mechanism
exhibited substantial generalization, achieving 85% accuracy on stimuli for which it had no prior same-task
experience. This network produced only one-third as many errors on novel stimuli than any other network.
Interestingly, however, this was only the case for the Full PFC network trained in the all-tasks regimen. Even
though the network configuration was identical, training on pairs of tasks exhibited more than four times
as many generalization errors as training on all tasks. This indicates that breadth of experience interacted
with and was critical for exploiting the mechanisms present in the PFC. Note that the increased breadth of
experience in the all-tasks condition was strictly with respect to the range of task contexts in which stimuli
were experienced, and was not confounded by the total amount of training or range of stimulus inputs
presented during training (see (23) for details).
An important goal of this study was to examine the types of representations that develop in the PFC
layer of the network, and relate these to the capacity for generalization. Figure 2 shows examples of representations that developed in four different network configurations. In the full model trained on all tasks,
distinct representations for each stimulus dimension developed in the PFC layer. That is, each PFC unit
came to represent a single dimension, and to represent all features in that dimension. More precisely, these
representations collectively formed a basis set of orthogonal vectors that spanned the space of task-relevant
stimuli, and that were aligned with the dimensions along which features had to be distinguished for task
performance. This effect was only partially apparent in the configuration having a PFC but lacking an adaptive gating mechanism, as well as the full PFC model trained only on task pairs, and was essentially absent
from the model entirely lacking a PFC. Furthermore, the degree to which representations developed that
were dimensional in character was closely related to the network’s generalization performance, as shown
in Figure 2e. A measure of the orthogonality of the PFC representations and their alignment with feature
a) Posterior – All Tasks
b) No Gate – All Tasks
c) Full PFC – Task Pairs
d) Full PFC – All Tasks
e)
f)
texture
position
shape
size
color
Generalization % Correct
Rule−Likeness Predicts Generalization
FullPFC−All
75
r=.97
NoGate−All
50
Posterior−All
25
FullPFC−Pair
0.25
0.50
0.75
Rule Representation Measure
Figure 2: Representations that developed in four different network configurations: a) Posterior cortex (no PFC)
trained on all tasks; b) PFC without the adaptive gating mechanism (all tasks); c) Full PFC trained only on task pairs
(NF & MF in this case); and d) Full PFC (all tasks). Each panel shows the weights from the hidden units (a) or PFC
(b–d) to the response layer. Larger squares correspond to units (all 30 in the PFC, and a random and representative
subset of 30 from the 145 hidden units in the posterior model), and the smaller squares designate the strength of the
connection (lighter = stronger) from that unit to each of the units in the response layer. Note that each row designates
connections to response units representing features in the same stimulus dimension (as illustrated in e) and Figure 1).
It is evident, therefore, that each of the PFC units in the full model (d) represents a single dimension and, conversely,
that each dimension is represented by a distinct subset of PFC units. This pattern is less evident in the model lacking an
adaptive gating mechanism (b), and in the PFC model trained only on task pairs (c), and is almost entirely absent in the
posterior model (a) in which the hidden units appear to encode arbitrary combinations of features across dimensions.
Panel (f) shows the correlation of generalization performance in these cases with the extent to which the units distinctly
and orthogonally encode stimulus dimensions (the rule representation measure, described in (24)).
dimensions was highly correlated with generalization performance across network configurations (r=0.97).
(24) These findings suggest that the model’s ability to generalize task performance to new stimuli is closely
related to the extent to which it developed representations that abstractly and discretely encoded the relevant
stimulus dimensions needed to perform the task (if the tasks emphasized different distinctions, the system
would learn different dimensional representations aligned with these distinctions). These results agree with
those of recent neurophysiological studies, in which PFC neurons have been identified that seem to encode,
in similarly explicit form, the rules that govern the current task (3).
Examining the model, and the conditions under which it was trained, provides insights into why these
rule-like representations developed and were related to generalization performance. During training, when
the network produced an incorrect response, the adaptive gating mechanism temporarily inhibited PFC units
that were active during that response, favoring activation of other units on the next trial. This implemented a
simple form of search (random sampling with delayed replacement). The influence of this search mechanism
interacted with the requirement that, for each task, the network had to attend to features in one dimension and
ignore the others (the “rule” for the task). Together, these put pressure on the network to commit individual
units to a single dimension. If a unit represented several dimensions, then even if one of these was the
correct one, nevertheless that unit would often be inhibited because the other dimensions it represented were
incorrect (and supported spurious responses). As a result of this inhibition, connection weights supporting
the activation of these units were weakened. In contrast, units that represented a single dimension were
allowed to remain fully active as long as that was the correct dimension, and thus their connection weights
were strengthened. As a consequence, the network developed PFC representations in which each unit was
committed to a single dimension.
Complementing this pressure to represent a single dimension, training also put pressure on each PFC
unit to represent all of the features in a given dimension, constituting an abstract representation of that
dimension. This is because within a block of trials the relevant dimension remained the same, while the
specific features within that dimension to which the network had to respond varied. This interacted with the
capacity of the PFC to actively maintain dimension representations. Networks with this capacity learned
that it was more efficient to use a single pattern of activity to represent an entire dimension, and maintain
this pattern of activity across a block of trials, than to switch the pattern activity for each new stimulus (as
the specific features changed). Interestingly, this effect was accentuated by training over all tasks versus just
pairs of tasks: The orthogonality and dimensional alignment of PFC representations was greater for the all
task regimen than the two task regimen, paralleling the effects on generalization. Note that for both training
regimens non-identical subsets of features were used for the different tasks, so that there was no direct
pressure on the network to develop the same dimensional representations for both (or all) tasks. In principle,
the network could have developed dimensional representations dedicated to each task, that spanned only
Normalized Epoch Wt Diff
Relative Stability of PFC & Hidden
1.00
0.75
PFC
Hidden
0.50
0.25
0.00
0 10 20 30 40 50 60 70 80 90 100
Epochs
Figure 3: Relative stability of PFC and hidden layer (posterior cortex) in the model, as indexed by euclidean distance
between weight states at the end of subsequent epochs (one epoch is 2,000 trials). The PFC takes longer to stabilize
(i.e., exhibits greater levels of weight change across epochs) than the posterior cortex. For PFC, within-PFC recurrent
weights were used. For Hidden, weights from stimulus input to Hidden were used. Both sets of weights are an
equivalent distance from error signals at the output layer. The learning rate is reduced at 10 epochs, producing a blip
at that point.
those features needed to perform that task. However, with an increased number of tasks, the overlap among
different subsets of features across tasks, and the constancy of dimensions within task blocks, combined to
support the development of representations that spanned full dimensions across all tasks. This then led to
greater generalization across tasks.
In summary, the development within the PFC layer of an orthogonal basis set corresponding to the taskrelevant dimensions made it possible for the network to identify the appropriate dimension for a given task,
maintain this while needed, and then switch to another one when called upon to do so. These functional
characteristics correspond well to those of rule-governed behavior, and the role ascribed to PFC in such
behaviors. Task rules are typically well-defined and stable over extended periods of time. Thus, the model
illustrates how the mechanisms of active maintenance and adaptive gating, together with appropriate training, can give rise to representations sufficient to support rule-governed behavior. An important feature of
our results is that this ability depended critically upon an interaction of processing mechanisms with an
appropriate range of task experience. We suggest that this accords well with the protracted developmental
course in humans of the PFC and associated dopaminergic systems (extending into late adolescence), allowing a broad range of experience to shape PFC representations (19). Accordingly, we hypothesized that PFC
representations in our model should stabilize later in development (training) than those in posterior areas.
We tested this by measuring the average magnitude of weight changes from projections into the main hidden
layer and in the PFC layer. The hidden layer stabilized within 20 epochs (2,000 trials per epoch), while the
PFC did not stabilize until 70 epochs (Figure 3).
Simulation Study 2
One important question that might be asked is whether the rule-like PFC representations learned by our
model will produce the appropriate patterns of performance in tasks specifically associated with prefrontal
function. To address this question, we conducted simulations of the Stroop task and the Wisconsin Card
Sort Task (WCST), two tasks that have been used widely as benchmarks of prefrontal function (25, 26).
In the Stroop task (27), participants are presented with color words printed in various colors, and are
asked to either read the word or name the color in which it is printed. A universal finding is that, in adults,
color naming is slower than word reading, and that an incongruent word (e.g., “green” displayed in red)
interferes with color naming (saying “red”) while word reading is relatively unaffected by the color of the
display (28). A common interpretation of these findings is that word reading is more automatic (“overlearned”) than color naming, and therefore that color naming demands more cognitive control, especially
when the word conflicts with the response. This demand for cognitive control has been linked directly to
prefrontal cortex function (26). To simulate Stroop task performance, one of the stimulus dimensions was
trained less than the other four dimensions, with all other factors unchanged from the first study. As in previous models of the Stroop task (10), this simulated the asymmetries of experience human participants have
had with color naming versus word reading. Figure 4a compares the model to human performance (data
from 29) in the principal conditions of the Stroop task. The model captures not only the slower responses
for color naming versus word reading, but also the asymmetry of interference effects. These results replicate those of previous modeling work, which have shown that the effects can be explained in terms of an
interaction between PFC representations of the dimensions that define each task (colors vs. words) and the
relative strength of the posterior pathways that map the stimuli onto their corresponding responses (30, 10,
13). Thus, our findings suggest that the present model has discovered representations that function similarly
to those that have been manually imposed in previous models of the role of the PFC in performance on the
Stroop task (10, 13).
The Stroop task illustrates the capacity for cognitive control (i.e., to respond to one stimulus dimension
a)
Reaction Time (msec)
Stroop Task
900
People: Color
People: Word
Model: Lo Freq
Model: Hi Freq
800
700
600
500
Neutral
Conflict
Perseverations/Errors
b)
WCST Task
30
Perseverations
Other Errors
20
10
0
0
25
50
75
PFC Damage (%)
Figure 4: a) Performance of the full PFC network on a simulated Stroop task, demonstrating the classic pattern of
conflict effects on the subordinate task of color naming with unaffected performance on the dominant word reading
task (human data from 29). This was simulated by training one dimension (A) with 1/3 the frequency of the others,
making it weaker. Dimension B was used for the strong (word reading) pathway. To simulate the neutral condition,
inputs were presented with a single stimulus feature active in either the weak or strong dimension, and reaction time
was measured as the number of cycles to name this feature in the response layer with activation > .75. In the
conflict condition, features in both dimensions were activated, and the task input specified which dimension was to be
named. b) Performance of the full PFC network with various levels of damage (units removed) in a simulated WCST
task, demonstrating the classic pattern of increasing perseveration with increased damage. We simplified the task by
measuring for each input “card” the model’s feature naming response, without giving it any cue input specifying the
currently-relevant dimension. Therefore, the model had to discover this dimension based solely on error-feedback
provided to the AG unit (if it responded correctly a reward value of 1 was delivered to this unit; otherwise, a 0
was delivered). This AG error feedback would destabilize the PFC representation until it represented the currently
relevant dimension. After 8 correct trials in a row, the dimension was switched. Performance is plotted in terms of the
number of sequential productions of feature names corresponding to the previously-relevant dimension after a switch
(perseverations). Clearly, the simulated PFC is critical for rapid, flexible switching.
and override responses associated with a more compelling one), but what about the flexibility of control
(i.e., the ability to switch rapidly from processing one dimension to another)? This is illustrated by the
WCST, a task that has long been recognized as being sensitive to frontal lobe damage. In the WCST task
(31), participants are provided with a deck of cards bearing multidimensional stimuli that vary in shape,
size, color, and number. These must be sorted, one by one, according to a particular dimension (the “rule”).
Critically, the participant is not told the rule, and so must discover this from feedback provided about whether
or not each card was sorted correctly. The rule is held constant until the participant makes a criterion number
of correct responses in sequence (e.g, 8), at which point the rule is switched (again without informing the
participant) and the procedure repeated. Neurologically intact individuals rapidly discover and switch to
the new sorting dimension. Patients with frontal damage typically are able to discover the first rule without
difficulty. However, once it switches, they perseverate in sorting according to the previous rule. This has led
many authors to conclude that PFC plays a critical role in the cognitive flexibility required to switch “mental
set” from one rule to another (32).
The WCST is analogous to the form of the feature naming task learned by our model where dimension
cues are not presented to the network: The model is presented with a stimulus and must identify its feature
in the relevant dimension (i.e., discover the appropriate rule). We used this task to compare the performance
of the full model under varying degrees of degradation to the PFC layer. Figure 4b shows that such damage
produces a disproportionate increase in perseverative responding relative to other types of errors. This is
consistent with findings from previous modeling studies, which have accounted for performance deficits in
the WCST in terms of a degradation of rule-like representations in PFC (14, 16, 17). Again, we emphasize
that our model learned the relevant representations in PFC through learning mechanisms, instead of having
them manually imposed as in most of these existing models.
Discussion
The findings reported here support the conjecture that the contribution made by PFC to cognitive control
rests on a critical combination of anatomic and functional specializations coupled with sufficiently broad
learning experiences. When these conditions are satisfied, a “vocabulary” of rule-like representations develop within the PFC that contribute to the flexibility of cognitive control. Our simulations shed light on
several issues that are central in the effort to understand the neural mechanisms underlying cognitive control. First, they illustrate that breadth of training experience has a critical influence on the development of
representations in PFC. This may be one reason why development of the PFC and associated dopaminergic systems is so protracted relative to other brain areas, extending into late adolescence. Our examination
of the developmental trajectory of representations in the model suggested a corollary reason: It is necessary for representations in posterior systems to stabilize before the PFC can extract the dimensions of these
representations relevant to task performance.
Although we found that abstract, rule-like PFC representations supported good generalization in the fully
regular domains that we explored here, we do not claim that these representations are universally beneficial.
In particular, it is unlikely that such discrete, abstract representations are as useful in domains characterized
by more graded knowledge structures. Here distributed representations may perform better (for example, in
the coding and retrieval of semantic knowledge in posterior cortical structures). Thus, there may be a tradeoff
between PFC and posterior cortical types of representations, in which each is better suited for a different
kind of domain. This is consistent with data showing that posterior cortex may be better at learning complex,
similarity-based categories, whereas PFC can more quickly acquire simple rule-based categories (33). More
work is needed to explore these potential tradeoffs, for example in the richness of complex domains such
as language, wherein our model may provide a productive middle ground between the neural network and
symbolic modeling perspectives in the long-standing “rules and regularities in language processing” debates
(e.g., (34).
The model illustrates another critical factor that contributes to flexibility of control: The use of patterns
of activity rather than changes in synaptic weights as a means of exerting control over processing (35). As
with learning in all neural network models, PFC representations developed through synaptic modification.
However, once these were learned, then adaptive behavior in novel circumstances was mediated by a search
for the appropriate pattern of activity, rather than the need to learn a new set of connection strengths. This
may clarify the mechanisms underlying the adaptive coding hypothesis (36), which holds that PFC dynamically reconfigures itself for the task at hand. Whereas this could be construed as involving rapid learning
processes, our theory instead suggests that it depends on flexible updating of activity patterns in PFC. Similarly, this activation-based processing differs fundamentally from the kinds of arbitrary symbol binding
mechanisms present in traditional symbolic models, where the meaning of the underlying representations
can be arbitrarily bound to novel inputs to achieve flexible performance.
Our primary focus in this work has been on one of the factors that contributes to the flexibility of cognitive
control: the ability to generalize task performance to novel environments, and to switch task representations
dynamically in response to changing demands (as in the WCST task). However, there is another equally
important factor in the flexibility of control: the ability to generate entirely new behaviors. Our model does
not directly address this capacity for generativity. While the model was tested on its ability to perform tasks
on new combinations of features, it was never asked to perform entirely new tasks. One might contend
that the same is true for humans — the flexibility of cognitive control has its limits. People cannot actually
perform entirely new tasks without training, at least not with any alacrity (e.g., one cannot play complex
card games such as bridge very well when first learning them). Nevertheless, people can recombine familiar
behaviors in novel ways. Such recombination, when built upon a rich vocabulary of primitives, may support
a broad range of behavior and thus contribute to generativity. Indeed, it is this ability to flexibly recombine
learned behaviors that may best distinguish humans from other species, and account for characteristically
human abilities such as planning and problem solving.
We believe that the PFC representations learned by our model provide a rudimentary example of the type
of task or rule “vocabulary” that can support recombination, and thus constitute an important step toward
understanding the neural mechanisms underlying generativity. In particular, the discrete, orthogonal nature
of these representations should facilitate the formation of novel combinations. As a result, the mechanisms
and processes that we have identified as critical to generalization may also set the stage for understanding
generativity. Generativity also highlights the centrality of search processes to find and activate the appropriate combination of representations for a given task, a point that has long been recognized by symbolic
models of cognition (37). Such models have explored a range of search algorithms and heuristics that vary
in complexity and power (e.g., grid search, hill-climbing, means-ends, etc.). Our model implemented a relatively simple form of search: random sampling with delayed replacement. This appeared to be sufficient
for performance of the simple tasks addressed, consistent with previous results (14, 16). An important focus
of future research will be to identify neural mechanisms that implement more sophisticated forms of search.
This is likely to be necessary to account for human performance in more complex problem solving domains,
in which the PFC plays a critical role (38).
Finally, it should be clear that our model of PFC function relies heavily on the nature of its representations, which contrasts with other approaches that focus instead on processes (e.g., manipulation, inhibition,
planning). Although it is often difficult to clearly distinguish these approaches, it has been argued that
representational accounts provide a better fit with the body of existing data regarding how the brain stores
and processes information (39). We concur with this view, and add that neural network models provide a
particularly clear mechanistic perspective on these issues. Nevertheless, there may be important differences
in the types of representations developed in our model compared to others (e.g., 39), that we hope to explore
in future modeling and empirical work.
In conclusion, we have argued that the capacity of the PFC to actively maintain rule-like task representations, update these as circumstances demand, generalize them to novel environments, and recombine them
in novel ways, are critical factors in the flexibility of cognitive control. However, it is important to recognize
that while these factors may be necessary to understand the richness of human cognitive function, they are
certainly not sufficient. No doubt, an understanding of how these mechanisms interact with other systems
(such as those supporting episodic memory, language function, and affect) will be equally important in developing a full understanding of how cognitive control is implemented in the brain. We hope that the work
presented here contributes to this understanding, and will help foster more detailed studies of this fascinating
and important domain.
*
References and Notes
1. S. Monsell. Control of mental processes. In V. Bruce, editor, Unsolved mysteries of the mind: Tutorial
essays in cognition, pages 93–148. Psychology press, Hove, UK, (1996).
2. J. R. Anderson and C. Lebiere. The Atomic Components of Thought. Erlbaum, Mahwah, NJ, (1998).
A. Newell. Unified Theories of Cognition. Harvard University Press, Cambridge, MA, (1990).
3. I. M. White and S. P. Wise, Experimental Brain Research, 126, 315, (1999). D. Freedman, M. Riesenhuber, T. Poggio, and E. Miller, Science, 291, 312–316, (2001). J. D. Wallis, K. C. Anderson, and E. K.
Miller, Nature, 411, 953–956, (2001). K. Sakai and R. E. Passingham, Nature Neuroscience, 6, 75–81,
(2003). M. Watanabe, Experimental Brain Research, 89, 233–247, (1992).
4. J. M. Fuster. The Prefrontal Cortex: Anatomy, Physiology and Neuropsychology of the Frontal Lobe,
3rd Edition. Lippincott-Raven, New York, (1997).
5. K. Kubota and H. Niki, Journal of Neurophysiology, 34, 337–347, (1971). P. S. Goldman-Rakic,
Handbook of Physiology — The Nervous System, 5, 373–417, (1987). E. K. Miller, C. A. Erickson, and
R. Desimone, Journal of Neuroscience, 16, 5154, (1996).
6. J. M. Fellous, X. J. Wang, and J. E. Lisman, Nature Neuroscience, 1, 273–275, (1998). X.-J. Wang,
Journal of Neuroscience, 19, 9587, (1999). D. Durstewitz, J. K. Seamans, and T. J. Sejnowski, Journal
of Neurophysiology, 83, 1733, (2000).
7. J. D. Cohen, T. S. Braver, and R. C. O’Reilly, Philosophical Transactions of the Royal Society (London)
B, 351, 1515–1527, (1996). S. Hochreiter and J. Schmidhuber, Neural Computation, 9, 1735–1780,
(1997).
8. R. C. O’Reilly, T. S. Braver, and J. D. Cohen. A biologically based computational model of working
memory.
In A. Miyake and P. Shah, editors, Models of Working Memory: Mechanisms of Active
Maintenance and Executive Control., pages 375–411. Cambridge University Press, New York, (1999).
9. T. S. Braver and J. D. Cohen. On the control of control: The role of dopamine in regulating prefrontal
function and working memory. In S. Monsell and J. Driver, editors, Control of Cognitive Processes:
Attention and Performance XVIII, pages 713–737. MIT Press, Cambridge, MA, (2000).
10. J. D. Cohen, K. Dunbar, and J. L. McClelland, Psychological Review, 97, 332–361, (1990).
11. E. K. Miller and J. D. Cohen, Annual Review of Neuroscience, 24, 167–202, (2001).
12. J. D. Cohen, D. Servan-Schreiber, and J. L. McClelland, American Journal of Psychology, 105, 239–
269, (1992)., S. Dehaene and J. P. Changeux, Journal of Cognitive Neuroscience, 1, 244–261, (1989).
13. R. C. O’Reilly and Y. Munakata. Computational Explorations in Cognitive Neuroscience: Understanding the Mind by Simulating the Brain. MIT Press, Cambridge, MA, (2000).
14. S. Dehaene and J. P. Changeux, Cerebral Cortex, 1, 62–79, (1991).
15. S. J. Gilbert and T. Shallice, Cognitive Psychology, 44, 297–337, (2002).
16. R. C. O’Reilly, D. Noelle, T. S. Braver, and J. D. Cohen, Cerebral Cortex, 12, 246–257, (2002).
17. N. P. Rougier and R. C. O‘Reilly, Cognitive Science, 26, 503–520, (2002).
18. J. D. Cohen and D. Servan-Schreiber, Psychological Review, 99, 45–77, (1992). T. S. Braver and J. D.
Cohen, Cognitive Processing, 2, 25–55, (2001). E. Ruppin, Network : Computation in Neural Systems,
6, 635, (1995).
19. A. Diamond and P. S. Goldman-Rakic, Society for Neuroscience Abstracts, 12, 742, (1986). P. R.
Huttenlocher, Neuropsychologia, 28, 517–527, (1990). D. A. Lewis, Neuropsychopharmacology, 16,
385–398, (1997). B. J. Casey, S. Durston, and J. A. Fossella, Clinical Neuroscience Research, 1,
267–282, (2001). J. B. Morton and Y. Munakata, Developmental Science, 5, 435–440, (2002).
20. M. J. Frank, B. Loughry, and R. C. O’Reilly, Cognitive, Affective, and Behavioral Neuroscience, 1,
137–160, (2001). D. Zipser, Neural Computation, 3, 179–193, (1991).
21. P. R. Montague, P. Dayan, and T. J. Sejnowski, Journal of Neuroscience, 16, 1936–1947, (1996).
22. R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction. MIT Press, Cambridge, MA,
(1998).
23. See online supplemental materials for complete details of the network’s architecture and training and
testing procedures. See Figure 1b for a description of the network architecture. Each of the four
tasks required the network to respond to feature values in one dimension and ignore the other four
dimensions. In the naming task (NF), the network was provided with a single stimulus, and required
to identify (“name”) the value of the feature in the selected dimension. In the matching task (MF),
the network was presented with two stimuli and, if they shared the same feature value in the selected
dimension, it had to identify this value, otherwise respond “no.” Finally, two magnitude comparison
tasks required the network to compare the feature values of the two stimuli in the selected dimension
and, in one case identify the larger value (LF) and in the other the smaller value (SF) (with value
arbitrarily assigned according to increasing feature number from left to right across the dimensional
row). Each task was presented in an instructed and an uninstructed form. In the instructed form,
the input units corresponding to both the current task and the relevant dimension were activated. In
the uninstructed form, information about the task but not the relevant dimension was provided to the
network. In this case, the relevant dimension remained the same for sequences of trials, and the network
had to use the feedback it received after each trial to discover the relevant dimension. Note that this
procedure required the network, once it discovered the relevant dimension, to “remember” this for
subsequent trials in the sequence, until the dimension changed, at which point it had to discover the
next dimension.
For a given task the model saw only a subset of the feature values along each dimension, and a relatively
small fraction (about 30%) of all possible stimuli (i.e., combinations of features across dimensions).
A given training run consisted of 100 epochs of 2,000 trials per epoch; it took the networks only
roughly 10 epochs to achieve near-perfect performance on the training items, but we measured crosstask generalization performance every 5 epochs throughout the duration to find the best generalization
for each network, unconfounded by any differences in architecture or in the raw amount of exposure
to features across different training scenarios. Generalization testing measured the network’s ability to
respond to stimuli it had not seen in that task. To evaluate how the range of task experience influenced
performance, the model was trained with two regimens: one involving all possible pairs of tasks, and
the other involving all four tasks.
Finally, we trained and tested different network configurations, in order to test the contribution made
by constituent mechanisms to learning and performance. All network configurations had the same total
number of processing units, in order to control for the effects of overall computing resources. The only
differences among configurations were the patterns of connectivity and the presence or absence of the
adaptive gating mechanism. The various configurations are described in Figure 1. These ranged from
a simple feedforward network with 145 hidden units (equaling the number of hidden plus PFC units in
the full PFC model) to the complete model including full recurrent connectivity within the PFC and an
adaptive gating mechanism.
24. The rule-like representation measure was computed by comparing each unit’s pattern of weights to
the set of 5 orthogonal, complete dimensional target patterns (i.e., the A dimension target pattern has
a 1 for each A feature, and 0’s for the features in all other dimensions, etc.). A numeric value between
0 and 1, where 1 represents a completely orthogonal and complete dimensional representation was
P k |wi ·tk | where tk is the dimensional target pattern k and w i is the
computed for unit i as: di = max
|w ·t |
k
i
k
weight vector for unit i, and |wi · tk | represents the normalized dot product of the two vectors (i.e., the
cosine). This value was then averaged across all units in the layer (PFC or Hidden, as shown in Figure
2) and then correlated with that network’s generalization performance.
25. D. R. Weinberger, K. F. Berman, and D. G. Daniel. Prefrontal cortex dysfunction in schizophrenia.
In H. S. Levin, H. M. Eisenberg, and A. L. Benton, editors, Frontal Lobe Function and Dysfunction,
pages 276–285. Oxford University Press, New York, (1991). P. Vendrell, C. Junque, and J. Grafman,
Neuropsychologia, 33, 341, (1995). D. T. Stuss, B. Levine, M. P. Alexander, J. Hong, C. Palumbo,
L. Hamer, K. J. Murphy, and D. Izukawa, Neuropsychologia, 38, 388–402, (2000). B. Milner, Archives
of Neurology, 9, 90–100, (1963).
26. A. W. MacDonald III, J. D. Cohen, and C. S. Carter, Science, 288, 1835, (2000). M. T. Banich,
M. P. Milham, R. Atchley, N. J. Cohen, A. Webb, T. Wszalek, A. F. Kramer, Z. P. Liang, A. Wright,
J. Shenker, and R. Magin, Journal of Cognitive Neuroscience, 12, 988–1000, (2000).
27. J. R. Stroop, Journal of Experimental Psychology, 18, 643–662, (1935).
28. C. M. MacLeod, Psychological Bulletin, 109, 163–203, (1991).
29. K. Dunbar and C. M. MacLeod, Journal of Experimental Psychology: Human Perception and Performance, 10, 622–639, (1984).
30. R. H. Phaf, A. Heijden, and P. T. Hudson, Cognitive Psychology, 22, 273–341, (1990).
31. D. A. Grant and E. A. Berg, Journal of Experimental Psychology, 38, 404–411, (1948).
32. T. Shallice. From Neuropsychology to Mental Structure. Cambridge University Press, New York,
(1988). J. Duncan, H. Emslie, P. Williams, R. Johnson, and C. Freer, Cognitive Psychology, 30, 257–
303, (1996).
33. F. G. Ashby, L. A. Alfonso-Reese, and E. M. Waldron, Psychological Review, 105, 442, (1998). E. E.
Smith, A. L. Patalano, and J. Jonides, Cognition, 65, 167–196, (1998).
34. J. L. McClelland and K. Patterson, Trends in Cognitive Sciences, 6, 465–472, (2002). S. Pinker and
M. T. Ullman, Trends in Cognitive Sciences, 6, 456–463, (2002).
35. R. C. O’Reilly and Y. Munakata. Computational Explorations in Cognitive Neuroscience: Understanding the Mind by Simulating the Brain. MIT Press, Cambridge, MA, (2000), pp 380-381. Y. Munakata,
Developmental Science, 1, 161–184, (1998).
36. J. Duncan, Nature Reviews Neuroscience, 2, 820–829, (2001).
37. A. Newell and H. A. Simon. Human Problem Solving. Prentice-Hall, Englewood Cliffs, NJ, (1972).
38. S. C. Baker, R. D. Rogers, A. M. Owen, C. D. Frith, R. J. Dolan, R. S. J. Frackowiak, and T. W.
Robbins, Neuropsychologia, 34, 515, (1996).
39. J. N. Wood and J. Grafman, Nature Reviews Neuroscience, 4, 139–147, (2003).
Prefrontal Cortex and the Flexibility of Cognitive Control:
Rules Without Symbols
Supplemental Material
Nicolas P. Rougier
David C. Noelle
Todd S. Braver
Dept of Psychology, CU Boulder
Department of Computer Science
Department of Psychology
INRIA Lorraine, France
Vanderbilt University
Washington University
Jonathan D. Cohen†
Randall C. O’Reilly†
Department of Psychology
Department of Psychology
Princeton University
University of Colorado Boulder
Abstract:
.
† corresponding authors: [email protected], 345 UCB,
Boulder, CO 80309-0345 or [email protected], Green Hall, Princeton,
NJ 08544. Supported by ONR grants N00014-00-1-0246 and N0001403-1-0428, and NIH grants MH64445. Last authorship reflects equal
contribution; order was determined by a flip of coin. We thank Tim
Curran, Michael Frank, Tom Hazy, Dave Jilk, Ken Norman, Yuko Munakata and members of the CCN lab for helpful comments.
No
E1 E2 E3 E4
AG
Res
pon
se
D1 D2 D3 D4
C1 C2 C3 C4
B1 B2 B3 B4
E1 E2 E3 E4
D1 D2 D3 D4
C1 C2 C3 C4
B1 B2 B3 B4
A1 A2 A3 A4
Left Stimulus
Cu
(16 e Hidd
unit en
s)
Hid
den
Tas
(16 k Hidd
unit en
s)
(83
unit
s)
PF
(30 C Con
unit text
s)
A1 A2 A3 A4
E1 E2 E3 E4
D1 D2 D3 D4
C1 C2 C3 C4
B1 B2 B3 B4
A1 A2 A3 A4
Right Stimulus
NF MF SF LF
Task
A
B
C
D
E
Dimension Cue
Figure 1: Full PFC model (as was presented in Figure 1 of the main text). Blue represents input/output layers (two Stimulus
locations, Task and Dimension Cue inputs, and Response output), green represents hidden layers (Task & Cue Hidden, each 16
units, and main Hidden layer with 83 units), and red represents PFC active maintenance layers and related units (i.e., the AG
adaptive gating unit).
This supplementary material provides a detailed explication of the model architectures, equations, and training/testing procedures used. A number of different variations of model architectures and training/testing procedures
were used to determine the basis for the results reported in the main paper.
Model Architectures
The architecture of the full PFC model is shown in Figure 1 of the main paper, and reproduced here (Figure 1). All
models contained the same input/output layers (two Stimulus locations, Task and Dimension Cue inputs, and Response
output). The full model has three hidden layers (Task & Cue Hidden, each 16 units, and main Hidden layer with 83
units) which serve to encode the relevant stimuli and map them onto the appropriate responses. These are thought to
correspond to association areas in posterior cortex. The Task and Cue hidden layers enable the network to develop
internal representations of these inputs that are influenced by the other layers of the network, in particular the PFC.
For example, the PFC representation of a stimulus dimension can activate the Cue hidden layer with a corresponding
representation during the uninstructed task (when there is no dimension cue input), and this helps to support the
network in focusing on the appropriate dimension, through weights that were trained on the instructed cases when a
dimension cue is present. We think of this as representing the process of subvocalization (inner speech), where people
generate internal verbal representations to support task performance (1). As described in the main paper, the PFC and
associated adaptive gating unit (AG) support robust active maintenance that is also capable of rapid updating. Note that
the PFC layer (30 units) had full self connectivity (each unit connected to all other units in the layer), which supported
maintenance in conjunction with a bistable intracellular “latching” mechanism that is described in detail later.
We tested a number of variations on this basic architecture to determine the importance of different features. In all
No
E1 E2 E3 E4
Res
pon
se
D1 D2 D3 D4
C1 C2 C3 C4
Context (30 units)
B1 B2 B3 B4
Context at (t−1)
(30 units)
E1 E2 E3 E4
D1 D2 D3 D4
C1 C2 C3 C4
B1 B2 B3 B4
A1 A2 A3 A4
Left Stimulus
Cu
(16 e Hidd
unit en
s)
Hid
den
Tas
(16 k Hidd
unit en
s)
(83
unit
s)
A1 A2 A3 A4
E1 E2 E3 E4
D1 D2 D3 D4
C1 C2 C3 C4
B1 B2 B3 B4
A1 A2 A3 A4
Right Stimulus
NF MF SF LF
Task
A
B
C
D
E
Dimension Cue
Figure 2: SRN-PFC model: PFC layer has associated context layer that is updated as a simple recurrent network (SRN). The
context is copied from the PFC layer after each trial of processing, providing a maintained activity pattern representing the prior
trial of processing on the next trial.
cases, the total number of internal (hidden + PFC) units was the same, to rule out changes in overall representational
power as a confounding factor. In all cases, we found that these alternative architectures generalized significantly
worse than the full PFC model (Figure 1b in the main paper). This is despite the fact that all network architectures
were capable of learning the training patterns to a similarly high level of performance.
PFC No Gate
This variation was identical to the full PFC model, except it lacked the AG adaptive gating unit. The PFC layer
was still capable of maintenance through its recurrent excitatory connections, but this maintenance was not subject to
adaptive gating signals in response to differences between predicted and actual rewards (as computed by the AG unit).
SRN-PFC
The simple recurrent network (SRN) architecture (2) is perhaps the most widely used mechanism in cognitive
models for simulating the maintenance of contextual information for temporally extended tasks. The SRN uses a
context layer that is updated from an associated hidden layer in an alternating, out-of-phase manner. That is, after a
trial of processing, the hidden layer activations are copied to the context layer, which provides input to the hidden layer
on the next trial. In this way, processing on the next trial can be contextualized by information present on the previous
trial. We have noted that this regular context updating represents a simple form of a gating mechanism, where gating
occurs on every trial (3). To determine how effective this form of gating is compared to the adaptive gating provided
by the AG unit, the SRN-PFC model provides an SRN context layer for the PFC layer (Figure 2).
No
E1 E2 E3 E4
R es
pon
se
D1 D2 D3 D4
C1 C2 C3 C4
B1 B2 B3 B4
Co
nte
xt
(14
5u
nits
)
Hid
den
(14
5u
nits
)
A1 A2 A3 A4
E1 E2 E3 E4
E1 E2 E3 E4
D1 D2 D3 D4
D1 D2 D3 D4
C1 C2 C3 C4
C1 C2 C3 C4
B1 B2 B3 B4
A1 A2 A3 A4
Left Stimulus
B1 B2 B3 B4
NF MF SF LF
A1 A2 A3 A4
Right Stimulus
A
Task
B
C
D
E
Dimension Cue
Figure 3: SRN (No PFC) model: A more standard type of SRN network without the PFC or other hidden layers. The SRN Context
layer is a copy of a single Hidden processing layer that has the same number of total units as those in all the Hidden and PFC layers
of the full PFC model.
No
E1 E2 E3 E4
C1 C2 C3 C4
B1 B2 B3 B4
A1 A2 A3 A4
Hid
den
(14
5u
nits
)
Res
pon
se
D1 D2 D3 D4
E1 E2 E3 E4
D1 D2 D3 D4
C1 C2 C3 C4
B1 B2 B3 B4
A1 A2 A3 A4
Left Stimulus
E1 E2 E3 E4
D1 D2 D3 D4
C1 C2 C3 C4
B1 B2 B3 B4
A1 A2 A3 A4
Right Stimulus
NF MF SF LF
Task
A
B
C
D
E
Dimension Cue
Figure 4: Simple Posterior-Cortex model, with a single Hidden layer having no specialized active maintenance abilities mediating
the input-output mapping. Variations of this model included some maintenance abilities through either self (each unit connected to
itself) or full (each unit connected to all other units) recurrent connections.
SRN (No PFC)
The SRN architecture is usually implemented where the SRN context layer copies a single hidden layer. In many
respects, the SRN context layer can be considered as a simple version of the PFC layer in the original full PFC model.
Therefore, we tested a more standard SRN architecture with a single Hidden layer having the same number of units
present in all the hidden and PFC layers in the original full PFC model (145 units), and an SRN context layer that
copies this Hidden layer (Figure 3).
Posterior-Cortex Model and Variations
The final set of architectures consisted of a single large (145 units) hidden layer, representing posterior cortical
areas, mediating the input-output mapping. This represents a standard three-layer architecture as commonly used in
cognitive modeling. In its most basic form, the hidden layer has no capability of maintenance whatsoever. We also
explored two variations having either self recurrent connectivity (each hidden unit connected to itself) or full recurrent
Feature level 1 2 3 4
Dimension
E
D
C
B
A
Figure 5: Stimulus representation for multi-dimensional stimuli. There were five dimensions, which are thought to correspond to
things like size, color, shape, etc. Each dimension had four feature levels (e.g., small, medium, large, XL for the size dimension). For
the naming feature (NF) task, the network would output the name of the feature along one of the five dimensions (e.g., responding
2 (medium) for the A (size) dimension).
connectivity (each hidden unit connected to all other hidden units).
Given that generalization is of primary concern in these networks, and that in some cases larger networks have been
shown to generalize worse than smaller ones (i.e., due to overfitting from too many degrees of freedom), one might
want to also see results from networks with smaller hidden layers. These networks were run and performed worse than
those with larger hidden layers. This is consistent with the fact that the present task is completely deterministic and
has no opportunities for overfitting to represent noise. In such conditions, larger networks are better because the larger
sample of hidden unit activities reduces the chances that the idiosyncrasies of individual units dominate the overall
responding (i.e., a law of large numbers effect). See (4) for further discussion, results, and analysis.
Training and Testing
All models were trained on four types of tasks, with each task type having instructed and uninstructed versions, for
a total of eight different task conditions. Each task involved performing a simple operation on one or two multi-featured
stimulus inputs (Figure 5), such as naming one stimulus feature (NF), matching features across two stimuli (MF),
and comparing the relative magnitude of features across two stimuli (smaller, SF or larger, LF). Each stimulus was
composed by selecting one of four feature values along five different dimensions, which we think of as corresponding
to simple perceptual dimensions such as number, size, color, shape, and texture (e.g., as in the cards used in the WCST
task). For convenience, the five dimensions were abstractly labeled with the letters A-E, and the features numbered 1-4
(e.g., A2 refers to the 2nd feature in dimension A). No attempt was made to impose any real-world semantic constraints
on these stimuli. There are a total of 45 = 1024 different stimuli. Stimuli could be presented in two different locations
in the input, meaning that there are more than one million distinct stimulus input patterns (1024×1024). The Response
output layer contained the same stimulus representation as the input, with an additional “No” output response.
In each task, one of the five dimensions is relevant at any given time. The relevant dimension can either be
explicitly specified in the Dimension Cue input (instructed task) or it must be guessed by the network through trial and
error (uninstructed task). The task type is always specified via the Task input layer. For example, in the NF-I (naming
feature-instructed) task, the single input stimulus (presented in either input location) might correspond to three large,
rough, blue, squares, and if the relevant dimension in the Dimension Cue is size, the correct output would be “large”.
In the other three task types, the input display consists of two stimuli (in the left and right input locations), and these
stimuli are compared along the relevant dimension. Consistent with the NF task, the output for these tasks is either the
name of a feature, or “No.” For example, if the left stimulus and relevant dimension (size) are as in the prior example,
and the right stimulus is medium sized, then the correct output for MF (matching) is “No”, SF (smaller) is “medium”,
and LF (larger) is “large”.
Each network was trained 10 times with random initial weights for 100 epochs consisting of 2,000 individual
training trials per epoch. Testing for cross-task generalization performance occurred every 5 epochs throughout training, and the best level of generalization performance for each training run was recorded. This procedure eliminated
possible confounds associated with differential amounts of feature exposure per epoch, etc depending on the different
training conditions. The presence of the uninstructed versions of tasks required that relevant dimensions be organized
into contiguous blocks of trials where a given dimension was always relevant. The size and ordering of these blocks
was systematically varied in different “training protocols”. Testing was always performed on the instructed tasks, so
that the differential memory abilities of the various model architectures would not confound the results; no memory of
the relevant dimension is required in the instructed version.
Cross-task generalization testing requires that certain stimuli are not experienced on a given task during training;
these stimuli are then used during testing. Specifically, features B3, C2, D3 and E2 were always excluded from training
on all the instructed tasks, and testing consisted only of instructed task cases where one of these non-trained stimuli
was the relevant stimulus. This means that the instructed tasks were trained on
324
1024
= 31.64% of the full stimulus
space (324 = 4 × 3 × 3 × 3 × 3). The uninstructed versions of each of the four task types were trained with one omitted
stimulus feature. Specifically, the B3 stimulus was never seen during the NF-U task, the C2 stimulus was omitted for
the MF-U, etc. Generalization therefore required the learning on uninstructed tasks to transfer to instructed tasks.
Although some of the generalization tested was within the same task type between uninstructed and instructed (e.g.,
training on NF-U C2 could promote generalization of NF-I on C2), these task types were actually quite different from
the network’s perspective. Specifically, the critical relevant dimension must come from two different sources between
the instructed and uninstructed versions of the tasks, and this means that different representations and weights are
involved. The ability to produce an appropriate response across both cases requires that the internal representation
of the relevant dimension be consistent, and consistently used, across both cases. This generalization of dimensional
representation across tasks was exactly what we sought to test, and it applies equally well to generalization across
instructed and uninstructed as to generalization across naming and matching.
Two training/testing conditions were run; one in which all four task types were used, and another in which only 2
out of the four were used for any given network, with results averaged across all possible combinations of such task
pairs. Due to the structuring of the omitted features, these two conditions did not differ in the range of features trained,
but only in the extent to which a given feature appeared across multiple tasks. Thus, the differences in generalization
between these two conditions reveal the benefits of having processed a given feature in two versus four different task
contexts.
Finally, the sequencing of tasks and relevant dimensions was organized in a hierarchically blocked fashion, with
the outermost block consisting of a loop through all instructed tasks followed by all uninstructed tasks, with each
task being performed for a block of 25 input/output trials. The relevant dimension was switched after every 2 of the
outermost blocks. This relatively low rate of switching allows the networks without adaptive gating mechanisms plenty
of time to adapt to the relevant dimension. In other preliminary experiments, we tested a variety of blocking strategies,
including: a) interleaved instructed and uninstructed tasks within a 50 trial block, followed by a dimension switch; b)
the protocol described above (the “standard” protocol, outer blocks of tasks, instructed first, with inner-blocks of 25
trials); b) the standard protocol but with uninstructed cases first; c) standard but with uninstructed first; d) standard
but with inner-block sizes of 50 with 1 outer-block per dimension switch; e) like d but with uninstructed first; f) loop
through all instructed tasks with 50 trials per relevant dimension, followed by a loop through all uninstructed tasks
with 50 trials per dimension. g) like f but with uninstructed first. The impact of these manipulations never made more
than a 10% impact on performance in tests run on the generalization between U and I versions of the NF task (e.g., the
Full PFC model performance varied between 13.6% generalization error to 23.3% generalization error across these
versions).
Model Equations
All of the models were implemented using the Leabra framework, which incorporates a number of standard neural network mechanisms (e.g., point-neuron biophysical activation function, bidirectional excitatory connectivity, inhibitory competition, Hebbian and error-driven learning) (5, 6). This framework has been used to simulate over 40
different models in (6), and a number of other research models, including some that are closely related to that used
here (7, 8). Thus, models can be viewed as instantiations of a systematic modeling framework using standardized
mechanisms, instead of constructing new mechanisms for each model. All models and supplementary code can be
obtained by email at [email protected].
Pseudocode
The pseudocode for Leabra is given here, showing exactly how the pieces of the algorithm described in more detail
in the subsequent sections fit together.
Outer loop: Iterate over events (trials) within an epoch. For each event:
1. Iterate over minus and plus phases of settling for each event.
(a) At start of settling, for all units:
i. Initialize all state variables (activation, v m, etc).
ii. Apply external patterns (clamp input in minus, input & output in plus).
(b) During each cycle of settling, for all non-clamped units:
i. Compute excitatory netinput (ge (t) or ηj , eq 2).
ii. Compute kWTA inhibition for each layer, based on giΘ (eq 6):
A. Sort units into two groups based on giΘ : top k and remaining k + 1 to n.
B. If basic, find k and k + 1th highest; if avg-based, compute avg of 1 → k & k + 1 → n.
Θ
C. Set inhibitory conductance gi from gkΘ and gk+1
(eq 5).
iii. Compute point-neuron activation combining excitatory input and inhibition (eq 1).
(c) After settling, for all units:
i. Record final settling activations as either minus or plus phase (yj− or yj+ ).
2. After both phases update the weights (based on linear current weight values), for all connections:
(a) Compute error-driven weight changes (eq 8) with soft weight bounding (eq 9).
(b) Compute Hebbian weight changes from plus-phase activations (eq 7).
(c) Compute net weight change as weighted sum of error-driven and Hebbian (eq 10).
(d) Increment the weights according to net weight change.
Point Neuron Activation Function
Leabra uses a point neuron activation function that models the electrophysiological properties of real neurons,
while simplifying their geometry to a single point. This function is nearly as simple computationally as the standard sigmoidal activation function, but the more biologically-based implementation makes it considerably easier to
Parameter
El
Ei
Ee
Vrest
τ
k Hidden
k PFC
Value
0.15
0.15
1.00
0.15
.02
15%
10%
.01
Parameter
gl
gi
ge
Θ
γ
k Output
khebb
to AG Value
0.10
1.0
1.0
0.25
600
1 unit*
.01
.04*
Table 1: Parameters for the simulation (see equations in text for explanations of parameters). All are standard default parameter
values except for those with a * (most of which have no default because they are intrinsically task-dependent). The faster learning
rate () for connections into the AG was important for ensuring rapid learning of reward.
model inhibitory competition, as described below. Further, using this function enables cognitive models to be more
easily related to more physiologically detailed simulations, thereby facilitating bridge-building between biology and
cognition.
The membrane potential Vm is updated as a function of ionic conductances g with reversal (driving) potentials E
as follows:
∆Vm (t) = τ
X
c
gc (t)gc (Ec − Vm (t))
(1)
with 3 channels (c) corresponding to: e excitatory input; l leak current; and i inhibitory input. Following electrophysiological convention, the overall conductance is decomposed into a time-varying component g c (t) computed as
a function of the dynamic state of the network, and a constant g c that controls the relative influence of the different
conductances.
The excitatory net input/conductance ge (t) or ηj is computed as the proportion of open excitatory channels as a
function of sending activations times the weight values:
ηj = ge (t) = hxi wij i =
1X
xi wij
n i
(2)
The inhibitory conductance is computed via the kWTA function described in the next section, and leak is a constant.
Activation communicated to other cells (yj ) is a thresholded (Θ) sigmoidal function of the membrane potential
with gain parameter γ:
yj (t) = 1
1+
1
γ[Vm (t)−Θ]+
(3)
where [x]+ is a threshold function that returns 0 if x < 0 and x if X > 0. Note that if it returns 0, we assume
yj (t) = 0, to avoid dividing by 0. As it is, this function has a very sharp threshold, which interferes with graded
learning learning mechanisms (e.g., gradient descent). To produce a less discontinuous deterministic function with a
softer threshold, the function is convolved with a Gaussian noise kernel (µ = 0, σ = .005), which reflects the intrinsic
processing noise of biological neurons:
Z
yj∗ (x) =
∞
−∞
√
2
2
1
e−z /(2σ ) yj (z − x)dz
2πσ
(4)
where x represents the [Vm (t)−Θ]+ value, and yj∗ (x) is the noise-convolved activation for that value. In the simulation,
this function is implemented using a numerical lookup table.
k-Winners-Take-All Inhibition
Leabra uses a kWTA (k-Winners-Take-All) function to achieve inhibitory competition among units within a layer
(area). The kWTA function computes a uniform level of inhibitory current for all units in the layer, such that the k+1th
most excited unit within a layer is below its firing threshold, while the kth is above threshold. Activation dynamics
similar to those produced by the kWTA function have been shown to result from simulated inhibitory interneurons
that project both feedforward and feedback inhibition (6). Thus, although the kWTA function is somewhat biologically implausible in its implementation (e.g., requiring global information about activation states and using sorting
mechanisms), it provides a computationally effective approximation to biologically plausible inhibitory dynamics.
kWTA is computed via a uniform level of inhibitory current for all units in the layer as follows:
Θ
Θ
gi = gk+1
+ q(gkΘ − gk+1
)
(5)
where 0 < q < 1 (.25 default used here) is a parameter for setting the inhibition between the upper bound of g kΘ
Θ
and the lower bound of gk+1
. These boundary inhibition values are computed as a function of the level of inhibition
necessary to keep a unit right at threshold:
giΘ =
ge∗ g¯e (Ee − Θ) + gl g¯l (El − Θ)
Θ − Ei
(6)
where ge∗ is the excitatory net input without the bias weight contribution — this allows the bias weights to override the
kWTA constraint.
Hebbian and Error-Driven Learning
For learning, Leabra uses a combination of error-driven and Hebbian learning. The error-driven component is
the symmetric midpoint version of the GeneRec algorithm (9), which is functionally equivalent to the deterministic Boltzmann machine and contrastive Hebbian learning (CHL). The network settles in two phases, an expectation
(minus) phase where the network’s actual output is produced, and an outcome (plus) phase where the target output is experienced, and then computes a simple difference of a pre and postsynaptic activation product across these
two phases. For Hebbian learning, Leabra uses essentially the same learning rule used in competitive learning or
mixtures-of-Gaussians which can be seen as a variant of the Oja normalization. The error-driven and Hebbian learning
components are combined additively at each connection to produce a net weight change.
The equation for the Hebbian weight change is:
+
+
+ +
∆hebb wij = x+
i yj − yj wij = yj (xi − wij )
(7)
and for error-driven learning using CHL:
+
− −
∆err wij = (x+
i yj ) − (xi yj )
(8)
which is subject to a soft-weight bounding to keep within the 0 − 1 range:
∆sberr wij = [∆err ]+ (1 − wij ) + [∆err ]− wij
(9)
The two terms are then combined additively with a normalized mixing constant k hebb :
∆wij = [khebb (∆hebb ) + (1 − khebb )(∆sberr )]
(10)
Temporal Differences and Adaptive Critic Gating Mechanisms
The adaptive gating mechanism was based on an adaptive critic unit (AG) that updated its activations according
to the temporal-differences (TD) algorithm. The specific mechanism and its motivations are discussed in detail in (8).
It was implemented in Leabra using plus-minus phase states that correspond to the expected reward at the previous
time step (minus) and the current time step (plus). The difference between these two states is the TD error δ, which
is essentially equivalent to the more standard kinds of error signals computed by the error-driven learning component
of Leabra, except that it represents an error of prediction over time, instead of an instantaneous error in the network
output. This δ value then modulates the strength of an excitatory ionic current (labeled m here) that helps to maintain
PFC activations:
gm (t − 1) = 0 if |δ(t)| > r
(11)
gm (t)j = gm (t − 1) + δ(t)yj (t)
(12)
Thus, a positive δ increases this maintenance current for active units, and a negative δ decreases it. Furthermore, if δ
is sufficiently large in magnitude (greater than the reset threshold r = .5), it resets any existing currents.
To prevent gating from occurring for occasional errors, we included a fault-tolerant mechanism where two errors
in a row had to be made before the AG unit was delivered a 0 reward value. Furthermore, we included rapidly
adapting and decaying negative-only bias weights, which serve to inhibit prior task representations (again see 8) for
more detailed motivations and explorations of this mechanism). These bias weights in the PFC layer learn on the
plus-minus phase difference in unit activations, as normal bias weights do, but they only learn negative values, with
a learning rate of .2 (compared to the standard .01 for the remainder of the network). These biases decay with a time
constant of .5 (i.e., .5 of the existing weight value is subtracted on each trial).
*
References and Notes
1. M. J. Emerson and A. Miyake, Journal of Memory and Language, 48, 148–168, (2003). A. R. Luria. The role
of speech in the regulation of normal and abnormal behavior. Liveright Publishing Corporation, New York,
(1961). L. Vygotsky. Thought and Language. MIT Press (Original work published in 1934), Cambridge, MA,
(1934/1986).
2. J. L. Elman, Cognitive Science, 14, 179–211, (1990).
3. R. C. O’Reilly, Neural Computation, (submitted).
4. R. C. O’Reilly, Neural Computation, 13, 1199–1242, (2001).
5. R. C. O’Reilly, Trends in Cognitive Sciences, 2, 455–462, (1998).
6. R. C. O’Reilly and Y. Munakata. Computational Explorations in Cognitive Neuroscience: Understanding the
Mind by Simulating the Brain. MIT Press, Cambridge, MA, (2000).
7. R. C. O’Reilly, D. Noelle, T. S. Braver, and J. D. Cohen, Cerebral Cortex, 12, 246–257, (2002).
8. N. P. Rougier and R. C. O‘Reilly, Cognitive Science, 26, 503–520, (2002).
9. R. C. O’Reilly, Neural Computation, 8, 895–938, (1996).